{ "diffoscope-json-version": 1, "source1": "/srv/reproducible-results/rbuild-debian/tmp.6K4Of6sBVj/b1/mpi-testsuite_3.2+dfsg-1_amd64.changes", "source2": "/srv/reproducible-results/rbuild-debian/tmp.6K4Of6sBVj/b2/mpi-testsuite_3.2+dfsg-1_amd64.changes", "unified_diff": null, "details": [ { "source1": "Files", "source2": "Files", "unified_diff": "@@ -1,2 +1,2 @@\n \n- bc23c8e1e14a316f9a737f1ee5e1f59b 860956 science optional mpi-testsuite_3.2+dfsg-1_amd64.deb\n+ fac0cf090653a55459044c11606b86b3 835952 science optional mpi-testsuite_3.2+dfsg-1_amd64.deb\n" }, { "source1": "mpi-testsuite_3.2+dfsg-1_amd64.deb", "source2": "mpi-testsuite_3.2+dfsg-1_amd64.deb", "unified_diff": null, "details": [ { "source1": "file list", "source2": "file list", "unified_diff": "@@ -1,3 +1,3 @@\n -rw-r--r-- 0 0 0 4 2016-07-07 08:50:33.000000 debian-binary\n--rw-r--r-- 0 0 0 968 2016-07-07 08:50:33.000000 control.tar.xz\n--rw-r--r-- 0 0 0 859796 2016-07-07 08:50:33.000000 data.tar.xz\n+-rw-r--r-- 0 0 0 972 2016-07-07 08:50:33.000000 control.tar.xz\n+-rw-r--r-- 0 0 0 834788 2016-07-07 08:50:33.000000 data.tar.xz\n" }, { "source1": "control.tar.xz", "source2": "control.tar.xz", "unified_diff": null, "details": [ { "source1": "control.tar", "source2": "control.tar", "unified_diff": null, "details": [ { "source1": "./control", "source2": "./control", "unified_diff": "@@ -1,12 +1,12 @@\n Package: mpi-testsuite\n Version: 3.2+dfsg-1\n Architecture: amd64\n Maintainer: Debian Science Maintainers \n-Installed-Size: 108782\n+Installed-Size: 108654\n Section: science\n Priority: optional\n Homepage: http://trac.mpich.org/projects/mpich/browser/test/mpi\n Description: MPI testsuite results\n The MPI testsuite is taken from the MPICH source tree. It exercises a wide\n range of MPI tests written in C, C++, Fortran77 and Fortran90. Both the MPI-2\n and the MPI-3 standard is targetted. Currently the following MPI\n" }, { "source1": "./md5sums", "source2": "./md5sums", "unified_diff": null, "details": [ { "source1": "./md5sums", "source2": "./md5sums", "comments": [ "Files differ" ], "unified_diff": null } ] } ] } ] }, { "source1": "data.tar.xz", "source2": "data.tar.xz", "unified_diff": null, "details": [ { "source1": "data.tar", "source2": "data.tar", "unified_diff": null, "details": [ { "source1": "file list", "source2": "file list", "unified_diff": "@@ -4,14 +4,14 @@\n drwxr-xr-x 0 root (0) root (0) 0 2016-07-07 08:50:33.000000 ./usr/share/doc/\n drwxr-xr-x 0 root (0) root (0) 0 2016-07-07 08:50:33.000000 ./usr/share/doc/mpi-testsuite/\n -rw-r--r-- 0 root (0) root (0) 164 2016-07-07 08:50:33.000000 ./usr/share/doc/mpi-testsuite/changelog.Debian.gz\n -rw-r--r-- 0 root (0) root (0) 4381 2016-06-19 11:25:41.000000 ./usr/share/doc/mpi-testsuite/copyright\n drwxr-xr-x 0 root (0) root (0) 0 2016-07-07 08:50:33.000000 ./usr/share/mpi-testsuite/\n drwxr-xr-x 0 root (0) root (0) 0 2016-07-07 08:50:33.000000 ./usr/share/mpi-testsuite/results/\n drwxr-xr-x 0 root (0) root (0) 0 2016-07-07 08:50:33.000000 ./usr/share/mpi-testsuite/results/mpich/\n--rw-r--r-- 0 root (0) root (0) 876250 2016-07-07 08:50:33.000000 ./usr/share/mpi-testsuite/results/mpich/summary.junit.xml\n--rw-r--r-- 0 root (0) root (0) 698506 2016-07-07 08:50:33.000000 ./usr/share/mpi-testsuite/results/mpich/summary.tap\n--rw-r--r-- 0 root (0) root (0) 630276 2016-07-07 08:50:33.000000 ./usr/share/mpi-testsuite/results/mpich/summary.xml\n+-rw-r--r-- 0 root (0) root (0) 827503 2016-07-07 08:50:33.000000 ./usr/share/mpi-testsuite/results/mpich/summary.junit.xml\n+-rw-r--r-- 0 root (0) root (0) 649759 2016-07-07 08:50:33.000000 ./usr/share/mpi-testsuite/results/mpich/summary.tap\n+-rw-r--r-- 0 root (0) root (0) 584013 2016-07-07 08:50:33.000000 ./usr/share/mpi-testsuite/results/mpich/summary.xml\n drwxr-xr-x 0 root (0) root (0) 0 2016-07-07 08:50:33.000000 ./usr/share/mpi-testsuite/results/openmpi/\n--rw-r--r-- 0 root (0) root (0) 37426616 2016-07-07 08:50:33.000000 ./usr/share/mpi-testsuite/results/openmpi/summary.junit.xml\n--rw-r--r-- 0 root (0) root (0) 37353470 2016-07-07 08:50:33.000000 ./usr/share/mpi-testsuite/results/openmpi/summary.tap\n--rw-r--r-- 0 root (0) root (0) 34388644 2016-07-07 08:50:33.000000 ./usr/share/mpi-testsuite/results/openmpi/summary.xml\n+-rw-r--r-- 0 root (0) root (0) 37431598 2016-07-07 08:50:33.000000 ./usr/share/mpi-testsuite/results/openmpi/summary.junit.xml\n+-rw-r--r-- 0 root (0) root (0) 37357186 2016-07-07 08:50:33.000000 ./usr/share/mpi-testsuite/results/openmpi/summary.tap\n+-rw-r--r-- 0 root (0) root (0) 34391173 2016-07-07 08:50:33.000000 ./usr/share/mpi-testsuite/results/openmpi/summary.xml\n" }, { "source1": "./usr/share/mpi-testsuite/results/mpich/summary.junit.xml", "source2": "./usr/share/mpi-testsuite/results/mpich/summary.junit.xml", "unified_diff": null, "details": [ { "source1": "./usr/share/mpi-testsuite/results/mpich/summary.junit.xml", "source2": "./usr/share/mpi-testsuite/results/mpich/summary.junit.xml", "unified_diff": "@@ -1,12751 +1,12295 @@\n \n \n- \n+ \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n message="requires MPI version 3.1">\n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n message="requires MPI version 3.1">\n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n message="requires MPI version 3.1">\n \n \n@@ -12807,1477 +12351,1389 @@\n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n message="requires MPI version 3.1">\n \n \n@@ -14289,3702 +13745,3474 @@\n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n message="requires MPI version 3.1">\n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n message="requires MPI version 3.1">\n \n \n@@ -17996,1418 +17224,1386 @@\n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n message="requires MPI version 3.1">\n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n message="non-strict test, strict MPI mode requested">\n \n \n@@ -19434,1508 +18630,1484 @@\n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n" } ] }, { "source1": "./usr/share/mpi-testsuite/results/mpich/summary.tap", "source2": "./usr/share/mpi-testsuite/results/mpich/summary.tap", "unified_diff": "@@ -1,10497 +1,10041 @@\n TAP version 13\n # MPICH test suite results (TAP format)\n-# date 2020-01-05-01-58\n+# date 2018-12-03-17-38\n not ok 1 - ./attr/attrt 2\n ---\n Directory: ./attr\n File: attrt\n Num-procs: 2\n- Date: \"Sun Jan 5 01:58:32 2020\"\n+ Date: \"Mon Dec 3 17:38:42 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n-not ok 2 - ./attr/attric 2\n- ---\n- Directory: ./attr\n- File: attric\n- Num-procs: 2\n- Date: \"Sun Jan 5 01:58:33 2020\"\n- ...\n-## Test output (expected 'No Errors'):\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n+not ok 2 - ./attr/attric 2\n+ ---\n+ Directory: ./attr\n+ File: attric\n+ Num-procs: 2\n+ Date: \"Mon Dec 3 17:38:43 2018\"\n+ ...\n+## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 3 - ./attr/attrerr 1\n ---\n Directory: ./attr\n File: attrerr\n Num-procs: 1\n- Date: \"Sun Jan 5 01:58:34 2020\"\n+ Date: \"Mon Dec 3 17:38:43 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 4 - ./attr/attrend 1\n ---\n Directory: ./attr\n File: attrend\n Num-procs: 1\n- Date: \"Sun Jan 5 01:58:35 2020\"\n+ Date: \"Mon Dec 3 17:38:43 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 5 - ./attr/attrend 2\n ---\n Directory: ./attr\n File: attrend\n Num-procs: 2\n- Date: \"Sun Jan 5 01:58:37 2020\"\n+ Date: \"Mon Dec 3 17:38:44 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 6 - ./attr/attrend2 1\n ---\n Directory: ./attr\n File: attrend2\n Num-procs: 1\n- Date: \"Sun Jan 5 01:58:38 2020\"\n+ Date: \"Mon Dec 3 17:38:44 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 7 - ./attr/attrend2 2\n ---\n Directory: ./attr\n File: attrend2\n Num-procs: 2\n- Date: \"Sun Jan 5 01:58:39 2020\"\n+ Date: \"Mon Dec 3 17:38:45 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 8 - ./attr/attrerrcomm 1\n ---\n Directory: ./attr\n File: attrerrcomm\n Num-procs: 1\n- Date: \"Sun Jan 5 01:58:40 2020\"\n+ Date: \"Mon Dec 3 17:38:45 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 9 - ./attr/attrerrtype 1\n ---\n Directory: ./attr\n File: attrerrtype\n Num-procs: 1\n- Date: \"Sun Jan 5 01:58:41 2020\"\n+ Date: \"Mon Dec 3 17:38:45 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 10 - ./attr/attrdeleteget 1\n ---\n Directory: ./attr\n File: attrdeleteget\n Num-procs: 1\n- Date: \"Sun Jan 5 01:58:42 2020\"\n+ Date: \"Mon Dec 3 17:38:46 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 11 - ./attr/attr2type 1\n ---\n Directory: ./attr\n File: attr2type\n Num-procs: 1\n- Date: \"Sun Jan 5 01:58:43 2020\"\n+ Date: \"Mon Dec 3 17:38:46 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 12 - ./attr/attrorder 1\n ---\n Directory: ./attr\n File: attrorder\n Num-procs: 1\n- Date: \"Sun Jan 5 01:58:43 2020\"\n+ Date: \"Mon Dec 3 17:38:46 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 13 - ./attr/attrordercomm 1\n ---\n Directory: ./attr\n File: attrordercomm\n Num-procs: 1\n- Date: \"Sun Jan 5 01:58:44 2020\"\n+ Date: \"Mon Dec 3 17:38:47 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 14 - ./attr/attrordertype 1\n ---\n Directory: ./attr\n File: attrordertype\n Num-procs: 1\n- Date: \"Sun Jan 5 01:58:45 2020\"\n+ Date: \"Mon Dec 3 17:38:47 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 15 - ./attr/baseattr2 1\n ---\n Directory: ./attr\n File: baseattr2\n Num-procs: 1\n- Date: \"Sun Jan 5 01:58:46 2020\"\n+ Date: \"Mon Dec 3 17:38:47 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 16 - ./attr/baseattrcomm 1\n ---\n Directory: ./attr\n File: baseattrcomm\n Num-procs: 1\n- Date: \"Sun Jan 5 01:58:46 2020\"\n+ Date: \"Mon Dec 3 17:38:48 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 17 - ./attr/fkeyval 1\n ---\n Directory: ./attr\n File: fkeyval\n Num-procs: 1\n- Date: \"Sun Jan 5 01:58:47 2020\"\n+ Date: \"Mon Dec 3 17:38:48 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 18 - ./attr/fkeyvalcomm 1\n ---\n Directory: ./attr\n File: fkeyvalcomm\n Num-procs: 1\n- Date: \"Sun Jan 5 01:58:47 2020\"\n+ Date: \"Mon Dec 3 17:38:48 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 19 - ./attr/fkeyvaltype 1\n ---\n Directory: ./attr\n File: fkeyvaltype\n Num-procs: 1\n- Date: \"Sun Jan 5 01:58:48 2020\"\n+ Date: \"Mon Dec 3 17:38:49 2018\"\n ...\n ## Test output (expected 'No Errors'):\n+## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n+## MPIR_Init_thread(565)..............: \n+## MPID_Init(224).....................: channel initialization failed\n+## MPIDI_CH3_Init(105)................: \n+## MPID_nem_init(324).................: \n+## MPID_nem_tcp_init(178).............: \n+## MPID_nem_tcp_get_business_card(425): \n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 20 - ./attr/keyval_double_free 1\n ---\n Directory: ./attr\n File: keyval_double_free\n Num-procs: 1\n- Date: \"Sun Jan 5 01:58:49 2020\"\n+ Date: \"Mon Dec 3 17:38:49 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 21 - ./attr/keyval_double_free_comm 1\n ---\n Directory: ./attr\n File: keyval_double_free_comm\n Num-procs: 1\n- Date: \"Sun Jan 5 01:58:49 2020\"\n+ Date: \"Mon Dec 3 17:38:49 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 22 - ./attr/keyval_double_free_type 1\n ---\n Directory: ./attr\n File: keyval_double_free_type\n Num-procs: 1\n- Date: \"Sun Jan 5 01:58:50 2020\"\n+ Date: \"Mon Dec 3 17:38:50 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 23 - ./attr/keyval_double_free_win 1\n ---\n Directory: ./attr\n File: keyval_double_free_win\n Num-procs: 1\n- Date: \"Sun Jan 5 01:58:51 2020\"\n+ Date: \"Mon Dec 3 17:38:50 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 24 - ./coll/allred 2\n ---\n Directory: ./coll\n File: allred\n Num-procs: 2\n- Date: \"Sun Jan 5 01:58:51 2020\"\n+ Date: \"Mon Dec 3 17:38:51 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n-## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n-## MPIR_Init_thread(565)..............: \n-## MPID_Init(224).....................: channel initialization failed\n-## MPIDI_CH3_Init(105)................: \n-## MPID_nem_init(324).................: \n-## MPID_nem_tcp_init(178).............: \n-## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 25 - ./coll/allred 2\n ---\n Directory: ./coll\n File: allred\n Num-procs: 2\n- Date: \"Sun Jan 5 01:58:52 2020\"\n+ Date: \"Mon Dec 3 17:38:51 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n-## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n-## MPIR_Init_thread(565)..............: \n-## MPID_Init(224).....................: channel initialization failed\n-## MPIDI_CH3_Init(105)................: \n-## MPID_nem_init(324).................: \n-## MPID_nem_tcp_init(178).............: \n-## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 26 - ./coll/allred 2\n ---\n Directory: ./coll\n File: allred\n Num-procs: 2\n- Date: \"Sun Jan 5 01:58:53 2020\"\n+ Date: \"Mon Dec 3 17:38:52 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 27 - ./coll/allredmany 2\n ---\n Directory: ./coll\n File: allredmany\n Num-procs: 2\n- Date: \"Sun Jan 5 01:58:54 2020\"\n+ Date: \"Mon Dec 3 17:38:52 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n-## Fatal error in PMPI_Init: Other MPI error, error stack:\n-## MPIR_Init_thread(565)..............: \n-## MPID_Init(224).....................: channel initialization failed\n-## MPIDI_CH3_Init(105)................: \n-## MPID_nem_init(324).................: \n-## MPID_nem_tcp_init(178).............: \n-## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 28 - ./coll/allred2 2\n ---\n Directory: ./coll\n File: allred2\n Num-procs: 2\n- Date: \"Sun Jan 5 01:58:55 2020\"\n+ Date: \"Mon Dec 3 17:38:53 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 29 - ./coll/allred3 2\n ---\n Directory: ./coll\n File: allred3\n Num-procs: 2\n- Date: \"Sun Jan 5 01:58:55 2020\"\n+ Date: \"Mon Dec 3 17:38:53 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 30 - ./coll/allred4 2\n ---\n Directory: ./coll\n File: allred4\n Num-procs: 2\n- Date: \"Sun Jan 5 01:58:56 2020\"\n+ Date: \"Mon Dec 3 17:38:54 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 31 - ./coll/allred5 2\n ---\n Directory: ./coll\n File: allred5\n Num-procs: 2\n- Date: \"Sun Jan 5 01:58:57 2020\"\n+ Date: \"Mon Dec 3 17:38:54 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 32 - ./coll/allred5 2\n ---\n Directory: ./coll\n File: allred5\n Num-procs: 2\n- Date: \"Sun Jan 5 01:58:58 2020\"\n+ Date: \"Mon Dec 3 17:38:55 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 33 - ./coll/allred6 2\n ---\n Directory: ./coll\n File: allred6\n Num-procs: 2\n- Date: \"Sun Jan 5 01:58:58 2020\"\n+ Date: \"Mon Dec 3 17:38:55 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n-## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n-## MPIR_Init_thread(565)..............: \n-## MPID_Init(224).....................: channel initialization failed\n-## MPIDI_CH3_Init(105)................: \n-## MPID_nem_init(324).................: \n-## MPID_nem_tcp_init(178).............: \n-## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 34 - ./coll/allred6 2\n ---\n Directory: ./coll\n File: allred6\n Num-procs: 2\n- Date: \"Sun Jan 5 01:58:59 2020\"\n+ Date: \"Mon Dec 3 17:38:56 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n-not ok 35 - ./coll/reduce 2\n- ---\n- Directory: ./coll\n- File: reduce\n- Num-procs: 2\n- Date: \"Sun Jan 5 01:59:01 2020\"\n- ...\n-## Test output (expected 'No Errors'):\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n-not ok 36 - ./coll/reduce 2\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n+not ok 35 - ./coll/reduce 2\n ---\n Directory: ./coll\n File: reduce\n Num-procs: 2\n- Date: \"Sun Jan 5 01:59:02 2020\"\n+ Date: \"Mon Dec 3 17:38:56 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n-not ok 37 - ./coll/reduce_local 2\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n+not ok 36 - ./coll/reduce 2\n ---\n Directory: ./coll\n- File: reduce_local\n+ File: reduce\n Num-procs: 2\n- Date: \"Sun Jan 5 01:59:03 2020\"\n+ Date: \"Mon Dec 3 17:38:57 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n-not ok 38 - ./coll/op_commutative 2\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n+not ok 37 - ./coll/reduce_local 2\n ---\n Directory: ./coll\n- File: op_commutative\n+ File: reduce_local\n Num-procs: 2\n- Date: \"Sun Jan 5 01:59:05 2020\"\n+ Date: \"Mon Dec 3 17:38:57 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n+not ok 38 - ./coll/op_commutative 2\n+ ---\n+ Directory: ./coll\n+ File: op_commutative\n+ Num-procs: 2\n+ Date: \"Mon Dec 3 17:38:58 2018\"\n+ ...\n+## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 39 - ./coll/red3 2\n ---\n Directory: ./coll\n File: red3\n Num-procs: 2\n- Date: \"Sun Jan 5 01:59:05 2020\"\n+ Date: \"Mon Dec 3 17:38:59 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n-## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n-## MPIR_Init_thread(565)..............: \n-## MPID_Init(224).....................: channel initialization failed\n-## MPIDI_CH3_Init(105)................: \n-## MPID_nem_init(324).................: \n-## MPID_nem_tcp_init(178).............: \n-## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 40 - ./coll/red4 2\n ---\n Directory: ./coll\n File: red4\n Num-procs: 2\n- Date: \"Sun Jan 5 01:59:06 2020\"\n+ Date: \"Mon Dec 3 17:38:59 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 41 - ./coll/alltoall1 2\n ---\n Directory: ./coll\n File: alltoall1\n Num-procs: 2\n- Date: \"Sun Jan 5 01:59:08 2020\"\n+ Date: \"Mon Dec 3 17:39:00 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n-## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n-## MPIR_Init_thread(565)..............: \n-## MPID_Init(224).....................: channel initialization failed\n-## MPIDI_CH3_Init(105)................: \n-## MPID_nem_init(324).................: \n-## MPID_nem_tcp_init(178).............: \n-## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 42 - ./coll/alltoallv 2\n ---\n Directory: ./coll\n File: alltoallv\n Num-procs: 2\n- Date: \"Sun Jan 5 01:59:09 2020\"\n+ Date: \"Mon Dec 3 17:39:00 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 43 - ./coll/alltoallv0 2\n ---\n Directory: ./coll\n File: alltoallv0\n Num-procs: 2\n- Date: \"Sun Jan 5 01:59:10 2020\"\n+ Date: \"Mon Dec 3 17:39:01 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n-## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n-## MPIR_Init_thread(565)..............: \n-## MPID_Init(224).....................: channel initialization failed\n-## MPIDI_CH3_Init(105)................: \n-## MPID_nem_init(324).................: \n-## MPID_nem_tcp_init(178).............: \n-## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 44 - ./coll/alltoallw1 2\n ---\n Directory: ./coll\n File: alltoallw1\n Num-procs: 2\n- Date: \"Sun Jan 5 01:59:10 2020\"\n+ Date: \"Mon Dec 3 17:39:01 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 45 - ./coll/alltoallw2 2\n ---\n Directory: ./coll\n File: alltoallw2\n Num-procs: 2\n- Date: \"Sun Jan 5 01:59:11 2020\"\n+ Date: \"Mon Dec 3 17:39:02 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n+## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n+## MPIR_Init_thread(565)..............: \n+## MPID_Init(224).....................: channel initialization failed\n+## MPIDI_CH3_Init(105)................: \n+## MPID_nem_init(324).................: \n+## MPID_nem_tcp_init(178).............: \n+## MPID_nem_tcp_get_business_card(425): \n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 46 - ./coll/alltoallw_zeros 1\n ---\n Directory: ./coll\n File: alltoallw_zeros\n Num-procs: 1\n- Date: \"Sun Jan 5 01:59:12 2020\"\n+ Date: \"Mon Dec 3 17:39:02 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 47 - ./coll/alltoallw_zeros 2\n ---\n Directory: ./coll\n File: alltoallw_zeros\n Num-procs: 2\n- Date: \"Sun Jan 5 01:59:13 2020\"\n+ Date: \"Mon Dec 3 17:39:03 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 48 - ./coll/alltoallw_zeros 2\n ---\n Directory: ./coll\n File: alltoallw_zeros\n Num-procs: 2\n- Date: \"Sun Jan 5 01:59:13 2020\"\n+ Date: \"Mon Dec 3 17:39:04 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n-## Fatal error in PMPI_Init: Other MPI error, error stack:\n-## MPIR_Init_thread(565)..............: \n-## MPID_Init(224).....................: channel initialization failed\n-## MPIDI_CH3_Init(105)................: \n-## MPID_nem_init(324).................: \n-## MPID_nem_tcp_init(178).............: \n-## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 49 - ./coll/alltoallw_zeros 2\n ---\n Directory: ./coll\n File: alltoallw_zeros\n Num-procs: 2\n- Date: \"Sun Jan 5 01:59:14 2020\"\n+ Date: \"Mon Dec 3 17:39:04 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n+## Fatal error in PMPI_Init: Other MPI error, error stack:\n+## MPIR_Init_thread(565)..............: \n+## MPID_Init(224).....................: channel initialization failed\n+## MPIDI_CH3_Init(105)................: \n+## MPID_nem_init(324).................: \n+## MPID_nem_tcp_init(178).............: \n+## MPID_nem_tcp_get_business_card(425): \n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 50 - ./coll/allgather2 2\n ---\n Directory: ./coll\n File: allgather2\n Num-procs: 2\n- Date: \"Sun Jan 5 01:59:15 2020\"\n+ Date: \"Mon Dec 3 17:39:05 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 51 - ./coll/allgather3 2\n ---\n Directory: ./coll\n File: allgather3\n Num-procs: 2\n- Date: \"Sun Jan 5 01:59:17 2020\"\n+ Date: \"Mon Dec 3 17:39:05 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n-## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n-## MPIR_Init_thread(565)..............: \n-## MPID_Init(224).....................: channel initialization failed\n-## MPIDI_CH3_Init(105)................: \n-## MPID_nem_init(324).................: \n-## MPID_nem_tcp_init(178).............: \n-## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 52 - ./coll/allgatherv2 2\n ---\n Directory: ./coll\n File: allgatherv2\n Num-procs: 2\n- Date: \"Sun Jan 5 01:59:18 2020\"\n+ Date: \"Mon Dec 3 17:39:06 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n+not ok 53 - ./coll/allgatherv3 2\n+ ---\n+ Directory: ./coll\n+ File: allgatherv3\n+ Num-procs: 2\n+ Date: \"Mon Dec 3 17:39:06 2018\"\n+ ...\n+## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n-not ok 53 - ./coll/allgatherv3 2\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n+not ok 54 - ./coll/allgatherv4 2\n ---\n Directory: ./coll\n- File: allgatherv3\n+ File: allgatherv4\n Num-procs: 2\n- Date: \"Sun Jan 5 01:59:19 2020\"\n+ Date: \"Mon Dec 3 17:39:07 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n-not ok 54 - ./coll/allgatherv4 2\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n+not ok 55 - ./coll/allgather_struct 2\n ---\n Directory: ./coll\n- File: allgatherv4\n+ File: allgather_struct\n Num-procs: 2\n- Date: \"Sun Jan 5 01:59:20 2020\"\n+ Date: \"Mon Dec 3 17:39:07 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n-not ok 55 - ./coll/allgather_struct 2\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n+not ok 56 - ./coll/bcasttest 2\n ---\n Directory: ./coll\n- File: allgather_struct\n+ File: bcasttest\n Num-procs: 2\n- Date: \"Sun Jan 5 01:59:21 2020\"\n+ Date: \"Mon Dec 3 17:39:08 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n-not ok 56 - ./coll/bcasttest 2\n- ---\n- Directory: ./coll\n- File: bcasttest\n- Num-procs: 2\n- Date: \"Sun Jan 5 01:59:22 2020\"\n- ...\n-## Test output (expected 'No Errors'):\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 57 - ./coll/bcasttest 2\n ---\n Directory: ./coll\n File: bcasttest\n Num-procs: 2\n- Date: \"Sun Jan 5 01:59:23 2020\"\n+ Date: \"Mon Dec 3 17:39:09 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 58 - ./coll/bcasttest 2\n ---\n Directory: ./coll\n File: bcasttest\n Num-procs: 2\n- Date: \"Sun Jan 5 01:59:24 2020\"\n+ Date: \"Mon Dec 3 17:39:09 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 59 - ./coll/bcast_full 2\n ---\n Directory: ./coll\n File: bcast_full\n Num-procs: 2\n- Date: \"Sun Jan 5 01:59:25 2020\"\n+ Date: \"Mon Dec 3 17:39:10 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n-## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n-## MPIR_Init_thread(565)..............: \n-## MPID_Init(224).....................: channel initialization failed\n-## MPIDI_CH3_Init(105)................: \n-## MPID_nem_init(324).................: \n-## MPID_nem_tcp_init(178).............: \n-## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 60 - ./coll/bcast_min_datatypes 2\n ---\n Directory: ./coll\n File: bcast_min_datatypes\n Num-procs: 2\n- Date: \"Sun Jan 5 01:59:26 2020\"\n+ Date: \"Mon Dec 3 17:39:10 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n-## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n-## MPIR_Init_thread(565)..............: \n-## MPID_Init(224).....................: channel initialization failed\n-## MPIDI_CH3_Init(105)................: \n-## MPID_nem_init(324).................: \n-## MPID_nem_tcp_init(178).............: \n-## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 61 - ./coll/bcast_comm_world 2\n ---\n Directory: ./coll\n File: bcast_comm_world\n Num-procs: 2\n- Date: \"Sun Jan 5 01:59:27 2020\"\n+ Date: \"Mon Dec 3 17:39:11 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 62 - ./coll/bcastzerotype 1\n ---\n Directory: ./coll\n File: bcastzerotype\n Num-procs: 1\n- Date: \"Sun Jan 5 01:59:28 2020\"\n+ Date: \"Mon Dec 3 17:39:12 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 63 - ./coll/bcastzerotype 2\n ---\n Directory: ./coll\n File: bcastzerotype\n Num-procs: 2\n- Date: \"Sun Jan 5 01:59:29 2020\"\n+ Date: \"Mon Dec 3 17:39:12 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n+## Fatal error in PMPI_Init: Other MPI error, error stack:\n+## MPIR_Init_thread(565)..............: \n+## MPID_Init(224).....................: channel initialization failed\n+## MPIDI_CH3_Init(105)................: \n+## MPID_nem_init(324).................: \n+## MPID_nem_tcp_init(178).............: \n+## MPID_nem_tcp_get_business_card(425): \n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 64 - ./coll/bcastzerotype 2\n ---\n Directory: ./coll\n File: bcastzerotype\n Num-procs: 2\n- Date: \"Sun Jan 5 01:59:31 2020\"\n+ Date: \"Mon Dec 3 17:39:13 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n ## Fatal error in PMPI_Init: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 65 - ./coll/bcastzerotype 2\n ---\n Directory: ./coll\n File: bcastzerotype\n Num-procs: 2\n- Date: \"Sun Jan 5 01:59:32 2020\"\n+ Date: \"Mon Dec 3 17:39:13 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n+## Fatal error in PMPI_Init: Other MPI error, error stack:\n+## MPIR_Init_thread(565)..............: \n+## MPID_Init(224).....................: channel initialization failed\n+## MPIDI_CH3_Init(105)................: \n+## MPID_nem_init(324).................: \n+## MPID_nem_tcp_init(178).............: \n+## MPID_nem_tcp_get_business_card(425): \n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 66 - ./coll/coll2 2\n ---\n Directory: ./coll\n File: coll2\n Num-procs: 2\n- Date: \"Sun Jan 5 01:59:32 2020\"\n+ Date: \"Mon Dec 3 17:39:14 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 67 - ./coll/coll3 2\n ---\n Directory: ./coll\n File: coll3\n Num-procs: 2\n- Date: \"Sun Jan 5 01:59:34 2020\"\n+ Date: \"Mon Dec 3 17:39:15 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 68 - ./coll/coll4 2\n ---\n Directory: ./coll\n File: coll4\n Num-procs: 2\n- Date: \"Sun Jan 5 01:59:35 2020\"\n+ Date: \"Mon Dec 3 17:39:15 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 69 - ./coll/coll5 2\n ---\n Directory: ./coll\n File: coll5\n Num-procs: 2\n- Date: \"Sun Jan 5 01:59:36 2020\"\n+ Date: \"Mon Dec 3 17:39:16 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n-## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n-## MPIR_Init_thread(565)..............: \n-## MPID_Init(224).....................: channel initialization failed\n-## MPIDI_CH3_Init(105)................: \n-## MPID_nem_init(324).................: \n-## MPID_nem_tcp_init(178).............: \n-## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 70 - ./coll/coll6 2\n ---\n Directory: ./coll\n File: coll6\n Num-procs: 2\n- Date: \"Sun Jan 5 01:59:36 2020\"\n+ Date: \"Mon Dec 3 17:39:16 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 71 - ./coll/coll7 1\n ---\n Directory: ./coll\n File: coll7\n Num-procs: 1\n- Date: \"Sun Jan 5 01:59:37 2020\"\n+ Date: \"Mon Dec 3 17:39:17 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 72 - ./coll/coll7 2\n ---\n Directory: ./coll\n File: coll7\n Num-procs: 2\n- Date: \"Sun Jan 5 01:59:38 2020\"\n+ Date: \"Mon Dec 3 17:39:18 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n+## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n+## MPIR_Init_thread(565)..............: \n+## MPID_Init(224).....................: channel initialization failed\n+## MPIDI_CH3_Init(105)................: \n+## MPID_nem_init(324).................: \n+## MPID_nem_tcp_init(178).............: \n+## MPID_nem_tcp_get_business_card(425): \n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 73 - ./coll/coll7 2\n ---\n Directory: ./coll\n File: coll7\n Num-procs: 2\n- Date: \"Sun Jan 5 01:59:39 2020\"\n+ Date: \"Mon Dec 3 17:39:18 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n-not ok 74 - ./coll/coll8 2\n- ---\n- Directory: ./coll\n- File: coll8\n- Num-procs: 2\n- Date: \"Sun Jan 5 01:59:40 2020\"\n- ...\n-## Test output (expected 'No Errors'):\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n+not ok 74 - ./coll/coll8 2\n+ ---\n+ Directory: ./coll\n+ File: coll8\n+ Num-procs: 2\n+ Date: \"Mon Dec 3 17:39:19 2018\"\n+ ...\n+## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 75 - ./coll/coll9 2\n ---\n Directory: ./coll\n File: coll9\n Num-procs: 2\n- Date: \"Sun Jan 5 01:59:41 2020\"\n+ Date: \"Mon Dec 3 17:39:20 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 76 - ./coll/coll10 2\n ---\n Directory: ./coll\n File: coll10\n Num-procs: 2\n- Date: \"Sun Jan 5 01:59:43 2020\"\n+ Date: \"Mon Dec 3 17:39:20 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n+## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n+## MPIR_Init_thread(565)..............: \n+## MPID_Init(224).....................: channel initialization failed\n+## MPIDI_CH3_Init(105)................: \n+## MPID_nem_init(324).................: \n+## MPID_nem_tcp_init(178).............: \n+## MPID_nem_tcp_get_business_card(425): \n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 77 - ./coll/coll11 2\n ---\n Directory: ./coll\n File: coll11\n Num-procs: 2\n- Date: \"Sun Jan 5 01:59:44 2020\"\n+ Date: \"Mon Dec 3 17:39:21 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n+## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n+## MPIR_Init_thread(565)..............: \n+## MPID_Init(224).....................: channel initialization failed\n+## MPIDI_CH3_Init(105)................: \n+## MPID_nem_init(324).................: \n+## MPID_nem_tcp_init(178).............: \n+## MPID_nem_tcp_get_business_card(425): \n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 78 - ./coll/coll12 2\n ---\n Directory: ./coll\n File: coll12\n Num-procs: 2\n- Date: \"Sun Jan 5 01:59:45 2020\"\n+ Date: \"Mon Dec 3 17:39:22 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 79 - ./coll/coll13 2\n ---\n Directory: ./coll\n File: coll13\n Num-procs: 2\n- Date: \"Sun Jan 5 01:59:46 2020\"\n+ Date: \"Mon Dec 3 17:39:22 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n-## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n-## MPIR_Init_thread(565)..............: \n-## MPID_Init(224).....................: channel initialization failed\n-## MPIDI_CH3_Init(105)................: \n-## MPID_nem_init(324).................: \n-## MPID_nem_tcp_init(178).............: \n-## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 80 - ./coll/longuser 2\n ---\n Directory: ./coll\n File: longuser\n Num-procs: 2\n- Date: \"Sun Jan 5 01:59:47 2020\"\n+ Date: \"Mon Dec 3 17:39:23 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n-not ok 81 - ./coll/redscat 2\n- ---\n- Directory: ./coll\n- File: redscat\n- Num-procs: 2\n- Date: \"Sun Jan 5 01:59:47 2020\"\n- ...\n-## Test output (expected 'No Errors'):\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n ## Fatal error in PMPI_Init: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n+not ok 81 - ./coll/redscat 2\n+ ---\n+ Directory: ./coll\n+ File: redscat\n+ Num-procs: 2\n+ Date: \"Mon Dec 3 17:39:23 2018\"\n+ ...\n+## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 82 - ./coll/redscat 2\n ---\n Directory: ./coll\n File: redscat\n Num-procs: 2\n- Date: \"Sun Jan 5 01:59:49 2020\"\n+ Date: \"Mon Dec 3 17:39:24 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 83 - ./coll/redscat2 2\n ---\n Directory: ./coll\n File: redscat2\n Num-procs: 2\n- Date: \"Sun Jan 5 01:59:51 2020\"\n+ Date: \"Mon Dec 3 17:39:25 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n-## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n-## MPIR_Init_thread(565)..............: \n-## MPID_Init(224).....................: channel initialization failed\n-## MPIDI_CH3_Init(105)................: \n-## MPID_nem_init(324).................: \n-## MPID_nem_tcp_init(178).............: \n-## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 84 - ./coll/redscat2 2\n ---\n Directory: ./coll\n File: redscat2\n Num-procs: 2\n- Date: \"Sun Jan 5 01:59:52 2020\"\n+ Date: \"Mon Dec 3 17:39:25 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 85 - ./coll/redscat2 2\n ---\n Directory: ./coll\n File: redscat2\n Num-procs: 2\n- Date: \"Sun Jan 5 01:59:54 2020\"\n+ Date: \"Mon Dec 3 17:39:26 2018\"\n ...\n ## Test output (expected 'No Errors'):\n-## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n-## MPIR_Init_thread(565)..............: \n-## MPID_Init(224).....................: channel initialization failed\n-## MPIDI_CH3_Init(105)................: \n-## MPID_nem_init(324).................: \n-## MPID_nem_tcp_init(178).............: \n-## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n not ok 86 - ./coll/redscat3 2\n ---\n Directory: ./coll\n File: redscat3\n Num-procs: 2\n- Date: \"Sun Jan 5 01:59:55 2020\"\n+ Date: \"Mon Dec 3 17:39:27 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 87 - ./coll/redscatinter 2\n ---\n Directory: ./coll\n File: redscatinter\n Num-procs: 2\n- Date: \"Sun Jan 5 01:59:55 2020\"\n+ Date: \"Mon Dec 3 17:39:27 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n-## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n-## MPIR_Init_thread(565)..............: \n-## MPID_Init(224).....................: channel initialization failed\n-## MPIDI_CH3_Init(105)................: \n-## MPID_nem_init(324).................: \n-## MPID_nem_tcp_init(178).............: \n-## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 88 - ./coll/red_scat_block 2\n ---\n Directory: ./coll\n File: red_scat_block\n Num-procs: 2\n- Date: \"Sun Jan 5 01:59:56 2020\"\n+ Date: \"Mon Dec 3 17:39:28 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 89 - ./coll/red_scat_block 2\n ---\n Directory: ./coll\n File: red_scat_block\n Num-procs: 2\n- Date: \"Sun Jan 5 01:59:57 2020\"\n+ Date: \"Mon Dec 3 17:39:28 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n+## Fatal error in PMPI_Init: Other MPI error, error stack:\n+## MPIR_Init_thread(565)..............: \n+## MPID_Init(224).....................: channel initialization failed\n+## MPIDI_CH3_Init(105)................: \n+## MPID_nem_init(324).................: \n+## MPID_nem_tcp_init(178).............: \n+## MPID_nem_tcp_get_business_card(425): \n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 90 - ./coll/red_scat_block 2\n ---\n Directory: ./coll\n File: red_scat_block\n Num-procs: 2\n- Date: \"Sun Jan 5 01:59:58 2020\"\n+ Date: \"Mon Dec 3 17:39:29 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 91 - ./coll/red_scat_block2 2\n ---\n Directory: ./coll\n File: red_scat_block2\n Num-procs: 2\n- Date: \"Sun Jan 5 01:59:58 2020\"\n+ Date: \"Mon Dec 3 17:39:29 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n-## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n-## MPIR_Init_thread(565)..............: \n-## MPID_Init(224).....................: channel initialization failed\n-## MPIDI_CH3_Init(105)................: \n-## MPID_nem_init(324).................: \n-## MPID_nem_tcp_init(178).............: \n-## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 92 - ./coll/red_scat_block2 2\n ---\n Directory: ./coll\n File: red_scat_block2\n Num-procs: 2\n- Date: \"Sun Jan 5 01:59:59 2020\"\n+ Date: \"Mon Dec 3 17:39:30 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n-## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n-## MPIR_Init_thread(565)..............: \n-## MPID_Init(224).....................: channel initialization failed\n-## MPIDI_CH3_Init(105)................: \n-## MPID_nem_init(324).................: \n-## MPID_nem_tcp_init(178).............: \n-## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 93 - ./coll/red_scat_block2 2\n ---\n Directory: ./coll\n File: red_scat_block2\n Num-procs: 2\n- Date: \"Sun Jan 5 02:00:01 2020\"\n+ Date: \"Mon Dec 3 17:39:30 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 94 - ./coll/redscatblk3 2\n ---\n Directory: ./coll\n File: redscatblk3\n Num-procs: 2\n- Date: \"Sun Jan 5 02:00:02 2020\"\n+ Date: \"Mon Dec 3 17:39:31 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 95 - ./coll/redscatblk3 2\n ---\n Directory: ./coll\n File: redscatblk3\n Num-procs: 2\n- Date: \"Sun Jan 5 02:00:03 2020\"\n+ Date: \"Mon Dec 3 17:39:31 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 96 - ./coll/redscatbkinter 2\n ---\n Directory: ./coll\n File: redscatbkinter\n Num-procs: 2\n- Date: \"Sun Jan 5 02:00:04 2020\"\n+ Date: \"Mon Dec 3 17:39:32 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n-## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n-## MPIR_Init_thread(565)..............: \n-## MPID_Init(224).....................: channel initialization failed\n-## MPIDI_CH3_Init(105)................: \n-## MPID_nem_init(324).................: \n-## MPID_nem_tcp_init(178).............: \n-## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 97 - ./coll/redscatbkinter 2\n ---\n Directory: ./coll\n File: redscatbkinter\n Num-procs: 2\n- Date: \"Sun Jan 5 02:00:05 2020\"\n+ Date: \"Mon Dec 3 17:39:33 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 98 - ./coll/scantst 2\n ---\n Directory: ./coll\n File: scantst\n Num-procs: 2\n- Date: \"Sun Jan 5 02:00:06 2020\"\n+ Date: \"Mon Dec 3 17:39:33 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 99 - ./coll/exscan 2\n ---\n Directory: ./coll\n File: exscan\n Num-procs: 2\n- Date: \"Sun Jan 5 02:00:08 2020\"\n+ Date: \"Mon Dec 3 17:39:34 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 100 - ./coll/exscan2 2\n ---\n Directory: ./coll\n File: exscan2\n Num-procs: 2\n- Date: \"Sun Jan 5 02:00:09 2020\"\n+ Date: \"Mon Dec 3 17:39:34 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 101 - ./coll/gather 2\n ---\n Directory: ./coll\n File: gather\n Num-procs: 2\n- Date: \"Sun Jan 5 02:00:10 2020\"\n+ Date: \"Mon Dec 3 17:39:35 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 102 - ./coll/gather2 2\n ---\n Directory: ./coll\n File: gather2\n Num-procs: 2\n- Date: \"Sun Jan 5 02:00:10 2020\"\n+ Date: \"Mon Dec 3 17:39:35 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 103 - ./coll/scattern 2\n ---\n Directory: ./coll\n File: scattern\n Num-procs: 2\n- Date: \"Sun Jan 5 02:00:11 2020\"\n+ Date: \"Mon Dec 3 17:39:36 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 104 - ./coll/scatter2 2\n ---\n Directory: ./coll\n File: scatter2\n Num-procs: 2\n- Date: \"Sun Jan 5 02:00:12 2020\"\n+ Date: \"Mon Dec 3 17:39:36 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 105 - ./coll/scatter3 2\n ---\n Directory: ./coll\n File: scatter3\n Num-procs: 2\n- Date: \"Sun Jan 5 02:00:13 2020\"\n+ Date: \"Mon Dec 3 17:39:37 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n-## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n-## MPIR_Init_thread(565)..............: \n-## MPID_Init(224).....................: channel initialization failed\n-## MPIDI_CH3_Init(105)................: \n-## MPID_nem_init(324).................: \n-## MPID_nem_tcp_init(178).............: \n-## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 106 - ./coll/scatterv 2\n ---\n Directory: ./coll\n File: scatterv\n Num-procs: 2\n- Date: \"Sun Jan 5 02:00:14 2020\"\n+ Date: \"Mon Dec 3 17:39:39 2018\"\n ...\n ## Test output (expected 'No Errors'):\n-## Fatal error in PMPI_Init: Other MPI error, error stack:\n-## MPIR_Init_thread(565)..............: \n-## MPID_Init(224).....................: channel initialization failed\n-## MPIDI_CH3_Init(105)................: \n-## MPID_nem_init(324).................: \n-## MPID_nem_tcp_init(178).............: \n-## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n not ok 107 - ./coll/icbcast 2\n ---\n Directory: ./coll\n File: icbcast\n Num-procs: 2\n- Date: \"Sun Jan 5 02:00:15 2020\"\n+ Date: \"Mon Dec 3 17:39:50 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n-## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n-## MPIR_Init_thread(565)..............: \n-## MPID_Init(224).....................: channel initialization failed\n-## MPIDI_CH3_Init(105)................: \n-## MPID_nem_init(324).................: \n-## MPID_nem_tcp_init(178).............: \n-## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 108 - ./coll/icbcast 2\n ---\n Directory: ./coll\n File: icbcast\n Num-procs: 2\n- Date: \"Sun Jan 5 02:00:16 2020\"\n+ Date: \"Mon Dec 3 17:39:56 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 109 - ./coll/icallreduce 2\n ---\n Directory: ./coll\n File: icallreduce\n Num-procs: 2\n- Date: \"Sun Jan 5 02:00:17 2020\"\n+ Date: \"Mon Dec 3 17:40:02 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n-## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n-## MPIR_Init_thread(565)..............: \n-## MPID_Init(224).....................: channel initialization failed\n-## MPIDI_CH3_Init(105)................: \n-## MPID_nem_init(324).................: \n-## MPID_nem_tcp_init(178).............: \n-## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 110 - ./coll/icallreduce 2\n ---\n Directory: ./coll\n File: icallreduce\n Num-procs: 2\n- Date: \"Sun Jan 5 02:00:18 2020\"\n+ Date: \"Mon Dec 3 17:40:07 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n-## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n-## MPIR_Init_thread(565)..............: \n-## MPID_Init(224).....................: channel initialization failed\n-## MPIDI_CH3_Init(105)................: \n-## MPID_nem_init(324).................: \n-## MPID_nem_tcp_init(178).............: \n-## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 111 - ./coll/icreduce 2\n ---\n Directory: ./coll\n File: icreduce\n Num-procs: 2\n- Date: \"Sun Jan 5 02:00:19 2020\"\n+ Date: \"Mon Dec 3 17:40:07 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 112 - ./coll/icreduce 2\n ---\n Directory: ./coll\n File: icreduce\n Num-procs: 2\n- Date: \"Sun Jan 5 02:00:20 2020\"\n+ Date: \"Mon Dec 3 17:40:08 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 113 - ./coll/icscatter 2\n ---\n Directory: ./coll\n File: icscatter\n Num-procs: 2\n- Date: \"Sun Jan 5 02:00:20 2020\"\n+ Date: \"Mon Dec 3 17:40:08 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 114 - ./coll/icscatter 2\n ---\n Directory: ./coll\n File: icscatter\n Num-procs: 2\n- Date: \"Sun Jan 5 02:00:21 2020\"\n+ Date: \"Mon Dec 3 17:40:15 2018\"\n ...\n ## Test output (expected 'No Errors'):\n-## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n-## MPIR_Init_thread(565)..............: \n-## MPID_Init(224).....................: channel initialization failed\n-## MPIDI_CH3_Init(105)................: \n-## MPID_nem_init(324).................: \n-## MPID_nem_tcp_init(178).............: \n-## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n not ok 115 - ./coll/icgather 2\n ---\n Directory: ./coll\n File: icgather\n Num-procs: 2\n- Date: \"Sun Jan 5 02:00:22 2020\"\n+ Date: \"Mon Dec 3 17:40:16 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 116 - ./coll/icgather 2\n ---\n Directory: ./coll\n File: icgather\n Num-procs: 2\n- Date: \"Sun Jan 5 02:00:23 2020\"\n+ Date: \"Mon Dec 3 17:40:17 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n+## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n+## MPIR_Init_thread(565)..............: \n+## MPID_Init(224).....................: channel initialization failed\n+## MPIDI_CH3_Init(105)................: \n+## MPID_nem_init(324).................: \n+## MPID_nem_tcp_init(178).............: \n+## MPID_nem_tcp_get_business_card(425): \n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 117 - ./coll/icallgather 2\n ---\n Directory: ./coll\n File: icallgather\n Num-procs: 2\n- Date: \"Sun Jan 5 02:00:23 2020\"\n+ Date: \"Mon Dec 3 17:40:18 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 118 - ./coll/icallgather 2\n ---\n Directory: ./coll\n File: icallgather\n Num-procs: 2\n- Date: \"Sun Jan 5 02:00:26 2020\"\n+ Date: \"Mon Dec 3 17:40:18 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 119 - ./coll/icbarrier 2\n ---\n Directory: ./coll\n File: icbarrier\n Num-procs: 2\n- Date: \"Sun Jan 5 02:00:27 2020\"\n+ Date: \"Mon Dec 3 17:40:19 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n+## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n+## MPIR_Init_thread(565)..............: \n+## MPID_Init(224).....................: channel initialization failed\n+## MPIDI_CH3_Init(105)................: \n+## MPID_nem_init(324).................: \n+## MPID_nem_tcp_init(178).............: \n+## MPID_nem_tcp_get_business_card(425): \n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 120 - ./coll/icbarrier 2\n ---\n Directory: ./coll\n File: icbarrier\n Num-procs: 2\n- Date: \"Sun Jan 5 02:00:28 2020\"\n+ Date: \"Mon Dec 3 17:40:19 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 121 - ./coll/icallgatherv 2\n ---\n Directory: ./coll\n File: icallgatherv\n Num-procs: 2\n- Date: \"Sun Jan 5 02:00:28 2020\"\n+ Date: \"Mon Dec 3 17:40:20 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 122 - ./coll/icallgatherv 2\n ---\n Directory: ./coll\n File: icallgatherv\n Num-procs: 2\n- Date: \"Sun Jan 5 02:00:29 2020\"\n+ Date: \"Mon Dec 3 17:40:21 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 123 - ./coll/icgatherv 2\n ---\n Directory: ./coll\n File: icgatherv\n Num-procs: 2\n- Date: \"Sun Jan 5 02:00:30 2020\"\n+ Date: \"Mon Dec 3 17:40:21 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n-not ok 124 - ./coll/icgatherv 2\n- ---\n- Directory: ./coll\n- File: icgatherv\n- Num-procs: 2\n- Date: \"Sun Jan 5 02:00:32 2020\"\n- ...\n-## Test output (expected 'No Errors'):\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n+not ok 124 - ./coll/icgatherv 2\n+ ---\n+ Directory: ./coll\n+ File: icgatherv\n+ Num-procs: 2\n+ Date: \"Mon Dec 3 17:40:22 2018\"\n+ ...\n+## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 125 - ./coll/icscatterv 2\n ---\n Directory: ./coll\n File: icscatterv\n Num-procs: 2\n- Date: \"Sun Jan 5 02:00:33 2020\"\n+ Date: \"Mon Dec 3 17:40:22 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n-## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n-## MPIR_Init_thread(565)..............: \n-## MPID_Init(224).....................: channel initialization failed\n-## MPIDI_CH3_Init(105)................: \n-## MPID_nem_init(324).................: \n-## MPID_nem_tcp_init(178).............: \n-## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 126 - ./coll/icscatterv 2\n ---\n Directory: ./coll\n File: icscatterv\n Num-procs: 2\n- Date: \"Sun Jan 5 02:00:35 2020\"\n+ Date: \"Mon Dec 3 17:40:23 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n-## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n-## MPIR_Init_thread(565)..............: \n-## MPID_Init(224).....................: channel initialization failed\n-## MPIDI_CH3_Init(105)................: \n-## MPID_nem_init(324).................: \n-## MPID_nem_tcp_init(178).............: \n-## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 127 - ./coll/icalltoall 2\n ---\n Directory: ./coll\n File: icalltoall\n Num-procs: 2\n- Date: \"Sun Jan 5 02:00:36 2020\"\n+ Date: \"Mon Dec 3 17:40:23 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 128 - ./coll/icalltoall 2\n ---\n Directory: ./coll\n File: icalltoall\n Num-procs: 2\n- Date: \"Sun Jan 5 02:00:38 2020\"\n+ Date: \"Mon Dec 3 17:40:24 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n-## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n-## MPIR_Init_thread(565)..............: \n-## MPID_Init(224).....................: channel initialization failed\n-## MPIDI_CH3_Init(105)................: \n-## MPID_nem_init(324).................: \n-## MPID_nem_tcp_init(178).............: \n-## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 129 - ./coll/icalltoallv 2\n ---\n Directory: ./coll\n File: icalltoallv\n Num-procs: 2\n- Date: \"Sun Jan 5 02:00:38 2020\"\n+ Date: \"Mon Dec 3 17:40:24 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 130 - ./coll/icalltoallv 2\n ---\n Directory: ./coll\n File: icalltoallv\n Num-procs: 2\n- Date: \"Sun Jan 5 02:00:40 2020\"\n+ Date: \"Mon Dec 3 17:40:25 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n+## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n+## MPIR_Init_thread(565)..............: \n+## MPID_Init(224).....................: channel initialization failed\n+## MPIDI_CH3_Init(105)................: \n+## MPID_nem_init(324).................: \n+## MPID_nem_tcp_init(178).............: \n+## MPID_nem_tcp_get_business_card(425): \n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 131 - ./coll/icalltoallw 2\n ---\n Directory: ./coll\n File: icalltoallw\n Num-procs: 2\n- Date: \"Sun Jan 5 02:00:41 2020\"\n+ Date: \"Mon Dec 3 17:40:25 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 132 - ./coll/icalltoallw 2\n ---\n Directory: ./coll\n File: icalltoallw\n Num-procs: 2\n- Date: \"Sun Jan 5 02:00:42 2020\"\n+ Date: \"Mon Dec 3 17:40:26 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n-## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n-## MPIR_Init_thread(565)..............: \n-## MPID_Init(224).....................: channel initialization failed\n-## MPIDI_CH3_Init(105)................: \n-## MPID_nem_init(324).................: \n-## MPID_nem_tcp_init(178).............: \n-## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 133 - ./coll/opland 2\n ---\n Directory: ./coll\n File: opland\n Num-procs: 2\n- Date: \"Sun Jan 5 02:00:43 2020\"\n+ Date: \"Mon Dec 3 17:40:27 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 134 - ./coll/oplor 2\n ---\n Directory: ./coll\n File: oplor\n Num-procs: 2\n- Date: \"Sun Jan 5 02:00:44 2020\"\n+ Date: \"Mon Dec 3 17:40:27 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n-## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n-## MPIR_Init_thread(565)..............: \n-## MPID_Init(224).....................: channel initialization failed\n-## MPIDI_CH3_Init(105)................: \n-## MPID_nem_init(324).................: \n-## MPID_nem_tcp_init(178).............: \n-## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 135 - ./coll/oplxor 2\n ---\n Directory: ./coll\n File: oplxor\n Num-procs: 2\n- Date: \"Sun Jan 5 02:00:44 2020\"\n+ Date: \"Mon Dec 3 17:40:28 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 136 - ./coll/oplxor 2\n ---\n Directory: ./coll\n File: oplxor\n Num-procs: 2\n- Date: \"Sun Jan 5 02:00:45 2020\"\n+ Date: \"Mon Dec 3 17:40:28 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n-not ok 137 - ./coll/opband 2\n- ---\n- Directory: ./coll\n- File: opband\n- Num-procs: 2\n- Date: \"Sun Jan 5 02:00:47 2020\"\n- ...\n-## Test output (expected 'No Errors'):\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n+not ok 137 - ./coll/opband 2\n+ ---\n+ Directory: ./coll\n+ File: opband\n+ Num-procs: 2\n+ Date: \"Mon Dec 3 17:40:28 2018\"\n+ ...\n+## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 138 - ./coll/opbor 2\n ---\n Directory: ./coll\n File: opbor\n Num-procs: 2\n- Date: \"Sun Jan 5 02:00:48 2020\"\n+ Date: \"Mon Dec 3 17:40:29 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n-## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n-## MPIR_Init_thread(565)..............: \n-## MPID_Init(224).....................: channel initialization failed\n-## MPIDI_CH3_Init(105)................: \n-## MPID_nem_init(324).................: \n-## MPID_nem_tcp_init(178).............: \n-## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 139 - ./coll/opbxor 2\n ---\n Directory: ./coll\n File: opbxor\n Num-procs: 2\n- Date: \"Sun Jan 5 02:00:49 2020\"\n+ Date: \"Mon Dec 3 17:40:29 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n-## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n-## MPIR_Init_thread(565)..............: \n-## MPID_Init(224).....................: channel initialization failed\n-## MPIDI_CH3_Init(105)................: \n-## MPID_nem_init(324).................: \n-## MPID_nem_tcp_init(178).............: \n-## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 140 - ./coll/opbxor 2\n ---\n Directory: ./coll\n File: opbxor\n Num-procs: 2\n- Date: \"Sun Jan 5 02:00:49 2020\"\n+ Date: \"Mon Dec 3 17:40:30 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 141 - ./coll/opprod 2\n ---\n Directory: ./coll\n File: opprod\n Num-procs: 2\n- Date: \"Sun Jan 5 02:00:50 2020\"\n+ Date: \"Mon Dec 3 17:40:30 2018\"\n ...\n ## Test output (expected 'No Errors'):\n-## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n-## MPIR_Init_thread(565)..............: \n-## MPID_Init(224).....................: channel initialization failed\n-## MPIDI_CH3_Init(105)................: \n-## MPID_nem_init(324).................: \n-## MPID_nem_tcp_init(178).............: \n-## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n-## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n-## MPIR_Init_thread(565)..............: \n-## MPID_Init(224).....................: channel initialization failed\n-## MPIDI_CH3_Init(105)................: \n-## MPID_nem_init(324).................: \n-## MPID_nem_tcp_init(178).............: \n-## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n not ok 142 - ./coll/opprod 2\n ---\n Directory: ./coll\n File: opprod\n Num-procs: 2\n- Date: \"Sun Jan 5 02:00:51 2020\"\n+ Date: \"Mon Dec 3 17:40:31 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n+## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n+## MPIR_Init_thread(565)..............: \n+## MPID_Init(224).....................: channel initialization failed\n+## MPIDI_CH3_Init(105)................: \n+## MPID_nem_init(324).................: \n+## MPID_nem_tcp_init(178).............: \n+## MPID_nem_tcp_get_business_card(425): \n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 143 - ./coll/opsum 2\n ---\n Directory: ./coll\n File: opsum\n Num-procs: 2\n- Date: \"Sun Jan 5 02:00:52 2020\"\n+ Date: \"Mon Dec 3 17:40:31 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 144 - ./coll/opmin 2\n ---\n Directory: ./coll\n File: opmin\n Num-procs: 2\n- Date: \"Sun Jan 5 02:00:54 2020\"\n+ Date: \"Mon Dec 3 17:40:43 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 145 - ./coll/opminloc 2\n ---\n Directory: ./coll\n File: opminloc\n Num-procs: 2\n- Date: \"Sun Jan 5 02:00:55 2020\"\n+ Date: \"Mon Dec 3 17:40:50 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n-## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n-## MPIR_Init_thread(565)..............: \n-## MPID_Init(224).....................: channel initialization failed\n-## MPIDI_CH3_Init(105)................: \n-## MPID_nem_init(324).................: \n-## MPID_nem_tcp_init(178).............: \n-## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 146 - ./coll/opmax 2\n ---\n Directory: ./coll\n File: opmax\n Num-procs: 2\n- Date: \"Sun Jan 5 02:00:56 2020\"\n+ Date: \"Mon Dec 3 17:40:58 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n-## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n-## MPIR_Init_thread(565)..............: \n-## MPID_Init(224).....................: channel initialization failed\n-## MPIDI_CH3_Init(105)................: \n-## MPID_nem_init(324).................: \n-## MPID_nem_tcp_init(178).............: \n-## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 147 - ./coll/opmaxloc 2\n ---\n Directory: ./coll\n File: opmaxloc\n Num-procs: 2\n- Date: \"Sun Jan 5 02:00:57 2020\"\n+ Date: \"Mon Dec 3 17:40:59 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 148 - ./coll/uoplong 2\n ---\n Directory: ./coll\n File: uoplong\n Num-procs: 2\n- Date: \"Sun Jan 5 02:00:58 2020\"\n+ Date: \"Mon Dec 3 17:41:02 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n-## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n-## MPIR_Init_thread(565)..............: \n-## MPID_Init(224).....................: channel initialization failed\n-## MPIDI_CH3_Init(105)................: \n-## MPID_nem_init(324).................: \n-## MPID_nem_tcp_init(178).............: \n-## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 149 - ./coll/uoplong 2\n ---\n Directory: ./coll\n File: uoplong\n Num-procs: 2\n- Date: \"Sun Jan 5 02:00:59 2020\"\n+ Date: \"Mon Dec 3 17:41:04 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 150 - ./coll/uoplong 2\n ---\n Directory: ./coll\n File: uoplong\n Num-procs: 2\n- Date: \"Sun Jan 5 02:01:00 2020\"\n+ Date: \"Mon Dec 3 17:41:13 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 151 - ./coll/nonblocking 2\n ---\n Directory: ./coll\n File: nonblocking\n Num-procs: 2\n- Date: \"Sun Jan 5 02:01:01 2020\"\n+ Date: \"Mon Dec 3 17:41:17 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 152 - ./coll/nonblocking 2\n ---\n Directory: ./coll\n File: nonblocking\n Num-procs: 2\n- Date: \"Sun Jan 5 02:01:02 2020\"\n+ Date: \"Mon Dec 3 17:41:18 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n-## Fatal error in PMPI_Init: Other MPI error, error stack:\n-## MPIR_Init_thread(565)..............: \n-## MPID_Init(224).....................: channel initialization failed\n-## MPIDI_CH3_Init(105)................: \n-## MPID_nem_init(324).................: \n-## MPID_nem_tcp_init(178).............: \n-## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 153 - ./coll/nonblocking 2\n ---\n Directory: ./coll\n File: nonblocking\n Num-procs: 2\n- Date: \"Sun Jan 5 02:01:03 2020\"\n+ Date: \"Mon Dec 3 17:41:18 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 154 - ./coll/nonblocking2 1\n ---\n Directory: ./coll\n File: nonblocking2\n Num-procs: 1\n- Date: \"Sun Jan 5 02:01:04 2020\"\n+ Date: \"Mon Dec 3 17:41:19 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 155 - ./coll/nonblocking2 2\n ---\n Directory: ./coll\n File: nonblocking2\n Num-procs: 2\n- Date: \"Sun Jan 5 02:01:05 2020\"\n+ Date: \"Mon Dec 3 17:41:19 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 156 - ./coll/nonblocking2 2\n ---\n Directory: ./coll\n File: nonblocking2\n Num-procs: 2\n- Date: \"Sun Jan 5 02:01:07 2020\"\n+ Date: \"Mon Dec 3 17:41:20 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n+## Fatal error in PMPI_Init: Other MPI error, error stack:\n+## MPIR_Init_thread(565)..............: \n+## MPID_Init(224).....................: channel initialization failed\n+## MPIDI_CH3_Init(105)................: \n+## MPID_nem_init(324).................: \n+## MPID_nem_tcp_init(178).............: \n+## MPID_nem_tcp_get_business_card(425): \n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 157 - ./coll/nonblocking2 2\n ---\n Directory: ./coll\n File: nonblocking2\n Num-procs: 2\n- Date: \"Sun Jan 5 02:01:08 2020\"\n+ Date: \"Mon Dec 3 17:41:20 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 158 - ./coll/nonblocking3 1\n ---\n Directory: ./coll\n File: nonblocking3\n Num-procs: 1\n- Date: \"Sun Jan 5 02:01:08 2020\"\n+ Date: \"Mon Dec 3 17:41:21 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 159 - ./coll/nonblocking3 2\n ---\n Directory: ./coll\n File: nonblocking3\n Num-procs: 2\n- Date: \"Sun Jan 5 02:01:10 2020\"\n+ Date: \"Mon Dec 3 17:41:21 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n-## Fatal error in PMPI_Init: Other MPI error, error stack:\n-## MPIR_Init_thread(565)..............: \n-## MPID_Init(224).....................: channel initialization failed\n-## MPIDI_CH3_Init(105)................: \n-## MPID_nem_init(324).................: \n-## MPID_nem_tcp_init(178).............: \n-## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 160 - ./coll/nonblocking3 2\n ---\n Directory: ./coll\n File: nonblocking3\n Num-procs: 2\n- Date: \"Sun Jan 5 02:01:11 2020\"\n+ Date: \"Mon Dec 3 17:41:22 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n+## Fatal error in PMPI_Init: Other MPI error, error stack:\n+## MPIR_Init_thread(565)..............: \n+## MPID_Init(224).....................: channel initialization failed\n+## MPIDI_CH3_Init(105)................: \n+## MPID_nem_init(324).................: \n+## MPID_nem_tcp_init(178).............: \n+## MPID_nem_tcp_get_business_card(425): \n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 161 - ./coll/nonblocking3 2\n ---\n Directory: ./coll\n File: nonblocking3\n Num-procs: 2\n- Date: \"Sun Jan 5 02:01:13 2020\"\n+ Date: \"Mon Dec 3 17:41:23 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n+## Fatal error in PMPI_Init: Other MPI error, error stack:\n+## MPIR_Init_thread(565)..............: \n+## MPID_Init(224).....................: channel initialization failed\n+## MPIDI_CH3_Init(105)................: \n+## MPID_nem_init(324).................: \n+## MPID_nem_tcp_init(178).............: \n+## MPID_nem_tcp_get_business_card(425): \n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 162 - ./coll/iallred 2\n ---\n Directory: ./coll\n File: iallred\n Num-procs: 2\n- Date: \"Sun Jan 5 02:01:13 2020\"\n+ Date: \"Mon Dec 3 17:41:23 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 163 - ./coll/ibarrier 2\n ---\n Directory: ./coll\n File: ibarrier\n Num-procs: 2\n- Date: \"Sun Jan 5 02:01:14 2020\"\n+ Date: \"Mon Dec 3 17:41:24 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 164 - ./coll/nballtoall1 2\n ---\n Directory: ./coll\n File: nballtoall1\n Num-procs: 2\n- Date: \"Sun Jan 5 02:01:16 2020\"\n+ Date: \"Mon Dec 3 17:41:25 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 165 - ./coll/nbcoll2 2\n ---\n Directory: ./coll\n File: nbcoll2\n Num-procs: 2\n- Date: \"Sun Jan 5 02:01:17 2020\"\n+ Date: \"Mon Dec 3 17:41:25 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n-## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n-## MPIR_Init_thread(565)..............: \n-## MPID_Init(224).....................: channel initialization failed\n-## MPIDI_CH3_Init(105)................: \n-## MPID_nem_init(324).................: \n-## MPID_nem_tcp_init(178).............: \n-## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 166 - ./coll/nbredscat 2\n ---\n Directory: ./coll\n File: nbredscat\n Num-procs: 2\n- Date: \"Sun Jan 5 02:01:18 2020\"\n+ Date: \"Mon Dec 3 17:41:26 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n-## Fatal error in PMPI_Init: Other MPI error, error stack:\n-## MPIR_Init_thread(565)..............: \n-## MPID_Init(224).....................: channel initialization failed\n-## MPIDI_CH3_Init(105)................: \n-## MPID_nem_init(324).................: \n-## MPID_nem_tcp_init(178).............: \n-## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 167 - ./coll/nbredscat 2\n ---\n Directory: ./coll\n File: nbredscat\n Num-procs: 2\n- Date: \"Sun Jan 5 02:01:18 2020\"\n+ Date: \"Mon Dec 3 17:41:27 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 168 - ./coll/nbredscat3 2\n ---\n Directory: ./coll\n File: nbredscat3\n Num-procs: 2\n- Date: \"Sun Jan 5 02:01:20 2020\"\n+ Date: \"Mon Dec 3 17:41:27 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 169 - ./coll/nbredscatinter 2\n ---\n Directory: ./coll\n File: nbredscatinter\n Num-procs: 2\n- Date: \"Sun Jan 5 02:01:21 2020\"\n+ Date: \"Mon Dec 3 17:41:28 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 170 - ./coll/nbicbcast 2\n ---\n Directory: ./coll\n File: nbicbcast\n Num-procs: 2\n- Date: \"Sun Jan 5 02:01:22 2020\"\n+ Date: \"Mon Dec 3 17:41:29 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 171 - ./coll/nbicallreduce 2\n ---\n Directory: ./coll\n File: nbicallreduce\n Num-procs: 2\n- Date: \"Sun Jan 5 02:01:23 2020\"\n+ Date: \"Mon Dec 3 17:41:29 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 172 - ./coll/nbicreduce 2\n ---\n Directory: ./coll\n File: nbicreduce\n Num-procs: 2\n- Date: \"Sun Jan 5 02:01:24 2020\"\n+ Date: \"Mon Dec 3 17:41:30 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 173 - ./coll/nbicscatter 2\n ---\n Directory: ./coll\n File: nbicscatter\n Num-procs: 2\n- Date: \"Sun Jan 5 02:01:24 2020\"\n+ Date: \"Mon Dec 3 17:41:30 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n+## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n+## MPIR_Init_thread(565)..............: \n+## MPID_Init(224).....................: channel initialization failed\n+## MPIDI_CH3_Init(105)................: \n+## MPID_nem_init(324).................: \n+## MPID_nem_tcp_init(178).............: \n+## MPID_nem_tcp_get_business_card(425): \n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 174 - ./coll/nbicgather 2\n ---\n Directory: ./coll\n File: nbicgather\n Num-procs: 2\n- Date: \"Sun Jan 5 02:01:25 2020\"\n+ Date: \"Mon Dec 3 17:41:31 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 175 - ./coll/nbicallgather 2\n ---\n Directory: ./coll\n File: nbicallgather\n Num-procs: 2\n- Date: \"Sun Jan 5 02:01:26 2020\"\n+ Date: \"Mon Dec 3 17:41:32 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 176 - ./coll/nbicbarrier 2\n ---\n Directory: ./coll\n File: nbicbarrier\n Num-procs: 2\n- Date: \"Sun Jan 5 02:01:29 2020\"\n+ Date: \"Mon Dec 3 17:41:32 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 177 - ./coll/nbicallgatherv 2\n ---\n Directory: ./coll\n File: nbicallgatherv\n Num-procs: 2\n- Date: \"Sun Jan 5 02:01:31 2020\"\n+ Date: \"Mon Dec 3 17:41:33 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 178 - ./coll/nbicgatherv 2\n ---\n Directory: ./coll\n File: nbicgatherv\n Num-procs: 2\n- Date: \"Sun Jan 5 02:01:34 2020\"\n+ Date: \"Mon Dec 3 17:41:34 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n+## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n+## MPIR_Init_thread(565)..............: \n+## MPID_Init(224).....................: channel initialization failed\n+## MPIDI_CH3_Init(105)................: \n+## MPID_nem_init(324).................: \n+## MPID_nem_tcp_init(178).............: \n+## MPID_nem_tcp_get_business_card(425): \n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 179 - ./coll/nbicscatterv 2\n ---\n Directory: ./coll\n File: nbicscatterv\n Num-procs: 2\n- Date: \"Sun Jan 5 02:01:35 2020\"\n+ Date: \"Mon Dec 3 17:41:34 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 180 - ./coll/nbicalltoall 2\n ---\n Directory: ./coll\n File: nbicalltoall\n Num-procs: 2\n- Date: \"Sun Jan 5 02:01:36 2020\"\n+ Date: \"Mon Dec 3 17:41:35 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 181 - ./coll/nbicalltoallv 2\n ---\n Directory: ./coll\n File: nbicalltoallv\n Num-procs: 2\n- Date: \"Sun Jan 5 02:01:37 2020\"\n+ Date: \"Mon Dec 3 17:41:35 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n-## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n-## MPIR_Init_thread(565)..............: \n-## MPID_Init(224).....................: channel initialization failed\n-## MPIDI_CH3_Init(105)................: \n-## MPID_nem_init(324).................: \n-## MPID_nem_tcp_init(178).............: \n-## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 182 - ./coll/nbicalltoallw 2\n ---\n Directory: ./coll\n File: nbicalltoallw\n Num-procs: 2\n- Date: \"Sun Jan 5 02:01:38 2020\"\n+ Date: \"Mon Dec 3 17:41:36 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 183 - ./comm/dup 2\n ---\n Directory: ./comm\n File: dup\n Num-procs: 2\n- Date: \"Sun Jan 5 02:01:39 2020\"\n+ Date: \"Mon Dec 3 17:41:36 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n-## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n-## MPIR_Init_thread(565)..............: \n-## MPID_Init(224).....................: channel initialization failed\n-## MPIDI_CH3_Init(105)................: \n-## MPID_nem_init(324).................: \n-## MPID_nem_tcp_init(178).............: \n-## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 184 - ./comm/dupic 2\n ---\n Directory: ./comm\n File: dupic\n Num-procs: 2\n- Date: \"Sun Jan 5 02:01:40 2020\"\n+ Date: \"Mon Dec 3 17:41:37 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 185 - ./comm/commcreate1 2\n ---\n Directory: ./comm\n File: commcreate1\n Num-procs: 2\n- Date: \"Sun Jan 5 02:01:41 2020\"\n+ Date: \"Mon Dec 3 17:41:37 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n-## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n-## MPIR_Init_thread(565)..............: \n-## MPID_Init(224).....................: channel initialization failed\n-## MPIDI_CH3_Init(105)................: \n-## MPID_nem_init(324).................: \n-## MPID_nem_tcp_init(178).............: \n-## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 186 - ./comm/commname 2\n ---\n Directory: ./comm\n File: commname\n Num-procs: 2\n- Date: \"Sun Jan 5 02:01:42 2020\"\n+ Date: \"Mon Dec 3 17:41:37 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 187 - ./comm/ic1 2\n ---\n Directory: ./comm\n File: ic1\n Num-procs: 2\n- Date: \"Sun Jan 5 02:01:44 2020\"\n+ Date: \"Mon Dec 3 17:41:38 2018\"\n ...\n ## Test output (expected 'No Errors'):\n-## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n-## MPIR_Init_thread(565)..............: \n-## MPID_Init(224).....................: channel initialization failed\n-## MPIDI_CH3_Init(105)................: \n-## MPID_nem_init(324).................: \n-## MPID_nem_tcp_init(178).............: \n-## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n-## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n-## MPIR_Init_thread(565)..............: \n-## MPID_Init(224).....................: channel initialization failed\n-## MPIDI_CH3_Init(105)................: \n-## MPID_nem_init(324).................: \n-## MPID_nem_tcp_init(178).............: \n-## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n not ok 188 - ./comm/ic2 2\n ---\n Directory: ./comm\n File: ic2\n Num-procs: 2\n- Date: \"Sun Jan 5 02:01:45 2020\"\n+ Date: \"Mon Dec 3 17:41:38 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 189 - ./comm/icgroup 2\n ---\n Directory: ./comm\n File: icgroup\n Num-procs: 2\n- Date: \"Sun Jan 5 02:01:46 2020\"\n+ Date: \"Mon Dec 3 17:41:39 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 190 - ./comm/icm 2\n ---\n Directory: ./comm\n File: icm\n Num-procs: 2\n- Date: \"Sun Jan 5 02:01:46 2020\"\n+ Date: \"Mon Dec 3 17:41:39 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 191 - ./comm/icsplit 2\n ---\n Directory: ./comm\n File: icsplit\n Num-procs: 2\n- Date: \"Sun Jan 5 02:01:47 2020\"\n+ Date: \"Mon Dec 3 17:41:40 2018\"\n ...\n ## Test output (expected 'No Errors'):\n+## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n+## MPIR_Init_thread(565)..............: \n+## MPID_Init(224).....................: channel initialization failed\n+## MPIDI_CH3_Init(105)................: \n+## MPID_nem_init(324).................: \n+## MPID_nem_tcp_init(178).............: \n+## MPID_nem_tcp_get_business_card(425): \n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 192 - ./comm/iccreate 2\n ---\n Directory: ./comm\n File: iccreate\n Num-procs: 2\n- Date: \"Sun Jan 5 02:01:48 2020\"\n+ Date: \"Mon Dec 3 17:41:40 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n+## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n+## MPIR_Init_thread(565)..............: \n+## MPID_Init(224).....................: channel initialization failed\n+## MPIDI_CH3_Init(105)................: \n+## MPID_nem_init(324).................: \n+## MPID_nem_tcp_init(178).............: \n+## MPID_nem_tcp_get_business_card(425): \n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 193 - ./comm/ctxalloc 2\n ---\n Directory: ./comm\n File: ctxalloc\n Num-procs: 2\n- Date: \"Sun Jan 5 02:01:49 2020\"\n+ Date: \"Mon Dec 3 17:41:41 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 194 - ./comm/ctxsplit 2\n ---\n Directory: ./comm\n File: ctxsplit\n Num-procs: 2\n- Date: \"Sun Jan 5 02:01:50 2020\"\n+ Date: \"Mon Dec 3 17:41:41 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n+## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n+## MPIR_Init_thread(565)..............: \n+## MPID_Init(224).....................: channel initialization failed\n+## MPIDI_CH3_Init(105)................: \n+## MPID_nem_init(324).................: \n+## MPID_nem_tcp_init(178).............: \n+## MPID_nem_tcp_get_business_card(425): \n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 195 - ./comm/cmfree 2\n ---\n Directory: ./comm\n File: cmfree\n Num-procs: 2\n- Date: \"Sun Jan 5 02:01:51 2020\"\n+ Date: \"Mon Dec 3 17:41:41 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 196 - ./comm/cmsplit 2\n ---\n Directory: ./comm\n File: cmsplit\n Num-procs: 2\n- Date: \"Sun Jan 5 02:01:52 2020\"\n+ Date: \"Mon Dec 3 17:41:42 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 197 - ./comm/cmsplit2 2\n ---\n Directory: ./comm\n File: cmsplit2\n Num-procs: 2\n- Date: \"Sun Jan 5 02:01:52 2020\"\n+ Date: \"Mon Dec 3 17:41:42 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n-## Fatal error in PMPI_Init: Other MPI error, error stack:\n-## MPIR_Init_thread(565)..............: \n-## MPID_Init(224).....................: channel initialization failed\n-## MPIDI_CH3_Init(105)................: \n-## MPID_nem_init(324).................: \n-## MPID_nem_tcp_init(178).............: \n-## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 198 - ./comm/probe-intercomm 2\n ---\n Directory: ./comm\n File: probe-intercomm\n Num-procs: 2\n- Date: \"Sun Jan 5 02:01:53 2020\"\n+ Date: \"Mon Dec 3 17:41:43 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 199 - ./comm/cmsplit_type 2\n ---\n Directory: ./comm\n File: cmsplit_type\n Num-procs: 2\n- Date: \"Sun Jan 5 02:01:54 2020\"\n+ Date: \"Mon Dec 3 17:41:43 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 200 - ./comm/comm_create_group 2\n ---\n Directory: ./comm\n File: comm_create_group\n Num-procs: 2\n- Date: \"Sun Jan 5 02:01:55 2020\"\n+ Date: \"Mon Dec 3 17:41:44 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 201 - ./comm/comm_create_group 2\n ---\n Directory: ./comm\n File: comm_create_group\n Num-procs: 2\n- Date: \"Sun Jan 5 02:01:56 2020\"\n+ Date: \"Mon Dec 3 17:41:44 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n-## Fatal error in PMPI_Init: Other MPI error, error stack:\n-## MPIR_Init_thread(565)..............: \n-## MPID_Init(224).....................: channel initialization failed\n-## MPIDI_CH3_Init(105)................: \n-## MPID_nem_init(324).................: \n-## MPID_nem_tcp_init(178).............: \n-## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 202 - ./comm/comm_group_half 2\n ---\n Directory: ./comm\n File: comm_group_half\n Num-procs: 2\n- Date: \"Sun Jan 5 02:01:57 2020\"\n+ Date: \"Mon Dec 3 17:41:44 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 203 - ./comm/comm_group_half 2\n ---\n Directory: ./comm\n File: comm_group_half\n Num-procs: 2\n- Date: \"Sun Jan 5 02:01:58 2020\"\n+ Date: \"Mon Dec 3 17:41:45 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n-## Fatal error in PMPI_Init: Other MPI error, error stack:\n-## MPIR_Init_thread(565)..............: \n-## MPID_Init(224).....................: channel initialization failed\n-## MPIDI_CH3_Init(105)................: \n-## MPID_nem_init(324).................: \n-## MPID_nem_tcp_init(178).............: \n-## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 204 - ./comm/comm_group_half 2\n ---\n Directory: ./comm\n File: comm_group_half\n Num-procs: 2\n- Date: \"Sun Jan 5 02:01:59 2020\"\n+ Date: \"Mon Dec 3 17:41:45 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n-not ok 205 - ./comm/comm_group_rand 2\n- ---\n- Directory: ./comm\n- File: comm_group_rand\n- Num-procs: 2\n- Date: \"Sun Jan 5 02:02:00 2020\"\n- ...\n-## Test output (expected 'No Errors'):\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n ## Fatal error in PMPI_Init: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n+not ok 205 - ./comm/comm_group_rand 2\n+ ---\n+ Directory: ./comm\n+ File: comm_group_rand\n+ Num-procs: 2\n+ Date: \"Mon Dec 3 17:41:46 2018\"\n+ ...\n+## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 206 - ./comm/comm_group_rand 2\n ---\n Directory: ./comm\n File: comm_group_rand\n Num-procs: 2\n- Date: \"Sun Jan 5 02:02:01 2020\"\n+ Date: \"Mon Dec 3 17:41:46 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n ## Fatal error in PMPI_Init: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 207 - ./comm/comm_group_rand 2\n ---\n Directory: ./comm\n File: comm_group_rand\n Num-procs: 2\n- Date: \"Sun Jan 5 02:02:02 2020\"\n+ Date: \"Mon Dec 3 17:41:47 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 208 - ./comm/comm_idup 2\n ---\n Directory: ./comm\n File: comm_idup\n Num-procs: 2\n- Date: \"Sun Jan 5 02:02:03 2020\"\n+ Date: \"Mon Dec 3 17:41:47 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 209 - ./comm/comm_idup 2\n ---\n Directory: ./comm\n File: comm_idup\n Num-procs: 2\n- Date: \"Sun Jan 5 02:02:04 2020\"\n+ Date: \"Mon Dec 3 17:41:48 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n-## Fatal error in PMPI_Init: Other MPI error, error stack:\n-## MPIR_Init_thread(565)..............: \n-## MPID_Init(224).....................: channel initialization failed\n-## MPIDI_CH3_Init(105)................: \n-## MPID_nem_init(324).................: \n-## MPID_nem_tcp_init(178).............: \n-## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 210 - ./comm/comm_idup 2\n ---\n Directory: ./comm\n File: comm_idup\n Num-procs: 2\n- Date: \"Sun Jan 5 02:02:04 2020\"\n+ Date: \"Mon Dec 3 17:41:48 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n+not ok 211 - ./comm/comm_idup_mul 2\n+ ---\n+ Directory: ./comm\n+ File: comm_idup_mul\n+ Num-procs: 2\n+ Date: \"Mon Dec 3 17:41:49 2018\"\n+ ...\n+## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n-not ok 211 - ./comm/comm_idup_mul 2\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n+not ok 212 - ./comm/comm_idup_overlap 2\n ---\n Directory: ./comm\n- File: comm_idup_mul\n+ File: comm_idup_overlap\n Num-procs: 2\n- Date: \"Sun Jan 5 02:02:05 2020\"\n+ Date: \"Mon Dec 3 17:41:49 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n ## Fatal error in PMPI_Init: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n-not ok 212 - ./comm/comm_idup_overlap 2\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n+not ok 213 - ./comm/comm_idup_iallreduce 2\n ---\n Directory: ./comm\n- File: comm_idup_overlap\n+ File: comm_idup_iallreduce\n Num-procs: 2\n- Date: \"Sun Jan 5 02:02:05 2020\"\n+ Date: \"Mon Dec 3 17:41:49 2018\"\n ...\n ## Test output (expected 'No Errors'):\n-## Fatal error in PMPI_Init: Other MPI error, error stack:\n+## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n-not ok 213 - ./comm/comm_idup_iallreduce 2\n- ---\n- Directory: ./comm\n- File: comm_idup_iallreduce\n- Num-procs: 2\n- Date: \"Sun Jan 5 02:02:05 2020\"\n- ...\n-## Test output (expected 'No Errors'):\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 214 - ./comm/comm_idup_nb 2\n ---\n Directory: ./comm\n File: comm_idup_nb\n Num-procs: 2\n- Date: \"Sun Jan 5 02:02:06 2020\"\n+ Date: \"Mon Dec 3 17:41:50 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 215 - ./comm/comm_idup_isend 2\n ---\n Directory: ./comm\n File: comm_idup_isend\n Num-procs: 2\n- Date: \"Sun Jan 5 02:02:06 2020\"\n+ Date: \"Mon Dec 3 17:41:50 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 216 - ./comm/comm_idup_comm 2\n ---\n Directory: ./comm\n File: comm_idup_comm\n Num-procs: 2\n- Date: \"Sun Jan 5 02:02:06 2020\"\n+ Date: \"Mon Dec 3 17:41:51 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n-## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n-## MPIR_Init_thread(565)..............: \n-## MPID_Init(224).....................: channel initialization failed\n-## MPIDI_CH3_Init(105)................: \n-## MPID_nem_init(324).................: \n-## MPID_nem_tcp_init(178).............: \n-## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 217 - ./comm/comm_idup_comm2 2\n ---\n Directory: ./comm\n File: comm_idup_comm2\n Num-procs: 2\n- Date: \"Sun Jan 5 02:02:07 2020\"\n+ Date: \"Mon Dec 3 17:41:51 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n-## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n-## MPIR_Init_thread(565)..............: \n-## MPID_Init(224).....................: channel initialization failed\n-## MPIDI_CH3_Init(105)................: \n-## MPID_nem_init(324).................: \n-## MPID_nem_tcp_init(178).............: \n-## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 218 - ./comm/dup_with_info 2\n ---\n Directory: ./comm\n File: dup_with_info\n Num-procs: 2\n- Date: \"Sun Jan 5 02:02:07 2020\"\n+ Date: \"Mon Dec 3 17:41:52 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n-## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n-## MPIR_Init_thread(565)..............: \n-## MPID_Init(224).....................: channel initialization failed\n-## MPIDI_CH3_Init(105)................: \n-## MPID_nem_init(324).................: \n-## MPID_nem_tcp_init(178).............: \n-## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 219 - ./comm/dup_with_info 2\n ---\n Directory: ./comm\n File: dup_with_info\n Num-procs: 2\n- Date: \"Sun Jan 5 02:02:07 2020\"\n+ Date: \"Mon Dec 3 17:41:52 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 220 - ./comm/dup_with_info 2\n ---\n Directory: ./comm\n File: dup_with_info\n Num-procs: 2\n- Date: \"Sun Jan 5 02:02:07 2020\"\n+ Date: \"Mon Dec 3 17:41:53 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n-## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n-## MPIR_Init_thread(565)..............: \n-## MPID_Init(224).....................: channel initialization failed\n-## MPIDI_CH3_Init(105)................: \n-## MPID_nem_init(324).................: \n-## MPID_nem_tcp_init(178).............: \n-## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 221 - ./comm/comm_info 2\n ---\n Directory: ./comm\n File: comm_info\n Num-procs: 2\n- Date: \"Sun Jan 5 02:02:08 2020\"\n+ Date: \"Mon Dec 3 17:41:53 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n-## Fatal error in PMPI_Init: Other MPI error, error stack:\n-## MPIR_Init_thread(565)..............: \n-## MPID_Init(224).....................: channel initialization failed\n-## MPIDI_CH3_Init(105)................: \n-## MPID_nem_init(324).................: \n-## MPID_nem_tcp_init(178).............: \n-## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 222 - ./comm/comm_create_group_idup 2\n ---\n Directory: ./comm\n File: comm_create_group_idup\n Num-procs: 2\n- Date: \"Sun Jan 5 02:02:08 2020\"\n+ Date: \"Mon Dec 3 17:41:53 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 223 - ./datatype/contents 1\n ---\n Directory: ./datatype\n File: contents\n Num-procs: 1\n- Date: \"Sun Jan 5 02:02:09 2020\"\n+ Date: \"Mon Dec 3 17:41:54 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 224 - ./datatype/gaddress 1\n ---\n Directory: ./datatype\n File: gaddress\n Num-procs: 1\n- Date: \"Sun Jan 5 02:02:09 2020\"\n+ Date: \"Mon Dec 3 17:41:54 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 225 - ./datatype/lbub 1\n ---\n Directory: ./datatype\n File: lbub\n Num-procs: 1\n- Date: \"Sun Jan 5 02:02:09 2020\"\n+ Date: \"Mon Dec 3 17:41:55 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 226 - ./datatype/localpack 1\n ---\n Directory: ./datatype\n File: localpack\n Num-procs: 1\n- Date: \"Sun Jan 5 02:02:09 2020\"\n+ Date: \"Mon Dec 3 17:41:55 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 227 - ./datatype/simple-pack 1\n ---\n Directory: ./datatype\n File: simple-pack\n Num-procs: 1\n- Date: \"Sun Jan 5 02:02:10 2020\"\n+ Date: \"Mon Dec 3 17:41:56 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 228 - ./datatype/simple-pack-external 1\n ---\n Directory: ./datatype\n File: simple-pack-external\n Num-procs: 1\n- Date: \"Sun Jan 5 02:02:10 2020\"\n+ Date: \"Mon Dec 3 17:41:56 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 229 - ./datatype/transpose-pack 1\n ---\n Directory: ./datatype\n File: transpose-pack\n Num-procs: 1\n- Date: \"Sun Jan 5 02:02:10 2020\"\n+ Date: \"Mon Dec 3 17:41:57 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 230 - ./datatype/slice-pack 1\n ---\n Directory: ./datatype\n File: slice-pack\n Num-procs: 1\n- Date: \"Sun Jan 5 02:02:10 2020\"\n+ Date: \"Mon Dec 3 17:41:57 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 231 - ./datatype/struct-pack 1\n ---\n Directory: ./datatype\n File: struct-pack\n Num-procs: 1\n- Date: \"Sun Jan 5 02:02:11 2020\"\n+ Date: \"Mon Dec 3 17:41:57 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 232 - ./datatype/structpack2 1\n ---\n Directory: ./datatype\n File: structpack2\n Num-procs: 1\n- Date: \"Sun Jan 5 02:02:11 2020\"\n+ Date: \"Mon Dec 3 17:41:58 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 233 - ./datatype/typecommit 1\n ---\n Directory: ./datatype\n File: typecommit\n Num-procs: 1\n- Date: \"Sun Jan 5 02:02:11 2020\"\n+ Date: \"Mon Dec 3 17:41:58 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 234 - ./datatype/typename 1\n ---\n Directory: ./datatype\n File: typename\n Num-procs: 1\n- Date: \"Sun Jan 5 02:02:12 2020\"\n+ Date: \"Mon Dec 3 17:41:59 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 235 - ./datatype/typefree 1\n ---\n Directory: ./datatype\n File: typefree\n Num-procs: 1\n- Date: \"Sun Jan 5 02:02:12 2020\"\n+ Date: \"Mon Dec 3 17:41:59 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 236 - ./datatype/zeroparms 1\n ---\n Directory: ./datatype\n File: zeroparms\n Num-procs: 1\n- Date: \"Sun Jan 5 02:02:13 2020\"\n+ Date: \"Mon Dec 3 17:41:59 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 237 - ./datatype/getpartelm 2\n ---\n Directory: ./datatype\n File: getpartelm\n Num-procs: 2\n- Date: \"Sun Jan 5 02:02:13 2020\"\n+ Date: \"Mon Dec 3 17:42:00 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n-## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n-## MPIR_Init_thread(565)..............: \n-## MPID_Init(224).....................: channel initialization failed\n-## MPIDI_CH3_Init(105)................: \n-## MPID_nem_init(324).................: \n-## MPID_nem_tcp_init(178).............: \n-## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 238 - ./datatype/tresized 2\n ---\n Directory: ./datatype\n File: tresized\n Num-procs: 2\n- Date: \"Sun Jan 5 02:02:13 2020\"\n+ Date: \"Mon Dec 3 17:42:00 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 239 - ./datatype/tresized2 2\n ---\n Directory: ./datatype\n File: tresized2\n Num-procs: 2\n- Date: \"Sun Jan 5 02:02:14 2020\"\n+ Date: \"Mon Dec 3 17:42:01 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n-## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n-## MPIR_Init_thread(565)..............: \n-## MPID_Init(224).....................: channel initialization failed\n-## MPIDI_CH3_Init(105)................: \n-## MPID_nem_init(324).................: \n-## MPID_nem_tcp_init(178).............: \n-## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 240 - ./datatype/sendrecvt2 2\n ---\n Directory: ./datatype\n File: sendrecvt2\n Num-procs: 2\n- Date: \"Sun Jan 5 02:02:14 2020\"\n+ Date: \"Mon Dec 3 17:42:02 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n-## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n-## MPIR_Init_thread(565)..............: \n-## MPID_Init(224).....................: channel initialization failed\n-## MPIDI_CH3_Init(105)................: \n-## MPID_nem_init(324).................: \n-## MPID_nem_tcp_init(178).............: \n-## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 241 - ./datatype/sendrecvt4 2\n ---\n Directory: ./datatype\n File: sendrecvt4\n Num-procs: 2\n- Date: \"Sun Jan 5 02:02:15 2020\"\n+ Date: \"Mon Dec 3 17:42:02 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n-## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n-## MPIR_Init_thread(565)..............: \n-## MPID_Init(224).....................: channel initialization failed\n-## MPIDI_CH3_Init(105)................: \n-## MPID_nem_init(324).................: \n-## MPID_nem_tcp_init(178).............: \n-## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 242 - ./datatype/tmatchsize 1\n ---\n Directory: ./datatype\n File: tmatchsize\n Num-procs: 1\n- Date: \"Sun Jan 5 02:02:15 2020\"\n+ Date: \"Mon Dec 3 17:42:02 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 243 - ./datatype/tfree 2\n ---\n Directory: ./datatype\n File: tfree\n Num-procs: 2\n- Date: \"Sun Jan 5 02:02:16 2020\"\n+ Date: \"Mon Dec 3 17:42:03 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 244 - ./datatype/typelb 1\n ---\n Directory: ./datatype\n File: typelb\n Num-procs: 1\n- Date: \"Sun Jan 5 02:02:16 2020\"\n+ Date: \"Mon Dec 3 17:42:03 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 245 - ./datatype/contigstruct 1\n ---\n Directory: ./datatype\n File: contigstruct\n Num-procs: 1\n- Date: \"Sun Jan 5 02:02:16 2020\"\n+ Date: \"Mon Dec 3 17:42:04 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 246 - ./datatype/struct-zero-count 1\n ---\n Directory: ./datatype\n File: struct-zero-count\n Num-procs: 1\n- Date: \"Sun Jan 5 02:02:16 2020\"\n+ Date: \"Mon Dec 3 17:42:04 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 247 - ./datatype/blockindexed-zero-count 1\n ---\n Directory: ./datatype\n File: blockindexed-zero-count\n Num-procs: 1\n- Date: \"Sun Jan 5 02:02:17 2020\"\n+ Date: \"Mon Dec 3 17:42:05 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 248 - ./datatype/blockindexed-misc 1\n ---\n Directory: ./datatype\n File: blockindexed-misc\n Num-procs: 1\n- Date: \"Sun Jan 5 02:02:17 2020\"\n+ Date: \"Mon Dec 3 17:42:05 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 249 - ./datatype/indexed-misc 1\n ---\n Directory: ./datatype\n File: indexed-misc\n Num-procs: 1\n- Date: \"Sun Jan 5 02:02:17 2020\"\n+ Date: \"Mon Dec 3 17:42:06 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 250 - ./datatype/subarray-pack 1\n ---\n Directory: ./datatype\n File: subarray-pack\n Num-procs: 1\n- Date: \"Sun Jan 5 02:02:17 2020\"\n+ Date: \"Mon Dec 3 17:42:06 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 251 - ./datatype/subarray 2\n ---\n Directory: ./datatype\n File: subarray\n Num-procs: 2\n- Date: \"Sun Jan 5 02:02:18 2020\"\n+ Date: \"Mon Dec 3 17:42:07 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n ## Fatal error in PMPI_Init: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 252 - ./datatype/darray-pack 1\n ---\n Directory: ./datatype\n File: darray-pack\n Num-procs: 1\n- Date: \"Sun Jan 5 02:02:18 2020\"\n+ Date: \"Mon Dec 3 17:42:07 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 253 - ./datatype/darray-pack 2\n ---\n Directory: ./datatype\n File: darray-pack\n Num-procs: 2\n- Date: \"Sun Jan 5 02:02:18 2020\"\n+ Date: \"Mon Dec 3 17:42:08 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 254 - ./datatype/darray-pack 2\n ---\n Directory: ./datatype\n File: darray-pack\n Num-procs: 2\n- Date: \"Sun Jan 5 02:02:18 2020\"\n+ Date: \"Mon Dec 3 17:42:08 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n+## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n+## MPIR_Init_thread(565)..............: \n+## MPID_Init(224).....................: channel initialization failed\n+## MPIDI_CH3_Init(105)................: \n+## MPID_nem_init(324).................: \n+## MPID_nem_tcp_init(178).............: \n+## MPID_nem_tcp_get_business_card(425): \n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 255 - ./datatype/darray-cyclic 2\n ---\n Directory: ./datatype\n File: darray-cyclic\n Num-procs: 2\n- Date: \"Sun Jan 5 02:02:19 2020\"\n+ Date: \"Mon Dec 3 17:42:09 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 256 - ./datatype/pairtype-size-extent 1\n ---\n Directory: ./datatype\n File: pairtype-size-extent\n Num-procs: 1\n- Date: \"Sun Jan 5 02:02:19 2020\"\n+ Date: \"Mon Dec 3 17:42:09 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 257 - ./datatype/simple-commit 1\n ---\n Directory: ./datatype\n File: simple-commit\n Num-procs: 1\n- Date: \"Sun Jan 5 02:02:19 2020\"\n+ Date: \"Mon Dec 3 17:42:10 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 258 - ./datatype/simple-size-extent 1\n ---\n Directory: ./datatype\n File: simple-size-extent\n Num-procs: 1\n- Date: \"Sun Jan 5 02:02:19 2020\"\n+ Date: \"Mon Dec 3 17:42:10 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 259 - ./datatype/struct-no-real-types 1\n ---\n Directory: ./datatype\n File: struct-no-real-types\n Num-procs: 1\n- Date: \"Sun Jan 5 02:02:19 2020\"\n+ Date: \"Mon Dec 3 17:42:11 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 260 - ./datatype/struct-empty-el 1\n ---\n Directory: ./datatype\n File: struct-empty-el\n Num-procs: 1\n- Date: \"Sun Jan 5 02:02:20 2020\"\n+ Date: \"Mon Dec 3 17:42:11 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 261 - ./datatype/contig-zero-count 1\n ---\n Directory: ./datatype\n File: contig-zero-count\n Num-procs: 1\n- Date: \"Sun Jan 5 02:02:20 2020\"\n+ Date: \"Mon Dec 3 17:42:12 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 262 - ./datatype/simple-resized 1\n ---\n Directory: ./datatype\n File: simple-resized\n Num-procs: 1\n- Date: \"Sun Jan 5 02:02:20 2020\"\n+ Date: \"Mon Dec 3 17:42:12 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 263 - ./datatype/unusual-noncontigs 1\n ---\n Directory: ./datatype\n File: unusual-noncontigs\n Num-procs: 1\n- Date: \"Sun Jan 5 02:02:20 2020\"\n+ Date: \"Mon Dec 3 17:42:13 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 264 - ./datatype/hindexed-zeros 1\n ---\n Directory: ./datatype\n File: hindexed-zeros\n Num-procs: 1\n- Date: \"Sun Jan 5 02:02:20 2020\"\n+ Date: \"Mon Dec 3 17:42:13 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 265 - ./datatype/lots-of-types 1\n ---\n Directory: ./datatype\n File: lots-of-types\n Num-procs: 1\n- Date: \"Sun Jan 5 02:02:20 2020\"\n+ Date: \"Mon Dec 3 17:42:14 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 266 - ./datatype/get-elements-pairtype 1\n ---\n Directory: ./datatype\n File: get-elements-pairtype\n Num-procs: 1\n- Date: \"Sun Jan 5 02:02:21 2020\"\n+ Date: \"Mon Dec 3 17:42:14 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 267 - ./datatype/unpack 1\n ---\n Directory: ./datatype\n File: unpack\n Num-procs: 1\n- Date: \"Sun Jan 5 02:02:21 2020\"\n+ Date: \"Mon Dec 3 17:42:15 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 268 - ./datatype/struct-ezhov 1\n ---\n Directory: ./datatype\n File: struct-ezhov\n Num-procs: 1\n- Date: \"Sun Jan 5 02:02:21 2020\"\n+ Date: \"Mon Dec 3 17:42:15 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 269 - ./datatype/zeroblks 1\n ---\n Directory: ./datatype\n File: zeroblks\n Num-procs: 1\n- Date: \"Sun Jan 5 02:02:21 2020\"\n+ Date: \"Mon Dec 3 17:42:16 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 270 - ./datatype/struct-derived-zeros 1\n ---\n Directory: ./datatype\n File: struct-derived-zeros\n Num-procs: 1\n- Date: \"Sun Jan 5 02:02:21 2020\"\n+ Date: \"Mon Dec 3 17:42:16 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 271 - ./datatype/struct-verydeep 1\n ---\n Directory: ./datatype\n File: struct-verydeep\n Num-procs: 1\n- Date: \"Sun Jan 5 02:02:21 2020\"\n+ Date: \"Mon Dec 3 17:42:17 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 272 - ./datatype/get-elements 1\n ---\n Directory: ./datatype\n File: get-elements\n Num-procs: 1\n- Date: \"Sun Jan 5 02:02:22 2020\"\n+ Date: \"Mon Dec 3 17:42:17 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 273 - ./datatype/hindexed_block 1\n ---\n Directory: ./datatype\n File: hindexed_block\n Num-procs: 1\n- Date: \"Sun Jan 5 02:02:22 2020\"\n+ Date: \"Mon Dec 3 17:42:18 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 274 - ./datatype/hindexed_block_contents 1\n ---\n Directory: ./datatype\n File: hindexed_block_contents\n Num-procs: 1\n- Date: \"Sun Jan 5 02:02:22 2020\"\n+ Date: \"Mon Dec 3 17:42:18 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 275 - ./datatype/vecblklen 1\n ---\n Directory: ./datatype\n File: vecblklen\n Num-procs: 1\n- Date: \"Sun Jan 5 02:02:22 2020\"\n+ Date: \"Mon Dec 3 17:42:18 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 276 - ./datatype/hvecblklen 1\n ---\n Directory: ./datatype\n File: hvecblklen\n Num-procs: 1\n- Date: \"Sun Jan 5 02:02:22 2020\"\n+ Date: \"Mon Dec 3 17:42:19 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 277 - ./datatype/longdouble 1\n ---\n Directory: ./datatype\n File: longdouble\n Num-procs: 1\n- Date: \"Sun Jan 5 02:02:22 2020\"\n+ Date: \"Mon Dec 3 17:42:19 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 278 - ./datatype/dataalign 2\n ---\n Directory: ./datatype\n File: dataalign\n Num-procs: 2\n- Date: \"Sun Jan 5 02:02:23 2020\"\n+ Date: \"Mon Dec 3 17:42:20 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n-## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n-## MPIR_Init_thread(565)..............: \n-## MPID_Init(224).....................: channel initialization failed\n-## MPIDI_CH3_Init(105)................: \n-## MPID_nem_init(324).................: \n-## MPID_nem_tcp_init(178).............: \n-## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 279 - ./datatype/cxx-types 1\n ---\n Directory: ./datatype\n File: cxx-types\n Num-procs: 1\n- Date: \"Sun Jan 5 02:02:23 2020\"\n+ Date: \"Mon Dec 3 17:42:20 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 280 - ./errhan/adderr 1\n ---\n Directory: ./errhan\n File: adderr\n Num-procs: 1\n- Date: \"Sun Jan 5 02:02:23 2020\"\n+ Date: \"Mon Dec 3 17:42:21 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 281 - ./errhan/commcall 2\n ---\n Directory: ./errhan\n File: commcall\n Num-procs: 2\n- Date: \"Sun Jan 5 02:02:23 2020\"\n+ Date: \"Mon Dec 3 17:42:21 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n ok 282 - ./errhan/errfatal 1\n not ok 283 - ./errhan/predef_eh 1\n ---\n Directory: ./errhan\n File: predef_eh\n Num-procs: 1\n- Date: \"Sun Jan 5 02:02:23 2020\"\n+ Date: \"Mon Dec 3 17:42:22 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 284 - ./errhan/predef_eh 2\n ---\n Directory: ./errhan\n File: predef_eh\n Num-procs: 2\n- Date: \"Sun Jan 5 02:02:24 2020\"\n+ Date: \"Mon Dec 3 17:42:22 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n-## Fatal error in PMPI_Init: Other MPI error, error stack:\n-## MPIR_Init_thread(565)..............: \n-## MPID_Init(224).....................: channel initialization failed\n-## MPIDI_CH3_Init(105)................: \n-## MPID_nem_init(324).................: \n-## MPID_nem_tcp_init(178).............: \n-## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 285 - ./errhan/errstring2 1\n ---\n Directory: ./errhan\n File: errstring2\n Num-procs: 1\n- Date: \"Sun Jan 5 02:02:24 2020\"\n+ Date: \"Mon Dec 3 17:42:22 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 286 - ./errhan/dynamic_errcode_predefined_errclass 1\n ---\n Directory: ./errhan\n File: dynamic_errcode_predefined_errclass\n Num-procs: 1\n- Date: \"Sun Jan 5 02:02:24 2020\"\n+ Date: \"Mon Dec 3 17:42:23 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 287 - ./group/@groupcreate \n ---\n Directory: ./group\n File: @groupcreate\n Num-procs: \n- Date: \"Sun Jan 5 02:02:24 2020\"\n+ Date: \"Mon Dec 3 17:42:23 2018\"\n ...\n ## Test output (expected 'No Errors'):\n-## Failed to build @groupcreate; make[2]: Entering directory '/build/1st/mpi-testsuite-3.2+dfsg/build-mpich/group'\n+## Failed to build @groupcreate; make[2]: Entering directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-mpich/group'\n ## make[2]: *** No rule to make target '@groupcreate'. Stop.\n-## make[2]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-mpich/group'\n+## make[2]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-mpich/group'\n not ok 288 - ./group/@grouptest \n ---\n Directory: ./group\n File: @grouptest\n Num-procs: \n- Date: \"Sun Jan 5 02:02:24 2020\"\n+ Date: \"Mon Dec 3 17:42:23 2018\"\n ...\n ## Test output (expected 'No Errors'):\n-## Failed to build @grouptest; make[2]: Entering directory '/build/1st/mpi-testsuite-3.2+dfsg/build-mpich/group'\n+## Failed to build @grouptest; make[2]: Entering directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-mpich/group'\n ## make[2]: *** No rule to make target '@grouptest'. Stop.\n-## make[2]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-mpich/group'\n+## make[2]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-mpich/group'\n not ok 289 - ./group/@grouptest2 \n ---\n Directory: ./group\n File: @grouptest2\n Num-procs: \n- Date: \"Sun Jan 5 02:02:24 2020\"\n+ Date: \"Mon Dec 3 17:42:23 2018\"\n ...\n ## Test output (expected 'No Errors'):\n-## Failed to build @grouptest2; make[2]: Entering directory '/build/1st/mpi-testsuite-3.2+dfsg/build-mpich/group'\n+## Failed to build @grouptest2; make[2]: Entering directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-mpich/group'\n ## make[2]: *** No rule to make target '@grouptest2'. Stop.\n-## make[2]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-mpich/group'\n+## make[2]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-mpich/group'\n not ok 290 - ./group/@groupnullincl \n ---\n Directory: ./group\n File: @groupnullincl\n Num-procs: \n- Date: \"Sun Jan 5 02:02:24 2020\"\n+ Date: \"Mon Dec 3 17:42:23 2018\"\n ...\n ## Test output (expected 'No Errors'):\n-## Failed to build @groupnullincl; make[2]: Entering directory '/build/1st/mpi-testsuite-3.2+dfsg/build-mpich/group'\n+## Failed to build @groupnullincl; make[2]: Entering directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-mpich/group'\n ## make[2]: *** No rule to make target '@groupnullincl'. Stop.\n-## make[2]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-mpich/group'\n+## make[2]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-mpich/group'\n not ok 291 - ./group/@gtranks \n ---\n Directory: ./group\n File: @gtranks\n Num-procs: \n- Date: \"Sun Jan 5 02:02:24 2020\"\n+ Date: \"Mon Dec 3 17:42:23 2018\"\n ...\n ## Test output (expected 'No Errors'):\n-## Failed to build @gtranks; make[2]: Entering directory '/build/1st/mpi-testsuite-3.2+dfsg/build-mpich/group'\n+## Failed to build @gtranks; make[2]: Entering directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-mpich/group'\n ## make[2]: *** No rule to make target '@gtranks'. Stop.\n-## make[2]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-mpich/group'\n+## make[2]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-mpich/group'\n not ok 292 - ./group/@gtranksperf \n ---\n Directory: ./group\n File: @gtranksperf\n Num-procs: \n- Date: \"Sun Jan 5 02:02:24 2020\"\n+ Date: \"Mon Dec 3 17:42:23 2018\"\n ...\n ## Test output (expected 'No Errors'):\n-## Failed to build @gtranksperf; make[2]: Entering directory '/build/1st/mpi-testsuite-3.2+dfsg/build-mpich/group'\n+## Failed to build @gtranksperf; make[2]: Entering directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-mpich/group'\n ## make[2]: *** No rule to make target '@gtranksperf'. Stop.\n-## make[2]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-mpich/group'\n+## make[2]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-mpich/group'\n not ok 293 - ./info/infodup 1\n ---\n Directory: ./info\n File: infodup\n Num-procs: 1\n- Date: \"Sun Jan 5 02:02:24 2020\"\n+ Date: \"Mon Dec 3 17:42:24 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 294 - ./info/infodel 1\n ---\n Directory: ./info\n File: infodel\n Num-procs: 1\n- Date: \"Sun Jan 5 02:02:24 2020\"\n+ Date: \"Mon Dec 3 17:42:24 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 295 - ./info/infovallen 1\n ---\n Directory: ./info\n File: infovallen\n Num-procs: 1\n- Date: \"Sun Jan 5 02:02:24 2020\"\n+ Date: \"Mon Dec 3 17:42:25 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 296 - ./info/infoorder 1\n ---\n Directory: ./info\n File: infoorder\n Num-procs: 1\n- Date: \"Sun Jan 5 02:02:24 2020\"\n+ Date: \"Mon Dec 3 17:42:25 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 297 - ./info/infomany 1\n ---\n Directory: ./info\n File: infomany\n Num-procs: 1\n- Date: \"Sun Jan 5 02:02:25 2020\"\n+ Date: \"Mon Dec 3 17:42:26 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 298 - ./info/infomany2 1\n ---\n Directory: ./info\n File: infomany2\n Num-procs: 1\n- Date: \"Sun Jan 5 02:02:25 2020\"\n+ Date: \"Mon Dec 3 17:42:26 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 299 - ./info/infotest 1\n ---\n Directory: ./info\n File: infotest\n Num-procs: 1\n- Date: \"Sun Jan 5 02:02:25 2020\"\n+ Date: \"Mon Dec 3 17:42:26 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 300 - ./info/infoget 1\n ---\n Directory: ./info\n File: infoget\n Num-procs: 1\n- Date: \"Sun Jan 5 02:02:25 2020\"\n+ Date: \"Mon Dec 3 17:42:27 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 301 - ./info/infoenv 1\n ---\n Directory: ./info\n File: infoenv\n Num-procs: 1\n- Date: \"Sun Jan 5 02:02:25 2020\"\n+ Date: \"Mon Dec 3 17:42:27 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n ok 302 - ./init/exitst1 2\n ok 303 - ./init/exitst2 2\n not ok 304 - ./init/initstat 1\n ---\n Directory: ./init\n File: initstat\n Num-procs: 1\n- Date: \"Sun Jan 5 02:02:26 2020\"\n+ Date: \"Mon Dec 3 17:42:28 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 305 - ./init/timeout 2\n ---\n Directory: ./init\n File: timeout\n Num-procs: 2\n- Date: \"Sun Jan 5 02:02:26 2020\"\n+ Date: \"Mon Dec 3 17:42:29 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 306 - ./init/version 1\n ---\n Directory: ./init\n File: version\n Num-procs: 1\n- Date: \"Sun Jan 5 02:02:26 2020\"\n+ Date: \"Mon Dec 3 17:42:29 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 307 - ./init/finalized 1\n ---\n Directory: ./init\n File: finalized\n Num-procs: 1\n- Date: \"Sun Jan 5 02:02:26 2020\"\n+ Date: \"Mon Dec 3 17:42:29 2018\"\n ...\n ## Test output (expected 'No Errors'):\n-## Fatal error in PMPI_Init: Other MPI error, error stack:\n-## MPIR_Init_thread(565)..............: \n-## MPID_Init(224).....................: channel initialization failed\n-## MPIDI_CH3_Init(105)................: \n-## MPID_nem_init(324).................: \n-## MPID_nem_tcp_init(178).............: \n-## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n not ok 308 - ./init/attrself 1\n ---\n Directory: ./init\n File: attrself\n Num-procs: 1\n- Date: \"Sun Jan 5 02:02:26 2020\"\n+ Date: \"Mon Dec 3 17:42:30 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 309 - ./init/library_version 1\n ---\n Directory: ./init\n File: library_version\n Num-procs: 1\n- Date: \"Sun Jan 5 02:02:27 2020\"\n+ Date: \"Mon Dec 3 17:42:30 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 310 - ./mpi_t/mpi_t_str 1\n ---\n Directory: ./mpi_t\n File: mpi_t_str\n Num-procs: 1\n- Date: \"Sun Jan 5 02:02:27 2020\"\n+ Date: \"Mon Dec 3 17:42:30 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 311 - ./mpi_t/mpit_vars 1\n ---\n Directory: ./mpi_t\n File: mpit_vars\n Num-procs: 1\n- Date: \"Sun Jan 5 02:02:27 2020\"\n+ Date: \"Mon Dec 3 17:42:31 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 312 - ./mpi_t/cvarwrite 1\n ---\n Directory: ./mpi_t\n File: cvarwrite\n Num-procs: 1\n- Date: \"Sun Jan 5 02:02:27 2020\"\n+ Date: \"Mon Dec 3 17:42:31 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n ok 313 - ./mpi_t/getindex 1 # SKIP requires MPI version 3.1\n not ok 313 - ./pt2pt/sendrecv1 2\n ---\n Directory: ./pt2pt\n File: sendrecv1\n Num-procs: 2\n- Date: \"Sun Jan 5 02:02:27 2020\"\n+ Date: \"Mon Dec 3 17:42:32 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 314 - ./pt2pt/sendrecv2 2\n ---\n Directory: ./pt2pt\n File: sendrecv2\n Num-procs: 2\n- Date: \"Sun Jan 5 02:02:27 2020\"\n+ Date: \"Mon Dec 3 17:42:32 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n ## Fatal error in PMPI_Init: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 315 - ./pt2pt/sendrecv3 2\n ---\n Directory: ./pt2pt\n File: sendrecv3\n Num-procs: 2\n- Date: \"Sun Jan 5 02:02:28 2020\"\n+ Date: \"Mon Dec 3 17:42:33 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 316 - ./pt2pt/sendflood 2\n ---\n Directory: ./pt2pt\n File: sendflood\n Num-procs: 2\n- Date: \"Sun Jan 5 02:02:28 2020\"\n+ Date: \"Mon Dec 3 17:42:33 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 317 - ./pt2pt/sendself 1\n ---\n Directory: ./pt2pt\n File: sendself\n Num-procs: 1\n- Date: \"Sun Jan 5 02:02:28 2020\"\n+ Date: \"Mon Dec 3 17:42:33 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 318 - ./pt2pt/sendall 2\n ---\n Directory: ./pt2pt\n File: sendall\n Num-procs: 2\n- Date: \"Sun Jan 5 02:02:28 2020\"\n+ Date: \"Mon Dec 3 17:42:34 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 319 - ./pt2pt/anyall 2\n ---\n Directory: ./pt2pt\n File: anyall\n Num-procs: 2\n- Date: \"Sun Jan 5 02:02:28 2020\"\n+ Date: \"Mon Dec 3 17:42:34 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n-## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n-## MPIR_Init_thread(565)..............: \n-## MPID_Init(224).....................: channel initialization failed\n-## MPIDI_CH3_Init(105)................: \n-## MPID_nem_init(324).................: \n-## MPID_nem_tcp_init(178).............: \n-## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 320 - ./pt2pt/eagerdt 2\n ---\n Directory: ./pt2pt\n File: eagerdt\n Num-procs: 2\n- Date: \"Sun Jan 5 02:02:29 2020\"\n+ Date: \"Mon Dec 3 17:42:35 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 321 - ./pt2pt/pingping 2\n ---\n Directory: ./pt2pt\n File: pingping\n Num-procs: 2\n- Date: \"Sun Jan 5 02:02:29 2020\"\n+ Date: \"Mon Dec 3 17:42:35 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n-## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n-## MPIR_Init_thread(565)..............: \n-## MPID_Init(224).....................: channel initialization failed\n-## MPIDI_CH3_Init(105)................: \n-## MPID_nem_init(324).................: \n-## MPID_nem_tcp_init(178).............: \n-## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 322 - ./pt2pt/bottom 2\n ---\n Directory: ./pt2pt\n File: bottom\n Num-procs: 2\n- Date: \"Sun Jan 5 02:02:29 2020\"\n+ Date: \"Mon Dec 3 17:42:35 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 323 - ./pt2pt/bsend1 1\n ---\n Directory: ./pt2pt\n File: bsend1\n Num-procs: 1\n- Date: \"Sun Jan 5 02:02:29 2020\"\n+ Date: \"Mon Dec 3 17:42:36 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 324 - ./pt2pt/bsend2 1\n ---\n Directory: ./pt2pt\n File: bsend2\n Num-procs: 1\n- Date: \"Sun Jan 5 02:02:29 2020\"\n+ Date: \"Mon Dec 3 17:42:36 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 325 - ./pt2pt/bsend3 1\n ---\n Directory: ./pt2pt\n File: bsend3\n Num-procs: 1\n- Date: \"Sun Jan 5 02:02:29 2020\"\n+ Date: \"Mon Dec 3 17:42:37 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 326 - ./pt2pt/bsend4 1\n ---\n Directory: ./pt2pt\n File: bsend4\n Num-procs: 1\n- Date: \"Sun Jan 5 02:02:30 2020\"\n+ Date: \"Mon Dec 3 17:42:37 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 327 - ./pt2pt/bsend5 2\n ---\n Directory: ./pt2pt\n File: bsend5\n Num-procs: 2\n- Date: \"Sun Jan 5 02:02:30 2020\"\n+ Date: \"Mon Dec 3 17:42:37 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 328 - ./pt2pt/bsendalign 2\n ---\n Directory: ./pt2pt\n File: bsendalign\n Num-procs: 2\n- Date: \"Sun Jan 5 02:02:30 2020\"\n+ Date: \"Mon Dec 3 17:42:38 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n+## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n+## MPIR_Init_thread(565)..............: \n+## MPID_Init(224).....................: channel initialization failed\n+## MPIDI_CH3_Init(105)................: \n+## MPID_nem_init(324).................: \n+## MPID_nem_tcp_init(178).............: \n+## MPID_nem_tcp_get_business_card(425): \n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 329 - ./pt2pt/bsendpending 2\n ---\n Directory: ./pt2pt\n File: bsendpending\n Num-procs: 2\n- Date: \"Sun Jan 5 02:02:30 2020\"\n+ Date: \"Mon Dec 3 17:42:38 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 330 - ./pt2pt/isendself 1\n ---\n Directory: ./pt2pt\n File: isendself\n Num-procs: 1\n- Date: \"Sun Jan 5 02:02:31 2020\"\n+ Date: \"Mon Dec 3 17:42:39 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 331 - ./pt2pt/issendselfcancel 1\n ---\n Directory: ./pt2pt\n File: issendselfcancel\n Num-procs: 1\n- Date: \"Sun Jan 5 02:02:32 2020\"\n+ Date: \"Mon Dec 3 17:42:39 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 332 - ./pt2pt/isendirecv 2\n ---\n Directory: ./pt2pt\n File: isendirecv\n Num-procs: 2\n- Date: \"Sun Jan 5 02:02:33 2020\"\n+ Date: \"Mon Dec 3 17:42:39 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n+## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n+## MPIR_Init_thread(565)..............: \n+## MPID_Init(224).....................: channel initialization failed\n+## MPIDI_CH3_Init(105)................: \n+## MPID_nem_init(324).................: \n+## MPID_nem_tcp_init(178).............: \n+## MPID_nem_tcp_get_business_card(425): \n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 333 - ./pt2pt/bsendfrag 2\n ---\n Directory: ./pt2pt\n File: bsendfrag\n Num-procs: 2\n- Date: \"Sun Jan 5 02:02:34 2020\"\n+ Date: \"Mon Dec 3 17:42:40 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 334 - ./pt2pt/icsend 2\n ---\n Directory: ./pt2pt\n File: icsend\n Num-procs: 2\n- Date: \"Sun Jan 5 02:02:34 2020\"\n+ Date: \"Mon Dec 3 17:42:40 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n+## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n+## MPIR_Init_thread(565)..............: \n+## MPID_Init(224).....................: channel initialization failed\n+## MPIDI_CH3_Init(105)................: \n+## MPID_nem_init(324).................: \n+## MPID_nem_tcp_init(178).............: \n+## MPID_nem_tcp_get_business_card(425): \n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 335 - ./pt2pt/rqstatus 2\n ---\n Directory: ./pt2pt\n File: rqstatus\n Num-procs: 2\n- Date: \"Sun Jan 5 02:02:36 2020\"\n+ Date: \"Mon Dec 3 17:42:41 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 336 - ./pt2pt/rqfreeb 2\n ---\n Directory: ./pt2pt\n File: rqfreeb\n Num-procs: 2\n- Date: \"Sun Jan 5 02:02:36 2020\"\n+ Date: \"Mon Dec 3 17:42:41 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n-## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n-## MPIR_Init_thread(565)..............: \n-## MPID_Init(224).....................: channel initialization failed\n-## MPIDI_CH3_Init(105)................: \n-## MPID_nem_init(324).................: \n-## MPID_nem_tcp_init(178).............: \n-## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 337 - ./pt2pt/greq1 1\n ---\n Directory: ./pt2pt\n File: greq1\n Num-procs: 1\n- Date: \"Sun Jan 5 02:02:37 2020\"\n+ Date: \"Mon Dec 3 17:42:42 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 338 - ./pt2pt/probe-unexp 2\n ---\n Directory: ./pt2pt\n File: probe-unexp\n Num-procs: 2\n- Date: \"Sun Jan 5 02:02:38 2020\"\n+ Date: \"Mon Dec 3 17:42:42 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 339 - ./pt2pt/probenull 1\n ---\n Directory: ./pt2pt\n File: probenull\n Num-procs: 1\n- Date: \"Sun Jan 5 02:02:38 2020\"\n+ Date: \"Mon Dec 3 17:42:43 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 340 - ./pt2pt/scancel 2\n ---\n Directory: ./pt2pt\n File: scancel\n Num-procs: 2\n- Date: \"Sun Jan 5 02:02:39 2020\"\n+ Date: \"Mon Dec 3 17:42:43 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 341 - ./pt2pt/scancel2 2\n ---\n Directory: ./pt2pt\n File: scancel2\n Num-procs: 2\n- Date: \"Sun Jan 5 02:02:39 2020\"\n+ Date: \"Mon Dec 3 17:42:43 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 342 - ./pt2pt/pscancel 2\n ---\n Directory: ./pt2pt\n File: pscancel\n Num-procs: 2\n- Date: \"Sun Jan 5 02:02:41 2020\"\n+ Date: \"Mon Dec 3 17:42:44 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n+## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n+## MPIR_Init_thread(565)..............: \n+## MPID_Init(224).....................: channel initialization failed\n+## MPIDI_CH3_Init(105)................: \n+## MPID_nem_init(324).................: \n+## MPID_nem_tcp_init(178).............: \n+## MPID_nem_tcp_get_business_card(425): \n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 343 - ./pt2pt/rcancel 2\n ---\n Directory: ./pt2pt\n File: rcancel\n Num-procs: 2\n- Date: \"Sun Jan 5 02:02:42 2020\"\n+ Date: \"Mon Dec 3 17:42:44 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 344 - ./pt2pt/cancelrecv 2\n ---\n Directory: ./pt2pt\n File: cancelrecv\n Num-procs: 2\n- Date: \"Sun Jan 5 02:02:43 2020\"\n+ Date: \"Mon Dec 3 17:42:45 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 345 - ./pt2pt/scancel_unmatch 2\n ---\n Directory: ./pt2pt\n File: scancel_unmatch\n Num-procs: 2\n- Date: \"Sun Jan 5 02:02:44 2020\"\n+ Date: \"Mon Dec 3 17:42:45 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n ## Fatal error in PMPI_Init: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 346 - ./pt2pt/cancelanysrc 2\n ---\n Directory: ./pt2pt\n File: cancelanysrc\n Num-procs: 2\n- Date: \"Sun Jan 5 02:02:45 2020\"\n+ Date: \"Mon Dec 3 17:42:46 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n-## Fatal error in PMPI_Init: Other MPI error, error stack:\n-## MPIR_Init_thread(565)..............: \n-## MPID_Init(224).....................: channel initialization failed\n-## MPIDI_CH3_Init(105)................: \n-## MPID_nem_init(324).................: \n-## MPID_nem_tcp_init(178).............: \n-## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 347 - ./pt2pt/isendselfprobe 1\n ---\n Directory: ./pt2pt\n File: isendselfprobe\n Num-procs: 1\n- Date: \"Sun Jan 5 02:02:46 2020\"\n+ Date: \"Mon Dec 3 17:42:46 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 348 - ./pt2pt/inactivereq 1\n ---\n Directory: ./pt2pt\n File: inactivereq\n Num-procs: 1\n- Date: \"Sun Jan 5 02:02:47 2020\"\n+ Date: \"Mon Dec 3 17:42:46 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 349 - ./pt2pt/waittestnull 1\n ---\n Directory: ./pt2pt\n File: waittestnull\n Num-procs: 1\n- Date: \"Sun Jan 5 02:02:48 2020\"\n+ Date: \"Mon Dec 3 17:42:47 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 350 - ./pt2pt/waitany-null 1\n ---\n Directory: ./pt2pt\n File: waitany-null\n Num-procs: 1\n- Date: \"Sun Jan 5 02:02:49 2020\"\n+ Date: \"Mon Dec 3 17:42:47 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 351 - ./pt2pt/mprobe 2\n ---\n Directory: ./pt2pt\n File: mprobe\n Num-procs: 2\n- Date: \"Sun Jan 5 02:02:49 2020\"\n+ Date: \"Mon Dec 3 17:42:48 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n-## Fatal error in PMPI_Init: Other MPI error, error stack:\n-## MPIR_Init_thread(565)..............: \n-## MPID_Init(224).....................: channel initialization failed\n-## MPIDI_CH3_Init(105)................: \n-## MPID_nem_init(324).................: \n-## MPID_nem_tcp_init(178).............: \n-## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 352 - ./pt2pt/big_count_status 1\n ---\n Directory: ./pt2pt\n File: big_count_status\n Num-procs: 1\n- Date: \"Sun Jan 5 02:02:51 2020\"\n+ Date: \"Mon Dec 3 17:42:48 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 353 - ./rma/winname 2\n ---\n Directory: ./rma\n File: winname\n Num-procs: 2\n- Date: \"Sun Jan 5 02:02:51 2020\"\n+ Date: \"Mon Dec 3 17:42:49 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n-## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n-## MPIR_Init_thread(565)..............: \n-## MPID_Init(224).....................: channel initialization failed\n-## MPIDI_CH3_Init(105)................: \n-## MPID_nem_init(324).................: \n-## MPID_nem_tcp_init(178).............: \n-## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 354 - ./rma/allocmem 2\n ---\n Directory: ./rma\n File: allocmem\n Num-procs: 2\n- Date: \"Sun Jan 5 02:02:52 2020\"\n+ Date: \"Mon Dec 3 17:42:49 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n-## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n-## MPIR_Init_thread(565)..............: \n-## MPID_Init(224).....................: channel initialization failed\n-## MPIDI_CH3_Init(105)................: \n-## MPID_nem_init(324).................: \n-## MPID_nem_tcp_init(178).............: \n-## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 355 - ./rma/putfidx 2\n ---\n Directory: ./rma\n File: putfidx\n Num-procs: 2\n- Date: \"Sun Jan 5 02:02:53 2020\"\n+ Date: \"Mon Dec 3 17:42:50 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 356 - ./rma/accfence1 2\n ---\n Directory: ./rma\n File: accfence1\n Num-procs: 2\n- Date: \"Sun Jan 5 02:02:55 2020\"\n+ Date: \"Mon Dec 3 17:42:51 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n-## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n-## MPIR_Init_thread(565)..............: \n-## MPID_Init(224).....................: channel initialization failed\n-## MPIDI_CH3_Init(105)................: \n-## MPID_nem_init(324).................: \n-## MPID_nem_tcp_init(178).............: \n-## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 357 - ./rma/adlb_mimic1 2\n ---\n Directory: ./rma\n File: adlb_mimic1\n Num-procs: 2\n- Date: \"Sun Jan 5 02:02:57 2020\"\n+ Date: \"Mon Dec 3 17:42:51 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 358 - ./rma/accfence2 2\n ---\n Directory: ./rma\n File: accfence2\n Num-procs: 2\n- Date: \"Sun Jan 5 02:02:58 2020\"\n+ Date: \"Mon Dec 3 17:42:52 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n-not ok 359 - ./rma/putpscw1 2\n- ---\n- Directory: ./rma\n- File: putpscw1\n- Num-procs: 2\n- Date: \"Sun Jan 5 02:02:58 2020\"\n- ...\n-## Test output (expected 'No Errors'):\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n+not ok 359 - ./rma/putpscw1 2\n+ ---\n+ Directory: ./rma\n+ File: putpscw1\n+ Num-procs: 2\n+ Date: \"Mon Dec 3 17:42:53 2018\"\n+ ...\n+## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 360 - ./rma/accpscw1 2\n ---\n Directory: ./rma\n File: accpscw1\n Num-procs: 2\n- Date: \"Sun Jan 5 02:02:59 2020\"\n+ Date: \"Mon Dec 3 17:42:53 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 361 - ./rma/getgroup 2\n ---\n Directory: ./rma\n File: getgroup\n Num-procs: 2\n- Date: \"Sun Jan 5 02:02:59 2020\"\n+ Date: \"Mon Dec 3 17:42:54 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 362 - ./rma/transpose1 2\n ---\n Directory: ./rma\n File: transpose1\n Num-procs: 2\n- Date: \"Sun Jan 5 02:03:00 2020\"\n+ Date: \"Mon Dec 3 17:42:55 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n+## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n+## MPIR_Init_thread(565)..............: \n+## MPID_Init(224).....................: channel initialization failed\n+## MPIDI_CH3_Init(105)................: \n+## MPID_nem_init(324).................: \n+## MPID_nem_tcp_init(178).............: \n+## MPID_nem_tcp_get_business_card(425): \n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 363 - ./rma/transpose2 2\n ---\n Directory: ./rma\n File: transpose2\n Num-procs: 2\n- Date: \"Sun Jan 5 02:03:00 2020\"\n+ Date: \"Mon Dec 3 17:42:55 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 364 - ./rma/transpose3 2\n ---\n Directory: ./rma\n File: transpose3\n Num-procs: 2\n- Date: \"Sun Jan 5 02:03:01 2020\"\n+ Date: \"Mon Dec 3 17:42:56 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n-## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n-## MPIR_Init_thread(565)..............: \n-## MPID_Init(224).....................: channel initialization failed\n-## MPIDI_CH3_Init(105)................: \n-## MPID_nem_init(324).................: \n-## MPID_nem_tcp_init(178).............: \n-## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 365 - ./rma/transpose3_shm 2\n ---\n Directory: ./rma\n File: transpose3_shm\n Num-procs: 2\n- Date: \"Sun Jan 5 02:03:01 2020\"\n+ Date: \"Mon Dec 3 17:42:57 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 366 - ./rma/transpose5 2\n ---\n Directory: ./rma\n File: transpose5\n Num-procs: 2\n- Date: \"Sun Jan 5 02:03:01 2020\"\n+ Date: \"Mon Dec 3 17:42:57 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 367 - ./rma/transpose6 1\n ---\n Directory: ./rma\n File: transpose6\n Num-procs: 1\n- Date: \"Sun Jan 5 02:03:02 2020\"\n+ Date: \"Mon Dec 3 17:42:58 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 368 - ./rma/transpose7 2\n ---\n Directory: ./rma\n File: transpose7\n Num-procs: 2\n- Date: \"Sun Jan 5 02:03:02 2020\"\n+ Date: \"Mon Dec 3 17:42:59 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 369 - ./rma/test1 2\n ---\n Directory: ./rma\n File: test1\n Num-procs: 2\n- Date: \"Sun Jan 5 02:03:02 2020\"\n+ Date: \"Mon Dec 3 17:42:59 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n-## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n-## MPIR_Init_thread(565)..............: \n-## MPID_Init(224).....................: channel initialization failed\n-## MPIDI_CH3_Init(105)................: \n-## MPID_nem_init(324).................: \n-## MPID_nem_tcp_init(178).............: \n-## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 370 - ./rma/test2 2\n ---\n Directory: ./rma\n File: test2\n Num-procs: 2\n- Date: \"Sun Jan 5 02:03:03 2020\"\n+ Date: \"Mon Dec 3 17:43:00 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n+## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n+## MPIR_Init_thread(565)..............: \n+## MPID_Init(224).....................: channel initialization failed\n+## MPIDI_CH3_Init(105)................: \n+## MPID_nem_init(324).................: \n+## MPID_nem_tcp_init(178).............: \n+## MPID_nem_tcp_get_business_card(425): \n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 371 - ./rma/test2_shm 2\n ---\n Directory: ./rma\n File: test2_shm\n Num-procs: 2\n- Date: \"Sun Jan 5 02:03:03 2020\"\n+ Date: \"Mon Dec 3 17:43:00 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n+## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n+## MPIR_Init_thread(565)..............: \n+## MPID_Init(224).....................: channel initialization failed\n+## MPIDI_CH3_Init(105)................: \n+## MPID_nem_init(324).................: \n+## MPID_nem_tcp_init(178).............: \n+## MPID_nem_tcp_get_business_card(425): \n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 372 - ./rma/test3 2\n ---\n Directory: ./rma\n File: test3\n Num-procs: 2\n- Date: \"Sun Jan 5 02:03:03 2020\"\n+ Date: \"Mon Dec 3 17:43:01 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 373 - ./rma/test3_shm 2\n ---\n Directory: ./rma\n File: test3_shm\n Num-procs: 2\n- Date: \"Sun Jan 5 02:03:04 2020\"\n+ Date: \"Mon Dec 3 17:43:01 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n+## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n+## MPIR_Init_thread(565)..............: \n+## MPID_Init(224).....................: channel initialization failed\n+## MPIDI_CH3_Init(105)................: \n+## MPID_nem_init(324).................: \n+## MPID_nem_tcp_init(178).............: \n+## MPID_nem_tcp_get_business_card(425): \n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 374 - ./rma/test4 2\n ---\n Directory: ./rma\n File: test4\n Num-procs: 2\n- Date: \"Sun Jan 5 02:03:04 2020\"\n+ Date: \"Mon Dec 3 17:43:02 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 375 - ./rma/test5 2\n ---\n Directory: ./rma\n File: test5\n Num-procs: 2\n- Date: \"Sun Jan 5 02:03:04 2020\"\n+ Date: \"Mon Dec 3 17:43:03 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 376 - ./rma/lockcontention 2\n ---\n Directory: ./rma\n File: lockcontention\n Num-procs: 2\n- Date: \"Sun Jan 5 02:03:05 2020\"\n+ Date: \"Mon Dec 3 17:43:04 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n-## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n-## MPIR_Init_thread(565)..............: \n-## MPID_Init(224).....................: channel initialization failed\n-## MPIDI_CH3_Init(105)................: \n-## MPID_nem_init(324).................: \n-## MPID_nem_tcp_init(178).............: \n-## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 377 - ./rma/lockcontention2 2\n ---\n Directory: ./rma\n File: lockcontention2\n Num-procs: 2\n- Date: \"Sun Jan 5 02:03:05 2020\"\n+ Date: \"Mon Dec 3 17:43:04 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 378 - ./rma/lockcontention2 2\n ---\n Directory: ./rma\n File: lockcontention2\n Num-procs: 2\n- Date: \"Sun Jan 5 02:03:05 2020\"\n+ Date: \"Mon Dec 3 17:43:05 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n-## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n-## MPIR_Init_thread(565)..............: \n-## MPID_Init(224).....................: channel initialization failed\n-## MPIDI_CH3_Init(105)................: \n-## MPID_nem_init(324).................: \n-## MPID_nem_tcp_init(178).............: \n-## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 379 - ./rma/lockcontention3 2\n ---\n Directory: ./rma\n File: lockcontention3\n Num-procs: 2\n- Date: \"Sun Jan 5 02:03:06 2020\"\n+ Date: \"Mon Dec 3 17:43:05 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n+## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n+## MPIR_Init_thread(565)..............: \n+## MPID_Init(224).....................: channel initialization failed\n+## MPIDI_CH3_Init(105)................: \n+## MPID_nem_init(324).................: \n+## MPID_nem_tcp_init(178).............: \n+## MPID_nem_tcp_get_business_card(425): \n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 380 - ./rma/lockopts 2\n ---\n Directory: ./rma\n File: lockopts\n Num-procs: 2\n- Date: \"Sun Jan 5 02:03:06 2020\"\n+ Date: \"Mon Dec 3 17:43:06 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 381 - ./rma/lock_dt 2\n ---\n Directory: ./rma\n File: lock_dt\n Num-procs: 2\n- Date: \"Sun Jan 5 02:03:06 2020\"\n+ Date: \"Mon Dec 3 17:43:07 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 382 - ./rma/lock_dt_flush 2\n ---\n Directory: ./rma\n File: lock_dt_flush\n Num-procs: 2\n- Date: \"Sun Jan 5 02:03:06 2020\"\n+ Date: \"Mon Dec 3 17:43:08 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n+## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n+## MPIR_Init_thread(565)..............: \n+## MPID_Init(224).....................: channel initialization failed\n+## MPIDI_CH3_Init(105)................: \n+## MPID_nem_init(324).................: \n+## MPID_nem_tcp_init(178).............: \n+## MPID_nem_tcp_get_business_card(425): \n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 383 - ./rma/lock_dt_flushlocal 2\n ---\n Directory: ./rma\n File: lock_dt_flushlocal\n Num-procs: 2\n- Date: \"Sun Jan 5 02:03:07 2020\"\n+ Date: \"Mon Dec 3 17:43:08 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 384 - ./rma/lockall_dt 2\n ---\n Directory: ./rma\n File: lockall_dt\n Num-procs: 2\n- Date: \"Sun Jan 5 02:03:07 2020\"\n+ Date: \"Mon Dec 3 17:43:09 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 385 - ./rma/lockall_dt_flush 2\n ---\n Directory: ./rma\n File: lockall_dt_flush\n Num-procs: 2\n- Date: \"Sun Jan 5 02:03:07 2020\"\n+ Date: \"Mon Dec 3 17:43:10 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n-## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n-## MPIR_Init_thread(565)..............: \n-## MPID_Init(224).....................: channel initialization failed\n-## MPIDI_CH3_Init(105)................: \n-## MPID_nem_init(324).................: \n-## MPID_nem_tcp_init(178).............: \n-## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 386 - ./rma/lockall_dt_flushall 2\n ---\n Directory: ./rma\n File: lockall_dt_flushall\n Num-procs: 2\n- Date: \"Sun Jan 5 02:03:07 2020\"\n+ Date: \"Mon Dec 3 17:43:10 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 387 - ./rma/lockall_dt_flushlocal 2\n ---\n Directory: ./rma\n File: lockall_dt_flushlocal\n Num-procs: 2\n- Date: \"Sun Jan 5 02:03:08 2020\"\n+ Date: \"Mon Dec 3 17:43:11 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n-not ok 388 - ./rma/lockall_dt_flushlocalall 2\n- ---\n- Directory: ./rma\n- File: lockall_dt_flushlocalall\n- Num-procs: 2\n- Date: \"Sun Jan 5 02:03:08 2020\"\n- ...\n-## Test output (expected 'No Errors'):\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n+not ok 388 - ./rma/lockall_dt_flushlocalall 2\n+ ---\n+ Directory: ./rma\n+ File: lockall_dt_flushlocalall\n+ Num-procs: 2\n+ Date: \"Mon Dec 3 17:43:11 2018\"\n+ ...\n+## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 389 - ./rma/lock_contention_dt 2\n ---\n Directory: ./rma\n File: lock_contention_dt\n Num-procs: 2\n- Date: \"Sun Jan 5 02:03:08 2020\"\n+ Date: \"Mon Dec 3 17:43:12 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n-## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n-## MPIR_Init_thread(565)..............: \n-## MPID_Init(224).....................: channel initialization failed\n-## MPIDI_CH3_Init(105)................: \n-## MPID_nem_init(324).................: \n-## MPID_nem_tcp_init(178).............: \n-## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 390 - ./rma/transpose4 2\n ---\n Directory: ./rma\n File: transpose4\n Num-procs: 2\n- Date: \"Sun Jan 5 02:03:08 2020\"\n+ Date: \"Mon Dec 3 17:43:12 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n-## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n-## MPIR_Init_thread(565)..............: \n-## MPID_Init(224).....................: channel initialization failed\n-## MPIDI_CH3_Init(105)................: \n-## MPID_nem_init(324).................: \n-## MPID_nem_tcp_init(178).............: \n-## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 391 - ./rma/fetchandadd 2\n ---\n Directory: ./rma\n File: fetchandadd\n Num-procs: 2\n- Date: \"Sun Jan 5 02:03:09 2020\"\n+ Date: \"Mon Dec 3 17:43:13 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 392 - ./rma/fetchandadd_tree 2\n ---\n Directory: ./rma\n File: fetchandadd_tree\n Num-procs: 2\n- Date: \"Sun Jan 5 02:03:09 2020\"\n+ Date: \"Mon Dec 3 17:43:13 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n+## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n+## MPIR_Init_thread(565)..............: \n+## MPID_Init(224).....................: channel initialization failed\n+## MPIDI_CH3_Init(105)................: \n+## MPID_nem_init(324).................: \n+## MPID_nem_tcp_init(178).............: \n+## MPID_nem_tcp_get_business_card(425): \n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 393 - ./rma/wintest 2\n ---\n Directory: ./rma\n File: wintest\n Num-procs: 2\n- Date: \"Sun Jan 5 02:03:09 2020\"\n+ Date: \"Mon Dec 3 17:43:14 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 394 - ./rma/wintest_shm 2\n ---\n Directory: ./rma\n File: wintest_shm\n Num-procs: 2\n- Date: \"Sun Jan 5 02:03:09 2020\"\n+ Date: \"Mon Dec 3 17:43:15 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n-## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n-## MPIR_Init_thread(565)..............: \n-## MPID_Init(224).....................: channel initialization failed\n-## MPIDI_CH3_Init(105)................: \n-## MPID_nem_init(324).................: \n-## MPID_nem_tcp_init(178).............: \n-## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 395 - ./rma/contig_displ 1\n ---\n Directory: ./rma\n File: contig_displ\n Num-procs: 1\n- Date: \"Sun Jan 5 02:03:09 2020\"\n+ Date: \"Mon Dec 3 17:43:15 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 396 - ./rma/test1_am 2\n ---\n Directory: ./rma\n File: test1_am\n Num-procs: 2\n- Date: \"Sun Jan 5 02:03:10 2020\"\n+ Date: \"Mon Dec 3 17:43:16 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 397 - ./rma/test2_am 2\n ---\n Directory: ./rma\n File: test2_am\n Num-procs: 2\n- Date: \"Sun Jan 5 02:03:10 2020\"\n+ Date: \"Mon Dec 3 17:43:16 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n-## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n-## MPIR_Init_thread(565)..............: \n-## MPID_Init(224).....................: channel initialization failed\n-## MPIDI_CH3_Init(105)................: \n-## MPID_nem_init(324).................: \n-## MPID_nem_tcp_init(178).............: \n-## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 398 - ./rma/test2_am_shm 2\n ---\n Directory: ./rma\n File: test2_am_shm\n Num-procs: 2\n- Date: \"Sun Jan 5 02:03:10 2020\"\n+ Date: \"Mon Dec 3 17:43:17 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n-## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n-## MPIR_Init_thread(565)..............: \n-## MPID_Init(224).....................: channel initialization failed\n-## MPIDI_CH3_Init(105)................: \n-## MPID_nem_init(324).................: \n-## MPID_nem_tcp_init(178).............: \n-## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 399 - ./rma/test3_am 2\n ---\n Directory: ./rma\n File: test3_am\n Num-procs: 2\n- Date: \"Sun Jan 5 02:03:10 2020\"\n+ Date: \"Mon Dec 3 17:43:18 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 400 - ./rma/test3_am_shm 2\n ---\n Directory: ./rma\n File: test3_am_shm\n Num-procs: 2\n- Date: \"Sun Jan 5 02:03:11 2020\"\n+ Date: \"Mon Dec 3 17:43:18 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 401 - ./rma/test4_am 2\n ---\n Directory: ./rma\n File: test4_am\n Num-procs: 2\n- Date: \"Sun Jan 5 02:03:11 2020\"\n+ Date: \"Mon Dec 3 17:43:19 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 402 - ./rma/test5_am 2\n ---\n Directory: ./rma\n File: test5_am\n Num-procs: 2\n- Date: \"Sun Jan 5 02:03:11 2020\"\n+ Date: \"Mon Dec 3 17:43:19 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n-## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n-## MPIR_Init_thread(565)..............: \n-## MPID_Init(224).....................: channel initialization failed\n-## MPIDI_CH3_Init(105)................: \n-## MPID_nem_init(324).................: \n-## MPID_nem_tcp_init(178).............: \n-## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 403 - ./rma/fetchandadd_am 2\n ---\n Directory: ./rma\n File: fetchandadd_am\n Num-procs: 2\n- Date: \"Sun Jan 5 02:03:11 2020\"\n+ Date: \"Mon Dec 3 17:43:20 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 404 - ./rma/fetchandadd_tree_am 2\n ---\n Directory: ./rma\n File: fetchandadd_tree_am\n Num-procs: 2\n- Date: \"Sun Jan 5 02:03:11 2020\"\n+ Date: \"Mon Dec 3 17:43:20 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n-## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n-## MPIR_Init_thread(565)..............: \n-## MPID_Init(224).....................: channel initialization failed\n-## MPIDI_CH3_Init(105)................: \n-## MPID_nem_init(324).................: \n-## MPID_nem_tcp_init(178).............: \n-## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 405 - ./rma/accfence2_am 2\n ---\n Directory: ./rma\n File: accfence2_am\n Num-procs: 2\n- Date: \"Sun Jan 5 02:03:12 2020\"\n+ Date: \"Mon Dec 3 17:43:21 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n+## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n+## MPIR_Init_thread(565)..............: \n+## MPID_Init(224).....................: channel initialization failed\n+## MPIDI_CH3_Init(105)................: \n+## MPID_nem_init(324).................: \n+## MPID_nem_tcp_init(178).............: \n+## MPID_nem_tcp_get_business_card(425): \n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 406 - ./rma/test1_dt 2\n ---\n Directory: ./rma\n File: test1_dt\n Num-procs: 2\n- Date: \"Sun Jan 5 02:03:12 2020\"\n+ Date: \"Mon Dec 3 17:43:22 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n+## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n+## MPIR_Init_thread(565)..............: \n+## MPID_Init(224).....................: channel initialization failed\n+## MPIDI_CH3_Init(105)................: \n+## MPID_nem_init(324).................: \n+## MPID_nem_tcp_init(178).............: \n+## MPID_nem_tcp_get_business_card(425): \n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 407 - ./rma/nullpscw 2\n ---\n Directory: ./rma\n File: nullpscw\n Num-procs: 2\n- Date: \"Sun Jan 5 02:03:12 2020\"\n+ Date: \"Mon Dec 3 17:43:22 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 408 - ./rma/nullpscw_shm 2\n ---\n Directory: ./rma\n File: nullpscw_shm\n Num-procs: 2\n- Date: \"Sun Jan 5 02:03:12 2020\"\n+ Date: \"Mon Dec 3 17:43:23 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 409 - ./rma/attrorderwin 1\n ---\n Directory: ./rma\n File: attrorderwin\n Num-procs: 1\n- Date: \"Sun Jan 5 02:03:12 2020\"\n+ Date: \"Mon Dec 3 17:43:23 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 410 - ./rma/wincall 2\n ---\n Directory: ./rma\n File: wincall\n Num-procs: 2\n- Date: \"Sun Jan 5 02:03:13 2020\"\n+ Date: \"Mon Dec 3 17:43:24 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n-## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n-## MPIR_Init_thread(565)..............: \n-## MPID_Init(224).....................: channel initialization failed\n-## MPIDI_CH3_Init(105)................: \n-## MPID_nem_init(324).................: \n-## MPID_nem_tcp_init(178).............: \n-## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 411 - ./rma/baseattrwin 1\n ---\n Directory: ./rma\n File: baseattrwin\n Num-procs: 1\n- Date: \"Sun Jan 5 02:03:13 2020\"\n+ Date: \"Mon Dec 3 17:43:24 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 412 - ./rma/fkeyvalwin 1\n ---\n Directory: ./rma\n File: fkeyvalwin\n Num-procs: 1\n- Date: \"Sun Jan 5 02:03:13 2020\"\n+ Date: \"Mon Dec 3 17:43:25 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 413 - ./rma/selfrma 1\n ---\n Directory: ./rma\n File: selfrma\n Num-procs: 1\n- Date: \"Sun Jan 5 02:03:13 2020\"\n+ Date: \"Mon Dec 3 17:43:25 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 414 - ./rma/mixedsync 2\n ---\n Directory: ./rma\n File: mixedsync\n Num-procs: 2\n- Date: \"Sun Jan 5 02:03:14 2020\"\n+ Date: \"Mon Dec 3 17:43:26 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 415 - ./rma/epochtest 2\n ---\n Directory: ./rma\n File: epochtest\n Num-procs: 2\n- Date: \"Sun Jan 5 02:03:14 2020\"\n+ Date: \"Mon Dec 3 17:43:27 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n+## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n+## MPIR_Init_thread(565)..............: \n+## MPID_Init(224).....................: channel initialization failed\n+## MPIDI_CH3_Init(105)................: \n+## MPID_nem_init(324).................: \n+## MPID_nem_tcp_init(178).............: \n+## MPID_nem_tcp_get_business_card(425): \n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 416 - ./rma/locknull 2\n ---\n Directory: ./rma\n File: locknull\n Num-procs: 2\n- Date: \"Sun Jan 5 02:03:14 2020\"\n+ Date: \"Mon Dec 3 17:43:27 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 417 - ./rma/rmanull 2\n ---\n Directory: ./rma\n File: rmanull\n Num-procs: 2\n- Date: \"Sun Jan 5 02:03:14 2020\"\n+ Date: \"Mon Dec 3 17:43:28 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n-## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n-## MPIR_Init_thread(565)..............: \n-## MPID_Init(224).....................: channel initialization failed\n-## MPIDI_CH3_Init(105)................: \n-## MPID_nem_init(324).................: \n-## MPID_nem_tcp_init(178).............: \n-## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 418 - ./rma/rmazero 2\n ---\n Directory: ./rma\n File: rmazero\n Num-procs: 2\n- Date: \"Sun Jan 5 02:03:15 2020\"\n+ Date: \"Mon Dec 3 17:43:28 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 419 - ./rma/strided_acc_indexed 2\n ---\n Directory: ./rma\n File: strided_acc_indexed\n Num-procs: 2\n- Date: \"Sun Jan 5 02:03:15 2020\"\n+ Date: \"Mon Dec 3 17:43:29 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n-## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n-## MPIR_Init_thread(565)..............: \n-## MPID_Init(224).....................: channel initialization failed\n-## MPIDI_CH3_Init(105)................: \n-## MPID_nem_init(324).................: \n-## MPID_nem_tcp_init(178).............: \n-## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 420 - ./rma/strided_acc_onelock 2\n ---\n Directory: ./rma\n File: strided_acc_onelock\n Num-procs: 2\n- Date: \"Sun Jan 5 02:03:15 2020\"\n+ Date: \"Mon Dec 3 17:43:30 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 421 - ./rma/strided_acc_subarray 2\n ---\n Directory: ./rma\n File: strided_acc_subarray\n Num-procs: 2\n- Date: \"Sun Jan 5 02:03:15 2020\"\n+ Date: \"Mon Dec 3 17:43:30 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n-## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n-## MPIR_Init_thread(565)..............: \n-## MPID_Init(224).....................: channel initialization failed\n-## MPIDI_CH3_Init(105)................: \n-## MPID_nem_init(324).................: \n-## MPID_nem_tcp_init(178).............: \n-## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 422 - ./rma/strided_get_indexed 2\n ---\n Directory: ./rma\n File: strided_get_indexed\n Num-procs: 2\n- Date: \"Sun Jan 5 02:03:16 2020\"\n+ Date: \"Mon Dec 3 17:43:31 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n-## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n-## MPIR_Init_thread(565)..............: \n-## MPID_Init(224).....................: channel initialization failed\n-## MPIDI_CH3_Init(105)................: \n-## MPID_nem_init(324).................: \n-## MPID_nem_tcp_init(178).............: \n-## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 423 - ./rma/strided_putget_indexed 2\n ---\n Directory: ./rma\n File: strided_putget_indexed\n Num-procs: 2\n- Date: \"Sun Jan 5 02:03:16 2020\"\n+ Date: \"Mon Dec 3 17:43:31 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 424 - ./rma/strided_putget_indexed_shared 2\n ---\n Directory: ./rma\n File: strided_putget_indexed_shared\n Num-procs: 2\n- Date: \"Sun Jan 5 02:03:16 2020\"\n+ Date: \"Mon Dec 3 17:43:32 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 425 - ./rma/strided_getacc_indexed 2\n ---\n Directory: ./rma\n File: strided_getacc_indexed\n Num-procs: 2\n- Date: \"Sun Jan 5 02:03:17 2020\"\n+ Date: \"Mon Dec 3 17:43:32 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 426 - ./rma/strided_getacc_indexed_shared 2\n ---\n Directory: ./rma\n File: strided_getacc_indexed_shared\n Num-procs: 2\n- Date: \"Sun Jan 5 02:03:17 2020\"\n+ Date: \"Mon Dec 3 17:43:33 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n-## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n-## MPIR_Init_thread(565)..............: \n-## MPID_Init(224).....................: channel initialization failed\n-## MPIDI_CH3_Init(105)................: \n-## MPID_nem_init(324).................: \n-## MPID_nem_tcp_init(178).............: \n-## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 427 - ./rma/window_creation 2\n ---\n Directory: ./rma\n File: window_creation\n Num-procs: 2\n- Date: \"Sun Jan 5 02:03:17 2020\"\n+ Date: \"Mon Dec 3 17:43:34 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n+## Fatal error in PMPI_Init: Other MPI error, error stack:\n+## MPIR_Init_thread(565)..............: \n+## MPID_Init(224).....................: channel initialization failed\n+## MPIDI_CH3_Init(105)................: \n+## MPID_nem_init(324).................: \n+## MPID_nem_tcp_init(178).............: \n+## MPID_nem_tcp_get_business_card(425): \n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 428 - ./rma/contention_put 2\n ---\n Directory: ./rma\n File: contention_put\n Num-procs: 2\n- Date: \"Sun Jan 5 02:03:18 2020\"\n+ Date: \"Mon Dec 3 17:43:34 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 429 - ./rma/contention_putget 2\n ---\n Directory: ./rma\n File: contention_putget\n Num-procs: 2\n- Date: \"Sun Jan 5 02:03:18 2020\"\n+ Date: \"Mon Dec 3 17:43:35 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n-## Fatal error in PMPI_Init: Other MPI error, error stack:\n-## MPIR_Init_thread(565)..............: \n-## MPID_Init(224).....................: channel initialization failed\n-## MPIDI_CH3_Init(105)................: \n-## MPID_nem_init(324).................: \n-## MPID_nem_tcp_init(178).............: \n-## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 430 - ./rma/put_base 2\n ---\n Directory: ./rma\n File: put_base\n Num-procs: 2\n- Date: \"Sun Jan 5 02:03:18 2020\"\n+ Date: \"Mon Dec 3 17:43:35 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n-## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n-## MPIR_Init_thread(565)..............: \n-## MPID_Init(224).....................: channel initialization failed\n-## MPIDI_CH3_Init(105)................: \n-## MPID_nem_init(324).................: \n-## MPID_nem_tcp_init(178).............: \n-## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 431 - ./rma/put_bottom 2\n ---\n Directory: ./rma\n File: put_bottom\n Num-procs: 2\n- Date: \"Sun Jan 5 02:03:18 2020\"\n+ Date: \"Mon Dec 3 17:43:36 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n-## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n-## MPIR_Init_thread(565)..............: \n-## MPID_Init(224).....................: channel initialization failed\n-## MPIDI_CH3_Init(105)................: \n-## MPID_nem_init(324).................: \n-## MPID_nem_tcp_init(178).............: \n-## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 432 - ./rma/win_flavors 2\n ---\n Directory: ./rma\n File: win_flavors\n Num-procs: 2\n- Date: \"Sun Jan 5 02:03:19 2020\"\n+ Date: \"Mon Dec 3 17:43:37 2018\"\n ...\n ## Test output (expected 'No Errors'):\n-## Fatal error in PMPI_Init: Other MPI error, error stack:\n-## MPIR_Init_thread(565)..............: \n-## MPID_Init(224).....................: channel initialization failed\n-## MPIDI_CH3_Init(105)................: \n-## MPID_nem_init(324).................: \n-## MPID_nem_tcp_init(178).............: \n-## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n not ok 433 - ./rma/win_flavors 2\n ---\n Directory: ./rma\n File: win_flavors\n Num-procs: 2\n- Date: \"Sun Jan 5 02:03:19 2020\"\n+ Date: \"Mon Dec 3 17:43:37 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n-## Fatal error in PMPI_Init: Other MPI error, error stack:\n-## MPIR_Init_thread(565)..............: \n-## MPID_Init(224).....................: channel initialization failed\n-## MPIDI_CH3_Init(105)................: \n-## MPID_nem_init(324).................: \n-## MPID_nem_tcp_init(178).............: \n-## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 434 - ./rma/manyrma2 2\n ---\n Directory: ./rma\n File: manyrma2\n Num-procs: 2\n- Date: \"Sun Jan 5 02:03:19 2020\"\n+ Date: \"Mon Dec 3 17:43:38 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n-not ok 435 - ./rma/manyrma2_shm 2\n- ---\n- Directory: ./rma\n- File: manyrma2_shm\n- Num-procs: 2\n- Date: \"Sun Jan 5 02:03:19 2020\"\n- ...\n-## Test output (expected 'No Errors'):\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n ## Fatal error in PMPI_Init: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n-not ok 436 - ./rma/manyrma3 2\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n+not ok 435 - ./rma/manyrma2_shm 2\n ---\n Directory: ./rma\n- File: manyrma3\n+ File: manyrma2_shm\n Num-procs: 2\n- Date: \"Sun Jan 5 02:03:20 2020\"\n+ Date: \"Mon Dec 3 17:43:38 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n ## Fatal error in PMPI_Init: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n-not ok 437 - ./rma/win_shared 2\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n+not ok 436 - ./rma/manyrma3 2\n ---\n Directory: ./rma\n- File: win_shared\n+ File: manyrma3\n Num-procs: 2\n- Date: \"Sun Jan 5 02:03:20 2020\"\n+ Date: \"Mon Dec 3 17:43:39 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n+not ok 437 - ./rma/win_shared 2\n+ ---\n+ Directory: ./rma\n+ File: win_shared\n+ Num-procs: 2\n+ Date: \"Mon Dec 3 17:43:40 2018\"\n+ ...\n+## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 438 - ./rma/win_shared_create_allocshm 2\n ---\n Directory: ./rma\n File: win_shared_create_allocshm\n Num-procs: 2\n- Date: \"Sun Jan 5 02:03:20 2020\"\n+ Date: \"Mon Dec 3 17:43:40 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n+## Fatal error in PMPI_Init: Other MPI error, error stack:\n+## MPIR_Init_thread(565)..............: \n+## MPID_Init(224).....................: channel initialization failed\n+## MPIDI_CH3_Init(105)................: \n+## MPID_nem_init(324).................: \n+## MPID_nem_tcp_init(178).............: \n+## MPID_nem_tcp_get_business_card(425): \n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 439 - ./rma/win_shared_create_no_allocshm 2\n ---\n Directory: ./rma\n File: win_shared_create_no_allocshm\n Num-procs: 2\n- Date: \"Sun Jan 5 02:03:20 2020\"\n+ Date: \"Mon Dec 3 17:43:41 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n ## Fatal error in PMPI_Init: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 440 - ./rma/win_shared_noncontig 2\n ---\n Directory: ./rma\n File: win_shared_noncontig\n Num-procs: 2\n- Date: \"Sun Jan 5 02:03:21 2020\"\n+ Date: \"Mon Dec 3 17:43:42 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n-## Fatal error in PMPI_Init: Other MPI error, error stack:\n-## MPIR_Init_thread(565)..............: \n-## MPID_Init(224).....................: channel initialization failed\n-## MPIDI_CH3_Init(105)................: \n-## MPID_nem_init(324).................: \n-## MPID_nem_tcp_init(178).............: \n-## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 441 - ./rma/win_shared_noncontig_put 2\n ---\n Directory: ./rma\n File: win_shared_noncontig_put\n Num-procs: 2\n- Date: \"Sun Jan 5 02:03:21 2020\"\n+ Date: \"Mon Dec 3 17:43:42 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n ## Fatal error in PMPI_Init: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 442 - ./rma/win_zero 2\n ---\n Directory: ./rma\n File: win_zero\n Num-procs: 2\n- Date: \"Sun Jan 5 02:03:21 2020\"\n+ Date: \"Mon Dec 3 17:43:43 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n-## Fatal error in PMPI_Init: Other MPI error, error stack:\n-## MPIR_Init_thread(565)..............: \n-## MPID_Init(224).....................: channel initialization failed\n-## MPIDI_CH3_Init(105)................: \n-## MPID_nem_init(324).................: \n-## MPID_nem_tcp_init(178).............: \n-## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 443 - ./rma/win_dynamic_acc 2\n ---\n Directory: ./rma\n File: win_dynamic_acc\n Num-procs: 2\n- Date: \"Sun Jan 5 02:03:21 2020\"\n+ Date: \"Mon Dec 3 17:43:44 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n ## Fatal error in PMPI_Init: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 444 - ./rma/get_acc_local 1\n ---\n Directory: ./rma\n File: get_acc_local\n Num-procs: 1\n- Date: \"Sun Jan 5 02:03:22 2020\"\n+ Date: \"Mon Dec 3 17:43:44 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 445 - ./rma/linked_list 2\n ---\n Directory: ./rma\n File: linked_list\n Num-procs: 2\n- Date: \"Sun Jan 5 02:03:22 2020\"\n+ Date: \"Mon Dec 3 17:43:45 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n ## Fatal error in PMPI_Init: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 446 - ./rma/linked_list_fop 2\n ---\n Directory: ./rma\n File: linked_list_fop\n Num-procs: 2\n- Date: \"Sun Jan 5 02:03:22 2020\"\n+ Date: \"Mon Dec 3 17:43:45 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 447 - ./rma/compare_and_swap 2\n ---\n Directory: ./rma\n File: compare_and_swap\n Num-procs: 2\n- Date: \"Sun Jan 5 02:03:22 2020\"\n+ Date: \"Mon Dec 3 17:43:46 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n ## Fatal error in PMPI_Init: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 448 - ./rma/fetch_and_op_char 2\n ---\n Directory: ./rma\n File: fetch_and_op_char\n Num-procs: 2\n- Date: \"Sun Jan 5 02:03:22 2020\"\n+ Date: \"Mon Dec 3 17:43:47 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 449 - ./rma/fetch_and_op_short 2\n ---\n Directory: ./rma\n File: fetch_and_op_short\n Num-procs: 2\n- Date: \"Sun Jan 5 02:03:23 2020\"\n+ Date: \"Mon Dec 3 17:43:47 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n+## Fatal error in PMPI_Init: Other MPI error, error stack:\n+## MPIR_Init_thread(565)..............: \n+## MPID_Init(224).....................: channel initialization failed\n+## MPIDI_CH3_Init(105)................: \n+## MPID_nem_init(324).................: \n+## MPID_nem_tcp_init(178).............: \n+## MPID_nem_tcp_get_business_card(425): \n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 450 - ./rma/fetch_and_op_int 2\n ---\n Directory: ./rma\n File: fetch_and_op_int\n Num-procs: 2\n- Date: \"Sun Jan 5 02:03:23 2020\"\n+ Date: \"Mon Dec 3 17:43:48 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n-not ok 451 - ./rma/fetch_and_op_long 2\n- ---\n- Directory: ./rma\n- File: fetch_and_op_long\n- Num-procs: 2\n- Date: \"Sun Jan 5 02:03:23 2020\"\n- ...\n-## Test output (expected 'No Errors'):\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n ## Fatal error in PMPI_Init: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n+not ok 451 - ./rma/fetch_and_op_long 2\n+ ---\n+ Directory: ./rma\n+ File: fetch_and_op_long\n+ Num-procs: 2\n+ Date: \"Mon Dec 3 17:43:49 2018\"\n+ ...\n+## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 452 - ./rma/fetch_and_op_double 2\n ---\n Directory: ./rma\n File: fetch_and_op_double\n Num-procs: 2\n- Date: \"Sun Jan 5 02:03:24 2020\"\n+ Date: \"Mon Dec 3 17:43:49 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 453 - ./rma/fetch_and_op_long_double 2\n ---\n Directory: ./rma\n File: fetch_and_op_long_double\n Num-procs: 2\n- Date: \"Sun Jan 5 02:03:25 2020\"\n+ Date: \"Mon Dec 3 17:43:50 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 454 - ./rma/get_accumulate_double 2\n ---\n Directory: ./rma\n File: get_accumulate_double\n Num-procs: 2\n- Date: \"Sun Jan 5 02:03:26 2020\"\n+ Date: \"Mon Dec 3 17:43:50 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n+## Fatal error in PMPI_Init: Other MPI error, error stack:\n+## MPIR_Init_thread(565)..............: \n+## MPID_Init(224).....................: channel initialization failed\n+## MPIDI_CH3_Init(105)................: \n+## MPID_nem_init(324).................: \n+## MPID_nem_tcp_init(178).............: \n+## MPID_nem_tcp_get_business_card(425): \n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 455 - ./rma/get_accumulate_double_derived 2\n ---\n Directory: ./rma\n File: get_accumulate_double_derived\n Num-procs: 2\n- Date: \"Sun Jan 5 02:03:27 2020\"\n+ Date: \"Mon Dec 3 17:43:51 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 456 - ./rma/get_accumulate_int 2\n ---\n Directory: ./rma\n File: get_accumulate_int\n Num-procs: 2\n- Date: \"Sun Jan 5 02:03:28 2020\"\n+ Date: \"Mon Dec 3 17:43:52 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n-## Fatal error in PMPI_Init: Other MPI error, error stack:\n-## MPIR_Init_thread(565)..............: \n-## MPID_Init(224).....................: channel initialization failed\n-## MPIDI_CH3_Init(105)................: \n-## MPID_nem_init(324).................: \n-## MPID_nem_tcp_init(178).............: \n-## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 457 - ./rma/get_accumulate_int_derived 2\n ---\n Directory: ./rma\n File: get_accumulate_int_derived\n Num-procs: 2\n- Date: \"Sun Jan 5 02:03:29 2020\"\n+ Date: \"Mon Dec 3 17:43:52 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 458 - ./rma/get_accumulate_long 2\n ---\n Directory: ./rma\n File: get_accumulate_long\n Num-procs: 2\n- Date: \"Sun Jan 5 02:03:30 2020\"\n+ Date: \"Mon Dec 3 17:43:53 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n+## Fatal error in PMPI_Init: Other MPI error, error stack:\n+## MPIR_Init_thread(565)..............: \n+## MPID_Init(224).....................: channel initialization failed\n+## MPIDI_CH3_Init(105)................: \n+## MPID_nem_init(324).................: \n+## MPID_nem_tcp_init(178).............: \n+## MPID_nem_tcp_get_business_card(425): \n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 459 - ./rma/get_accumulate_long_derived 2\n ---\n Directory: ./rma\n File: get_accumulate_long_derived\n Num-procs: 2\n- Date: \"Sun Jan 5 02:03:31 2020\"\n+ Date: \"Mon Dec 3 17:43:53 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n+## Fatal error in PMPI_Init: Other MPI error, error stack:\n+## MPIR_Init_thread(565)..............: \n+## MPID_Init(224).....................: channel initialization failed\n+## MPIDI_CH3_Init(105)................: \n+## MPID_nem_init(324).................: \n+## MPID_nem_tcp_init(178).............: \n+## MPID_nem_tcp_get_business_card(425): \n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 460 - ./rma/get_accumulate_short 2\n ---\n Directory: ./rma\n File: get_accumulate_short\n Num-procs: 2\n- Date: \"Sun Jan 5 02:03:32 2020\"\n+ Date: \"Mon Dec 3 17:43:54 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 461 - ./rma/get_accumulate_short_derived 2\n ---\n Directory: ./rma\n File: get_accumulate_short_derived\n Num-procs: 2\n- Date: \"Sun Jan 5 02:03:33 2020\"\n+ Date: \"Mon Dec 3 17:43:55 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n-## Fatal error in PMPI_Init: Other MPI error, error stack:\n-## MPIR_Init_thread(565)..............: \n-## MPID_Init(224).....................: channel initialization failed\n-## MPIDI_CH3_Init(105)................: \n-## MPID_nem_init(324).................: \n-## MPID_nem_tcp_init(178).............: \n-## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 462 - ./rma/flush 2\n ---\n Directory: ./rma\n File: flush\n Num-procs: 2\n- Date: \"Sun Jan 5 02:03:34 2020\"\n+ Date: \"Mon Dec 3 17:43:55 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n+## Fatal error in PMPI_Init: Other MPI error, error stack:\n+## MPIR_Init_thread(565)..............: \n+## MPID_Init(224).....................: channel initialization failed\n+## MPIDI_CH3_Init(105)................: \n+## MPID_nem_init(324).................: \n+## MPID_nem_tcp_init(178).............: \n+## MPID_nem_tcp_get_business_card(425): \n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 463 - ./rma/reqops 2\n ---\n Directory: ./rma\n File: reqops\n Num-procs: 2\n- Date: \"Sun Jan 5 02:03:35 2020\"\n+ Date: \"Mon Dec 3 17:43:56 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n-not ok 464 - ./rma/req_example 2\n- ---\n- Directory: ./rma\n- File: req_example\n- Num-procs: 2\n- Date: \"Sun Jan 5 02:03:36 2020\"\n- ...\n-## Test output (expected 'No Errors'):\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n ## Fatal error in PMPI_Init: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n-not ok 465 - ./rma/req_example_shm 2\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n+not ok 464 - ./rma/req_example 2\n ---\n Directory: ./rma\n- File: req_example_shm\n+ File: req_example\n Num-procs: 2\n- Date: \"Sun Jan 5 02:03:37 2020\"\n+ Date: \"Mon Dec 3 17:43:56 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n ## Fatal error in PMPI_Init: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n-not ok 466 - ./rma/rput_local_comp 2\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n+not ok 465 - ./rma/req_example_shm 2\n ---\n Directory: ./rma\n- File: rput_local_comp\n+ File: req_example_shm\n Num-procs: 2\n- Date: \"Sun Jan 5 02:03:38 2020\"\n+ Date: \"Mon Dec 3 17:43:57 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n-## Fatal error in PMPI_Init: Other MPI error, error stack:\n-## MPIR_Init_thread(565)..............: \n-## MPID_Init(224).....................: channel initialization failed\n-## MPIDI_CH3_Init(105)................: \n-## MPID_nem_init(324).................: \n-## MPID_nem_tcp_init(178).............: \n-## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n-not ok 467 - ./rma/racc_local_comp 2\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n+not ok 466 - ./rma/rput_local_comp 2\n ---\n Directory: ./rma\n- File: racc_local_comp\n+ File: rput_local_comp\n Num-procs: 2\n- Date: \"Sun Jan 5 02:03:39 2020\"\n+ Date: \"Mon Dec 3 17:43:57 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n-not ok 468 - ./rma/win_info 2\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n+not ok 467 - ./rma/racc_local_comp 2\n ---\n Directory: ./rma\n- File: win_info\n+ File: racc_local_comp\n Num-procs: 2\n- Date: \"Sun Jan 5 02:03:40 2020\"\n+ Date: \"Mon Dec 3 17:43:58 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n ## Fatal error in PMPI_Init: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n-not ok 469 - ./rma/linked_list_lockall 2\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n+not ok 468 - ./rma/win_info 2\n ---\n Directory: ./rma\n- File: linked_list_lockall\n+ File: win_info\n Num-procs: 2\n- Date: \"Sun Jan 5 02:03:40 2020\"\n+ Date: \"Mon Dec 3 17:43:58 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n+not ok 469 - ./rma/linked_list_lockall 2\n+ ---\n+ Directory: ./rma\n+ File: linked_list_lockall\n+ Num-procs: 2\n+ Date: \"Mon Dec 3 17:43:59 2018\"\n+ ...\n+## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 470 - ./rma/pscw_ordering 2\n ---\n Directory: ./rma\n File: pscw_ordering\n Num-procs: 2\n- Date: \"Sun Jan 5 02:03:42 2020\"\n+ Date: \"Mon Dec 3 17:44:00 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 471 - ./rma/pscw_ordering_shm 2\n ---\n Directory: ./rma\n File: pscw_ordering_shm\n Num-procs: 2\n- Date: \"Sun Jan 5 02:03:43 2020\"\n+ Date: \"Mon Dec 3 17:44:00 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n+## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n+## MPIR_Init_thread(565)..............: \n+## MPID_Init(224).....................: channel initialization failed\n+## MPIDI_CH3_Init(105)................: \n+## MPID_nem_init(324).................: \n+## MPID_nem_tcp_init(178).............: \n+## MPID_nem_tcp_get_business_card(425): \n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 472 - ./rma/linked_list_bench_lock_all 2\n ---\n Directory: ./rma\n File: linked_list_bench_lock_all\n Num-procs: 2\n- Date: \"Sun Jan 5 02:03:44 2020\"\n+ Date: \"Mon Dec 3 17:44:01 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n+## Fatal error in PMPI_Init: Other MPI error, error stack:\n+## MPIR_Init_thread(565)..............: \n+## MPID_Init(224).....................: channel initialization failed\n+## MPIDI_CH3_Init(105)................: \n+## MPID_nem_init(324).................: \n+## MPID_nem_tcp_init(178).............: \n+## MPID_nem_tcp_get_business_card(425): \n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 473 - ./rma/linked_list_bench_lock_excl 2\n ---\n Directory: ./rma\n File: linked_list_bench_lock_excl\n Num-procs: 2\n- Date: \"Sun Jan 5 02:03:45 2020\"\n+ Date: \"Mon Dec 3 17:44:02 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n ## Fatal error in PMPI_Init: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 474 - ./rma/linked_list_bench_lock_shr 2\n ---\n Directory: ./rma\n File: linked_list_bench_lock_shr\n Num-procs: 2\n- Date: \"Sun Jan 5 02:03:45 2020\"\n+ Date: \"Mon Dec 3 17:44:02 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n+## Fatal error in PMPI_Init: Other MPI error, error stack:\n+## MPIR_Init_thread(565)..............: \n+## MPID_Init(224).....................: channel initialization failed\n+## MPIDI_CH3_Init(105)................: \n+## MPID_nem_init(324).................: \n+## MPID_nem_tcp_init(178).............: \n+## MPID_nem_tcp_get_business_card(425): \n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 475 - ./rma/linked_list_bench_lock_shr_nocheck 2\n ---\n Directory: ./rma\n File: linked_list_bench_lock_shr_nocheck\n Num-procs: 2\n- Date: \"Sun Jan 5 02:03:46 2020\"\n+ Date: \"Mon Dec 3 17:44:03 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 476 - ./rma/mutex_bench 2\n ---\n Directory: ./rma\n File: mutex_bench\n Num-procs: 2\n- Date: \"Sun Jan 5 02:03:47 2020\"\n+ Date: \"Mon Dec 3 17:44:04 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n-not ok 477 - ./rma/mutex_bench_shared 2\n- ---\n- Directory: ./rma\n- File: mutex_bench_shared\n- Num-procs: 2\n- Date: \"Sun Jan 5 02:03:48 2020\"\n- ...\n-## Test output (expected 'No Errors'):\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n ## Fatal error in PMPI_Init: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n+not ok 477 - ./rma/mutex_bench_shared 2\n+ ---\n+ Directory: ./rma\n+ File: mutex_bench_shared\n+ Num-procs: 2\n+ Date: \"Mon Dec 3 17:44:04 2018\"\n+ ...\n+## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 478 - ./rma/mutex_bench_shm 2\n ---\n Directory: ./rma\n File: mutex_bench_shm\n Num-procs: 2\n- Date: \"Sun Jan 5 02:03:49 2020\"\n+ Date: \"Mon Dec 3 17:44:05 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 479 - ./rma/mutex_bench_shm_ordered 2\n ---\n Directory: ./rma\n File: mutex_bench_shm_ordered\n Num-procs: 2\n- Date: \"Sun Jan 5 02:03:49 2020\"\n+ Date: \"Mon Dec 3 17:44:05 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 480 - ./rma/rma-contig 2\n ---\n Directory: ./rma\n File: rma-contig\n Num-procs: 2\n- Date: \"Sun Jan 5 02:03:50 2020\"\n+ Date: \"Mon Dec 3 17:44:06 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n ## Fatal error in PMPI_Init: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 481 - ./rma/badrma 2\n ---\n Directory: ./rma\n File: badrma\n Num-procs: 2\n- Date: \"Sun Jan 5 02:03:51 2020\"\n+ Date: \"Mon Dec 3 17:44:06 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n ## Fatal error in PMPI_Init: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 482 - ./rma/acc-loc 2\n ---\n Directory: ./rma\n File: acc-loc\n Num-procs: 2\n- Date: \"Sun Jan 5 02:03:52 2020\"\n+ Date: \"Mon Dec 3 17:44:07 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n-## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n-## MPIR_Init_thread(565)..............: \n-## MPID_Init(224).....................: channel initialization failed\n-## MPIDI_CH3_Init(105)................: \n-## MPID_nem_init(324).................: \n-## MPID_nem_tcp_init(178).............: \n-## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 483 - ./rma/fence_shm 2\n ---\n Directory: ./rma\n File: fence_shm\n Num-procs: 2\n- Date: \"Sun Jan 5 02:03:52 2020\"\n+ Date: \"Mon Dec 3 17:44:08 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 484 - ./rma/win_shared_zerobyte 2\n ---\n Directory: ./rma\n File: win_shared_zerobyte\n Num-procs: 2\n- Date: \"Sun Jan 5 02:03:54 2020\"\n+ Date: \"Mon Dec 3 17:44:08 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n ## Fatal error in PMPI_Init: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 485 - ./rma/win_shared_put_flush_get 2\n ---\n Directory: ./rma\n File: win_shared_put_flush_get\n Num-procs: 2\n- Date: \"Sun Jan 5 02:03:55 2020\"\n+ Date: \"Mon Dec 3 17:44:09 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n-## Fatal error in PMPI_Init: Other MPI error, error stack:\n-## MPIR_Init_thread(565)..............: \n-## MPID_Init(224).....................: channel initialization failed\n-## MPIDI_CH3_Init(105)................: \n-## MPID_nem_init(324).................: \n-## MPID_nem_tcp_init(178).............: \n-## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 486 - ./rma/get-struct 2\n ---\n Directory: ./rma\n File: get-struct\n Num-procs: 2\n- Date: \"Sun Jan 5 02:03:56 2020\"\n+ Date: \"Mon Dec 3 17:44:09 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n+## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n+## MPIR_Init_thread(565)..............: \n+## MPID_Init(224).....................: channel initialization failed\n+## MPIDI_CH3_Init(105)................: \n+## MPID_nem_init(324).................: \n+## MPID_nem_tcp_init(178).............: \n+## MPID_nem_tcp_get_business_card(425): \n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 487 - ./rma/at_complete 2\n ---\n Directory: ./rma\n File: at_complete\n Num-procs: 2\n- Date: \"Sun Jan 5 02:03:57 2020\"\n+ Date: \"Mon Dec 3 17:44:10 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n-## Fatal error in PMPI_Init: Other MPI error, error stack:\n-## MPIR_Init_thread(565)..............: \n-## MPID_Init(224).....................: channel initialization failed\n-## MPIDI_CH3_Init(105)................: \n-## MPID_nem_init(324).................: \n-## MPID_nem_tcp_init(178).............: \n-## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 488 - ./rma/atomic_rmw_fop 2\n ---\n Directory: ./rma\n File: atomic_rmw_fop\n Num-procs: 2\n- Date: \"Sun Jan 5 02:03:57 2020\"\n+ Date: \"Mon Dec 3 17:44:11 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n-## Fatal error in PMPI_Init: Other MPI error, error stack:\n-## MPIR_Init_thread(565)..............: \n-## MPID_Init(224).....................: channel initialization failed\n-## MPIDI_CH3_Init(105)................: \n-## MPID_nem_init(324).................: \n-## MPID_nem_tcp_init(178).............: \n-## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 489 - ./rma/atomic_rmw_cas 2\n ---\n Directory: ./rma\n File: atomic_rmw_cas\n Num-procs: 2\n- Date: \"Sun Jan 5 02:03:58 2020\"\n+ Date: \"Mon Dec 3 17:44:11 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 490 - ./rma/atomic_rmw_gacc 2\n ---\n Directory: ./rma\n File: atomic_rmw_gacc\n Num-procs: 2\n- Date: \"Sun Jan 5 02:03:59 2020\"\n+ Date: \"Mon Dec 3 17:44:12 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 491 - ./rma/atomic_get 2\n ---\n Directory: ./rma\n File: atomic_get\n Num-procs: 2\n- Date: \"Sun Jan 5 02:04:01 2020\"\n+ Date: \"Mon Dec 3 17:44:13 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n ok 493 - ./rma/aint 2 # SKIP requires MPI version 3.1\n not ok 492 - ./rma/acc-pairtype 2\n ---\n Directory: ./rma\n File: acc-pairtype\n Num-procs: 2\n- Date: \"Sun Jan 5 02:04:01 2020\"\n+ Date: \"Mon Dec 3 17:44:13 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n-## Fatal error in PMPI_Init: Other MPI error, error stack:\n-## MPIR_Init_thread(565)..............: \n-## MPID_Init(224).....................: channel initialization failed\n-## MPIDI_CH3_Init(105)................: \n-## MPID_nem_init(324).................: \n-## MPID_nem_tcp_init(178).............: \n-## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 493 - ./rma/manyget 2\n ---\n Directory: ./rma\n File: manyget\n Num-procs: 2\n- Date: \"Sun Jan 5 02:04:02 2020\"\n+ Date: \"Mon Dec 3 17:44:14 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n ## Fatal error in PMPI_Init: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 494 - ./rma/derived-acc-flush_local 2\n ---\n Directory: ./rma\n File: derived-acc-flush_local\n Num-procs: 2\n- Date: \"Sun Jan 5 02:04:03 2020\"\n+ Date: \"Mon Dec 3 17:44:14 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n-## Fatal error in PMPI_Init: Other MPI error, error stack:\n-## MPIR_Init_thread(565)..............: \n-## MPID_Init(224).....................: channel initialization failed\n-## MPIDI_CH3_Init(105)................: \n-## MPID_nem_init(324).................: \n-## MPID_nem_tcp_init(178).............: \n-## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 495 - ./rma/large-acc-flush_local 2\n ---\n Directory: ./rma\n File: large-acc-flush_local\n Num-procs: 2\n- Date: \"Sun Jan 5 02:04:05 2020\"\n+ Date: \"Mon Dec 3 17:44:15 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n-not ok 496 - ./rma/large-small-acc 2\n- ---\n- Directory: ./rma\n- File: large-small-acc\n- Num-procs: 2\n- Date: \"Sun Jan 5 02:04:05 2020\"\n- ...\n-## Test output (expected 'No Errors'):\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n ## Fatal error in PMPI_Init: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n+not ok 496 - ./rma/large-small-acc 2\n+ ---\n+ Directory: ./rma\n+ File: large-small-acc\n+ Num-procs: 2\n+ Date: \"Mon Dec 3 17:44:15 2018\"\n+ ...\n+## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 497 - ./rma/win_shared_put_flush_load 2\n ---\n Directory: ./rma\n File: win_shared_put_flush_load\n Num-procs: 2\n- Date: \"Sun Jan 5 02:04:06 2020\"\n+ Date: \"Mon Dec 3 17:44:16 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n-## Fatal error in PMPI_Init: Other MPI error, error stack:\n-## MPIR_Init_thread(565)..............: \n-## MPID_Init(224).....................: channel initialization failed\n-## MPIDI_CH3_Init(105)................: \n-## MPID_nem_init(324).................: \n-## MPID_nem_tcp_init(178).............: \n-## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 498 - ./rma/win_shared_acc_flush_load 2\n ---\n Directory: ./rma\n File: win_shared_acc_flush_load\n Num-procs: 2\n- Date: \"Sun Jan 5 02:04:07 2020\"\n+ Date: \"Mon Dec 3 17:44:17 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n ## Fatal error in PMPI_Init: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 499 - ./rma/win_shared_gacc_flush_load 2\n ---\n Directory: ./rma\n File: win_shared_gacc_flush_load\n Num-procs: 2\n- Date: \"Sun Jan 5 02:04:08 2020\"\n+ Date: \"Mon Dec 3 17:44:17 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n-not ok 500 - ./rma/win_shared_fop_flush_load 2\n- ---\n- Directory: ./rma\n- File: win_shared_fop_flush_load\n- Num-procs: 2\n- Date: \"Sun Jan 5 02:04:09 2020\"\n- ...\n-## Test output (expected 'No Errors'):\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n ## Fatal error in PMPI_Init: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n+not ok 500 - ./rma/win_shared_fop_flush_load 2\n+ ---\n+ Directory: ./rma\n+ File: win_shared_fop_flush_load\n+ Num-procs: 2\n+ Date: \"Mon Dec 3 17:44:18 2018\"\n+ ...\n+## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 501 - ./rma/win_shared_cas_flush_load 2\n ---\n Directory: ./rma\n File: win_shared_cas_flush_load\n Num-procs: 2\n- Date: \"Sun Jan 5 02:04:10 2020\"\n+ Date: \"Mon Dec 3 17:44:18 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 502 - ./rma/put_flush_get 2\n ---\n Directory: ./rma\n File: put_flush_get\n Num-procs: 2\n- Date: \"Sun Jan 5 02:04:11 2020\"\n+ Date: \"Mon Dec 3 17:44:19 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n ## Fatal error in PMPI_Init: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 503 - ./rma/acc_flush_get 2\n ---\n Directory: ./rma\n File: acc_flush_get\n Num-procs: 2\n- Date: \"Sun Jan 5 02:04:12 2020\"\n+ Date: \"Mon Dec 3 17:44:20 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n+not ok 504 - ./rma/gacc_flush_get 2\n+ ---\n+ Directory: ./rma\n+ File: gacc_flush_get\n+ Num-procs: 2\n+ Date: \"Mon Dec 3 17:44:20 2018\"\n+ ...\n+## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n-not ok 504 - ./rma/gacc_flush_get 2\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n+not ok 505 - ./rma/fop_flush_get 2\n ---\n Directory: ./rma\n- File: gacc_flush_get\n+ File: fop_flush_get\n Num-procs: 2\n- Date: \"Sun Jan 5 02:04:13 2020\"\n+ Date: \"Mon Dec 3 17:44:21 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n ## Fatal error in PMPI_Init: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n-not ok 505 - ./rma/fop_flush_get 2\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n+not ok 506 - ./rma/cas_flush_get 2\n ---\n Directory: ./rma\n- File: fop_flush_get\n+ File: cas_flush_get\n Num-procs: 2\n- Date: \"Sun Jan 5 02:04:14 2020\"\n+ Date: \"Mon Dec 3 17:44:21 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n-not ok 506 - ./rma/cas_flush_get 2\n- ---\n- Directory: ./rma\n- File: cas_flush_get\n- Num-procs: 2\n- Date: \"Sun Jan 5 02:04:16 2020\"\n- ...\n-## Test output (expected 'No Errors'):\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n ## Fatal error in PMPI_Init: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 507 - ./spawn/namepub 2\n ---\n Directory: ./spawn\n File: namepub\n Num-procs: 2\n- Date: \"Sun Jan 5 02:04:17 2020\"\n+ Date: \"Mon Dec 3 17:44:22 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 508 - ./spawn/spawn1 1\n ---\n Directory: ./spawn\n File: spawn1\n Num-procs: 1\n- Date: \"Sun Jan 5 02:04:18 2020\"\n+ Date: \"Mon Dec 3 17:44:22 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 509 - ./spawn/spawn2 1\n ---\n Directory: ./spawn\n File: spawn2\n Num-procs: 1\n- Date: \"Sun Jan 5 02:04:19 2020\"\n+ Date: \"Mon Dec 3 17:44:23 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 510 - ./spawn/spawninfo1 1\n ---\n Directory: ./spawn\n File: spawninfo1\n Num-procs: 1\n- Date: \"Sun Jan 5 02:04:20 2020\"\n+ Date: \"Mon Dec 3 17:44:23 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 511 - ./spawn/spawnminfo1 1\n ---\n Directory: ./spawn\n File: spawnminfo1\n Num-procs: 1\n- Date: \"Sun Jan 5 02:04:21 2020\"\n+ Date: \"Mon Dec 3 17:44:24 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 512 - ./spawn/spawnintra 1\n ---\n Directory: ./spawn\n File: spawnintra\n Num-procs: 1\n- Date: \"Sun Jan 5 02:04:22 2020\"\n+ Date: \"Mon Dec 3 17:44:24 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 513 - ./spawn/spawnintra 2\n ---\n Directory: ./spawn\n File: spawnintra\n Num-procs: 2\n- Date: \"Sun Jan 5 02:04:23 2020\"\n+ Date: \"Mon Dec 3 17:44:25 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 514 - ./spawn/spawnargv 1\n ---\n Directory: ./spawn\n File: spawnargv\n Num-procs: 1\n- Date: \"Sun Jan 5 02:04:23 2020\"\n+ Date: \"Mon Dec 3 17:44:25 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 515 - ./spawn/spawnmanyarg 1\n ---\n Directory: ./spawn\n File: spawnmanyarg\n Num-procs: 1\n- Date: \"Sun Jan 5 02:04:24 2020\"\n+ Date: \"Mon Dec 3 17:44:26 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 516 - ./spawn/spawnmult2 2\n ---\n Directory: ./spawn\n File: spawnmult2\n Num-procs: 2\n- Date: \"Sun Jan 5 02:04:25 2020\"\n+ Date: \"Mon Dec 3 17:44:26 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n+## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n+## MPIR_Init_thread(565)..............: \n+## MPID_Init(224).....................: channel initialization failed\n+## MPIDI_CH3_Init(105)................: \n+## MPID_nem_init(324).................: \n+## MPID_nem_tcp_init(178).............: \n+## MPID_nem_tcp_get_business_card(425): \n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 517 - ./spawn/spaconacc 1\n ---\n Directory: ./spawn\n File: spaconacc\n Num-procs: 1\n- Date: \"Sun Jan 5 02:04:26 2020\"\n+ Date: \"Mon Dec 3 17:44:27 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 518 - ./spawn/spaconacc2 1\n ---\n Directory: ./spawn\n File: spaconacc2\n Num-procs: 1\n- Date: \"Sun Jan 5 02:04:27 2020\"\n+ Date: \"Mon Dec 3 17:44:27 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 519 - ./spawn/selfconacc 2\n ---\n Directory: ./spawn\n File: selfconacc\n Num-procs: 2\n- Date: \"Sun Jan 5 02:04:27 2020\"\n+ Date: \"Mon Dec 3 17:44:28 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 520 - ./spawn/spaiccreate 2\n ---\n Directory: ./spawn\n File: spaiccreate\n Num-procs: 2\n- Date: \"Sun Jan 5 02:04:28 2020\"\n+ Date: \"Mon Dec 3 17:44:28 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n-## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n-## MPIR_Init_thread(565)..............: \n-## MPID_Init(224).....................: channel initialization failed\n-## MPIDI_CH3_Init(105)................: \n-## MPID_nem_init(324).................: \n-## MPID_nem_tcp_init(178).............: \n-## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 521 - ./spawn/taskmaster 1\n ---\n Directory: ./spawn\n File: taskmaster\n Num-procs: 1\n- Date: \"Sun Jan 5 02:04:29 2020\"\n+ Date: \"Mon Dec 3 17:44:29 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 522 - ./spawn/taskmaster 2\n ---\n Directory: ./spawn\n File: taskmaster\n Num-procs: 2\n- Date: \"Sun Jan 5 02:04:30 2020\"\n+ Date: \"Mon Dec 3 17:44:29 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n+## Fatal error in PMPI_Init: Other MPI error, error stack:\n+## MPIR_Init_thread(565)..............: \n+## MPID_Init(224).....................: channel initialization failed\n+## MPIDI_CH3_Init(105)................: \n+## MPID_nem_init(324).................: \n+## MPID_nem_tcp_init(178).............: \n+## MPID_nem_tcp_get_business_card(425): \n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 523 - ./spawn/join 2\n ---\n Directory: ./spawn\n File: join\n Num-procs: 2\n- Date: \"Sun Jan 5 02:04:30 2020\"\n+ Date: \"Mon Dec 3 17:44:29 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 524 - ./spawn/disconnect_reconnect 2\n ---\n Directory: ./spawn\n File: disconnect_reconnect\n Num-procs: 2\n- Date: \"Sun Jan 5 02:04:32 2020\"\n+ Date: \"Mon Dec 3 17:44:30 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 525 - ./spawn/disconnect_reconnect2 2\n ---\n Directory: ./spawn\n File: disconnect_reconnect2\n Num-procs: 2\n- Date: \"Sun Jan 5 02:04:33 2020\"\n+ Date: \"Mon Dec 3 17:44:31 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n+## Fatal error in PMPI_Init: Other MPI error, error stack:\n+## MPIR_Init_thread(565)..............: \n+## MPID_Init(224).....................: channel initialization failed\n+## MPIDI_CH3_Init(105)................: \n+## MPID_nem_init(324).................: \n+## MPID_nem_tcp_init(178).............: \n+## MPID_nem_tcp_get_business_card(425): \n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 526 - ./spawn/disconnect_reconnect3 2\n ---\n Directory: ./spawn\n File: disconnect_reconnect3\n Num-procs: 2\n- Date: \"Sun Jan 5 02:04:33 2020\"\n+ Date: \"Mon Dec 3 17:44:31 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n-## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n-## MPIR_Init_thread(565)..............: \n-## MPID_Init(224).....................: channel initialization failed\n-## MPIDI_CH3_Init(105)................: \n-## MPID_nem_init(324).................: \n-## MPID_nem_tcp_init(178).............: \n-## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 527 - ./spawn/multiple_ports 2\n ---\n Directory: ./spawn\n File: multiple_ports\n Num-procs: 2\n- Date: \"Sun Jan 5 02:04:34 2020\"\n+ Date: \"Mon Dec 3 17:44:32 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n+not ok 528 - ./spawn/multiple_ports2 2\n+ ---\n+ Directory: ./spawn\n+ File: multiple_ports2\n+ Num-procs: 2\n+ Date: \"Mon Dec 3 17:44:32 2018\"\n+ ...\n+## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n-not ok 528 - ./spawn/multiple_ports2 2\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n+not ok 529 - ./spawn/disconnect 2\n ---\n Directory: ./spawn\n- File: multiple_ports2\n+ File: disconnect\n Num-procs: 2\n- Date: \"Sun Jan 5 02:04:34 2020\"\n+ Date: \"Mon Dec 3 17:44:33 2018\"\n ...\n ## Test output (expected 'No Errors'):\n-## Fatal error in PMPI_Init: Other MPI error, error stack:\n+## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n-## Fatal error in PMPI_Init: Other MPI error, error stack:\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n+## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n-not ok 529 - ./spawn/disconnect 2\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n+not ok 530 - ./spawn/disconnect2 2\n ---\n Directory: ./spawn\n- File: disconnect\n+ File: disconnect2\n Num-procs: 2\n- Date: \"Sun Jan 5 02:04:35 2020\"\n+ Date: \"Mon Dec 3 17:44:33 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n-not ok 530 - ./spawn/disconnect2 2\n- ---\n- Directory: ./spawn\n- File: disconnect2\n- Num-procs: 2\n- Date: \"Sun Jan 5 02:04:35 2020\"\n- ...\n-## Test output (expected 'No Errors'):\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 531 - ./spawn/disconnect3 2\n ---\n Directory: ./spawn\n File: disconnect3\n Num-procs: 2\n- Date: \"Sun Jan 5 02:04:36 2020\"\n+ Date: \"Mon Dec 3 17:44:34 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n+## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n+## MPIR_Init_thread(565)..............: \n+## MPID_Init(224).....................: channel initialization failed\n+## MPIDI_CH3_Init(105)................: \n+## MPID_nem_init(324).................: \n+## MPID_nem_tcp_init(178).............: \n+## MPID_nem_tcp_get_business_card(425): \n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 532 - ./spawn/concurrent_spawns 1\n ---\n Directory: ./spawn\n File: concurrent_spawns\n Num-procs: 1\n- Date: \"Sun Jan 5 02:04:36 2020\"\n+ Date: \"Mon Dec 3 17:44:34 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 533 - ./spawn/pgroup_connect_test 2\n ---\n Directory: ./spawn\n File: pgroup_connect_test\n Num-procs: 2\n- Date: \"Sun Jan 5 02:04:37 2020\"\n+ Date: \"Mon Dec 3 17:44:35 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n+## Fatal error in PMPI_Init: Other MPI error, error stack:\n+## MPIR_Init_thread(565)..............: \n+## MPID_Init(224).....................: channel initialization failed\n+## MPIDI_CH3_Init(105)................: \n+## MPID_nem_init(324).................: \n+## MPID_nem_tcp_init(178).............: \n+## MPID_nem_tcp_get_business_card(425): \n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 534 - ./spawn/pgroup_intercomm_test 2\n ---\n Directory: ./spawn\n File: pgroup_intercomm_test\n Num-procs: 2\n- Date: \"Sun Jan 5 02:04:38 2020\"\n+ Date: \"Mon Dec 3 17:44:35 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n+## Fatal error in PMPI_Init: Other MPI error, error stack:\n+## MPIR_Init_thread(565)..............: \n+## MPID_Init(224).....................: channel initialization failed\n+## MPIDI_CH3_Init(105)................: \n+## MPID_nem_init(324).................: \n+## MPID_nem_tcp_init(178).............: \n+## MPID_nem_tcp_get_business_card(425): \n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 535 - ./spawn/spawn-rootargs 2\n ---\n Directory: ./spawn\n File: spawn-rootargs\n Num-procs: 2\n- Date: \"Sun Jan 5 02:04:38 2020\"\n+ Date: \"Mon Dec 3 17:44:36 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 536 - ./topo/cartmap1 2\n ---\n Directory: ./topo\n File: cartmap1\n Num-procs: 2\n- Date: \"Sun Jan 5 02:04:39 2020\"\n+ Date: \"Mon Dec 3 17:44:36 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 537 - ./topo/cartzero 2\n ---\n Directory: ./topo\n File: cartzero\n Num-procs: 2\n- Date: \"Sun Jan 5 02:04:39 2020\"\n+ Date: \"Mon Dec 3 17:44:37 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 538 - ./topo/cartshift1 2\n ---\n Directory: ./topo\n File: cartshift1\n Num-procs: 2\n- Date: \"Sun Jan 5 02:04:40 2020\"\n+ Date: \"Mon Dec 3 17:44:37 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 539 - ./topo/cartsuball 2\n ---\n Directory: ./topo\n File: cartsuball\n Num-procs: 2\n- Date: \"Sun Jan 5 02:04:41 2020\"\n+ Date: \"Mon Dec 3 17:44:38 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n-## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n-## MPIR_Init_thread(565)..............: \n-## MPID_Init(224).....................: channel initialization failed\n-## MPIDI_CH3_Init(105)................: \n-## MPID_nem_init(324).................: \n-## MPID_nem_tcp_init(178).............: \n-## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 540 - ./topo/cartcreates 2\n ---\n Directory: ./topo\n File: cartcreates\n Num-procs: 2\n- Date: \"Sun Jan 5 02:04:41 2020\"\n+ Date: \"Mon Dec 3 17:44:38 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n-## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n-## MPIR_Init_thread(565)..............: \n-## MPID_Init(224).....................: channel initialization failed\n-## MPIDI_CH3_Init(105)................: \n-## MPID_nem_init(324).................: \n-## MPID_nem_tcp_init(178).............: \n-## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 541 - ./topo/dims1 2\n ---\n Directory: ./topo\n File: dims1\n Num-procs: 2\n- Date: \"Sun Jan 5 02:04:42 2020\"\n+ Date: \"Mon Dec 3 17:44:38 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n-## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n-## MPIR_Init_thread(565)..............: \n-## MPID_Init(224).....................: channel initialization failed\n-## MPIDI_CH3_Init(105)................: \n-## MPID_nem_init(324).................: \n-## MPID_nem_tcp_init(178).............: \n-## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 542 - ./topo/dims2 1\n ---\n Directory: ./topo\n File: dims2\n Num-procs: 1\n- Date: \"Sun Jan 5 02:04:42 2020\"\n+ Date: \"Mon Dec 3 17:44:39 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 543 - ./topo/dims3 1\n ---\n Directory: ./topo\n File: dims3\n Num-procs: 1\n- Date: \"Sun Jan 5 02:04:43 2020\"\n+ Date: \"Mon Dec 3 17:44:39 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 544 - ./topo/dims4 1\n ---\n Directory: ./topo\n File: dims4\n Num-procs: 1\n- Date: \"Sun Jan 5 02:04:43 2020\"\n+ Date: \"Mon Dec 3 17:44:40 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 545 - ./topo/graphmap1 2\n ---\n Directory: ./topo\n File: graphmap1\n Num-procs: 2\n- Date: \"Sun Jan 5 02:04:44 2020\"\n+ Date: \"Mon Dec 3 17:44:40 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n+## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n+## MPIR_Init_thread(565)..............: \n+## MPID_Init(224).....................: channel initialization failed\n+## MPIDI_CH3_Init(105)................: \n+## MPID_nem_init(324).................: \n+## MPID_nem_tcp_init(178).............: \n+## MPID_nem_tcp_get_business_card(425): \n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 546 - ./topo/topotest 2\n ---\n Directory: ./topo\n File: topotest\n Num-procs: 2\n- Date: \"Sun Jan 5 02:04:45 2020\"\n+ Date: \"Mon Dec 3 17:44:41 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 547 - ./topo/topodup 2\n ---\n Directory: ./topo\n File: topodup\n Num-procs: 2\n- Date: \"Sun Jan 5 02:04:45 2020\"\n+ Date: \"Mon Dec 3 17:44:41 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 548 - ./topo/graphcr 2\n ---\n Directory: ./topo\n File: graphcr\n Num-procs: 2\n- Date: \"Sun Jan 5 02:04:46 2020\"\n+ Date: \"Mon Dec 3 17:44:42 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n-not ok 549 - ./topo/graphcr2 2\n- ---\n- Directory: ./topo\n- File: graphcr2\n- Num-procs: 2\n- Date: \"Sun Jan 5 02:04:47 2020\"\n- ...\n-## Test output (expected 'No Errors'):\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n+not ok 549 - ./topo/graphcr2 2\n+ ---\n+ Directory: ./topo\n+ File: graphcr2\n+ Num-procs: 2\n+ Date: \"Mon Dec 3 17:44:42 2018\"\n+ ...\n+## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 550 - ./topo/distgraph1 2\n ---\n Directory: ./topo\n File: distgraph1\n Num-procs: 2\n- Date: \"Sun Jan 5 02:04:48 2020\"\n+ Date: \"Mon Dec 3 17:44:43 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n-not ok 551 - ./topo/dgraph_unwgt 2\n- ---\n- Directory: ./topo\n- File: dgraph_unwgt\n- Num-procs: 2\n- Date: \"Sun Jan 5 02:04:49 2020\"\n- ...\n-## Test output (expected 'No Errors'):\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n+not ok 551 - ./topo/dgraph_unwgt 2\n+ ---\n+ Directory: ./topo\n+ File: dgraph_unwgt\n+ Num-procs: 2\n+ Date: \"Mon Dec 3 17:44:43 2018\"\n+ ...\n+## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 552 - ./topo/neighb_coll 2\n ---\n Directory: ./topo\n File: neighb_coll\n Num-procs: 2\n- Date: \"Sun Jan 5 02:04:49 2020\"\n+ Date: \"Mon Dec 3 17:44:44 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 553 - ./io/rdwrord 2\n ---\n Directory: ./io\n File: rdwrord\n Num-procs: 2\n- Date: \"Sun Jan 5 02:04:50 2020\"\n+ Date: \"Mon Dec 3 17:44:44 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 554 - ./io/rdwrzero 2\n ---\n Directory: ./io\n File: rdwrzero\n Num-procs: 2\n- Date: \"Sun Jan 5 02:04:51 2020\"\n+ Date: \"Mon Dec 3 17:44:45 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n-## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n-## MPIR_Init_thread(565)..............: \n-## MPID_Init(224).....................: channel initialization failed\n-## MPIDI_CH3_Init(105)................: \n-## MPID_nem_init(324).................: \n-## MPID_nem_tcp_init(178).............: \n-## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 555 - ./io/getextent 2\n ---\n Directory: ./io\n File: getextent\n Num-procs: 2\n- Date: \"Sun Jan 5 02:04:51 2020\"\n+ Date: \"Mon Dec 3 17:44:45 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n-## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n-## MPIR_Init_thread(565)..............: \n-## MPID_Init(224).....................: channel initialization failed\n-## MPIDI_CH3_Init(105)................: \n-## MPID_nem_init(324).................: \n-## MPID_nem_tcp_init(178).............: \n-## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 556 - ./io/setinfo 2\n ---\n Directory: ./io\n File: setinfo\n Num-procs: 2\n- Date: \"Sun Jan 5 02:04:52 2020\"\n+ Date: \"Mon Dec 3 17:44:45 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n-## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n-## MPIR_Init_thread(565)..............: \n-## MPID_Init(224).....................: channel initialization failed\n-## MPIDI_CH3_Init(105)................: \n-## MPID_nem_init(324).................: \n-## MPID_nem_tcp_init(178).............: \n-## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 557 - ./io/setviewcur 2\n ---\n Directory: ./io\n File: setviewcur\n Num-procs: 2\n- Date: \"Sun Jan 5 02:04:53 2020\"\n+ Date: \"Mon Dec 3 17:44:46 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n+## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n+## MPIR_Init_thread(565)..............: \n+## MPID_Init(224).....................: channel initialization failed\n+## MPIDI_CH3_Init(105)................: \n+## MPID_nem_init(324).................: \n+## MPID_nem_tcp_init(178).............: \n+## MPID_nem_tcp_get_business_card(425): \n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 558 - ./io/i_noncontig 2\n ---\n Directory: ./io\n File: i_noncontig\n Num-procs: 2\n- Date: \"Sun Jan 5 02:04:53 2020\"\n+ Date: \"Mon Dec 3 17:44:46 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 559 - ./io/async 2\n ---\n Directory: ./io\n File: async\n Num-procs: 2\n- Date: \"Sun Jan 5 02:04:54 2020\"\n+ Date: \"Mon Dec 3 17:44:47 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 560 - ./io/async_any 2\n ---\n Directory: ./io\n File: async_any\n Num-procs: 2\n- Date: \"Sun Jan 5 02:04:55 2020\"\n+ Date: \"Mon Dec 3 17:44:47 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n+## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n+## MPIR_Init_thread(565)..............: \n+## MPID_Init(224).....................: channel initialization failed\n+## MPIDI_CH3_Init(105)................: \n+## MPID_nem_init(324).................: \n+## MPID_nem_tcp_init(178).............: \n+## MPID_nem_tcp_get_business_card(425): \n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 561 - ./io/userioerr 1\n ---\n Directory: ./io\n File: userioerr\n Num-procs: 1\n- Date: \"Sun Jan 5 02:04:55 2020\"\n+ Date: \"Mon Dec 3 17:44:48 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 562 - ./io/resized 1\n ---\n Directory: ./io\n File: resized\n Num-procs: 1\n- Date: \"Sun Jan 5 02:04:55 2020\"\n+ Date: \"Mon Dec 3 17:44:48 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 563 - ./io/resized2 1\n ---\n Directory: ./io\n File: resized2\n Num-procs: 1\n- Date: \"Sun Jan 5 02:04:56 2020\"\n+ Date: \"Mon Dec 3 17:44:49 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 564 - ./io/hindexed_io 1\n ---\n Directory: ./io\n File: hindexed_io\n Num-procs: 1\n- Date: \"Sun Jan 5 02:04:56 2020\"\n+ Date: \"Mon Dec 3 17:44:49 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 565 - ./io/simple_collective 1\n ---\n Directory: ./io\n File: simple_collective\n Num-procs: 1\n- Date: \"Sun Jan 5 02:04:56 2020\"\n+ Date: \"Mon Dec 3 17:44:49 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n ok 568 - ./io/i_bigtype 1 # SKIP requires MPI version 3.1\n ok 569 - ./io/i_hindexed_io 1 # SKIP requires MPI version 3.1\n ok 570 - ./io/i_rdwrord 2 # SKIP requires MPI version 3.1\n ok 571 - ./io/i_setviewcur 2 # SKIP requires MPI version 3.1\n ok 572 - ./io/i_aggregation1 2 # SKIP requires MPI version 3.1\n ok 573 - ./io/i_aggregation2 2 # SKIP requires MPI version 3.1\n ok 574 - ./io/i_coll_test 2 # SKIP requires MPI version 3.1\n@@ -10501,6650 +10045,6278 @@\n ok 578 - ./io/i_noncontig_coll2 2 # SKIP requires MPI version 3.1\n ok 579 - ./io/i_types_with_zeros 2 # SKIP requires MPI version 3.1\n not ok 566 - ./f77/attr/attrmpi1f 1\n ---\n Directory: ./f77/attr\n File: attrmpi1f\n Num-procs: 1\n- Date: \"Sun Jan 5 02:04:57 2020\"\n+ Date: \"Mon Dec 3 17:44:50 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 567 - ./f77/attr/baseattrf 1\n ---\n Directory: ./f77/attr\n File: baseattrf\n Num-procs: 1\n- Date: \"Sun Jan 5 02:04:57 2020\"\n+ Date: \"Mon Dec 3 17:44:50 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 568 - ./f77/attr/baseattr2f 1\n ---\n Directory: ./f77/attr\n File: baseattr2f\n Num-procs: 1\n- Date: \"Sun Jan 5 02:04:57 2020\"\n+ Date: \"Mon Dec 3 17:44:51 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 569 - ./f77/attr/commattrf 1\n ---\n Directory: ./f77/attr\n File: commattrf\n Num-procs: 1\n- Date: \"Sun Jan 5 02:04:57 2020\"\n+ Date: \"Mon Dec 3 17:44:51 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 570 - ./f77/attr/commattr2f 1\n ---\n Directory: ./f77/attr\n File: commattr2f\n Num-procs: 1\n- Date: \"Sun Jan 5 02:04:57 2020\"\n+ Date: \"Mon Dec 3 17:44:51 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 571 - ./f77/attr/commattr3f 1\n ---\n Directory: ./f77/attr\n File: commattr3f\n Num-procs: 1\n- Date: \"Sun Jan 5 02:04:58 2020\"\n+ Date: \"Mon Dec 3 17:44:52 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 572 - ./f77/attr/commattr4f 1\n ---\n Directory: ./f77/attr\n File: commattr4f\n Num-procs: 1\n- Date: \"Sun Jan 5 02:04:58 2020\"\n+ Date: \"Mon Dec 3 17:44:52 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 573 - ./f77/attr/typeattrf 1\n ---\n Directory: ./f77/attr\n File: typeattrf\n Num-procs: 1\n- Date: \"Sun Jan 5 02:04:58 2020\"\n+ Date: \"Mon Dec 3 17:44:52 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 574 - ./f77/attr/typeattr2f 1\n ---\n Directory: ./f77/attr\n File: typeattr2f\n Num-procs: 1\n- Date: \"Sun Jan 5 02:04:58 2020\"\n+ Date: \"Mon Dec 3 17:44:53 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 575 - ./f77/attr/typeattr3f 1\n ---\n Directory: ./f77/attr\n File: typeattr3f\n Num-procs: 1\n- Date: \"Sun Jan 5 02:04:59 2020\"\n+ Date: \"Mon Dec 3 17:44:53 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 576 - ./f77/coll/uallreducef 2\n ---\n Directory: ./f77/coll\n File: uallreducef\n Num-procs: 2\n- Date: \"Sun Jan 5 02:04:59 2020\"\n+ Date: \"Mon Dec 3 17:44:54 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 577 - ./f77/coll/exscanf 2\n ---\n Directory: ./f77/coll\n File: exscanf\n Num-procs: 2\n- Date: \"Sun Jan 5 02:04:59 2020\"\n+ Date: \"Mon Dec 3 17:44:54 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 578 - ./f77/coll/alltoallwf 2\n ---\n Directory: ./f77/coll\n File: alltoallwf\n Num-procs: 2\n- Date: \"Sun Jan 5 02:04:59 2020\"\n+ Date: \"Mon Dec 3 17:44:55 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n ## Fatal error in PMPI_Init: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 579 - ./f77/coll/alltoallvf 2\n ---\n Directory: ./f77/coll\n File: alltoallvf\n Num-procs: 2\n- Date: \"Sun Jan 5 02:05:00 2020\"\n+ Date: \"Mon Dec 3 17:44:55 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n ## Fatal error in PMPI_Init: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 580 - ./f77/coll/inplacef 2\n ---\n Directory: ./f77/coll\n File: inplacef\n Num-procs: 2\n- Date: \"Sun Jan 5 02:05:00 2020\"\n+ Date: \"Mon Dec 3 17:44:55 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n-## Fatal error in PMPI_Init: Other MPI error, error stack:\n-## MPIR_Init_thread(565)..............: \n-## MPID_Init(224).....................: channel initialization failed\n-## MPIDI_CH3_Init(105)................: \n-## MPID_nem_init(324).................: \n-## MPID_nem_tcp_init(178).............: \n-## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 581 - ./f77/coll/reducelocalf 2\n ---\n Directory: ./f77/coll\n File: reducelocalf\n Num-procs: 2\n- Date: \"Sun Jan 5 02:05:00 2020\"\n+ Date: \"Mon Dec 3 17:44:56 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 582 - ./f77/coll/redscatf 2\n ---\n Directory: ./f77/coll\n File: redscatf\n Num-procs: 2\n- Date: \"Sun Jan 5 02:05:00 2020\"\n+ Date: \"Mon Dec 3 17:44:56 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 583 - ./f77/coll/split_typef 2\n ---\n Directory: ./f77/coll\n File: split_typef\n Num-procs: 2\n- Date: \"Sun Jan 5 02:05:01 2020\"\n+ Date: \"Mon Dec 3 17:44:57 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 584 - ./f77/coll/nonblockingf 2\n ---\n Directory: ./f77/coll\n File: nonblockingf\n Num-procs: 2\n- Date: \"Sun Jan 5 02:05:01 2020\"\n+ Date: \"Mon Dec 3 17:44:57 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n-not ok 585 - ./f77/coll/vw_inplacef 2\n- ---\n- Directory: ./f77/coll\n- File: vw_inplacef\n- Num-procs: 2\n- Date: \"Sun Jan 5 02:05:01 2020\"\n- ...\n-## Test output (expected 'No Errors'):\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n ## Fatal error in PMPI_Init: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n+not ok 585 - ./f77/coll/vw_inplacef 2\n+ ---\n+ Directory: ./f77/coll\n+ File: vw_inplacef\n+ Num-procs: 2\n+ Date: \"Mon Dec 3 17:44:58 2018\"\n+ ...\n+## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 586 - ./f77/coll/red_scat_blockf 2\n ---\n Directory: ./f77/coll\n File: red_scat_blockf\n Num-procs: 2\n- Date: \"Sun Jan 5 02:05:01 2020\"\n+ Date: \"Mon Dec 3 17:44:58 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n-## Fatal error in PMPI_Init: Other MPI error, error stack:\n-## MPIR_Init_thread(565)..............: \n-## MPID_Init(224).....................: channel initialization failed\n-## MPIDI_CH3_Init(105)................: \n-## MPID_nem_init(324).................: \n-## MPID_nem_tcp_init(178).............: \n-## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 587 - ./f77/coll/nonblocking_inpf 2\n ---\n Directory: ./f77/coll\n File: nonblocking_inpf\n Num-procs: 2\n- Date: \"Sun Jan 5 02:05:02 2020\"\n+ Date: \"Mon Dec 3 17:44:59 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n+## Fatal error in PMPI_Init: Other MPI error, error stack:\n+## MPIR_Init_thread(565)..............: \n+## MPID_Init(224).....................: channel initialization failed\n+## MPIDI_CH3_Init(105)................: \n+## MPID_nem_init(324).................: \n+## MPID_nem_tcp_init(178).............: \n+## MPID_nem_tcp_get_business_card(425): \n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 588 - ./f77/datatype/typenamef 1\n ---\n Directory: ./f77/datatype\n File: typenamef\n Num-procs: 1\n- Date: \"Sun Jan 5 02:05:02 2020\"\n+ Date: \"Mon Dec 3 17:44:59 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 589 - ./f77/datatype/typename3f 1\n ---\n Directory: ./f77/datatype\n File: typename3f\n Num-procs: 1\n- Date: \"Sun Jan 5 02:05:02 2020\"\n+ Date: \"Mon Dec 3 17:44:59 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 590 - ./f77/datatype/typesnamef 1\n ---\n Directory: ./f77/datatype\n File: typesnamef\n Num-procs: 1\n- Date: \"Sun Jan 5 02:05:02 2020\"\n+ Date: \"Mon Dec 3 17:45:00 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 591 - ./f77/datatype/typecntsf 1\n ---\n Directory: ./f77/datatype\n File: typecntsf\n Num-procs: 1\n- Date: \"Sun Jan 5 02:05:02 2020\"\n+ Date: \"Mon Dec 3 17:45:00 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 592 - ./f77/datatype/typem2f 1\n ---\n Directory: ./f77/datatype\n File: typem2f\n Num-procs: 1\n- Date: \"Sun Jan 5 02:05:02 2020\"\n+ Date: \"Mon Dec 3 17:45:00 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 593 - ./f77/datatype/typesubf 1\n ---\n Directory: ./f77/datatype\n File: typesubf\n Num-procs: 1\n- Date: \"Sun Jan 5 02:05:03 2020\"\n+ Date: \"Mon Dec 3 17:45:01 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 594 - ./f77/datatype/packef 1\n ---\n Directory: ./f77/datatype\n File: packef\n Num-procs: 1\n- Date: \"Sun Jan 5 02:05:03 2020\"\n+ Date: \"Mon Dec 3 17:45:01 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 595 - ./f77/datatype/gaddressf 1\n ---\n Directory: ./f77/datatype\n File: gaddressf\n Num-procs: 1\n- Date: \"Sun Jan 5 02:05:03 2020\"\n+ Date: \"Mon Dec 3 17:45:02 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 596 - ./f77/datatype/allctypesf 1\n ---\n Directory: ./f77/datatype\n File: allctypesf\n Num-procs: 1\n- Date: \"Sun Jan 5 02:05:03 2020\"\n+ Date: \"Mon Dec 3 17:45:02 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 597 - ./f77/datatype/hindex1f 1\n ---\n Directory: ./f77/datatype\n File: hindex1f\n Num-procs: 1\n- Date: \"Sun Jan 5 02:05:03 2020\"\n+ Date: \"Mon Dec 3 17:45:02 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 598 - ./f77/datatype/hindexed_blockf 1\n ---\n Directory: ./f77/datatype\n File: hindexed_blockf\n Num-procs: 1\n- Date: \"Sun Jan 5 02:05:03 2020\"\n+ Date: \"Mon Dec 3 17:45:03 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 599 - ./f77/datatype/bottom 2\n ---\n Directory: ./f77/datatype\n File: bottom\n Num-procs: 2\n- Date: \"Sun Jan 5 02:05:04 2020\"\n+ Date: \"Mon Dec 3 17:45:03 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n-## Fatal error in PMPI_Init: Other MPI error, error stack:\n-## MPIR_Init_thread(565)..............: \n-## MPID_Init(224).....................: channel initialization failed\n-## MPIDI_CH3_Init(105)................: \n-## MPID_nem_init(324).................: \n-## MPID_nem_tcp_init(178).............: \n-## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 600 - ./f77/pt2pt/statusesf 1\n ---\n Directory: ./f77/pt2pt\n File: statusesf\n Num-procs: 1\n- Date: \"Sun Jan 5 02:05:04 2020\"\n+ Date: \"Mon Dec 3 17:45:04 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 601 - ./f77/pt2pt/greqf 1\n ---\n Directory: ./f77/pt2pt\n File: greqf\n Num-procs: 1\n- Date: \"Sun Jan 5 02:05:04 2020\"\n+ Date: \"Mon Dec 3 17:45:04 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 602 - ./f77/pt2pt/allpairf 2\n ---\n Directory: ./f77/pt2pt\n File: allpairf\n Num-procs: 2\n- Date: \"Sun Jan 5 02:05:04 2020\"\n+ Date: \"Mon Dec 3 17:45:05 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n-## Fatal error in PMPI_Init: Other MPI error, error stack:\n-## MPIR_Init_thread(565)..............: \n-## MPID_Init(224).....................: channel initialization failed\n-## MPIDI_CH3_Init(105)................: \n-## MPID_nem_init(324).................: \n-## MPID_nem_tcp_init(178).............: \n-## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 603 - ./f77/pt2pt/mprobef 2\n ---\n Directory: ./f77/pt2pt\n File: mprobef\n Num-procs: 2\n- Date: \"Sun Jan 5 02:05:05 2020\"\n+ Date: \"Mon Dec 3 17:45:05 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 604 - ./f77/info/infotestf 1\n ---\n Directory: ./f77/info\n File: infotestf\n Num-procs: 1\n- Date: \"Sun Jan 5 02:05:05 2020\"\n+ Date: \"Mon Dec 3 17:45:06 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 605 - ./f77/info/infotest2f 1\n ---\n Directory: ./f77/info\n File: infotest2f\n Num-procs: 1\n- Date: \"Sun Jan 5 02:05:05 2020\"\n+ Date: \"Mon Dec 3 17:45:06 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 606 - ./f77/spawn/namepubf 2\n ---\n Directory: ./f77/spawn\n File: namepubf\n Num-procs: 2\n- Date: \"Sun Jan 5 02:05:05 2020\"\n+ Date: \"Mon Dec 3 17:45:06 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n+## Fatal error in PMPI_Init: Other MPI error, error stack:\n+## MPIR_Init_thread(565)..............: \n+## MPID_Init(224).....................: channel initialization failed\n+## MPIDI_CH3_Init(105)................: \n+## MPID_nem_init(324).................: \n+## MPID_nem_tcp_init(178).............: \n+## MPID_nem_tcp_get_business_card(425): \n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 607 - ./f77/spawn/spawnf 1\n ---\n Directory: ./f77/spawn\n File: spawnf\n Num-procs: 1\n- Date: \"Sun Jan 5 02:05:05 2020\"\n+ Date: \"Mon Dec 3 17:45:07 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 608 - ./f77/spawn/spawnargvf 1\n ---\n Directory: ./f77/spawn\n File: spawnargvf\n Num-procs: 1\n- Date: \"Sun Jan 5 02:05:05 2020\"\n+ Date: \"Mon Dec 3 17:45:07 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 609 - ./f77/spawn/connaccf 2\n ---\n Directory: ./f77/spawn\n File: connaccf\n Num-procs: 2\n- Date: \"Sun Jan 5 02:05:05 2020\"\n+ Date: \"Mon Dec 3 17:45:08 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n-## Fatal error in PMPI_Init: Other MPI error, error stack:\n-## MPIR_Init_thread(565)..............: \n-## MPID_Init(224).....................: channel initialization failed\n-## MPIDI_CH3_Init(105)................: \n-## MPID_nem_init(324).................: \n-## MPID_nem_tcp_init(178).............: \n-## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 610 - ./f77/spawn/spawnmultf 1\n ---\n Directory: ./f77/spawn\n File: spawnmultf\n Num-procs: 1\n- Date: \"Sun Jan 5 02:05:06 2020\"\n+ Date: \"Mon Dec 3 17:45:08 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 611 - ./f77/spawn/spawnmult2f 2\n ---\n Directory: ./f77/spawn\n File: spawnmult2f\n Num-procs: 2\n- Date: \"Sun Jan 5 02:05:06 2020\"\n+ Date: \"Mon Dec 3 17:45:08 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n-## Fatal error in PMPI_Init: Other MPI error, error stack:\n-## MPIR_Init_thread(565)..............: \n-## MPID_Init(224).....................: channel initialization failed\n-## MPIDI_CH3_Init(105)................: \n-## MPID_nem_init(324).................: \n-## MPID_nem_tcp_init(178).............: \n-## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 612 - ./f77/io/iwriteatf 2\n ---\n Directory: ./f77/io\n File: iwriteatf\n Num-procs: 2\n- Date: \"Sun Jan 5 02:05:06 2020\"\n+ Date: \"Mon Dec 3 17:45:09 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n-## Fatal error in PMPI_Init: Other MPI error, error stack:\n-## MPIR_Init_thread(565)..............: \n-## MPID_Init(224).....................: channel initialization failed\n-## MPIDI_CH3_Init(105)................: \n-## MPID_nem_init(324).................: \n-## MPID_nem_tcp_init(178).............: \n-## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 613 - ./f77/io/iwritef 2\n ---\n Directory: ./f77/io\n File: iwritef\n Num-procs: 2\n- Date: \"Sun Jan 5 02:05:06 2020\"\n+ Date: \"Mon Dec 3 17:45:09 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n+## Fatal error in PMPI_Init: Other MPI error, error stack:\n+## MPIR_Init_thread(565)..............: \n+## MPID_Init(224).....................: channel initialization failed\n+## MPIDI_CH3_Init(105)................: \n+## MPID_nem_init(324).................: \n+## MPID_nem_tcp_init(178).............: \n+## MPID_nem_tcp_get_business_card(425): \n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 614 - ./f77/io/iwriteshf 2\n ---\n Directory: ./f77/io\n File: iwriteshf\n Num-procs: 2\n- Date: \"Sun Jan 5 02:05:07 2020\"\n+ Date: \"Mon Dec 3 17:45:10 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n ## Fatal error in PMPI_Init: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 615 - ./f77/io/writef 2\n ---\n Directory: ./f77/io\n File: writef\n Num-procs: 2\n- Date: \"Sun Jan 5 02:05:07 2020\"\n+ Date: \"Mon Dec 3 17:45:10 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 616 - ./f77/io/writeatf 2\n ---\n Directory: ./f77/io\n File: writeatf\n Num-procs: 2\n- Date: \"Sun Jan 5 02:05:07 2020\"\n+ Date: \"Mon Dec 3 17:45:11 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n+## Fatal error in PMPI_Init: Other MPI error, error stack:\n+## MPIR_Init_thread(565)..............: \n+## MPID_Init(224).....................: channel initialization failed\n+## MPIDI_CH3_Init(105)................: \n+## MPID_nem_init(324).................: \n+## MPID_nem_tcp_init(178).............: \n+## MPID_nem_tcp_get_business_card(425): \n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 617 - ./f77/io/writeallf 2\n ---\n Directory: ./f77/io\n File: writeallf\n Num-procs: 2\n- Date: \"Sun Jan 5 02:05:07 2020\"\n+ Date: \"Mon Dec 3 17:45:11 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 618 - ./f77/io/writeshf 2\n ---\n Directory: ./f77/io\n File: writeshf\n Num-procs: 2\n- Date: \"Sun Jan 5 02:05:07 2020\"\n+ Date: \"Mon Dec 3 17:45:11 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 619 - ./f77/io/writeordf 2\n ---\n Directory: ./f77/io\n File: writeordf\n Num-procs: 2\n- Date: \"Sun Jan 5 02:05:08 2020\"\n+ Date: \"Mon Dec 3 17:45:12 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n-## Fatal error in PMPI_Init: Other MPI error, error stack:\n-## MPIR_Init_thread(565)..............: \n-## MPID_Init(224).....................: channel initialization failed\n-## MPIDI_CH3_Init(105)................: \n-## MPID_nem_init(324).................: \n-## MPID_nem_tcp_init(178).............: \n-## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 620 - ./f77/io/writeatallf 2\n ---\n Directory: ./f77/io\n File: writeatallf\n Num-procs: 2\n- Date: \"Sun Jan 5 02:05:08 2020\"\n+ Date: \"Mon Dec 3 17:45:12 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 621 - ./f77/io/writeatallbef 2\n ---\n Directory: ./f77/io\n File: writeatallbef\n Num-procs: 2\n- Date: \"Sun Jan 5 02:05:08 2020\"\n+ Date: \"Mon Dec 3 17:45:12 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n-## Fatal error in PMPI_Init: Other MPI error, error stack:\n-## MPIR_Init_thread(565)..............: \n-## MPID_Init(224).....................: channel initialization failed\n-## MPIDI_CH3_Init(105)................: \n-## MPID_nem_init(324).................: \n-## MPID_nem_tcp_init(178).............: \n-## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 622 - ./f77/io/writeallbef 2\n ---\n Directory: ./f77/io\n File: writeallbef\n Num-procs: 2\n- Date: \"Sun Jan 5 02:05:08 2020\"\n+ Date: \"Mon Dec 3 17:45:13 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n-## Fatal error in PMPI_Init: Other MPI error, error stack:\n-## MPIR_Init_thread(565)..............: \n-## MPID_Init(224).....................: channel initialization failed\n-## MPIDI_CH3_Init(105)................: \n-## MPID_nem_init(324).................: \n-## MPID_nem_tcp_init(178).............: \n-## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 623 - ./f77/io/writeordbef 2\n ---\n Directory: ./f77/io\n File: writeordbef\n Num-procs: 2\n- Date: \"Sun Jan 5 02:05:09 2020\"\n+ Date: \"Mon Dec 3 17:45:13 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n-## Fatal error in PMPI_Init: Other MPI error, error stack:\n-## MPIR_Init_thread(565)..............: \n-## MPID_Init(224).....................: channel initialization failed\n-## MPIDI_CH3_Init(105)................: \n-## MPID_nem_init(324).................: \n-## MPID_nem_tcp_init(178).............: \n-## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 624 - ./f77/io/fileerrf 1\n ---\n Directory: ./f77/io\n File: fileerrf\n Num-procs: 1\n- Date: \"Sun Jan 5 02:05:09 2020\"\n+ Date: \"Mon Dec 3 17:45:14 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 625 - ./f77/io/fileinfof 2\n ---\n Directory: ./f77/io\n File: fileinfof\n Num-procs: 2\n- Date: \"Sun Jan 5 02:05:09 2020\"\n+ Date: \"Mon Dec 3 17:45:14 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n-## Fatal error in PMPI_Init: Other MPI error, error stack:\n-## MPIR_Init_thread(565)..............: \n-## MPID_Init(224).....................: channel initialization failed\n-## MPIDI_CH3_Init(105)................: \n-## MPID_nem_init(324).................: \n-## MPID_nem_tcp_init(178).............: \n-## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 626 - ./f77/io/shpositionf 2\n ---\n Directory: ./f77/io\n File: shpositionf\n Num-procs: 2\n- Date: \"Sun Jan 5 02:05:09 2020\"\n+ Date: \"Mon Dec 3 17:45:14 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n-## Fatal error in PMPI_Init: Other MPI error, error stack:\n-## MPIR_Init_thread(565)..............: \n-## MPID_Init(224).....................: channel initialization failed\n-## MPIDI_CH3_Init(105)................: \n-## MPID_nem_init(324).................: \n-## MPID_nem_tcp_init(178).............: \n-## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 627 - ./f77/io/atomicityf 2\n ---\n Directory: ./f77/io\n File: atomicityf\n Num-procs: 2\n- Date: \"Sun Jan 5 02:05:10 2020\"\n+ Date: \"Mon Dec 3 17:45:15 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n-## Fatal error in PMPI_Init: Other MPI error, error stack:\n-## MPIR_Init_thread(565)..............: \n-## MPID_Init(224).....................: channel initialization failed\n-## MPIDI_CH3_Init(105)................: \n-## MPID_nem_init(324).................: \n-## MPID_nem_tcp_init(178).............: \n-## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 628 - ./f77/io/miscfilef 2\n ---\n Directory: ./f77/io\n File: miscfilef\n Num-procs: 2\n- Date: \"Sun Jan 5 02:05:10 2020\"\n+ Date: \"Mon Dec 3 17:45:15 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n-## Fatal error in PMPI_Init: Other MPI error, error stack:\n-## MPIR_Init_thread(565)..............: \n-## MPID_Init(224).....................: channel initialization failed\n-## MPIDI_CH3_Init(105)................: \n-## MPID_nem_init(324).................: \n-## MPID_nem_tcp_init(178).............: \n-## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 629 - ./f77/io/setviewcurf 2\n ---\n Directory: ./f77/io\n File: setviewcurf\n Num-procs: 2\n- Date: \"Sun Jan 5 02:05:10 2020\"\n+ Date: \"Mon Dec 3 17:45:16 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 630 - ./f77/io/c2f2ciof 1\n ---\n Directory: ./f77/io\n File: c2f2ciof\n Num-procs: 1\n- Date: \"Sun Jan 5 02:05:10 2020\"\n+ Date: \"Mon Dec 3 17:45:16 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 631 - ./f77/io/c2fmultio 1\n ---\n Directory: ./f77/io\n File: c2fmultio\n Num-procs: 1\n- Date: \"Sun Jan 5 02:05:10 2020\"\n+ Date: \"Mon Dec 3 17:45:16 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n ok 646 - ./f77/io/i_setviewcurf 2 # SKIP requires MPI version 3.1\n ok 647 - ./f77/io/iwriteatallf 2 # SKIP requires MPI version 3.1\n not ok 632 - ./f77/rma/winscale1f 2\n ---\n Directory: ./f77/rma\n File: winscale1f\n Num-procs: 2\n- Date: \"Sun Jan 5 02:05:11 2020\"\n+ Date: \"Mon Dec 3 17:45:17 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n-## Fatal error in PMPI_Init: Other MPI error, error stack:\n-## MPIR_Init_thread(565)..............: \n-## MPID_Init(224).....................: channel initialization failed\n-## MPIDI_CH3_Init(105)................: \n-## MPID_nem_init(324).................: \n-## MPID_nem_tcp_init(178).............: \n-## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 633 - ./f77/rma/winfencef 2\n ---\n Directory: ./f77/rma\n File: winfencef\n Num-procs: 2\n- Date: \"Sun Jan 5 02:05:11 2020\"\n+ Date: \"Mon Dec 3 17:45:17 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n-## Fatal error in PMPI_Init: Other MPI error, error stack:\n-## MPIR_Init_thread(565)..............: \n-## MPID_Init(224).....................: channel initialization failed\n-## MPIDI_CH3_Init(105)................: \n-## MPID_nem_init(324).................: \n-## MPID_nem_tcp_init(178).............: \n-## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 634 - ./f77/rma/wingetf 2\n ---\n Directory: ./f77/rma\n File: wingetf\n Num-procs: 2\n- Date: \"Sun Jan 5 02:05:11 2020\"\n+ Date: \"Mon Dec 3 17:45:18 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n-## Fatal error in PMPI_Init: Other MPI error, error stack:\n-## MPIR_Init_thread(565)..............: \n-## MPID_Init(224).....................: channel initialization failed\n-## MPIDI_CH3_Init(105)................: \n-## MPID_nem_init(324).................: \n-## MPID_nem_tcp_init(178).............: \n-## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 635 - ./f77/rma/winscale2f 2\n ---\n Directory: ./f77/rma\n File: winscale2f\n Num-procs: 2\n- Date: \"Sun Jan 5 02:05:11 2020\"\n+ Date: \"Mon Dec 3 17:45:18 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 636 - ./f77/rma/winerrf 1\n ---\n Directory: ./f77/rma\n File: winerrf\n Num-procs: 1\n- Date: \"Sun Jan 5 02:05:11 2020\"\n+ Date: \"Mon Dec 3 17:45:18 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 637 - ./f77/rma/winnamef 1\n ---\n Directory: ./f77/rma\n File: winnamef\n Num-procs: 1\n- Date: \"Sun Jan 5 02:05:12 2020\"\n+ Date: \"Mon Dec 3 17:45:19 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 638 - ./f77/rma/wingroupf 2\n ---\n Directory: ./f77/rma\n File: wingroupf\n Num-procs: 2\n- Date: \"Sun Jan 5 02:05:12 2020\"\n+ Date: \"Mon Dec 3 17:45:19 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n-## Fatal error in PMPI_Init: Other MPI error, error stack:\n-## MPIR_Init_thread(565)..............: \n-## MPID_Init(224).....................: channel initialization failed\n-## MPIDI_CH3_Init(105)................: \n-## MPID_nem_init(324).................: \n-## MPID_nem_tcp_init(178).............: \n-## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 639 - ./f77/rma/winaccf 2\n ---\n Directory: ./f77/rma\n File: winaccf\n Num-procs: 2\n- Date: \"Sun Jan 5 02:05:12 2020\"\n+ Date: \"Mon Dec 3 17:45:20 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 640 - ./f77/rma/c2f2cwinf 1\n ---\n Directory: ./f77/rma\n File: c2f2cwinf\n Num-procs: 1\n- Date: \"Sun Jan 5 02:05:12 2020\"\n+ Date: \"Mon Dec 3 17:45:20 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 641 - ./f77/rma/baseattrwinf 1\n ---\n Directory: ./f77/rma\n File: baseattrwinf\n Num-procs: 1\n- Date: \"Sun Jan 5 02:05:12 2020\"\n+ Date: \"Mon Dec 3 17:45:20 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 642 - ./f77/rma/winattrf 1\n ---\n Directory: ./f77/rma\n File: winattrf\n Num-procs: 1\n- Date: \"Sun Jan 5 02:05:13 2020\"\n+ Date: \"Mon Dec 3 17:45:21 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 643 - ./f77/rma/winattr2f 1\n ---\n Directory: ./f77/rma\n File: winattr2f\n Num-procs: 1\n- Date: \"Sun Jan 5 02:05:13 2020\"\n+ Date: \"Mon Dec 3 17:45:21 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n ok 660 - ./f77/rma/aintf 2 # SKIP requires MPI version 3.1\n not ok 644 - ./f77/init/baseenvf 1\n ---\n Directory: ./f77/init\n File: baseenvf\n Num-procs: 1\n- Date: \"Sun Jan 5 02:05:13 2020\"\n+ Date: \"Mon Dec 3 17:45:21 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 645 - ./f77/comm/commnamef 2\n ---\n Directory: ./f77/comm\n File: commnamef\n Num-procs: 2\n- Date: \"Sun Jan 5 02:05:13 2020\"\n+ Date: \"Mon Dec 3 17:45:22 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n-not ok 646 - ./f77/comm/commerrf 2\n- ---\n- Directory: ./f77/comm\n- File: commerrf\n- Num-procs: 2\n- Date: \"Sun Jan 5 02:05:14 2020\"\n- ...\n-## Test output (expected 'No Errors'):\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n ## Fatal error in PMPI_Init: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n+not ok 646 - ./f77/comm/commerrf 2\n+ ---\n+ Directory: ./f77/comm\n+ File: commerrf\n+ Num-procs: 2\n+ Date: \"Mon Dec 3 17:45:22 2018\"\n+ ...\n+## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 647 - ./f77/ext/c2f2cf 1\n ---\n Directory: ./f77/ext\n File: c2f2cf\n Num-procs: 1\n- Date: \"Sun Jan 5 02:05:14 2020\"\n+ Date: \"Mon Dec 3 17:45:23 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 648 - ./f77/ext/c2fmult 1\n ---\n Directory: ./f77/ext\n File: c2fmult\n Num-procs: 1\n- Date: \"Sun Jan 5 02:05:14 2020\"\n+ Date: \"Mon Dec 3 17:45:23 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 649 - ./f77/ext/ctypesinf 1\n ---\n Directory: ./f77/ext\n File: ctypesinf\n Num-procs: 1\n- Date: \"Sun Jan 5 02:05:14 2020\"\n+ Date: \"Mon Dec 3 17:45:23 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 650 - ./f77/ext/allocmemf 1\n ---\n Directory: ./f77/ext\n File: allocmemf\n Num-procs: 1\n- Date: \"Sun Jan 5 02:05:15 2020\"\n+ Date: \"Mon Dec 3 17:45:34 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 651 - ./f77/topo/cartcrf 2\n ---\n Directory: ./f77/topo\n File: cartcrf\n Num-procs: 2\n- Date: \"Sun Jan 5 02:05:16 2020\"\n+ Date: \"Mon Dec 3 17:45:35 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n-## Fatal error in PMPI_Init: Other MPI error, error stack:\n-## MPIR_Init_thread(565)..............: \n-## MPID_Init(224).....................: channel initialization failed\n-## MPIDI_CH3_Init(105)................: \n-## MPID_nem_init(324).................: \n-## MPID_nem_tcp_init(178).............: \n-## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 652 - ./f77/topo/dgraph_wgtf 2\n ---\n Directory: ./f77/topo\n File: dgraph_wgtf\n Num-procs: 2\n- Date: \"Sun Jan 5 02:05:16 2020\"\n+ Date: \"Mon Dec 3 17:45:35 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n-## Fatal error in PMPI_Init: Other MPI error, error stack:\n-## MPIR_Init_thread(565)..............: \n-## MPID_Init(224).....................: channel initialization failed\n-## MPIDI_CH3_Init(105)................: \n-## MPID_nem_init(324).................: \n-## MPID_nem_tcp_init(178).............: \n-## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 653 - ./f77/topo/dgraph_unwgtf 2\n ---\n Directory: ./f77/topo\n File: dgraph_unwgtf\n Num-procs: 2\n- Date: \"Sun Jan 5 02:05:16 2020\"\n+ Date: \"Mon Dec 3 17:45:36 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n-## Fatal error in PMPI_Init: Other MPI error, error stack:\n-## MPIR_Init_thread(565)..............: \n-## MPID_Init(224).....................: channel initialization failed\n-## MPIDI_CH3_Init(105)................: \n-## MPID_nem_init(324).................: \n-## MPID_nem_tcp_init(178).............: \n-## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 654 - ./f77/profile/profile1f 2\n ---\n Directory: ./f77/profile\n File: profile1f\n Num-procs: 2\n- Date: \"Sun Jan 5 02:05:17 2020\"\n+ Date: \"Mon Dec 3 17:45:36 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n-## Fatal error in PMPI_Init: Other MPI error, error stack:\n-## MPIR_Init_thread(565)..............: \n-## MPID_Init(224).....................: channel initialization failed\n-## MPIDI_CH3_Init(105)................: \n-## MPID_nem_init(324).................: \n-## MPID_nem_tcp_init(178).............: \n-## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 655 - ./cxx/attr/attrtx 2\n ---\n Directory: ./cxx/attr\n File: attrtx\n Num-procs: 2\n- Date: \"Sun Jan 5 02:05:17 2020\"\n+ Date: \"Mon Dec 3 17:45:37 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n-not ok 656 - ./cxx/attr/attricx 2\n- ---\n- Directory: ./cxx/attr\n- File: attricx\n- Num-procs: 2\n- Date: \"Sun Jan 5 02:05:17 2020\"\n- ...\n-## Test output (expected 'No Errors'):\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n+not ok 656 - ./cxx/attr/attricx 2\n+ ---\n+ Directory: ./cxx/attr\n+ File: attricx\n+ Num-procs: 2\n+ Date: \"Mon Dec 3 17:45:37 2018\"\n+ ...\n+## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 657 - ./cxx/attr/baseattrcommx 1\n ---\n Directory: ./cxx/attr\n File: baseattrcommx\n Num-procs: 1\n- Date: \"Sun Jan 5 02:05:17 2020\"\n+ Date: \"Mon Dec 3 17:45:38 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 658 - ./cxx/attr/fkeyvalcommx 1\n ---\n Directory: ./cxx/attr\n File: fkeyvalcommx\n Num-procs: 1\n- Date: \"Sun Jan 5 02:05:18 2020\"\n+ Date: \"Mon Dec 3 17:45:38 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 659 - ./cxx/attr/fkeyvaltypex 1\n ---\n Directory: ./cxx/attr\n File: fkeyvaltypex\n Num-procs: 1\n- Date: \"Sun Jan 5 02:05:18 2020\"\n+ Date: \"Mon Dec 3 17:45:39 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 660 - ./cxx/pt2pt/bsend1cxx 2\n ---\n Directory: ./cxx/pt2pt\n File: bsend1cxx\n Num-procs: 2\n- Date: \"Sun Jan 5 02:05:18 2020\"\n+ Date: \"Mon Dec 3 17:45:39 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n-## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n-## MPIR_Init_thread(565)..............: \n-## MPID_Init(224).....................: channel initialization failed\n-## MPIDI_CH3_Init(105)................: \n-## MPID_nem_init(324).................: \n-## MPID_nem_tcp_init(178).............: \n-## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 661 - ./cxx/pt2pt/sendrecvx 2\n ---\n Directory: ./cxx/pt2pt\n File: sendrecvx\n Num-procs: 2\n- Date: \"Sun Jan 5 02:05:18 2020\"\n+ Date: \"Mon Dec 3 17:45:40 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n+## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n+## MPIR_Init_thread(565)..............: \n+## MPID_Init(224).....................: channel initialization failed\n+## MPIDI_CH3_Init(105)................: \n+## MPID_nem_init(324).................: \n+## MPID_nem_tcp_init(178).............: \n+## MPID_nem_tcp_get_business_card(425): \n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 662 - ./cxx/comm/commname2 2\n ---\n Directory: ./cxx/comm\n File: commname2\n Num-procs: 2\n- Date: \"Sun Jan 5 02:05:19 2020\"\n+ Date: \"Mon Dec 3 17:45:40 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 663 - ./cxx/coll/arcomplex 2\n ---\n Directory: ./cxx/coll\n File: arcomplex\n Num-procs: 2\n- Date: \"Sun Jan 5 02:05:19 2020\"\n+ Date: \"Mon Dec 3 17:45:41 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n-## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n-## MPIR_Init_thread(565)..............: \n-## MPID_Init(224).....................: channel initialization failed\n-## MPIDI_CH3_Init(105)................: \n-## MPID_nem_init(324).................: \n-## MPID_nem_tcp_init(178).............: \n-## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 664 - ./cxx/coll/uallredx 2\n ---\n Directory: ./cxx/coll\n File: uallredx\n Num-procs: 2\n- Date: \"Sun Jan 5 02:05:19 2020\"\n+ Date: \"Mon Dec 3 17:45:41 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n-## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n-## MPIR_Init_thread(565)..............: \n-## MPID_Init(224).....................: channel initialization failed\n-## MPIDI_CH3_Init(105)................: \n-## MPID_nem_init(324).................: \n-## MPID_nem_tcp_init(178).............: \n-## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 665 - ./cxx/coll/uallreduce 2\n ---\n Directory: ./cxx/coll\n File: uallreduce\n Num-procs: 2\n- Date: \"Sun Jan 5 02:05:19 2020\"\n+ Date: \"Mon Dec 3 17:45:42 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 666 - ./cxx/coll/ureduce 2\n ---\n Directory: ./cxx/coll\n File: ureduce\n Num-procs: 2\n- Date: \"Sun Jan 5 02:05:20 2020\"\n+ Date: \"Mon Dec 3 17:45:42 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n-## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n-## MPIR_Init_thread(565)..............: \n-## MPID_Init(224).....................: channel initialization failed\n-## MPIDI_CH3_Init(105)................: \n-## MPID_nem_init(324).................: \n-## MPID_nem_tcp_init(178).............: \n-## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 667 - ./cxx/coll/ureducelocal 2\n ---\n Directory: ./cxx/coll\n File: ureducelocal\n Num-procs: 2\n- Date: \"Sun Jan 5 02:05:20 2020\"\n+ Date: \"Mon Dec 3 17:45:43 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 668 - ./cxx/coll/uscan 2\n ---\n Directory: ./cxx/coll\n File: uscan\n Num-procs: 2\n- Date: \"Sun Jan 5 02:05:20 2020\"\n+ Date: \"Mon Dec 3 17:45:43 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 669 - ./cxx/coll/uexscan 2\n ---\n Directory: ./cxx/coll\n File: uexscan\n Num-procs: 2\n- Date: \"Sun Jan 5 02:05:21 2020\"\n+ Date: \"Mon Dec 3 17:45:44 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n-## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n-## MPIR_Init_thread(565)..............: \n-## MPID_Init(224).....................: channel initialization failed\n-## MPIDI_CH3_Init(105)................: \n-## MPID_nem_init(324).................: \n-## MPID_nem_tcp_init(178).............: \n-## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 670 - ./cxx/coll/alltoallw2x 2\n ---\n Directory: ./cxx/coll\n File: alltoallw2x\n Num-procs: 2\n- Date: \"Sun Jan 5 02:05:21 2020\"\n+ Date: \"Mon Dec 3 17:45:44 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 671 - ./cxx/coll/icbcastx 2\n ---\n Directory: ./cxx/coll\n File: icbcastx\n Num-procs: 2\n- Date: \"Sun Jan 5 02:05:21 2020\"\n+ Date: \"Mon Dec 3 17:45:44 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n-## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n-## MPIR_Init_thread(565)..............: \n-## MPID_Init(224).....................: channel initialization failed\n-## MPIDI_CH3_Init(105)................: \n-## MPID_nem_init(324).................: \n-## MPID_nem_tcp_init(178).............: \n-## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 672 - ./cxx/coll/icbcastx 2\n ---\n Directory: ./cxx/coll\n File: icbcastx\n Num-procs: 2\n- Date: \"Sun Jan 5 02:05:21 2020\"\n+ Date: \"Mon Dec 3 17:45:45 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n-## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n-## MPIR_Init_thread(565)..............: \n-## MPID_Init(224).....................: channel initialization failed\n-## MPIDI_CH3_Init(105)................: \n-## MPID_nem_init(324).................: \n-## MPID_nem_tcp_init(178).............: \n-## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 673 - ./cxx/coll/icallreducex 2\n ---\n Directory: ./cxx/coll\n File: icallreducex\n Num-procs: 2\n- Date: \"Sun Jan 5 02:05:22 2020\"\n+ Date: \"Mon Dec 3 17:45:45 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 674 - ./cxx/coll/icreducex 2\n ---\n Directory: ./cxx/coll\n File: icreducex\n Num-procs: 2\n- Date: \"Sun Jan 5 02:05:22 2020\"\n+ Date: \"Mon Dec 3 17:45:46 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n-## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n-## MPIR_Init_thread(565)..............: \n-## MPID_Init(224).....................: channel initialization failed\n-## MPIDI_CH3_Init(105)................: \n-## MPID_nem_init(324).................: \n-## MPID_nem_tcp_init(178).............: \n-## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 675 - ./cxx/coll/icscatterx 2\n ---\n Directory: ./cxx/coll\n File: icscatterx\n Num-procs: 2\n- Date: \"Sun Jan 5 02:05:22 2020\"\n+ Date: \"Mon Dec 3 17:45:46 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 676 - ./cxx/coll/icgatherx 2\n ---\n Directory: ./cxx/coll\n File: icgatherx\n Num-procs: 2\n- Date: \"Sun Jan 5 02:05:23 2020\"\n+ Date: \"Mon Dec 3 17:45:47 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 677 - ./cxx/coll/icallgatherx 2\n ---\n Directory: ./cxx/coll\n File: icallgatherx\n Num-procs: 2\n- Date: \"Sun Jan 5 02:05:23 2020\"\n+ Date: \"Mon Dec 3 17:45:47 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n-## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n-## MPIR_Init_thread(565)..............: \n-## MPID_Init(224).....................: channel initialization failed\n-## MPIDI_CH3_Init(105)................: \n-## MPID_nem_init(324).................: \n-## MPID_nem_tcp_init(178).............: \n-## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 678 - ./cxx/coll/icbarrierx 2\n ---\n Directory: ./cxx/coll\n File: icbarrierx\n Num-procs: 2\n- Date: \"Sun Jan 5 02:05:23 2020\"\n+ Date: \"Mon Dec 3 17:45:48 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 679 - ./cxx/coll/icallgathervx 2\n ---\n Directory: ./cxx/coll\n File: icallgathervx\n Num-procs: 2\n- Date: \"Sun Jan 5 02:05:23 2020\"\n+ Date: \"Mon Dec 3 17:45:48 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 680 - ./cxx/coll/icgathervx 2\n ---\n Directory: ./cxx/coll\n File: icgathervx\n Num-procs: 2\n- Date: \"Sun Jan 5 02:05:24 2020\"\n+ Date: \"Mon Dec 3 17:45:49 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n+## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n+## MPIR_Init_thread(565)..............: \n+## MPID_Init(224).....................: channel initialization failed\n+## MPIDI_CH3_Init(105)................: \n+## MPID_nem_init(324).................: \n+## MPID_nem_tcp_init(178).............: \n+## MPID_nem_tcp_get_business_card(425): \n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 681 - ./cxx/coll/icscattervx 2\n ---\n Directory: ./cxx/coll\n File: icscattervx\n Num-procs: 2\n- Date: \"Sun Jan 5 02:05:24 2020\"\n+ Date: \"Mon Dec 3 17:45:49 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 682 - ./cxx/coll/icalltoallx 2\n ---\n Directory: ./cxx/coll\n File: icalltoallx\n Num-procs: 2\n- Date: \"Sun Jan 5 02:05:24 2020\"\n+ Date: \"Mon Dec 3 17:45:49 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n+## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n+## MPIR_Init_thread(565)..............: \n+## MPID_Init(224).....................: channel initialization failed\n+## MPIDI_CH3_Init(105)................: \n+## MPID_nem_init(324).................: \n+## MPID_nem_tcp_init(178).............: \n+## MPID_nem_tcp_get_business_card(425): \n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 683 - ./cxx/coll/reduceboolx 2\n ---\n Directory: ./cxx/coll\n File: reduceboolx\n Num-procs: 2\n- Date: \"Sun Jan 5 02:05:25 2020\"\n+ Date: \"Mon Dec 3 17:45:50 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n+## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n+## MPIR_Init_thread(565)..............: \n+## MPID_Init(224).....................: channel initialization failed\n+## MPIDI_CH3_Init(105)................: \n+## MPID_nem_init(324).................: \n+## MPID_nem_tcp_init(178).............: \n+## MPID_nem_tcp_get_business_card(425): \n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 684 - ./cxx/coll/redscatblk 2\n ---\n Directory: ./cxx/coll\n File: redscatblk\n Num-procs: 2\n- Date: \"Sun Jan 5 02:05:25 2020\"\n+ Date: \"Mon Dec 3 17:45:50 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n-## Fatal error in PMPI_Init: Other MPI error, error stack:\n-## MPIR_Init_thread(565)..............: \n-## MPID_Init(224).....................: channel initialization failed\n-## MPIDI_CH3_Init(105)................: \n-## MPID_nem_init(324).................: \n-## MPID_nem_tcp_init(178).............: \n-## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 685 - ./cxx/errhan/commcallx 2\n ---\n Directory: ./cxx/errhan\n File: commcallx\n Num-procs: 2\n- Date: \"Sun Jan 5 02:05:25 2020\"\n+ Date: \"Mon Dec 3 17:45:51 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 686 - ./cxx/init/baseenv 1\n ---\n Directory: ./cxx/init\n File: baseenv\n Num-procs: 1\n- Date: \"Sun Jan 5 02:05:25 2020\"\n+ Date: \"Mon Dec 3 17:45:51 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 687 - ./cxx/init/initstatx 1\n ---\n Directory: ./cxx/init\n File: initstatx\n Num-procs: 1\n- Date: \"Sun Jan 5 02:05:25 2020\"\n+ Date: \"Mon Dec 3 17:45:52 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 688 - ./cxx/init/initstat2x 1\n ---\n Directory: ./cxx/init\n File: initstat2x\n Num-procs: 1\n- Date: \"Sun Jan 5 02:05:25 2020\"\n+ Date: \"Mon Dec 3 17:45:52 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 689 - ./cxx/info/infodupx 1\n ---\n Directory: ./cxx/info\n File: infodupx\n Num-procs: 1\n- Date: \"Sun Jan 5 02:05:26 2020\"\n+ Date: \"Mon Dec 3 17:45:52 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 690 - ./cxx/info/infodelx 1\n ---\n Directory: ./cxx/info\n File: infodelx\n Num-procs: 1\n- Date: \"Sun Jan 5 02:05:26 2020\"\n+ Date: \"Mon Dec 3 17:45:53 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 691 - ./cxx/info/infovallenx 1\n ---\n Directory: ./cxx/info\n File: infovallenx\n Num-procs: 1\n- Date: \"Sun Jan 5 02:05:26 2020\"\n+ Date: \"Mon Dec 3 17:45:53 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 692 - ./cxx/info/infoorderx 1\n ---\n Directory: ./cxx/info\n File: infoorderx\n Num-procs: 1\n- Date: \"Sun Jan 5 02:05:26 2020\"\n+ Date: \"Mon Dec 3 17:45:54 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 693 - ./cxx/datatype/packsizex 1\n ---\n Directory: ./cxx/datatype\n File: packsizex\n Num-procs: 1\n- Date: \"Sun Jan 5 02:05:27 2020\"\n+ Date: \"Mon Dec 3 17:45:54 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 694 - ./cxx/datatype/typecntsx 1\n ---\n Directory: ./cxx/datatype\n File: typecntsx\n Num-procs: 1\n- Date: \"Sun Jan 5 02:05:27 2020\"\n+ Date: \"Mon Dec 3 17:45:54 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 695 - ./cxx/datatype/typenamex 1\n ---\n Directory: ./cxx/datatype\n File: typenamex\n Num-procs: 1\n- Date: \"Sun Jan 5 02:05:27 2020\"\n+ Date: \"Mon Dec 3 17:45:55 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 696 - ./cxx/datatype/typemiscx 1\n ---\n Directory: ./cxx/datatype\n File: typemiscx\n Num-procs: 1\n- Date: \"Sun Jan 5 02:05:27 2020\"\n+ Date: \"Mon Dec 3 17:45:55 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 697 - ./cxx/io/iwriteatx 2\n ---\n Directory: ./cxx/io\n File: iwriteatx\n Num-procs: 2\n- Date: \"Sun Jan 5 02:05:28 2020\"\n+ Date: \"Mon Dec 3 17:45:56 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n-## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n-## MPIR_Init_thread(565)..............: \n-## MPID_Init(224).....................: channel initialization failed\n-## MPIDI_CH3_Init(105)................: \n-## MPID_nem_init(324).................: \n-## MPID_nem_tcp_init(178).............: \n-## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 698 - ./cxx/io/iwritex 2\n ---\n Directory: ./cxx/io\n File: iwritex\n Num-procs: 2\n- Date: \"Sun Jan 5 02:05:28 2020\"\n+ Date: \"Mon Dec 3 17:45:56 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n-## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n-## MPIR_Init_thread(565)..............: \n-## MPID_Init(224).....................: channel initialization failed\n-## MPIDI_CH3_Init(105)................: \n-## MPID_nem_init(324).................: \n-## MPID_nem_tcp_init(178).............: \n-## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 699 - ./cxx/io/iwriteshx 2\n ---\n Directory: ./cxx/io\n File: iwriteshx\n Num-procs: 2\n- Date: \"Sun Jan 5 02:05:28 2020\"\n+ Date: \"Mon Dec 3 17:45:57 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 700 - ./cxx/io/writex 2\n ---\n Directory: ./cxx/io\n File: writex\n Num-procs: 2\n- Date: \"Sun Jan 5 02:05:28 2020\"\n+ Date: \"Mon Dec 3 17:45:57 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n-## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n-## MPIR_Init_thread(565)..............: \n-## MPID_Init(224).....................: channel initialization failed\n-## MPIDI_CH3_Init(105)................: \n-## MPID_nem_init(324).................: \n-## MPID_nem_tcp_init(178).............: \n-## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 701 - ./cxx/io/writeatx 2\n ---\n Directory: ./cxx/io\n File: writeatx\n Num-procs: 2\n- Date: \"Sun Jan 5 02:05:29 2020\"\n+ Date: \"Mon Dec 3 17:45:58 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 702 - ./cxx/io/writeallx 2\n ---\n Directory: ./cxx/io\n File: writeallx\n Num-procs: 2\n- Date: \"Sun Jan 5 02:05:29 2020\"\n+ Date: \"Mon Dec 3 17:45:58 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n-## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n-## MPIR_Init_thread(565)..............: \n-## MPID_Init(224).....................: channel initialization failed\n-## MPIDI_CH3_Init(105)................: \n-## MPID_nem_init(324).................: \n-## MPID_nem_tcp_init(178).............: \n-## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 703 - ./cxx/io/writeshx 2\n ---\n Directory: ./cxx/io\n File: writeshx\n Num-procs: 2\n- Date: \"Sun Jan 5 02:05:29 2020\"\n+ Date: \"Mon Dec 3 17:45:59 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 704 - ./cxx/io/writeordx 2\n ---\n Directory: ./cxx/io\n File: writeordx\n Num-procs: 2\n- Date: \"Sun Jan 5 02:05:30 2020\"\n+ Date: \"Mon Dec 3 17:45:59 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 705 - ./cxx/io/writeatallx 2\n ---\n Directory: ./cxx/io\n File: writeatallx\n Num-procs: 2\n- Date: \"Sun Jan 5 02:05:30 2020\"\n+ Date: \"Mon Dec 3 17:46:00 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 706 - ./cxx/io/writeatallbex 2\n ---\n Directory: ./cxx/io\n File: writeatallbex\n Num-procs: 2\n- Date: \"Sun Jan 5 02:05:30 2020\"\n+ Date: \"Mon Dec 3 17:46:00 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 707 - ./cxx/io/writeallbex 2\n ---\n Directory: ./cxx/io\n File: writeallbex\n Num-procs: 2\n- Date: \"Sun Jan 5 02:05:31 2020\"\n+ Date: \"Mon Dec 3 17:46:01 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n+## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n+## MPIR_Init_thread(565)..............: \n+## MPID_Init(224).....................: channel initialization failed\n+## MPIDI_CH3_Init(105)................: \n+## MPID_nem_init(324).................: \n+## MPID_nem_tcp_init(178).............: \n+## MPID_nem_tcp_get_business_card(425): \n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 708 - ./cxx/io/writeordbex 2\n ---\n Directory: ./cxx/io\n File: writeordbex\n Num-procs: 2\n- Date: \"Sun Jan 5 02:05:31 2020\"\n+ Date: \"Mon Dec 3 17:46:01 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 709 - ./cxx/io/iwriteatnosx 2\n ---\n Directory: ./cxx/io\n File: iwriteatnosx\n Num-procs: 2\n- Date: \"Sun Jan 5 02:05:31 2020\"\n+ Date: \"Mon Dec 3 17:46:02 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 710 - ./cxx/io/iwritenosx 2\n ---\n Directory: ./cxx/io\n File: iwritenosx\n Num-procs: 2\n- Date: \"Sun Jan 5 02:05:31 2020\"\n+ Date: \"Mon Dec 3 17:46:02 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n-## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n-## MPIR_Init_thread(565)..............: \n-## MPID_Init(224).....................: channel initialization failed\n-## MPIDI_CH3_Init(105)................: \n-## MPID_nem_init(324).................: \n-## MPID_nem_tcp_init(178).............: \n-## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 711 - ./cxx/io/iwriteshnosx 2\n ---\n Directory: ./cxx/io\n File: iwriteshnosx\n Num-procs: 2\n- Date: \"Sun Jan 5 02:05:31 2020\"\n+ Date: \"Mon Dec 3 17:46:03 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n-not ok 712 - ./cxx/io/writenosx 2\n- ---\n- Directory: ./cxx/io\n- File: writenosx\n- Num-procs: 2\n- Date: \"Sun Jan 5 02:05:32 2020\"\n- ...\n-## Test output (expected 'No Errors'):\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n+not ok 712 - ./cxx/io/writenosx 2\n+ ---\n+ Directory: ./cxx/io\n+ File: writenosx\n+ Num-procs: 2\n+ Date: \"Mon Dec 3 17:46:03 2018\"\n+ ...\n+## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 713 - ./cxx/io/writeatnosx 2\n ---\n Directory: ./cxx/io\n File: writeatnosx\n Num-procs: 2\n- Date: \"Sun Jan 5 02:05:32 2020\"\n+ Date: \"Mon Dec 3 17:46:04 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 714 - ./cxx/io/writeallnosx 2\n ---\n Directory: ./cxx/io\n File: writeallnosx\n Num-procs: 2\n- Date: \"Sun Jan 5 02:05:32 2020\"\n+ Date: \"Mon Dec 3 17:46:04 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n-## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n-## MPIR_Init_thread(565)..............: \n-## MPID_Init(224).....................: channel initialization failed\n-## MPIDI_CH3_Init(105)................: \n-## MPID_nem_init(324).................: \n-## MPID_nem_tcp_init(178).............: \n-## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 715 - ./cxx/io/writeshnosx 2\n ---\n Directory: ./cxx/io\n File: writeshnosx\n Num-procs: 2\n- Date: \"Sun Jan 5 02:05:33 2020\"\n+ Date: \"Mon Dec 3 17:46:05 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 716 - ./cxx/io/writeordnosx 2\n ---\n Directory: ./cxx/io\n File: writeordnosx\n Num-procs: 2\n- Date: \"Sun Jan 5 02:05:33 2020\"\n+ Date: \"Mon Dec 3 17:46:05 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 717 - ./cxx/io/writeatallnosx 2\n ---\n Directory: ./cxx/io\n File: writeatallnosx\n Num-procs: 2\n- Date: \"Sun Jan 5 02:05:33 2020\"\n+ Date: \"Mon Dec 3 17:46:06 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 718 - ./cxx/io/writeatallbenosx 2\n ---\n Directory: ./cxx/io\n File: writeatallbenosx\n Num-procs: 2\n- Date: \"Sun Jan 5 02:05:33 2020\"\n+ Date: \"Mon Dec 3 17:46:06 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 719 - ./cxx/io/writeallbenosx 2\n ---\n Directory: ./cxx/io\n File: writeallbenosx\n Num-procs: 2\n- Date: \"Sun Jan 5 02:05:34 2020\"\n+ Date: \"Mon Dec 3 17:46:07 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 720 - ./cxx/io/writeordbenosx 2\n ---\n Directory: ./cxx/io\n File: writeordbenosx\n Num-procs: 2\n- Date: \"Sun Jan 5 02:05:34 2020\"\n+ Date: \"Mon Dec 3 17:46:07 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n+## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n+## MPIR_Init_thread(565)..............: \n+## MPID_Init(224).....................: channel initialization failed\n+## MPIDI_CH3_Init(105)................: \n+## MPID_nem_init(324).................: \n+## MPID_nem_tcp_init(178).............: \n+## MPID_nem_tcp_get_business_card(425): \n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 721 - ./cxx/io/fileerrx 1\n ---\n Directory: ./cxx/io\n File: fileerrx\n Num-procs: 1\n- Date: \"Sun Jan 5 02:05:34 2020\"\n+ Date: \"Mon Dec 3 17:46:08 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 722 - ./cxx/io/fileinfox 2\n ---\n Directory: ./cxx/io\n File: fileinfox\n Num-procs: 2\n- Date: \"Sun Jan 5 02:05:35 2020\"\n+ Date: \"Mon Dec 3 17:46:08 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 723 - ./cxx/io/filemiscx 2\n ---\n Directory: ./cxx/io\n File: filemiscx\n Num-procs: 2\n- Date: \"Sun Jan 5 02:05:35 2020\"\n+ Date: \"Mon Dec 3 17:46:09 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n-## Fatal error in PMPI_Init: Other MPI error, error stack:\n-## MPIR_Init_thread(565)..............: \n-## MPID_Init(224).....................: channel initialization failed\n-## MPIDI_CH3_Init(105)................: \n-## MPID_nem_init(324).................: \n-## MPID_nem_tcp_init(178).............: \n-## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 724 - ./cxx/io/shpositionx 2\n ---\n Directory: ./cxx/io\n File: shpositionx\n Num-procs: 2\n- Date: \"Sun Jan 5 02:05:35 2020\"\n+ Date: \"Mon Dec 3 17:46:09 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 725 - ./cxx/io/seekavail 1\n ---\n Directory: ./cxx/io\n File: seekavail\n Num-procs: 1\n- Date: \"Sun Jan 5 02:05:35 2020\"\n+ Date: \"Mon Dec 3 17:46:10 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 726 - ./cxx/spawn/namepubx 2\n ---\n Directory: ./cxx/spawn\n File: namepubx\n Num-procs: 2\n- Date: \"Sun Jan 5 02:05:36 2020\"\n+ Date: \"Mon Dec 3 17:46:10 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 727 - ./cxx/spawn/spawnintrax 1\n ---\n Directory: ./cxx/spawn\n File: spawnintrax\n Num-procs: 1\n- Date: \"Sun Jan 5 02:05:36 2020\"\n+ Date: \"Mon Dec 3 17:46:10 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 728 - ./cxx/spawn/spawnintrax 2\n ---\n Directory: ./cxx/spawn\n File: spawnintrax\n Num-procs: 2\n- Date: \"Sun Jan 5 02:05:36 2020\"\n+ Date: \"Mon Dec 3 17:46:11 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n-## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n-## MPIR_Init_thread(565)..............: \n-## MPID_Init(224).....................: channel initialization failed\n-## MPIDI_CH3_Init(105)................: \n-## MPID_nem_init(324).................: \n-## MPID_nem_tcp_init(178).............: \n-## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 729 - ./cxx/spawn/spawnargvx 1\n ---\n Directory: ./cxx/spawn\n File: spawnargvx\n Num-procs: 1\n- Date: \"Sun Jan 5 02:05:37 2020\"\n+ Date: \"Mon Dec 3 17:46:11 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 730 - ./cxx/spawn/selfconaccx 2\n ---\n Directory: ./cxx/spawn\n File: selfconaccx\n Num-procs: 2\n- Date: \"Sun Jan 5 02:05:37 2020\"\n+ Date: \"Mon Dec 3 17:46:12 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n-## Fatal error in PMPI_Init: Other MPI error, error stack:\n-## MPIR_Init_thread(565)..............: \n-## MPID_Init(224).....................: channel initialization failed\n-## MPIDI_CH3_Init(105)................: \n-## MPID_nem_init(324).................: \n-## MPID_nem_tcp_init(178).............: \n-## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 731 - ./cxx/rma/winnamex 1\n ---\n Directory: ./cxx/rma\n File: winnamex\n Num-procs: 1\n- Date: \"Sun Jan 5 02:05:37 2020\"\n+ Date: \"Mon Dec 3 17:46:12 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 732 - ./cxx/rma/wincallx 1\n ---\n Directory: ./cxx/rma\n File: wincallx\n Num-procs: 1\n- Date: \"Sun Jan 5 02:05:38 2020\"\n+ Date: \"Mon Dec 3 17:46:12 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 733 - ./cxx/rma/getgroupx 2\n ---\n Directory: ./cxx/rma\n File: getgroupx\n Num-procs: 2\n- Date: \"Sun Jan 5 02:05:38 2020\"\n+ Date: \"Mon Dec 3 17:46:13 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n-## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n-## MPIR_Init_thread(565)..............: \n-## MPID_Init(224).....................: channel initialization failed\n-## MPIDI_CH3_Init(105)................: \n-## MPID_nem_init(324).................: \n-## MPID_nem_tcp_init(178).............: \n-## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 734 - ./cxx/rma/winfencex 2\n ---\n Directory: ./cxx/rma\n File: winfencex\n Num-procs: 2\n- Date: \"Sun Jan 5 02:05:38 2020\"\n+ Date: \"Mon Dec 3 17:46:13 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n-## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n-## MPIR_Init_thread(565)..............: \n-## MPID_Init(224).....................: channel initialization failed\n-## MPIDI_CH3_Init(105)................: \n-## MPID_nem_init(324).................: \n-## MPID_nem_tcp_init(178).............: \n-## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 735 - ./cxx/rma/winscale1x 2\n ---\n Directory: ./cxx/rma\n File: winscale1x\n Num-procs: 2\n- Date: \"Sun Jan 5 02:05:38 2020\"\n+ Date: \"Mon Dec 3 17:46:14 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 736 - ./cxx/rma/winscale2x 2\n ---\n Directory: ./cxx/rma\n File: winscale2x\n Num-procs: 2\n- Date: \"Sun Jan 5 02:05:39 2020\"\n+ Date: \"Mon Dec 3 17:46:14 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n-## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n-## MPIR_Init_thread(565)..............: \n-## MPID_Init(224).....................: channel initialization failed\n-## MPIDI_CH3_Init(105)................: \n-## MPID_nem_init(324).................: \n-## MPID_nem_tcp_init(178).............: \n-## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 737 - ./cxx/rma/fkeyvalwinx 1\n ---\n Directory: ./cxx/rma\n File: fkeyvalwinx\n Num-procs: 1\n- Date: \"Sun Jan 5 02:05:39 2020\"\n+ Date: \"Mon Dec 3 17:46:14 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 738 - ./f90/attr/attrmpi1f90 1\n ---\n Directory: ./f90/attr\n File: attrmpi1f90\n Num-procs: 1\n- Date: \"Sun Jan 5 02:05:39 2020\"\n+ Date: \"Mon Dec 3 17:46:15 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 739 - ./f90/attr/baseattrf90 1\n ---\n Directory: ./f90/attr\n File: baseattrf90\n Num-procs: 1\n- Date: \"Sun Jan 5 02:05:39 2020\"\n+ Date: \"Mon Dec 3 17:46:15 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 740 - ./f90/attr/baseattr2f90 1\n ---\n Directory: ./f90/attr\n File: baseattr2f90\n Num-procs: 1\n- Date: \"Sun Jan 5 02:05:39 2020\"\n+ Date: \"Mon Dec 3 17:46:15 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 741 - ./f90/attr/commattrf90 1\n ---\n Directory: ./f90/attr\n File: commattrf90\n Num-procs: 1\n- Date: \"Sun Jan 5 02:05:40 2020\"\n+ Date: \"Mon Dec 3 17:46:15 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 742 - ./f90/attr/commattr2f90 1\n ---\n Directory: ./f90/attr\n File: commattr2f90\n Num-procs: 1\n- Date: \"Sun Jan 5 02:05:40 2020\"\n+ Date: \"Mon Dec 3 17:46:16 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 743 - ./f90/attr/commattr3f90 1\n ---\n Directory: ./f90/attr\n File: commattr3f90\n Num-procs: 1\n- Date: \"Sun Jan 5 02:05:40 2020\"\n+ Date: \"Mon Dec 3 17:46:16 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 744 - ./f90/attr/commattr4f90 1\n ---\n Directory: ./f90/attr\n File: commattr4f90\n Num-procs: 1\n- Date: \"Sun Jan 5 02:05:40 2020\"\n+ Date: \"Mon Dec 3 17:46:16 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 745 - ./f90/attr/typeattrf90 1\n ---\n Directory: ./f90/attr\n File: typeattrf90\n Num-procs: 1\n- Date: \"Sun Jan 5 02:05:41 2020\"\n+ Date: \"Mon Dec 3 17:46:17 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 746 - ./f90/attr/typeattr2f90 1\n ---\n Directory: ./f90/attr\n File: typeattr2f90\n Num-procs: 1\n- Date: \"Sun Jan 5 02:05:41 2020\"\n+ Date: \"Mon Dec 3 17:46:17 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 747 - ./f90/attr/typeattr3f90 1\n ---\n Directory: ./f90/attr\n File: typeattr3f90\n Num-procs: 1\n- Date: \"Sun Jan 5 02:05:41 2020\"\n+ Date: \"Mon Dec 3 17:46:17 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 748 - ./f90/attr/fandcattrf90 1\n ---\n Directory: ./f90/attr\n File: fandcattrf90\n Num-procs: 1\n- Date: \"Sun Jan 5 02:05:41 2020\"\n+ Date: \"Mon Dec 3 17:46:18 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 749 - ./f90/attr/baseattr3f90 1\n ---\n Directory: ./f90/attr\n File: baseattr3f90\n Num-procs: 1\n- Date: \"Sun Jan 5 02:05:42 2020\"\n+ Date: \"Mon Dec 3 17:46:18 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 750 - ./f90/attr/attrlangf90 1\n ---\n Directory: ./f90/attr\n File: attrlangf90\n Num-procs: 1\n- Date: \"Sun Jan 5 02:05:42 2020\"\n+ Date: \"Mon Dec 3 17:46:18 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 751 - ./f90/coll/uallreducef90 2\n ---\n Directory: ./f90/coll\n File: uallreducef90\n Num-procs: 2\n- Date: \"Sun Jan 5 02:05:42 2020\"\n+ Date: \"Mon Dec 3 17:46:19 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n-## Fatal error in PMPI_Init: Other MPI error, error stack:\n-## MPIR_Init_thread(565)..............: \n-## MPID_Init(224).....................: channel initialization failed\n-## MPIDI_CH3_Init(105)................: \n-## MPID_nem_init(324).................: \n-## MPID_nem_tcp_init(178).............: \n-## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 752 - ./f90/coll/exscanf90 2\n ---\n Directory: ./f90/coll\n File: exscanf90\n Num-procs: 2\n- Date: \"Sun Jan 5 02:05:43 2020\"\n- ...\n-## Test output (expected 'No Errors'):\n-## Fatal error in PMPI_Init: Other MPI error, error stack:\n-## MPIR_Init_thread(565)..............: \n-## MPID_Init(224).....................: channel initialization failed\n-## MPIDI_CH3_Init(105)................: \n-## MPID_nem_init(324).................: \n-## MPID_nem_tcp_init(178).............: \n-## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n-## [mpiexec@profitbricks-build5-amd64] control_cb (pm/pmiserv/pmiserv_cb.c:208): assert (!closed) failed\n-## [mpiexec@profitbricks-build5-amd64] HYDT_dmxu_poll_wait_for_event (tools/demux/demux_poll.c:77): callback returned error status\n-## [mpiexec@profitbricks-build5-amd64] HYDT_bscu_wait_for_completion (tools/bootstrap/utils/bscu_wait.c:60): error waiting for event\n-## [mpiexec@profitbricks-build5-amd64] HYDT_bsci_wait_for_completion (tools/bootstrap/src/bsci_wait.c:22): launcher returned error waiting for completion\n-not ok 753 - ./f90/coll/alltoallwf90 2\n- ---\n- Directory: ./f90/coll\n- File: alltoallwf90\n- Num-procs: 2\n- Date: \"Sun Jan 5 02:05:43 2020\"\n+ Date: \"Mon Dec 3 17:46:19 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n ## Fatal error in PMPI_Init: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n-not ok 754 - ./f90/coll/alltoallvf90 2\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n+not ok 753 - ./f90/coll/alltoallwf90 2\n ---\n Directory: ./f90/coll\n- File: alltoallvf90\n+ File: alltoallwf90\n Num-procs: 2\n- Date: \"Sun Jan 5 02:05:43 2020\"\n+ Date: \"Mon Dec 3 17:46:19 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n-not ok 755 - ./f90/coll/inplacef90 2\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n+not ok 754 - ./f90/coll/alltoallvf90 2\n ---\n Directory: ./f90/coll\n- File: inplacef90\n+ File: alltoallvf90\n Num-procs: 2\n- Date: \"Sun Jan 5 02:05:43 2020\"\n+ Date: \"Mon Dec 3 17:46:20 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n ## Fatal error in PMPI_Init: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n-not ok 756 - ./f90/coll/reducelocalf90 2\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n+not ok 755 - ./f90/coll/inplacef90 2\n ---\n Directory: ./f90/coll\n- File: reducelocalf90\n+ File: inplacef90\n Num-procs: 2\n- Date: \"Sun Jan 5 02:05:44 2020\"\n+ Date: \"Mon Dec 3 17:46:20 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n+not ok 756 - ./f90/coll/reducelocalf90 2\n+ ---\n+ Directory: ./f90/coll\n+ File: reducelocalf90\n+ Num-procs: 2\n+ Date: \"Mon Dec 3 17:46:20 2018\"\n+ ...\n+## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 757 - ./f90/coll/redscatf90 2\n ---\n Directory: ./f90/coll\n File: redscatf90\n Num-procs: 2\n- Date: \"Sun Jan 5 02:05:44 2020\"\n+ Date: \"Mon Dec 3 17:46:21 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 758 - ./f90/coll/split_typef90 2\n ---\n Directory: ./f90/coll\n File: split_typef90\n Num-procs: 2\n- Date: \"Sun Jan 5 02:05:44 2020\"\n+ Date: \"Mon Dec 3 17:46:21 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n-## Fatal error in PMPI_Init: Other MPI error, error stack:\n-## MPIR_Init_thread(565)..............: \n-## MPID_Init(224).....................: channel initialization failed\n-## MPIDI_CH3_Init(105)................: \n-## MPID_nem_init(324).................: \n-## MPID_nem_tcp_init(178).............: \n-## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 759 - ./f90/coll/nonblockingf90 2\n ---\n Directory: ./f90/coll\n File: nonblockingf90\n Num-procs: 2\n- Date: \"Sun Jan 5 02:05:45 2020\"\n+ Date: \"Mon Dec 3 17:46:21 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n-not ok 760 - ./f90/coll/vw_inplacef90 2\n- ---\n- Directory: ./f90/coll\n- File: vw_inplacef90\n- Num-procs: 2\n- Date: \"Sun Jan 5 02:05:45 2020\"\n- ...\n-## Test output (expected 'No Errors'):\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n ## Fatal error in PMPI_Init: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n+not ok 760 - ./f90/coll/vw_inplacef90 2\n+ ---\n+ Directory: ./f90/coll\n+ File: vw_inplacef90\n+ Num-procs: 2\n+ Date: \"Mon Dec 3 17:46:22 2018\"\n+ ...\n+## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 761 - ./f90/coll/red_scat_blockf90 2\n ---\n Directory: ./f90/coll\n File: red_scat_blockf90\n Num-procs: 2\n- Date: \"Sun Jan 5 02:05:45 2020\"\n+ Date: \"Mon Dec 3 17:46:22 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n ## Fatal error in PMPI_Init: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 762 - ./f90/coll/nonblocking_inpf90 2\n ---\n Directory: ./f90/coll\n File: nonblocking_inpf90\n Num-procs: 2\n- Date: \"Sun Jan 5 02:05:46 2020\"\n+ Date: \"Mon Dec 3 17:46:22 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 763 - ./f90/comm/commnamef90 2\n ---\n Directory: ./f90/comm\n File: commnamef90\n Num-procs: 2\n- Date: \"Sun Jan 5 02:05:47 2020\"\n+ Date: \"Mon Dec 3 17:46:23 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n ## Fatal error in PMPI_Init: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 764 - ./f90/comm/commerrf90 2\n ---\n Directory: ./f90/comm\n File: commerrf90\n Num-procs: 2\n- Date: \"Sun Jan 5 02:05:47 2020\"\n+ Date: \"Mon Dec 3 17:46:23 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n-## Fatal error in PMPI_Init: Other MPI error, error stack:\n-## MPIR_Init_thread(565)..............: \n-## MPID_Init(224).....................: channel initialization failed\n-## MPIDI_CH3_Init(105)................: \n-## MPID_nem_init(324).................: \n-## MPID_nem_tcp_init(178).............: \n-## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 765 - ./f90/ext/c2f2cf90 1\n ---\n Directory: ./f90/ext\n File: c2f2cf90\n Num-procs: 1\n- Date: \"Sun Jan 5 02:05:47 2020\"\n+ Date: \"Mon Dec 3 17:46:24 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 766 - ./f90/ext/c2f90mult 1\n ---\n Directory: ./f90/ext\n File: c2f90mult\n Num-procs: 1\n- Date: \"Sun Jan 5 02:05:48 2020\"\n+ Date: \"Mon Dec 3 17:46:24 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 767 - ./f90/ext/ctypesinf90 1\n ---\n Directory: ./f90/ext\n File: ctypesinf90\n Num-procs: 1\n- Date: \"Sun Jan 5 02:05:48 2020\"\n+ Date: \"Mon Dec 3 17:46:24 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 768 - ./f90/ext/allocmemf90 1\n ---\n Directory: ./f90/ext\n File: allocmemf90\n Num-procs: 1\n- Date: \"Sun Jan 5 02:05:49 2020\"\n+ Date: \"Mon Dec 3 17:46:26 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 769 - ./f90/info/infotestf90 1\n ---\n Directory: ./f90/info\n File: infotestf90\n Num-procs: 1\n- Date: \"Sun Jan 5 02:05:50 2020\"\n+ Date: \"Mon Dec 3 17:46:27 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 770 - ./f90/info/infotest2f90 1\n ---\n Directory: ./f90/info\n File: infotest2f90\n Num-procs: 1\n- Date: \"Sun Jan 5 02:05:50 2020\"\n+ Date: \"Mon Dec 3 17:46:27 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 771 - ./f90/init/baseenvf90 1\n ---\n Directory: ./f90/init\n File: baseenvf90\n Num-procs: 1\n- Date: \"Sun Jan 5 02:05:50 2020\"\n+ Date: \"Mon Dec 3 17:46:27 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 772 - ./f90/io/iwriteatf90 2\n ---\n Directory: ./f90/io\n File: iwriteatf90\n Num-procs: 2\n- Date: \"Sun Jan 5 02:05:51 2020\"\n+ Date: \"Mon Dec 3 17:46:28 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n+## Fatal error in PMPI_Init: Other MPI error, error stack:\n+## MPIR_Init_thread(565)..............: \n+## MPID_Init(224).....................: channel initialization failed\n+## MPIDI_CH3_Init(105)................: \n+## MPID_nem_init(324).................: \n+## MPID_nem_tcp_init(178).............: \n+## MPID_nem_tcp_get_business_card(425): \n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 773 - ./f90/io/iwritef90 2\n ---\n Directory: ./f90/io\n File: iwritef90\n Num-procs: 2\n- Date: \"Sun Jan 5 02:05:51 2020\"\n+ Date: \"Mon Dec 3 17:46:28 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n ## Fatal error in PMPI_Init: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 774 - ./f90/io/iwriteshf90 2\n ---\n Directory: ./f90/io\n File: iwriteshf90\n Num-procs: 2\n- Date: \"Sun Jan 5 02:05:51 2020\"\n+ Date: \"Mon Dec 3 17:46:29 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 775 - ./f90/io/writef90 2\n ---\n Directory: ./f90/io\n File: writef90\n Num-procs: 2\n- Date: \"Sun Jan 5 02:05:52 2020\"\n+ Date: \"Mon Dec 3 17:46:29 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n-## Fatal error in PMPI_Init: Other MPI error, error stack:\n-## MPIR_Init_thread(565)..............: \n-## MPID_Init(224).....................: channel initialization failed\n-## MPIDI_CH3_Init(105)................: \n-## MPID_nem_init(324).................: \n-## MPID_nem_tcp_init(178).............: \n-## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 776 - ./f90/io/writeatf90 2\n ---\n Directory: ./f90/io\n File: writeatf90\n Num-procs: 2\n- Date: \"Sun Jan 5 02:05:53 2020\"\n+ Date: \"Mon Dec 3 17:46:30 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n+## Fatal error in PMPI_Init: Other MPI error, error stack:\n+## MPIR_Init_thread(565)..............: \n+## MPID_Init(224).....................: channel initialization failed\n+## MPIDI_CH3_Init(105)................: \n+## MPID_nem_init(324).................: \n+## MPID_nem_tcp_init(178).............: \n+## MPID_nem_tcp_get_business_card(425): \n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 777 - ./f90/io/writeallf90 2\n ---\n Directory: ./f90/io\n File: writeallf90\n Num-procs: 2\n- Date: \"Sun Jan 5 02:05:53 2020\"\n+ Date: \"Mon Dec 3 17:46:30 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 778 - ./f90/io/writeshf90 2\n ---\n Directory: ./f90/io\n File: writeshf90\n Num-procs: 2\n- Date: \"Sun Jan 5 02:05:54 2020\"\n+ Date: \"Mon Dec 3 17:46:30 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 779 - ./f90/io/writeordf90 2\n ---\n Directory: ./f90/io\n File: writeordf90\n Num-procs: 2\n- Date: \"Sun Jan 5 02:05:55 2020\"\n+ Date: \"Mon Dec 3 17:46:31 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 780 - ./f90/io/writeatallf90 2\n ---\n Directory: ./f90/io\n File: writeatallf90\n Num-procs: 2\n- Date: \"Sun Jan 5 02:05:55 2020\"\n+ Date: \"Mon Dec 3 17:46:31 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 781 - ./f90/io/writeatallbef90 2\n ---\n Directory: ./f90/io\n File: writeatallbef90\n Num-procs: 2\n- Date: \"Sun Jan 5 02:05:55 2020\"\n+ Date: \"Mon Dec 3 17:46:31 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n-## Fatal error in PMPI_Init: Other MPI error, error stack:\n-## MPIR_Init_thread(565)..............: \n-## MPID_Init(224).....................: channel initialization failed\n-## MPIDI_CH3_Init(105)................: \n-## MPID_nem_init(324).................: \n-## MPID_nem_tcp_init(178).............: \n-## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 782 - ./f90/io/writeallbef90 2\n ---\n Directory: ./f90/io\n File: writeallbef90\n Num-procs: 2\n- Date: \"Sun Jan 5 02:05:56 2020\"\n+ Date: \"Mon Dec 3 17:46:32 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 783 - ./f90/io/writeordbef90 2\n ---\n Directory: ./f90/io\n File: writeordbef90\n Num-procs: 2\n- Date: \"Sun Jan 5 02:05:56 2020\"\n+ Date: \"Mon Dec 3 17:46:32 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n+## Fatal error in PMPI_Init: Other MPI error, error stack:\n+## MPIR_Init_thread(565)..............: \n+## MPID_Init(224).....................: channel initialization failed\n+## MPIDI_CH3_Init(105)................: \n+## MPID_nem_init(324).................: \n+## MPID_nem_tcp_init(178).............: \n+## MPID_nem_tcp_get_business_card(425): \n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 784 - ./f90/io/fileerrf90 1\n ---\n Directory: ./f90/io\n File: fileerrf90\n Num-procs: 1\n- Date: \"Sun Jan 5 02:05:57 2020\"\n+ Date: \"Mon Dec 3 17:46:32 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 785 - ./f90/io/fileinfof90 2\n ---\n Directory: ./f90/io\n File: fileinfof90\n Num-procs: 2\n- Date: \"Sun Jan 5 02:05:57 2020\"\n+ Date: \"Mon Dec 3 17:46:33 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 786 - ./f90/io/shpositionf90 2\n ---\n Directory: ./f90/io\n File: shpositionf90\n Num-procs: 2\n- Date: \"Sun Jan 5 02:05:58 2020\"\n+ Date: \"Mon Dec 3 17:46:33 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 787 - ./f90/io/atomicityf90 2\n ---\n Directory: ./f90/io\n File: atomicityf90\n Num-procs: 2\n- Date: \"Sun Jan 5 02:05:58 2020\"\n+ Date: \"Mon Dec 3 17:46:34 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n ## Fatal error in PMPI_Init: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 788 - ./f90/io/miscfilef90 2\n ---\n Directory: ./f90/io\n File: miscfilef90\n Num-procs: 2\n- Date: \"Sun Jan 5 02:05:59 2020\"\n+ Date: \"Mon Dec 3 17:46:34 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n-## Fatal error in PMPI_Init: Other MPI error, error stack:\n-## MPIR_Init_thread(565)..............: \n-## MPID_Init(224).....................: channel initialization failed\n-## MPIDI_CH3_Init(105)................: \n-## MPID_nem_init(324).................: \n-## MPID_nem_tcp_init(178).............: \n-## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 789 - ./f90/io/setviewcurf90 2\n ---\n Directory: ./f90/io\n File: setviewcurf90\n Num-procs: 2\n- Date: \"Sun Jan 5 02:05:59 2020\"\n+ Date: \"Mon Dec 3 17:46:34 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 790 - ./f90/io/c2f2ciof90 1\n ---\n Directory: ./f90/io\n File: c2f2ciof90\n Num-procs: 1\n- Date: \"Sun Jan 5 02:06:00 2020\"\n+ Date: \"Mon Dec 3 17:46:35 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 791 - ./f90/io/c2f90multio 1\n ---\n Directory: ./f90/io\n File: c2f90multio\n Num-procs: 1\n- Date: \"Sun Jan 5 02:06:01 2020\"\n+ Date: \"Mon Dec 3 17:46:35 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n ok 809 - ./f90/io/i_setviewcurf90 2 # SKIP requires MPI version 3.1\n ok 810 - ./f90/io/iwriteatallf90 2 # SKIP requires MPI version 3.1\n not ok 792 - ./f90/misc/sizeof2 1\n ---\n Directory: ./f90/misc\n File: sizeof2\n Num-procs: 1\n- Date: \"Sun Jan 5 02:06:01 2020\"\n+ Date: \"Mon Dec 3 17:46:36 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 793 - ./f90/pt2pt/statusesf90 1\n ---\n Directory: ./f90/pt2pt\n File: statusesf90\n Num-procs: 1\n- Date: \"Sun Jan 5 02:06:01 2020\"\n+ Date: \"Mon Dec 3 17:46:36 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 794 - ./f90/pt2pt/greqf90 1\n ---\n Directory: ./f90/pt2pt\n File: greqf90\n Num-procs: 1\n- Date: \"Sun Jan 5 02:06:02 2020\"\n+ Date: \"Mon Dec 3 17:46:36 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 795 - ./f90/pt2pt/allpairf90 2\n ---\n Directory: ./f90/pt2pt\n File: allpairf90\n Num-procs: 2\n- Date: \"Sun Jan 5 02:06:02 2020\"\n+ Date: \"Mon Dec 3 17:46:37 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 796 - ./f90/pt2pt/mprobef90 2\n ---\n Directory: ./f90/pt2pt\n File: mprobef90\n Num-procs: 2\n- Date: \"Sun Jan 5 02:06:03 2020\"\n+ Date: \"Mon Dec 3 17:46:37 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n-## Fatal error in PMPI_Init: Other MPI error, error stack:\n-## MPIR_Init_thread(565)..............: \n-## MPID_Init(224).....................: channel initialization failed\n-## MPIDI_CH3_Init(105)................: \n-## MPID_nem_init(324).................: \n-## MPID_nem_tcp_init(178).............: \n-## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 797 - ./f90/datatype/typenamef90 1\n ---\n Directory: ./f90/datatype\n File: typenamef90\n Num-procs: 1\n- Date: \"Sun Jan 5 02:06:03 2020\"\n+ Date: \"Mon Dec 3 17:46:37 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 798 - ./f90/datatype/typename3f90 1\n ---\n Directory: ./f90/datatype\n File: typename3f90\n Num-procs: 1\n- Date: \"Sun Jan 5 02:06:04 2020\"\n+ Date: \"Mon Dec 3 17:46:38 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 799 - ./f90/datatype/typesnamef90 1\n ---\n Directory: ./f90/datatype\n File: typesnamef90\n Num-procs: 1\n- Date: \"Sun Jan 5 02:06:04 2020\"\n+ Date: \"Mon Dec 3 17:46:38 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 800 - ./f90/datatype/typecntsf90 1\n ---\n Directory: ./f90/datatype\n File: typecntsf90\n Num-procs: 1\n- Date: \"Sun Jan 5 02:06:05 2020\"\n+ Date: \"Mon Dec 3 17:46:38 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 801 - ./f90/datatype/typem2f90 1\n ---\n Directory: ./f90/datatype\n File: typem2f90\n Num-procs: 1\n- Date: \"Sun Jan 5 02:06:05 2020\"\n+ Date: \"Mon Dec 3 17:46:39 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 802 - ./f90/datatype/typesubf90 1\n ---\n Directory: ./f90/datatype\n File: typesubf90\n Num-procs: 1\n- Date: \"Sun Jan 5 02:06:06 2020\"\n+ Date: \"Mon Dec 3 17:46:39 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 803 - ./f90/datatype/packef90 1\n ---\n Directory: ./f90/datatype\n File: packef90\n Num-procs: 1\n- Date: \"Sun Jan 5 02:06:06 2020\"\n+ Date: \"Mon Dec 3 17:46:39 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 804 - ./f90/datatype/gaddressf90 1\n ---\n Directory: ./f90/datatype\n File: gaddressf90\n Num-procs: 1\n- Date: \"Sun Jan 5 02:06:07 2020\"\n+ Date: \"Mon Dec 3 17:46:39 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 805 - ./f90/datatype/allctypesf90 1\n ---\n Directory: ./f90/datatype\n File: allctypesf90\n Num-procs: 1\n- Date: \"Sun Jan 5 02:06:07 2020\"\n+ Date: \"Mon Dec 3 17:46:40 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 806 - ./f90/datatype/hindex1f90 1\n ---\n Directory: ./f90/datatype\n File: hindex1f90\n Num-procs: 1\n- Date: \"Sun Jan 5 02:06:08 2020\"\n+ Date: \"Mon Dec 3 17:46:40 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 807 - ./f90/datatype/hindexed_blockf90 1\n ---\n Directory: ./f90/datatype\n File: hindexed_blockf90\n Num-procs: 1\n- Date: \"Sun Jan 5 02:06:08 2020\"\n+ Date: \"Mon Dec 3 17:46:40 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 808 - ./f90/datatype/bottom 2\n ---\n Directory: ./f90/datatype\n File: bottom\n Num-procs: 2\n- Date: \"Sun Jan 5 02:06:09 2020\"\n+ Date: \"Mon Dec 3 17:46:41 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 809 - ./f90/datatype/structf 2\n ---\n Directory: ./f90/datatype\n File: structf\n Num-procs: 2\n- Date: \"Sun Jan 5 02:06:10 2020\"\n+ Date: \"Mon Dec 3 17:46:41 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 810 - ./f90/datatype/indtype 2\n ---\n Directory: ./f90/datatype\n File: indtype\n Num-procs: 2\n- Date: \"Sun Jan 5 02:06:10 2020\"\n+ Date: \"Mon Dec 3 17:46:41 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 811 - ./f90/datatype/createf90 1\n ---\n Directory: ./f90/datatype\n File: createf90\n Num-procs: 1\n- Date: \"Sun Jan 5 02:06:11 2020\"\n+ Date: \"Mon Dec 3 17:46:42 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 812 - ./f90/datatype/sizeof 1\n ---\n Directory: ./f90/datatype\n File: sizeof\n Num-procs: 1\n- Date: \"Sun Jan 5 02:06:11 2020\"\n+ Date: \"Mon Dec 3 17:46:42 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 813 - ./f90/datatype/kinds 2\n ---\n Directory: ./f90/datatype\n File: kinds\n Num-procs: 2\n- Date: \"Sun Jan 5 02:06:12 2020\"\n+ Date: \"Mon Dec 3 17:46:42 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 814 - ./f90/datatype/trf90 1\n ---\n Directory: ./f90/datatype\n File: trf90\n Num-procs: 1\n- Date: \"Sun Jan 5 02:06:12 2020\"\n+ Date: \"Mon Dec 3 17:46:43 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 815 - ./f90/datatype/get_elem_d 2\n ---\n Directory: ./f90/datatype\n File: get_elem_d\n Num-procs: 2\n- Date: \"Sun Jan 5 02:06:13 2020\"\n+ Date: \"Mon Dec 3 17:46:43 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 816 - ./f90/datatype/get_elem_u 2\n ---\n Directory: ./f90/datatype\n File: get_elem_u\n Num-procs: 2\n- Date: \"Sun Jan 5 02:06:13 2020\"\n+ Date: \"Mon Dec 3 17:46:43 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 817 - ./f90/f90types/createf90types 1\n ---\n Directory: ./f90/f90types\n File: createf90types\n Num-procs: 1\n- Date: \"Sun Jan 5 02:06:14 2020\"\n+ Date: \"Mon Dec 3 17:46:44 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 818 - ./f90/f90types/createf90types 1\n ---\n Directory: ./f90/f90types\n File: createf90types\n Num-procs: 1\n- Date: \"Sun Jan 5 02:06:14 2020\"\n+ Date: \"Mon Dec 3 17:46:44 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 819 - ./f90/rma/winscale1f90 2\n ---\n Directory: ./f90/rma\n File: winscale1f90\n Num-procs: 2\n- Date: \"Sun Jan 5 02:06:15 2020\"\n+ Date: \"Mon Dec 3 17:46:44 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n-## Fatal error in PMPI_Init: Other MPI error, error stack:\n-## MPIR_Init_thread(565)..............: \n-## MPID_Init(224).....................: channel initialization failed\n-## MPIDI_CH3_Init(105)................: \n-## MPID_nem_init(324).................: \n-## MPID_nem_tcp_init(178).............: \n-## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 820 - ./f90/rma/winfencef90 2\n ---\n Directory: ./f90/rma\n File: winfencef90\n Num-procs: 2\n- Date: \"Sun Jan 5 02:06:15 2020\"\n+ Date: \"Mon Dec 3 17:46:45 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 821 - ./f90/rma/wingetf90 2\n ---\n Directory: ./f90/rma\n File: wingetf90\n Num-procs: 2\n- Date: \"Sun Jan 5 02:06:16 2020\"\n+ Date: \"Mon Dec 3 17:46:45 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 822 - ./f90/rma/winscale2f90 2\n ---\n Directory: ./f90/rma\n File: winscale2f90\n Num-procs: 2\n- Date: \"Sun Jan 5 02:06:16 2020\"\n+ Date: \"Mon Dec 3 17:46:45 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n ## Fatal error in PMPI_Init: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 823 - ./f90/rma/winerrf90 1\n ---\n Directory: ./f90/rma\n File: winerrf90\n Num-procs: 1\n- Date: \"Sun Jan 5 02:06:17 2020\"\n+ Date: \"Mon Dec 3 17:46:46 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 824 - ./f90/rma/winnamef90 1\n ---\n Directory: ./f90/rma\n File: winnamef90\n Num-procs: 1\n- Date: \"Sun Jan 5 02:06:18 2020\"\n+ Date: \"Mon Dec 3 17:46:46 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 825 - ./f90/rma/wingroupf90 2\n ---\n Directory: ./f90/rma\n File: wingroupf90\n Num-procs: 2\n- Date: \"Sun Jan 5 02:06:18 2020\"\n+ Date: \"Mon Dec 3 17:46:46 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 826 - ./f90/rma/winaccf90 2\n ---\n Directory: ./f90/rma\n File: winaccf90\n Num-procs: 2\n- Date: \"Sun Jan 5 02:06:19 2020\"\n+ Date: \"Mon Dec 3 17:46:47 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 827 - ./f90/rma/c2f2cwinf90 1\n ---\n Directory: ./f90/rma\n File: c2f2cwinf90\n Num-procs: 1\n- Date: \"Sun Jan 5 02:06:19 2020\"\n+ Date: \"Mon Dec 3 17:46:47 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 828 - ./f90/rma/baseattrwinf90 1\n ---\n Directory: ./f90/rma\n File: baseattrwinf90\n Num-procs: 1\n- Date: \"Sun Jan 5 02:06:19 2020\"\n+ Date: \"Mon Dec 3 17:46:47 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 829 - ./f90/rma/winattrf90 1\n ---\n Directory: ./f90/rma\n File: winattrf90\n Num-procs: 1\n- Date: \"Sun Jan 5 02:06:20 2020\"\n+ Date: \"Mon Dec 3 17:46:47 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 830 - ./f90/rma/winattr2f90 1\n ---\n Directory: ./f90/rma\n File: winattr2f90\n Num-procs: 1\n- Date: \"Sun Jan 5 02:06:20 2020\"\n+ Date: \"Mon Dec 3 17:46:48 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n ok 850 - ./f90/rma/aintf90 2 # SKIP requires MPI version 3.1\n not ok 831 - ./f90/spawn/namepubf90 2\n ---\n Directory: ./f90/spawn\n File: namepubf90\n Num-procs: 2\n- Date: \"Sun Jan 5 02:06:21 2020\"\n+ Date: \"Mon Dec 3 17:46:48 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 832 - ./f90/spawn/spawnf90 1\n ---\n Directory: ./f90/spawn\n File: spawnf90\n Num-procs: 1\n- Date: \"Sun Jan 5 02:06:21 2020\"\n+ Date: \"Mon Dec 3 17:46:48 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 833 - ./f90/spawn/spawnargvf90 1\n ---\n Directory: ./f90/spawn\n File: spawnargvf90\n Num-procs: 1\n- Date: \"Sun Jan 5 02:06:22 2020\"\n+ Date: \"Mon Dec 3 17:46:49 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 834 - ./f90/spawn/connaccf90 2\n ---\n Directory: ./f90/spawn\n File: connaccf90\n Num-procs: 2\n- Date: \"Sun Jan 5 02:06:22 2020\"\n+ Date: \"Mon Dec 3 17:46:49 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n ## Fatal error in PMPI_Init: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 835 - ./f90/spawn/spawnmultf90 1\n ---\n Directory: ./f90/spawn\n File: spawnmultf90\n Num-procs: 1\n- Date: \"Sun Jan 5 02:06:23 2020\"\n+ Date: \"Mon Dec 3 17:46:49 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 836 - ./f90/spawn/spawnmult2f90 2\n ---\n Directory: ./f90/spawn\n File: spawnmult2f90\n Num-procs: 2\n- Date: \"Sun Jan 5 02:06:23 2020\"\n+ Date: \"Mon Dec 3 17:46:50 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n-## Fatal error in PMPI_Init: Other MPI error, error stack:\n-## MPIR_Init_thread(565)..............: \n-## MPID_Init(224).....................: channel initialization failed\n-## MPIDI_CH3_Init(105)................: \n-## MPID_nem_init(324).................: \n-## MPID_nem_tcp_init(178).............: \n-## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 837 - ./f90/spawn/spawnargvf03 1\n ---\n Directory: ./f90/spawn\n File: spawnargvf03\n Num-procs: 1\n- Date: \"Sun Jan 5 02:06:24 2020\"\n+ Date: \"Mon Dec 3 17:46:50 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 838 - ./f90/spawn/spawnmultf03 1\n ---\n Directory: ./f90/spawn\n File: spawnmultf03\n Num-procs: 1\n- Date: \"Sun Jan 5 02:06:24 2020\"\n+ Date: \"Mon Dec 3 17:46:50 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 839 - ./f90/timer/wtimef90 1\n ---\n Directory: ./f90/timer\n File: wtimef90\n Num-procs: 1\n- Date: \"Sun Jan 5 02:06:25 2020\"\n+ Date: \"Mon Dec 3 17:46:51 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 840 - ./f90/topo/cartcrf90 2\n ---\n Directory: ./f90/topo\n File: cartcrf90\n Num-procs: 2\n- Date: \"Sun Jan 5 02:06:25 2020\"\n+ Date: \"Mon Dec 3 17:46:51 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n-## Fatal error in PMPI_Init: Other MPI error, error stack:\n-## MPIR_Init_thread(565)..............: \n-## MPID_Init(224).....................: channel initialization failed\n-## MPIDI_CH3_Init(105)................: \n-## MPID_nem_init(324).................: \n-## MPID_nem_tcp_init(178).............: \n-## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 841 - ./f90/topo/dgraph_wgtf90 2\n ---\n Directory: ./f90/topo\n File: dgraph_wgtf90\n Num-procs: 2\n- Date: \"Sun Jan 5 02:06:26 2020\"\n+ Date: \"Mon Dec 3 17:46:52 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 842 - ./f90/topo/dgraph_unwgtf90 2\n ---\n Directory: ./f90/topo\n File: dgraph_unwgtf90\n Num-procs: 2\n- Date: \"Sun Jan 5 02:06:27 2020\"\n+ Date: \"Mon Dec 3 17:46:52 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n-## Fatal error in PMPI_Init: Other MPI error, error stack:\n-## MPIR_Init_thread(565)..............: \n-## MPID_Init(224).....................: channel initialization failed\n-## MPIDI_CH3_Init(105)................: \n-## MPID_nem_init(324).................: \n-## MPID_nem_tcp_init(178).............: \n-## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 843 - ./f90/profile/profile1f90 2\n ---\n Directory: ./f90/profile\n File: profile1f90\n Num-procs: 2\n- Date: \"Sun Jan 5 02:06:27 2020\"\n+ Date: \"Mon Dec 3 17:46:52 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 844 - ./errors/attr/keyvalmis 1\n ---\n Directory: ./errors/attr\n File: keyvalmis\n Num-procs: 1\n- Date: \"Sun Jan 5 02:06:28 2020\"\n+ Date: \"Mon Dec 3 17:46:53 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 845 - ./errors/coll/noalias 2\n ---\n Directory: ./errors/coll\n File: noalias\n Num-procs: 2\n- Date: \"Sun Jan 5 02:06:29 2020\"\n+ Date: \"Mon Dec 3 17:46:53 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 846 - ./errors/coll/nb_noalias 2\n ---\n Directory: ./errors/coll\n File: nb_noalias\n Num-procs: 2\n- Date: \"Sun Jan 5 02:06:29 2020\"\n+ Date: \"Mon Dec 3 17:46:54 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 847 - ./errors/coll/noalias2 2\n ---\n Directory: ./errors/coll\n File: noalias2\n Num-procs: 2\n- Date: \"Sun Jan 5 02:06:30 2020\"\n+ Date: \"Mon Dec 3 17:46:54 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n-## Fatal error in PMPI_Init: Other MPI error, error stack:\n-## MPIR_Init_thread(565)..............: \n-## MPID_Init(224).....................: channel initialization failed\n-## MPIDI_CH3_Init(105)................: \n-## MPID_nem_init(324).................: \n-## MPID_nem_tcp_init(178).............: \n-## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 848 - ./errors/coll/noalias3 2\n ---\n Directory: ./errors/coll\n File: noalias3\n Num-procs: 2\n- Date: \"Sun Jan 5 02:06:30 2020\"\n+ Date: \"Mon Dec 3 17:46:54 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 849 - ./errors/coll/rerr 2\n ---\n Directory: ./errors/coll\n File: rerr\n Num-procs: 2\n- Date: \"Sun Jan 5 02:06:31 2020\"\n+ Date: \"Mon Dec 3 17:46:55 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 850 - ./errors/coll/nb_rerr 2\n ---\n Directory: ./errors/coll\n File: nb_rerr\n Num-procs: 2\n- Date: \"Sun Jan 5 02:06:32 2020\"\n+ Date: \"Mon Dec 3 17:46:55 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n+## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n+## MPIR_Init_thread(565)..............: \n+## MPID_Init(224).....................: channel initialization failed\n+## MPIDI_CH3_Init(105)................: \n+## MPID_nem_init(324).................: \n+## MPID_nem_tcp_init(178).............: \n+## MPID_nem_tcp_get_business_card(425): \n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 851 - ./errors/coll/reduce_local 1\n ---\n Directory: ./errors/coll\n File: reduce_local\n Num-procs: 1\n- Date: \"Sun Jan 5 02:06:32 2020\"\n+ Date: \"Mon Dec 3 17:46:55 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 852 - ./errors/coll/bcastlength 2\n ---\n Directory: ./errors/coll\n File: bcastlength\n Num-procs: 2\n- Date: \"Sun Jan 5 02:06:33 2020\"\n+ Date: \"Mon Dec 3 17:46:56 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n-## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n-## MPIR_Init_thread(565)..............: \n-## MPID_Init(224).....................: channel initialization failed\n-## MPIDI_CH3_Init(105)................: \n-## MPID_nem_init(324).................: \n-## MPID_nem_tcp_init(178).............: \n-## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 853 - ./errors/coll/ibcastlength 2\n ---\n Directory: ./errors/coll\n File: ibcastlength\n Num-procs: 2\n- Date: \"Sun Jan 5 02:06:34 2020\"\n+ Date: \"Mon Dec 3 17:46:56 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n+## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n+## MPIR_Init_thread(565)..............: \n+## MPID_Init(224).....................: channel initialization failed\n+## MPIDI_CH3_Init(105)................: \n+## MPID_nem_init(324).................: \n+## MPID_nem_tcp_init(178).............: \n+## MPID_nem_tcp_get_business_card(425): \n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 854 - ./errors/comm/cfree 2\n ---\n Directory: ./errors/comm\n File: cfree\n Num-procs: 2\n- Date: \"Sun Jan 5 02:06:34 2020\"\n+ Date: \"Mon Dec 3 17:46:57 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 855 - ./errors/comm/ccreate1 2\n ---\n Directory: ./errors/comm\n File: ccreate1\n Num-procs: 2\n- Date: \"Sun Jan 5 02:06:35 2020\"\n+ Date: \"Mon Dec 3 17:46:57 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 856 - ./errors/comm/userdup 2\n ---\n Directory: ./errors/comm\n File: userdup\n Num-procs: 2\n- Date: \"Sun Jan 5 02:06:36 2020\"\n+ Date: \"Mon Dec 3 17:46:57 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 857 - ./errors/comm/manysplit 2\n ---\n Directory: ./errors/comm\n File: manysplit\n Num-procs: 2\n- Date: \"Sun Jan 5 02:06:37 2020\"\n+ Date: \"Mon Dec 3 17:46:58 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n+## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n+## MPIR_Init_thread(565)..............: \n+## MPID_Init(224).....................: channel initialization failed\n+## MPIDI_CH3_Init(105)................: \n+## MPID_nem_init(324).................: \n+## MPID_nem_tcp_init(178).............: \n+## MPID_nem_tcp_get_business_card(425): \n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n ok 878 - ./errors/comm/too_many_comms 2 # SKIP non-strict test, strict MPI mode requested\n ok 879 - ./errors/comm/too_many_icomms 2 # SKIP non-strict test, strict MPI mode requested\n ok 880 - ./errors/comm/too_many_comms2 2 # SKIP non-strict test, strict MPI mode requested\n ok 881 - ./errors/comm/too_many_comms3 2 # SKIP non-strict test, strict MPI mode requested\n ok 882 - ./errors/comm/too_many_icomms2 2 # SKIP non-strict test, strict MPI mode requested\n not ok 858 - ./errors/group/gerr 1\n ---\n Directory: ./errors/group\n File: gerr\n Num-procs: 1\n- Date: \"Sun Jan 5 02:06:37 2020\"\n+ Date: \"Mon Dec 3 17:46:58 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 859 - ./errors/pt2pt/proberank 1\n ---\n Directory: ./errors/pt2pt\n File: proberank\n Num-procs: 1\n- Date: \"Sun Jan 5 02:06:38 2020\"\n+ Date: \"Mon Dec 3 17:46:58 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 860 - ./errors/pt2pt/truncmsg1 2\n ---\n Directory: ./errors/pt2pt\n File: truncmsg1\n Num-procs: 2\n- Date: \"Sun Jan 5 02:06:39 2020\"\n+ Date: \"Mon Dec 3 17:46:59 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 861 - ./errors/pt2pt/truncmsg2 2\n ---\n Directory: ./errors/pt2pt\n File: truncmsg2\n Num-procs: 2\n- Date: \"Sun Jan 5 02:06:39 2020\"\n+ Date: \"Mon Dec 3 17:46:59 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 862 - ./errors/pt2pt/errinstatts 2\n ---\n Directory: ./errors/pt2pt\n File: errinstatts\n Num-procs: 2\n- Date: \"Sun Jan 5 02:06:40 2020\"\n+ Date: \"Mon Dec 3 17:47:00 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 863 - ./errors/pt2pt/errinstatta 2\n ---\n Directory: ./errors/pt2pt\n File: errinstatta\n Num-procs: 2\n- Date: \"Sun Jan 5 02:06:41 2020\"\n+ Date: \"Mon Dec 3 17:47:00 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n-## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n-## MPIR_Init_thread(565)..............: \n-## MPID_Init(224).....................: channel initialization failed\n-## MPIDI_CH3_Init(105)................: \n-## MPID_nem_init(324).................: \n-## MPID_nem_tcp_init(178).............: \n-## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 864 - ./errors/pt2pt/errinstatws 2\n ---\n Directory: ./errors/pt2pt\n File: errinstatws\n Num-procs: 2\n- Date: \"Sun Jan 5 02:06:42 2020\"\n+ Date: \"Mon Dec 3 17:47:00 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 865 - ./errors/pt2pt/errinstatwa 2\n ---\n Directory: ./errors/pt2pt\n File: errinstatwa\n Num-procs: 2\n- Date: \"Sun Jan 5 02:06:43 2020\"\n+ Date: \"Mon Dec 3 17:47:01 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n-## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n-## MPIR_Init_thread(565)..............: \n-## MPID_Init(224).....................: channel initialization failed\n-## MPIDI_CH3_Init(105)................: \n-## MPID_nem_init(324).................: \n-## MPID_nem_tcp_init(178).............: \n-## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 866 - ./errors/topo/cartsmall 2\n ---\n Directory: ./errors/topo\n File: cartsmall\n Num-procs: 2\n- Date: \"Sun Jan 5 02:06:43 2020\"\n+ Date: \"Mon Dec 3 17:47:01 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 867 - ./errors/rma/winerr 2\n ---\n Directory: ./errors/rma\n File: winerr\n Num-procs: 2\n- Date: \"Sun Jan 5 02:06:44 2020\"\n+ Date: \"Mon Dec 3 17:47:02 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n+## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n+## MPIR_Init_thread(565)..............: \n+## MPID_Init(224).....................: channel initialization failed\n+## MPIDI_CH3_Init(105)................: \n+## MPID_nem_init(324).................: \n+## MPID_nem_tcp_init(178).............: \n+## MPID_nem_tcp_get_business_card(425): \n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 868 - ./errors/rma/winerr2 2\n ---\n Directory: ./errors/rma\n File: winerr2\n Num-procs: 2\n- Date: \"Sun Jan 5 02:06:45 2020\"\n+ Date: \"Mon Dec 3 17:47:02 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 869 - ./errors/rma/cas_type_check 2\n ---\n Directory: ./errors/rma\n File: cas_type_check\n Num-procs: 2\n- Date: \"Sun Jan 5 02:06:45 2020\"\n+ Date: \"Mon Dec 3 17:47:02 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n-## Fatal error in PMPI_Init: Other MPI error, error stack:\n-## MPIR_Init_thread(565)..............: \n-## MPID_Init(224).....................: channel initialization failed\n-## MPIDI_CH3_Init(105)................: \n-## MPID_nem_init(324).................: \n-## MPID_nem_tcp_init(178).............: \n-## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 870 - ./errors/rma/win_sync_unlock 2\n ---\n Directory: ./errors/rma\n File: win_sync_unlock\n Num-procs: 2\n- Date: \"Sun Jan 5 02:06:46 2020\"\n+ Date: \"Mon Dec 3 17:47:03 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n-## Fatal error in PMPI_Init: Other MPI error, error stack:\n-## MPIR_Init_thread(565)..............: \n-## MPID_Init(224).....................: channel initialization failed\n-## MPIDI_CH3_Init(105)................: \n-## MPID_nem_init(324).................: \n-## MPID_nem_tcp_init(178).............: \n-## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 871 - ./errors/rma/win_sync_free_pt 2\n ---\n Directory: ./errors/rma\n File: win_sync_free_pt\n Num-procs: 2\n- Date: \"Sun Jan 5 02:06:47 2020\"\n+ Date: \"Mon Dec 3 17:47:03 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 872 - ./errors/rma/win_sync_free_at 2\n ---\n Directory: ./errors/rma\n File: win_sync_free_at\n Num-procs: 2\n- Date: \"Sun Jan 5 02:06:48 2020\"\n+ Date: \"Mon Dec 3 17:47:03 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n ## Fatal error in PMPI_Init: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 873 - ./errors/rma/win_sync_complete 2\n ---\n Directory: ./errors/rma\n File: win_sync_complete\n Num-procs: 2\n- Date: \"Sun Jan 5 02:06:49 2020\"\n+ Date: \"Mon Dec 3 17:47:04 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 874 - ./errors/rma/win_sync_lock_at 2\n ---\n Directory: ./errors/rma\n File: win_sync_lock_at\n Num-procs: 2\n- Date: \"Sun Jan 5 02:06:49 2020\"\n+ Date: \"Mon Dec 3 17:47:04 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 875 - ./errors/rma/win_sync_lock_pt 2\n ---\n Directory: ./errors/rma\n File: win_sync_lock_pt\n Num-procs: 2\n- Date: \"Sun Jan 5 02:06:50 2020\"\n+ Date: \"Mon Dec 3 17:47:04 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n-## Fatal error in PMPI_Init: Other MPI error, error stack:\n-## MPIR_Init_thread(565)..............: \n-## MPID_Init(224).....................: channel initialization failed\n-## MPIDI_CH3_Init(105)................: \n-## MPID_nem_init(324).................: \n-## MPID_nem_tcp_init(178).............: \n-## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 876 - ./errors/rma/win_sync_lock_fence 2\n ---\n Directory: ./errors/rma\n File: win_sync_lock_fence\n Num-procs: 2\n- Date: \"Sun Jan 5 02:06:51 2020\"\n+ Date: \"Mon Dec 3 17:47:05 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 877 - ./errors/rma/win_sync_nested 2\n ---\n Directory: ./errors/rma\n File: win_sync_nested\n Num-procs: 2\n- Date: \"Sun Jan 5 02:06:52 2020\"\n+ Date: \"Mon Dec 3 17:47:05 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n ## Fatal error in PMPI_Init: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 878 - ./errors/rma/win_sync_op 2\n ---\n Directory: ./errors/rma\n File: win_sync_op\n Num-procs: 2\n- Date: \"Sun Jan 5 02:06:53 2020\"\n+ Date: \"Mon Dec 3 17:47:06 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 879 - ./errors/spawn/badport 2\n ---\n Directory: ./errors/spawn\n File: badport\n Num-procs: 2\n- Date: \"Sun Jan 5 02:06:53 2020\"\n+ Date: \"Mon Dec 3 17:47:06 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n-## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n-## MPIR_Init_thread(565)..............: \n-## MPID_Init(224).....................: channel initialization failed\n-## MPIDI_CH3_Init(105)................: \n-## MPID_nem_init(324).................: \n-## MPID_nem_tcp_init(178).............: \n-## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 880 - ./errors/spawn/unpub 1\n ---\n Directory: ./errors/spawn\n File: unpub\n Num-procs: 1\n- Date: \"Sun Jan 5 02:06:54 2020\"\n+ Date: \"Mon Dec 3 17:47:06 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 881 - ./errors/spawn/lookup_name 1\n ---\n Directory: ./errors/spawn\n File: lookup_name\n Num-procs: 1\n- Date: \"Sun Jan 5 02:06:54 2020\"\n+ Date: \"Mon Dec 3 17:47:07 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 882 - ./errors/io/fileerrret 1\n ---\n Directory: ./errors/io\n File: fileerrret\n Num-procs: 1\n- Date: \"Sun Jan 5 02:06:55 2020\"\n+ Date: \"Mon Dec 3 17:47:07 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 883 - ./errors/io/openerr 1\n ---\n Directory: ./errors/io\n File: openerr\n Num-procs: 1\n- Date: \"Sun Jan 5 02:06:56 2020\"\n+ Date: \"Mon Dec 3 17:47:07 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 884 - ./errors/io/file_errhdl 1\n ---\n Directory: ./errors/io\n File: file_errhdl\n Num-procs: 1\n- Date: \"Sun Jan 5 02:06:56 2020\"\n+ Date: \"Mon Dec 3 17:47:08 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 885 - ./errors/f77/io/uerrhandf 1\n ---\n Directory: ./errors/f77/io\n File: uerrhandf\n Num-procs: 1\n- Date: \"Sun Jan 5 02:06:57 2020\"\n+ Date: \"Mon Dec 3 17:47:08 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 886 - ./errors/cxx/errhan/errgetx 1\n ---\n Directory: ./errors/cxx/errhan\n File: errgetx\n Num-procs: 1\n- Date: \"Sun Jan 5 02:06:58 2020\"\n+ Date: \"Mon Dec 3 17:47:09 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 887 - ./errors/cxx/errhan/errsetx 1\n ---\n Directory: ./errors/cxx/errhan\n File: errsetx\n Num-procs: 1\n- Date: \"Sun Jan 5 02:06:59 2020\"\n+ Date: \"Mon Dec 3 17:47:09 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 888 - ./errors/cxx/errhan/throwtest 1\n ---\n Directory: ./errors/cxx/errhan\n File: throwtest\n Num-procs: 1\n- Date: \"Sun Jan 5 02:06:59 2020\"\n+ Date: \"Mon Dec 3 17:47:09 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 889 - ./errors/cxx/errhan/commerrx 2\n ---\n Directory: ./errors/cxx/errhan\n File: commerrx\n Num-procs: 2\n- Date: \"Sun Jan 5 02:07:00 2020\"\n+ Date: \"Mon Dec 3 17:47:10 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 890 - ./errors/cxx/io/fileerrretx 1\n ---\n Directory: ./errors/cxx/io\n File: fileerrretx\n Num-procs: 1\n- Date: \"Sun Jan 5 02:07:00 2020\"\n+ Date: \"Mon Dec 3 17:47:10 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 891 - ./errors/cxx/io/errgetfilex 1\n ---\n Directory: ./errors/cxx/io\n File: errgetfilex\n Num-procs: 1\n- Date: \"Sun Jan 5 02:07:01 2020\"\n+ Date: \"Mon Dec 3 17:47:11 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 892 - ./errors/cxx/io/throwtestfilex 1\n ---\n Directory: ./errors/cxx/io\n File: throwtestfilex\n Num-procs: 1\n- Date: \"Sun Jan 5 02:07:01 2020\"\n+ Date: \"Mon Dec 3 17:47:11 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 893 - ./errors/f90/io/uerrhandf90 1\n ---\n Directory: ./errors/f90/io\n File: uerrhandf90\n Num-procs: 1\n- Date: \"Sun Jan 5 02:07:01 2020\"\n+ Date: \"Mon Dec 3 17:47:12 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 894 - ./threads/pt2pt/threads 2\n ---\n Directory: ./threads/pt2pt\n File: threads\n Num-procs: 2\n- Date: \"Sun Jan 5 02:07:02 2020\"\n+ Date: \"Mon Dec 3 17:47:12 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n+## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n+## MPIR_Init_thread(565)..............: \n+## MPID_Init(224).....................: channel initialization failed\n+## MPIDI_CH3_Init(105)................: \n+## MPID_nem_init(324).................: \n+## MPID_nem_tcp_init(178).............: \n+## MPID_nem_tcp_get_business_card(425): \n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 895 - ./threads/pt2pt/threaded_sr 2\n ---\n Directory: ./threads/pt2pt\n File: threaded_sr\n Num-procs: 2\n- Date: \"Sun Jan 5 02:07:03 2020\"\n+ Date: \"Mon Dec 3 17:47:13 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 896 - ./threads/pt2pt/alltoall 2\n ---\n Directory: ./threads/pt2pt\n File: alltoall\n Num-procs: 2\n- Date: \"Sun Jan 5 02:07:03 2020\"\n+ Date: \"Mon Dec 3 17:47:13 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n-## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n-## MPIR_Init_thread(565)..............: \n-## MPID_Init(224).....................: channel initialization failed\n-## MPIDI_CH3_Init(105)................: \n-## MPID_nem_init(324).................: \n-## MPID_nem_tcp_init(178).............: \n-## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 897 - ./threads/pt2pt/sendselfth 1\n ---\n Directory: ./threads/pt2pt\n File: sendselfth\n Num-procs: 1\n- Date: \"Sun Jan 5 02:07:04 2020\"\n+ Date: \"Mon Dec 3 17:47:13 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 898 - ./threads/pt2pt/multisend 2\n ---\n Directory: ./threads/pt2pt\n File: multisend\n Num-procs: 2\n- Date: \"Sun Jan 5 02:07:04 2020\"\n+ Date: \"Mon Dec 3 17:47:14 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 899 - ./threads/pt2pt/multisend2 2\n ---\n Directory: ./threads/pt2pt\n File: multisend2\n Num-procs: 2\n- Date: \"Sun Jan 5 02:07:04 2020\"\n+ Date: \"Mon Dec 3 17:47:14 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n+## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n+## MPIR_Init_thread(565)..............: \n+## MPID_Init(224).....................: channel initialization failed\n+## MPIDI_CH3_Init(105)................: \n+## MPID_nem_init(324).................: \n+## MPID_nem_tcp_init(178).............: \n+## MPID_nem_tcp_get_business_card(425): \n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 900 - ./threads/pt2pt/multisend3 2\n ---\n Directory: ./threads/pt2pt\n File: multisend3\n Num-procs: 2\n- Date: \"Sun Jan 5 02:07:05 2020\"\n+ Date: \"Mon Dec 3 17:47:15 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 901 - ./threads/pt2pt/multisend4 2\n ---\n Directory: ./threads/pt2pt\n File: multisend4\n Num-procs: 2\n- Date: \"Sun Jan 5 02:07:05 2020\"\n+ Date: \"Mon Dec 3 17:47:15 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n-## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n-## MPIR_Init_thread(565)..............: \n-## MPID_Init(224).....................: channel initialization failed\n-## MPIDI_CH3_Init(105)................: \n-## MPID_nem_init(324).................: \n-## MPID_nem_tcp_init(178).............: \n-## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 902 - ./threads/pt2pt/greq_wait 1\n ---\n Directory: ./threads/pt2pt\n File: greq_wait\n Num-procs: 1\n- Date: \"Sun Jan 5 02:07:06 2020\"\n+ Date: \"Mon Dec 3 17:47:15 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 903 - ./threads/pt2pt/greq_test 1\n ---\n Directory: ./threads/pt2pt\n File: greq_test\n Num-procs: 1\n- Date: \"Sun Jan 5 02:07:06 2020\"\n+ Date: \"Mon Dec 3 17:47:15 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 904 - ./threads/pt2pt/ibsend 2\n ---\n Directory: ./threads/pt2pt\n File: ibsend\n Num-procs: 2\n- Date: \"Sun Jan 5 02:07:07 2020\"\n+ Date: \"Mon Dec 3 17:47:16 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n+## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n+## MPIR_Init_thread(565)..............: \n+## MPID_Init(224).....................: channel initialization failed\n+## MPIDI_CH3_Init(105)................: \n+## MPID_nem_init(324).................: \n+## MPID_nem_tcp_init(178).............: \n+## MPID_nem_tcp_get_business_card(425): \n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 905 - ./threads/comm/ctxdup 2\n ---\n Directory: ./threads/comm\n File: ctxdup\n Num-procs: 2\n- Date: \"Sun Jan 5 02:07:08 2020\"\n+ Date: \"Mon Dec 3 17:47:16 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 906 - ./threads/comm/dup_leak_test 2\n ---\n Directory: ./threads/comm\n File: dup_leak_test\n Num-procs: 2\n- Date: \"Sun Jan 5 02:07:08 2020\"\n+ Date: \"Mon Dec 3 17:47:17 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 907 - ./threads/comm/comm_dup_deadlock 2\n ---\n Directory: ./threads/comm\n File: comm_dup_deadlock\n Num-procs: 2\n- Date: \"Sun Jan 5 02:07:09 2020\"\n+ Date: \"Mon Dec 3 17:47:17 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n-## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n-## MPIR_Init_thread(565)..............: \n-## MPID_Init(224).....................: channel initialization failed\n-## MPIDI_CH3_Init(105)................: \n-## MPID_nem_init(324).................: \n-## MPID_nem_tcp_init(178).............: \n-## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 908 - ./threads/comm/comm_create_threads 2\n ---\n Directory: ./threads/comm\n File: comm_create_threads\n Num-procs: 2\n- Date: \"Sun Jan 5 02:07:09 2020\"\n+ Date: \"Mon Dec 3 17:47:17 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 909 - ./threads/comm/comm_create_group_threads 2\n ---\n Directory: ./threads/comm\n File: comm_create_group_threads\n Num-procs: 2\n- Date: \"Sun Jan 5 02:07:10 2020\"\n+ Date: \"Mon Dec 3 17:47:18 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 910 - ./threads/comm/comm_create_group_threads2 2\n ---\n Directory: ./threads/comm\n File: comm_create_group_threads2\n Num-procs: 2\n- Date: \"Sun Jan 5 02:07:11 2020\"\n+ Date: \"Mon Dec 3 17:47:18 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 911 - ./threads/comm/idup_deadlock 2\n ---\n Directory: ./threads/comm\n File: idup_deadlock\n Num-procs: 2\n- Date: \"Sun Jan 5 02:07:11 2020\"\n+ Date: \"Mon Dec 3 17:47:18 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n+## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n+## MPIR_Init_thread(565)..............: \n+## MPID_Init(224).....................: channel initialization failed\n+## MPIDI_CH3_Init(105)................: \n+## MPID_nem_init(324).................: \n+## MPID_nem_tcp_init(178).............: \n+## MPID_nem_tcp_get_business_card(425): \n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 912 - ./threads/comm/comm_idup 2\n ---\n Directory: ./threads/comm\n File: comm_idup\n Num-procs: 2\n- Date: \"Sun Jan 5 02:07:11 2020\"\n+ Date: \"Mon Dec 3 17:47:19 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 913 - ./threads/comm/ctxidup 2\n ---\n Directory: ./threads/comm\n File: ctxidup\n Num-procs: 2\n- Date: \"Sun Jan 5 02:07:12 2020\"\n+ Date: \"Mon Dec 3 17:47:19 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n-## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n-## MPIR_Init_thread(565)..............: \n-## MPID_Init(224).....................: channel initialization failed\n-## MPIDI_CH3_Init(105)................: \n-## MPID_nem_init(324).................: \n-## MPID_nem_tcp_init(178).............: \n-## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 914 - ./threads/comm/idup_nb 2\n ---\n Directory: ./threads/comm\n File: idup_nb\n Num-procs: 2\n- Date: \"Sun Jan 5 02:07:13 2020\"\n+ Date: \"Mon Dec 3 17:47:20 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 915 - ./threads/comm/idup_comm_gen 2\n ---\n Directory: ./threads/comm\n File: idup_comm_gen\n Num-procs: 2\n- Date: \"Sun Jan 5 02:07:13 2020\"\n+ Date: \"Mon Dec 3 17:47:20 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n+## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n+## MPIR_Init_thread(565)..............: \n+## MPID_Init(224).....................: channel initialization failed\n+## MPIDI_CH3_Init(105)................: \n+## MPID_nem_init(324).................: \n+## MPID_nem_tcp_init(178).............: \n+## MPID_nem_tcp_get_business_card(425): \n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 916 - ./threads/init/initth 1\n ---\n Directory: ./threads/init\n File: initth\n Num-procs: 1\n- Date: \"Sun Jan 5 02:07:14 2020\"\n+ Date: \"Mon Dec 3 17:47:20 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 917 - ./threads/init/initth 2\n ---\n Directory: ./threads/init\n File: initth\n Num-procs: 2\n- Date: \"Sun Jan 5 02:07:14 2020\"\n+ Date: \"Mon Dec 3 17:47:21 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 918 - ./threads/mpi_t/mpit_threading 1\n ---\n Directory: ./threads/mpi_t\n File: mpit_threading\n Num-procs: 1\n- Date: \"Sun Jan 5 02:07:15 2020\"\n+ Date: \"Mon Dec 3 17:47:21 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 919 - ./threads/spawn/multispawn 1\n ---\n Directory: ./threads/spawn\n File: multispawn\n Num-procs: 1\n- Date: \"Sun Jan 5 02:07:15 2020\"\n+ Date: \"Mon Dec 3 17:47:21 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 920 - ./threads/spawn/th_taskmaster 1\n ---\n Directory: ./threads/spawn\n File: th_taskmaster\n Num-procs: 1\n- Date: \"Sun Jan 5 02:07:16 2020\"\n+ Date: \"Mon Dec 3 17:47:22 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 921 - ./threads/spawn/th_taskmaster 2\n ---\n Directory: ./threads/spawn\n File: th_taskmaster\n Num-procs: 2\n- Date: \"Sun Jan 5 02:07:16 2020\"\n+ Date: \"Mon Dec 3 17:47:22 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n+## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n+## MPIR_Init_thread(565)..............: \n+## MPID_Init(224).....................: channel initialization failed\n+## MPIDI_CH3_Init(105)................: \n+## MPID_nem_init(324).................: \n+## MPID_nem_tcp_init(178).............: \n+## MPID_nem_tcp_get_business_card(425): \n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 922 - ./threads/rma/multirma 2\n ---\n Directory: ./threads/rma\n File: multirma\n Num-procs: 2\n- Date: \"Sun Jan 5 02:07:17 2020\"\n+ Date: \"Mon Dec 3 17:47:22 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 923 - ./threads/rma/multiget 2\n ---\n Directory: ./threads/rma\n File: multiget\n Num-procs: 2\n- Date: \"Sun Jan 5 02:07:17 2020\"\n+ Date: \"Mon Dec 3 17:47:23 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 924 - ./threads/coll/allred 2\n ---\n Directory: ./threads/coll\n File: allred\n Num-procs: 2\n- Date: \"Sun Jan 5 02:07:18 2020\"\n+ Date: \"Mon Dec 3 17:47:23 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n not ok 925 - ./threads/coll/iallred 2\n ---\n Directory: ./threads/coll\n File: iallred\n Num-procs: 2\n- Date: \"Sun Jan 5 02:07:18 2020\"\n+ Date: \"Mon Dec 3 17:47:24 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n ## MPIR_Init_thread(565)..............: \n ## MPID_Init(224).....................: channel initialization failed\n ## MPIDI_CH3_Init(105)................: \n ## MPID_nem_init(324).................: \n ## MPID_nem_tcp_init(178).............: \n ## MPID_nem_tcp_get_business_card(425): \n-## MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+## MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n 1..950\n" }, { "source1": "./usr/share/mpi-testsuite/results/mpich/summary.xml", "source2": "./usr/share/mpi-testsuite/results/mpich/summary.xml", "unified_diff": null, "details": [ { "source1": "./usr/share/mpi-testsuite/results/mpich/summary.xml", "source2": "./usr/share/mpi-testsuite/results/mpich/summary.xml", "unified_diff": "@@ -1,4611 +1,4310 @@\n \n \n \n- 2020-01-05-01-58\n+ 2018-12-03-17-38\n \n \n attrt\n 2\n ./attr\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n+Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n+MPIR_Init_thread(565)..............: \n+MPID_Init(224).....................: channel initialization failed\n+MPIDI_CH3_Init(105)................: \n+MPID_nem_init(324).................: \n+MPID_nem_tcp_init(178).............: \n+MPID_nem_tcp_get_business_card(425): \n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n attric\n 2\n ./attr\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n-Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n-MPIR_Init_thread(565)..............: \n-MPID_Init(224).....................: channel initialization failed\n-MPIDI_CH3_Init(105)................: \n-MPID_nem_init(324).................: \n-MPID_nem_tcp_init(178).............: \n-MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n attrerr\n 1\n ./attr\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n attrend\n 1\n ./attr\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n attrend\n 2\n ./attr\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n attrend2\n 1\n ./attr\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n attrend2\n 2\n ./attr\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n attrerrcomm\n 1\n ./attr\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n attrerrtype\n 1\n ./attr\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n attrdeleteget\n 1\n ./attr\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n attr2type\n 1\n ./attr\n fail\n Fatal error in PMPI_Init: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n attrorder\n 1\n ./attr\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n attrordercomm\n 1\n ./attr\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n attrordertype\n 1\n ./attr\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n baseattr2\n 1\n ./attr\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n baseattrcomm\n 1\n ./attr\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n fkeyval\n 1\n ./attr\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n fkeyvalcomm\n 1\n ./attr\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n fkeyvaltype\n 1\n ./attr\n fail\n- \n+ Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n+MPIR_Init_thread(565)..............: \n+MPID_Init(224).....................: channel initialization failed\n+MPIDI_CH3_Init(105)................: \n+MPID_nem_init(324).................: \n+MPID_nem_tcp_init(178).............: \n+MPID_nem_tcp_get_business_card(425): \n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n keyval_double_free\n 1\n ./attr\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n keyval_double_free_comm\n 1\n ./attr\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n keyval_double_free_type\n 1\n ./attr\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n keyval_double_free_win\n 1\n ./attr\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n allred\n 2\n ./coll\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n-Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n-MPIR_Init_thread(565)..............: \n-MPID_Init(224).....................: channel initialization failed\n-MPIDI_CH3_Init(105)................: \n-MPID_nem_init(324).................: \n-MPID_nem_tcp_init(178).............: \n-MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n allred\n 2\n ./coll\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n-Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n-MPIR_Init_thread(565)..............: \n-MPID_Init(224).....................: channel initialization failed\n-MPIDI_CH3_Init(105)................: \n-MPID_nem_init(324).................: \n-MPID_nem_tcp_init(178).............: \n-MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n allred\n 2\n ./coll\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n allredmany\n 2\n ./coll\n fail\n Fatal error in PMPI_Init: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n-Fatal error in PMPI_Init: Other MPI error, error stack:\n-MPIR_Init_thread(565)..............: \n-MPID_Init(224).....................: channel initialization failed\n-MPIDI_CH3_Init(105)................: \n-MPID_nem_init(324).................: \n-MPID_nem_tcp_init(178).............: \n-MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n allred2\n 2\n ./coll\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n allred3\n 2\n ./coll\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n allred4\n 2\n ./coll\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n allred5\n 2\n ./coll\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n allred5\n 2\n ./coll\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n allred6\n 2\n ./coll\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n-Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n-MPIR_Init_thread(565)..............: \n-MPID_Init(224).....................: channel initialization failed\n-MPIDI_CH3_Init(105)................: \n-MPID_nem_init(324).................: \n-MPID_nem_tcp_init(178).............: \n-MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n allred6\n 2\n ./coll\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n- \n- \n- reduce\n- 2\n- ./coll\n- fail\n- Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n+Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n reduce\n 2\n ./coll\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n- reduce_local\n+ reduce\n 2\n ./coll\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n- op_commutative\n+ reduce_local\n 2\n ./coll\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n-Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n+ \n+ \n+ op_commutative\n+ 2\n+ ./coll\n+ fail\n+ Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n red3\n 2\n ./coll\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n-Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n-MPIR_Init_thread(565)..............: \n-MPID_Init(224).....................: channel initialization failed\n-MPIDI_CH3_Init(105)................: \n-MPID_nem_init(324).................: \n-MPID_nem_tcp_init(178).............: \n-MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n red4\n 2\n ./coll\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n alltoall1\n 2\n ./coll\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n-Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n-MPIR_Init_thread(565)..............: \n-MPID_Init(224).....................: channel initialization failed\n-MPIDI_CH3_Init(105)................: \n-MPID_nem_init(324).................: \n-MPID_nem_tcp_init(178).............: \n-MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n alltoallv\n 2\n ./coll\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n alltoallv0\n 2\n ./coll\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n-Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n-MPIR_Init_thread(565)..............: \n-MPID_Init(224).....................: channel initialization failed\n-MPIDI_CH3_Init(105)................: \n-MPID_nem_init(324).................: \n-MPID_nem_tcp_init(178).............: \n-MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n alltoallw1\n 2\n ./coll\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n alltoallw2\n 2\n ./coll\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n+Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n+MPIR_Init_thread(565)..............: \n+MPID_Init(224).....................: channel initialization failed\n+MPIDI_CH3_Init(105)................: \n+MPID_nem_init(324).................: \n+MPID_nem_tcp_init(178).............: \n+MPID_nem_tcp_get_business_card(425): \n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n alltoallw_zeros\n 1\n ./coll\n fail\n Fatal error in PMPI_Init: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n alltoallw_zeros\n 2\n ./coll\n fail\n Fatal error in PMPI_Init: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n alltoallw_zeros\n 2\n ./coll\n fail\n Fatal error in PMPI_Init: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n-Fatal error in PMPI_Init: Other MPI error, error stack:\n-MPIR_Init_thread(565)..............: \n-MPID_Init(224).....................: channel initialization failed\n-MPIDI_CH3_Init(105)................: \n-MPID_nem_init(324).................: \n-MPID_nem_tcp_init(178).............: \n-MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n alltoallw_zeros\n 2\n ./coll\n fail\n Fatal error in PMPI_Init: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n+Fatal error in PMPI_Init: Other MPI error, error stack:\n+MPIR_Init_thread(565)..............: \n+MPID_Init(224).....................: channel initialization failed\n+MPIDI_CH3_Init(105)................: \n+MPID_nem_init(324).................: \n+MPID_nem_tcp_init(178).............: \n+MPID_nem_tcp_get_business_card(425): \n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n allgather2\n 2\n ./coll\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n allgather3\n 2\n ./coll\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n-Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n-MPIR_Init_thread(565)..............: \n-MPID_Init(224).....................: channel initialization failed\n-MPIDI_CH3_Init(105)................: \n-MPID_nem_init(324).................: \n-MPID_nem_tcp_init(178).............: \n-MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n allgatherv2\n 2\n ./coll\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n-Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n+ \n+ \n+ allgatherv3\n+ 2\n+ ./coll\n+ fail\n+ Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n- allgatherv3\n+ allgatherv4\n 2\n ./coll\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n- allgatherv4\n+ allgather_struct\n 2\n ./coll\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n- allgather_struct\n+ bcasttest\n 2\n ./coll\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n- \n- \n- bcasttest\n- 2\n- ./coll\n- fail\n- Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n+Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n bcasttest\n 2\n ./coll\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n bcasttest\n 2\n ./coll\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n bcast_full\n 2\n ./coll\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n-Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n-MPIR_Init_thread(565)..............: \n-MPID_Init(224).....................: channel initialization failed\n-MPIDI_CH3_Init(105)................: \n-MPID_nem_init(324).................: \n-MPID_nem_tcp_init(178).............: \n-MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n bcast_min_datatypes\n 2\n ./coll\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n-Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n-MPIR_Init_thread(565)..............: \n-MPID_Init(224).....................: channel initialization failed\n-MPIDI_CH3_Init(105)................: \n-MPID_nem_init(324).................: \n-MPID_nem_tcp_init(178).............: \n-MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n bcast_comm_world\n 2\n ./coll\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n bcastzerotype\n 1\n ./coll\n fail\n Fatal error in PMPI_Init: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n bcastzerotype\n 2\n ./coll\n fail\n Fatal error in PMPI_Init: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n+Fatal error in PMPI_Init: Other MPI error, error stack:\n+MPIR_Init_thread(565)..............: \n+MPID_Init(224).....................: channel initialization failed\n+MPIDI_CH3_Init(105)................: \n+MPID_nem_init(324).................: \n+MPID_nem_tcp_init(178).............: \n+MPID_nem_tcp_get_business_card(425): \n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n bcastzerotype\n 2\n ./coll\n fail\n Fatal error in PMPI_Init: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n Fatal error in PMPI_Init: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n bcastzerotype\n 2\n ./coll\n fail\n Fatal error in PMPI_Init: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n+Fatal error in PMPI_Init: Other MPI error, error stack:\n+MPIR_Init_thread(565)..............: \n+MPID_Init(224).....................: channel initialization failed\n+MPIDI_CH3_Init(105)................: \n+MPID_nem_init(324).................: \n+MPID_nem_tcp_init(178).............: \n+MPID_nem_tcp_get_business_card(425): \n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n coll2\n 2\n ./coll\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n coll3\n 2\n ./coll\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n coll4\n 2\n ./coll\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n coll5\n 2\n ./coll\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n-Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n-MPIR_Init_thread(565)..............: \n-MPID_Init(224).....................: channel initialization failed\n-MPIDI_CH3_Init(105)................: \n-MPID_nem_init(324).................: \n-MPID_nem_tcp_init(178).............: \n-MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n coll6\n 2\n ./coll\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n coll7\n 1\n ./coll\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n coll7\n 2\n ./coll\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n- \n- \n- coll7\n- 2\n- ./coll\n- fail\n- Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n+Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n- coll8\n+ coll7\n 2\n ./coll\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n- coll9\n+ coll8\n 2\n ./coll\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n- coll10\n+ coll9\n 2\n ./coll\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n- coll11\n+ coll10\n 2\n ./coll\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n- \n- \n- coll12\n- 2\n- ./coll\n- fail\n- Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n+Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n- coll13\n+ coll11\n 2\n ./coll\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n- longuser\n+ coll12\n 2\n ./coll\n fail\n- Fatal error in PMPI_Init: Other MPI error, error stack:\n+ Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n- redscat\n+ coll13\n+ 2\n+ ./coll\n+ fail\n+ Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n+MPIR_Init_thread(565)..............: \n+MPID_Init(224).....................: channel initialization failed\n+MPIDI_CH3_Init(105)................: \n+MPID_nem_init(324).................: \n+MPID_nem_tcp_init(178).............: \n+MPID_nem_tcp_get_business_card(425): \n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n+ \n+ \n+ longuser\n 2\n ./coll\n fail\n Fatal error in PMPI_Init: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n Fatal error in PMPI_Init: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n redscat\n 2\n ./coll\n fail\n Fatal error in PMPI_Init: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n- redscat2\n+ redscat\n 2\n ./coll\n fail\n- Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n+ Fatal error in PMPI_Init: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n-Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n+ \n+ \n+ redscat2\n+ 2\n+ ./coll\n+ fail\n+ Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n redscat2\n 2\n ./coll\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n redscat2\n 2\n ./coll\n fail\n- Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n-MPIR_Init_thread(565)..............: \n-MPID_Init(224).....................: channel initialization failed\n-MPIDI_CH3_Init(105)................: \n-MPID_nem_init(324).................: \n-MPID_nem_tcp_init(178).............: \n-MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+ \n \n \n redscat3\n 2\n ./coll\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n redscatinter\n 2\n ./coll\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n-Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n-MPIR_Init_thread(565)..............: \n-MPID_Init(224).....................: channel initialization failed\n-MPIDI_CH3_Init(105)................: \n-MPID_nem_init(324).................: \n-MPID_nem_tcp_init(178).............: \n-MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n red_scat_block\n 2\n ./coll\n fail\n Fatal error in PMPI_Init: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n red_scat_block\n 2\n ./coll\n fail\n Fatal error in PMPI_Init: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n+Fatal error in PMPI_Init: Other MPI error, error stack:\n+MPIR_Init_thread(565)..............: \n+MPID_Init(224).....................: channel initialization failed\n+MPIDI_CH3_Init(105)................: \n+MPID_nem_init(324).................: \n+MPID_nem_tcp_init(178).............: \n+MPID_nem_tcp_get_business_card(425): \n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n red_scat_block\n 2\n ./coll\n fail\n Fatal error in PMPI_Init: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n red_scat_block2\n 2\n ./coll\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n-Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n-MPIR_Init_thread(565)..............: \n-MPID_Init(224).....................: channel initialization failed\n-MPIDI_CH3_Init(105)................: \n-MPID_nem_init(324).................: \n-MPID_nem_tcp_init(178).............: \n-MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n red_scat_block2\n 2\n ./coll\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n-Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n-MPIR_Init_thread(565)..............: \n-MPID_Init(224).....................: channel initialization failed\n-MPIDI_CH3_Init(105)................: \n-MPID_nem_init(324).................: \n-MPID_nem_tcp_init(178).............: \n-MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n red_scat_block2\n 2\n ./coll\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n redscatblk3\n 2\n ./coll\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n redscatblk3\n 2\n ./coll\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n redscatbkinter\n 2\n ./coll\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n-Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n-MPIR_Init_thread(565)..............: \n-MPID_Init(224).....................: channel initialization failed\n-MPIDI_CH3_Init(105)................: \n-MPID_nem_init(324).................: \n-MPID_nem_tcp_init(178).............: \n-MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n redscatbkinter\n 2\n ./coll\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n scantst\n 2\n ./coll\n fail\n Fatal error in PMPI_Init: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n exscan\n 2\n ./coll\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n exscan2\n 2\n ./coll\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n gather\n 2\n ./coll\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n gather2\n 2\n ./coll\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n scattern\n 2\n ./coll\n fail\n Fatal error in PMPI_Init: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n scatter2\n 2\n ./coll\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n scatter3\n 2\n ./coll\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n-Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n-MPIR_Init_thread(565)..............: \n-MPID_Init(224).....................: channel initialization failed\n-MPIDI_CH3_Init(105)................: \n-MPID_nem_init(324).................: \n-MPID_nem_tcp_init(178).............: \n-MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n scatterv\n 2\n ./coll\n fail\n- Fatal error in PMPI_Init: Other MPI error, error stack:\n-MPIR_Init_thread(565)..............: \n-MPID_Init(224).....................: channel initialization failed\n-MPIDI_CH3_Init(105)................: \n-MPID_nem_init(324).................: \n-MPID_nem_tcp_init(178).............: \n-MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+ \n \n \n icbcast\n 2\n ./coll\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n-Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n-MPIR_Init_thread(565)..............: \n-MPID_Init(224).....................: channel initialization failed\n-MPIDI_CH3_Init(105)................: \n-MPID_nem_init(324).................: \n-MPID_nem_tcp_init(178).............: \n-MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n icbcast\n 2\n ./coll\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n icallreduce\n 2\n ./coll\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n-Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n-MPIR_Init_thread(565)..............: \n-MPID_Init(224).....................: channel initialization failed\n-MPIDI_CH3_Init(105)................: \n-MPID_nem_init(324).................: \n-MPID_nem_tcp_init(178).............: \n-MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n icallreduce\n 2\n ./coll\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n-Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n-MPIR_Init_thread(565)..............: \n-MPID_Init(224).....................: channel initialization failed\n-MPIDI_CH3_Init(105)................: \n-MPID_nem_init(324).................: \n-MPID_nem_tcp_init(178).............: \n-MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n icreduce\n 2\n ./coll\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n icreduce\n 2\n ./coll\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n icscatter\n 2\n ./coll\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n icscatter\n 2\n ./coll\n fail\n- Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n-MPIR_Init_thread(565)..............: \n-MPID_Init(224).....................: channel initialization failed\n-MPIDI_CH3_Init(105)................: \n-MPID_nem_init(324).................: \n-MPID_nem_tcp_init(178).............: \n-MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+ \n \n \n icgather\n 2\n ./coll\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n icgather\n 2\n ./coll\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n- \n- \n- icallgather\n- 2\n- ./coll\n- fail\n- Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n+Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n icallgather\n 2\n ./coll\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n- icbarrier\n+ icallgather\n 2\n ./coll\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n icbarrier\n 2\n ./coll\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n- icallgatherv\n+ icbarrier\n 2\n ./coll\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n+Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n+MPIR_Init_thread(565)..............: \n+MPID_Init(224).....................: channel initialization failed\n+MPIDI_CH3_Init(105)................: \n+MPID_nem_init(324).................: \n+MPID_nem_tcp_init(178).............: \n+MPID_nem_tcp_get_business_card(425): \n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n icallgatherv\n 2\n ./coll\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n- icgatherv\n+ icallgatherv\n 2\n ./coll\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n icgatherv\n 2\n ./coll\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n- icscatterv\n+ icgatherv\n 2\n ./coll\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n-Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n-MPIR_Init_thread(565)..............: \n-MPID_Init(224).....................: channel initialization failed\n-MPIDI_CH3_Init(105)................: \n-MPID_nem_init(324).................: \n-MPID_nem_tcp_init(178).............: \n-MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n icscatterv\n 2\n ./coll\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n-Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n+ \n+ \n+ icscatterv\n+ 2\n+ ./coll\n+ fail\n+ Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n icalltoall\n 2\n ./coll\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n icalltoall\n 2\n ./coll\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n-Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n-MPIR_Init_thread(565)..............: \n-MPID_Init(224).....................: channel initialization failed\n-MPIDI_CH3_Init(105)................: \n-MPID_nem_init(324).................: \n-MPID_nem_tcp_init(178).............: \n-MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n icalltoallv\n 2\n ./coll\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n icalltoallv\n 2\n ./coll\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n+Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n+MPIR_Init_thread(565)..............: \n+MPID_Init(224).....................: channel initialization failed\n+MPIDI_CH3_Init(105)................: \n+MPID_nem_init(324).................: \n+MPID_nem_tcp_init(178).............: \n+MPID_nem_tcp_get_business_card(425): \n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n icalltoallw\n 2\n ./coll\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n icalltoallw\n 2\n ./coll\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n-Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n-MPIR_Init_thread(565)..............: \n-MPID_Init(224).....................: channel initialization failed\n-MPIDI_CH3_Init(105)................: \n-MPID_nem_init(324).................: \n-MPID_nem_tcp_init(178).............: \n-MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n opland\n 2\n ./coll\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n oplor\n 2\n ./coll\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n-Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n-MPIR_Init_thread(565)..............: \n-MPID_Init(224).....................: channel initialization failed\n-MPIDI_CH3_Init(105)................: \n-MPID_nem_init(324).................: \n-MPID_nem_tcp_init(178).............: \n-MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n oplxor\n 2\n ./coll\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n oplxor\n 2\n ./coll\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n+Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n+MPIR_Init_thread(565)..............: \n+MPID_Init(224).....................: channel initialization failed\n+MPIDI_CH3_Init(105)................: \n+MPID_nem_init(324).................: \n+MPID_nem_tcp_init(178).............: \n+MPID_nem_tcp_get_business_card(425): \n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n opband\n 2\n ./coll\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n-Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n-MPIR_Init_thread(565)..............: \n-MPID_Init(224).....................: channel initialization failed\n-MPIDI_CH3_Init(105)................: \n-MPID_nem_init(324).................: \n-MPID_nem_tcp_init(178).............: \n-MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n opbor\n 2\n ./coll\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n-Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n-MPIR_Init_thread(565)..............: \n-MPID_Init(224).....................: channel initialization failed\n-MPIDI_CH3_Init(105)................: \n-MPID_nem_init(324).................: \n-MPID_nem_tcp_init(178).............: \n-MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n opbxor\n 2\n ./coll\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n-Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n-MPIR_Init_thread(565)..............: \n-MPID_Init(224).....................: channel initialization failed\n-MPIDI_CH3_Init(105)................: \n-MPID_nem_init(324).................: \n-MPID_nem_tcp_init(178).............: \n-MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n opbxor\n 2\n ./coll\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n+ \n+ \n+ opprod\n+ 2\n+ ./coll\n+ fail\n+ \n \n \n opprod\n 2\n ./coll\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n- \n- \n- opprod\n- 2\n- ./coll\n- fail\n- Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n-MPIR_Init_thread(565)..............: \n-MPID_Init(224).....................: channel initialization failed\n-MPIDI_CH3_Init(105)................: \n-MPID_nem_init(324).................: \n-MPID_nem_tcp_init(178).............: \n-MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n opsum\n 2\n ./coll\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n opmin\n 2\n ./coll\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n opminloc\n 2\n ./coll\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n-Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n-MPIR_Init_thread(565)..............: \n-MPID_Init(224).....................: channel initialization failed\n-MPIDI_CH3_Init(105)................: \n-MPID_nem_init(324).................: \n-MPID_nem_tcp_init(178).............: \n-MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n opmax\n 2\n ./coll\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n-Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n-MPIR_Init_thread(565)..............: \n-MPID_Init(224).....................: channel initialization failed\n-MPIDI_CH3_Init(105)................: \n-MPID_nem_init(324).................: \n-MPID_nem_tcp_init(178).............: \n-MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n opmaxloc\n 2\n ./coll\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n uoplong\n 2\n ./coll\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n-Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n-MPIR_Init_thread(565)..............: \n-MPID_Init(224).....................: channel initialization failed\n-MPIDI_CH3_Init(105)................: \n-MPID_nem_init(324).................: \n-MPID_nem_tcp_init(178).............: \n-MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n uoplong\n 2\n ./coll\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n uoplong\n 2\n ./coll\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n nonblocking\n 2\n ./coll\n fail\n Fatal error in PMPI_Init: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n nonblocking\n 2\n ./coll\n fail\n Fatal error in PMPI_Init: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n-Fatal error in PMPI_Init: Other MPI error, error stack:\n-MPIR_Init_thread(565)..............: \n-MPID_Init(224).....................: channel initialization failed\n-MPIDI_CH3_Init(105)................: \n-MPID_nem_init(324).................: \n-MPID_nem_tcp_init(178).............: \n-MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n nonblocking\n 2\n ./coll\n fail\n Fatal error in PMPI_Init: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n nonblocking2\n 1\n ./coll\n fail\n Fatal error in PMPI_Init: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n nonblocking2\n 2\n ./coll\n fail\n Fatal error in PMPI_Init: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n nonblocking2\n 2\n ./coll\n fail\n Fatal error in PMPI_Init: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n+Fatal error in PMPI_Init: Other MPI error, error stack:\n+MPIR_Init_thread(565)..............: \n+MPID_Init(224).....................: channel initialization failed\n+MPIDI_CH3_Init(105)................: \n+MPID_nem_init(324).................: \n+MPID_nem_tcp_init(178).............: \n+MPID_nem_tcp_get_business_card(425): \n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n nonblocking2\n 2\n ./coll\n fail\n Fatal error in PMPI_Init: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n nonblocking3\n 1\n ./coll\n fail\n Fatal error in PMPI_Init: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n nonblocking3\n 2\n ./coll\n fail\n Fatal error in PMPI_Init: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n-Fatal error in PMPI_Init: Other MPI error, error stack:\n-MPIR_Init_thread(565)..............: \n-MPID_Init(224).....................: channel initialization failed\n-MPIDI_CH3_Init(105)................: \n-MPID_nem_init(324).................: \n-MPID_nem_tcp_init(178).............: \n-MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n nonblocking3\n 2\n ./coll\n fail\n Fatal error in PMPI_Init: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n+Fatal error in PMPI_Init: Other MPI error, error stack:\n+MPIR_Init_thread(565)..............: \n+MPID_Init(224).....................: channel initialization failed\n+MPIDI_CH3_Init(105)................: \n+MPID_nem_init(324).................: \n+MPID_nem_tcp_init(178).............: \n+MPID_nem_tcp_get_business_card(425): \n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n nonblocking3\n 2\n ./coll\n fail\n Fatal error in PMPI_Init: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n+Fatal error in PMPI_Init: Other MPI error, error stack:\n+MPIR_Init_thread(565)..............: \n+MPID_Init(224).....................: channel initialization failed\n+MPIDI_CH3_Init(105)................: \n+MPID_nem_init(324).................: \n+MPID_nem_tcp_init(178).............: \n+MPID_nem_tcp_get_business_card(425): \n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n iallred\n 2\n ./coll\n fail\n Fatal error in PMPI_Init: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n ibarrier\n 2\n ./coll\n fail\n Fatal error in PMPI_Init: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n nballtoall1\n 2\n ./coll\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n nbcoll2\n 2\n ./coll\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n-Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n-MPIR_Init_thread(565)..............: \n-MPID_Init(224).....................: channel initialization failed\n-MPIDI_CH3_Init(105)................: \n-MPID_nem_init(324).................: \n-MPID_nem_tcp_init(178).............: \n-MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n nbredscat\n 2\n ./coll\n fail\n Fatal error in PMPI_Init: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n-Fatal error in PMPI_Init: Other MPI error, error stack:\n-MPIR_Init_thread(565)..............: \n-MPID_Init(224).....................: channel initialization failed\n-MPIDI_CH3_Init(105)................: \n-MPID_nem_init(324).................: \n-MPID_nem_tcp_init(178).............: \n-MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n nbredscat\n 2\n ./coll\n fail\n Fatal error in PMPI_Init: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n nbredscat3\n 2\n ./coll\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n nbredscatinter\n 2\n ./coll\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n nbicbcast\n 2\n ./coll\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n nbicallreduce\n 2\n ./coll\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n nbicreduce\n 2\n ./coll\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n nbicscatter\n 2\n ./coll\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n+Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n+MPIR_Init_thread(565)..............: \n+MPID_Init(224).....................: channel initialization failed\n+MPIDI_CH3_Init(105)................: \n+MPID_nem_init(324).................: \n+MPID_nem_tcp_init(178).............: \n+MPID_nem_tcp_get_business_card(425): \n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n nbicgather\n 2\n ./coll\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n nbicallgather\n 2\n ./coll\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n nbicbarrier\n 2\n ./coll\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n nbicallgatherv\n 2\n ./coll\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n nbicgatherv\n 2\n ./coll\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n+Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n+MPIR_Init_thread(565)..............: \n+MPID_Init(224).....................: channel initialization failed\n+MPIDI_CH3_Init(105)................: \n+MPID_nem_init(324).................: \n+MPID_nem_tcp_init(178).............: \n+MPID_nem_tcp_get_business_card(425): \n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n nbicscatterv\n 2\n ./coll\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n nbicalltoall\n 2\n ./coll\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n nbicalltoallv\n 2\n ./coll\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n-Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n-MPIR_Init_thread(565)..............: \n-MPID_Init(224).....................: channel initialization failed\n-MPIDI_CH3_Init(105)................: \n-MPID_nem_init(324).................: \n-MPID_nem_tcp_init(178).............: \n-MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n nbicalltoallw\n 2\n ./coll\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n dup\n 2\n ./comm\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n-Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n-MPIR_Init_thread(565)..............: \n-MPID_Init(224).....................: channel initialization failed\n-MPIDI_CH3_Init(105)................: \n-MPID_nem_init(324).................: \n-MPID_nem_tcp_init(178).............: \n-MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n dupic\n 2\n ./comm\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n commcreate1\n 2\n ./comm\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n-Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n-MPIR_Init_thread(565)..............: \n-MPID_Init(224).....................: channel initialization failed\n-MPIDI_CH3_Init(105)................: \n-MPID_nem_init(324).................: \n-MPID_nem_tcp_init(178).............: \n-MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n commname\n 2\n ./comm\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n ic1\n 2\n ./comm\n fail\n- Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n-MPIR_Init_thread(565)..............: \n-MPID_Init(224).....................: channel initialization failed\n-MPIDI_CH3_Init(105)................: \n-MPID_nem_init(324).................: \n-MPID_nem_tcp_init(178).............: \n-MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n-Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n-MPIR_Init_thread(565)..............: \n-MPID_Init(224).....................: channel initialization failed\n-MPIDI_CH3_Init(105)................: \n-MPID_nem_init(324).................: \n-MPID_nem_tcp_init(178).............: \n-MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+ \n \n \n ic2\n 2\n ./comm\n fail\n Fatal error in PMPI_Init: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n icgroup\n 2\n ./comm\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n icm\n 2\n ./comm\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n icsplit\n 2\n ./comm\n fail\n- \n+ Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n+MPIR_Init_thread(565)..............: \n+MPID_Init(224).....................: channel initialization failed\n+MPIDI_CH3_Init(105)................: \n+MPID_nem_init(324).................: \n+MPID_nem_tcp_init(178).............: \n+MPID_nem_tcp_get_business_card(425): \n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n iccreate\n 2\n ./comm\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n+Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n+MPIR_Init_thread(565)..............: \n+MPID_Init(224).....................: channel initialization failed\n+MPIDI_CH3_Init(105)................: \n+MPID_nem_init(324).................: \n+MPID_nem_tcp_init(178).............: \n+MPID_nem_tcp_get_business_card(425): \n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n ctxalloc\n 2\n ./comm\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n ctxsplit\n 2\n ./comm\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n+Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n+MPIR_Init_thread(565)..............: \n+MPID_Init(224).....................: channel initialization failed\n+MPIDI_CH3_Init(105)................: \n+MPID_nem_init(324).................: \n+MPID_nem_tcp_init(178).............: \n+MPID_nem_tcp_get_business_card(425): \n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n cmfree\n 2\n ./comm\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n cmsplit\n 2\n ./comm\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n cmsplit2\n 2\n ./comm\n fail\n Fatal error in PMPI_Init: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n-Fatal error in PMPI_Init: Other MPI error, error stack:\n-MPIR_Init_thread(565)..............: \n-MPID_Init(224).....................: channel initialization failed\n-MPIDI_CH3_Init(105)................: \n-MPID_nem_init(324).................: \n-MPID_nem_tcp_init(178).............: \n-MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n probe-intercomm\n 2\n ./comm\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n cmsplit_type\n 2\n ./comm\n fail\n Fatal error in PMPI_Init: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n comm_create_group\n 2\n ./comm\n fail\n Fatal error in PMPI_Init: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n comm_create_group\n 2\n ./comm\n fail\n Fatal error in PMPI_Init: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n-Fatal error in PMPI_Init: Other MPI error, error stack:\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n+ \n+ \n+ comm_group_half\n+ 2\n+ ./comm\n+ fail\n+ Fatal error in PMPI_Init: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n comm_group_half\n 2\n ./comm\n fail\n Fatal error in PMPI_Init: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n comm_group_half\n 2\n ./comm\n fail\n Fatal error in PMPI_Init: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n Fatal error in PMPI_Init: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n- comm_group_half\n+ comm_group_rand\n 2\n ./comm\n fail\n Fatal error in PMPI_Init: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n comm_group_rand\n 2\n ./comm\n fail\n Fatal error in PMPI_Init: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n Fatal error in PMPI_Init: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n comm_group_rand\n 2\n ./comm\n fail\n Fatal error in PMPI_Init: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n-Fatal error in PMPI_Init: Other MPI error, error stack:\n-MPIR_Init_thread(565)..............: \n-MPID_Init(224).....................: channel initialization failed\n-MPIDI_CH3_Init(105)................: \n-MPID_nem_init(324).................: \n-MPID_nem_tcp_init(178).............: \n-MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n- comm_group_rand\n+ comm_idup\n 2\n ./comm\n fail\n Fatal error in PMPI_Init: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n comm_idup\n 2\n ./comm\n fail\n Fatal error in PMPI_Init: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n comm_idup\n 2\n ./comm\n fail\n Fatal error in PMPI_Init: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n-Fatal error in PMPI_Init: Other MPI error, error stack:\n-MPIR_Init_thread(565)..............: \n-MPID_Init(224).....................: channel initialization failed\n-MPIDI_CH3_Init(105)................: \n-MPID_nem_init(324).................: \n-MPID_nem_tcp_init(178).............: \n-MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n- comm_idup\n+ comm_idup_mul\n 2\n ./comm\n fail\n Fatal error in PMPI_Init: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n-Fatal error in PMPI_Init: Other MPI error, error stack:\n-MPIR_Init_thread(565)..............: \n-MPID_Init(224).....................: channel initialization failed\n-MPIDI_CH3_Init(105)................: \n-MPID_nem_init(324).................: \n-MPID_nem_tcp_init(178).............: \n-MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n- comm_idup_mul\n+ comm_idup_overlap\n 2\n ./comm\n fail\n Fatal error in PMPI_Init: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n Fatal error in PMPI_Init: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n- comm_idup_overlap\n+ comm_idup_iallreduce\n 2\n ./comm\n fail\n- Fatal error in PMPI_Init: Other MPI error, error stack:\n+ Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n- \n- \n- comm_idup_iallreduce\n- 2\n- ./comm\n- fail\n- Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n+Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n comm_idup_nb\n 2\n ./comm\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n comm_idup_isend\n 2\n ./comm\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n comm_idup_comm\n 2\n ./comm\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n-Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n-MPIR_Init_thread(565)..............: \n-MPID_Init(224).....................: channel initialization failed\n-MPIDI_CH3_Init(105)................: \n-MPID_nem_init(324).................: \n-MPID_nem_tcp_init(178).............: \n-MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n comm_idup_comm2\n 2\n ./comm\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n-Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n-MPIR_Init_thread(565)..............: \n-MPID_Init(224).....................: channel initialization failed\n-MPIDI_CH3_Init(105)................: \n-MPID_nem_init(324).................: \n-MPID_nem_tcp_init(178).............: \n-MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n dup_with_info\n 2\n ./comm\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n-Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n-MPIR_Init_thread(565)..............: \n-MPID_Init(224).....................: channel initialization failed\n-MPIDI_CH3_Init(105)................: \n-MPID_nem_init(324).................: \n-MPID_nem_tcp_init(178).............: \n-MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n dup_with_info\n 2\n ./comm\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n dup_with_info\n 2\n ./comm\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n-Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n-MPIR_Init_thread(565)..............: \n-MPID_Init(224).....................: channel initialization failed\n-MPIDI_CH3_Init(105)................: \n-MPID_nem_init(324).................: \n-MPID_nem_tcp_init(178).............: \n-MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n comm_info\n 2\n ./comm\n fail\n Fatal error in PMPI_Init: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n-Fatal error in PMPI_Init: Other MPI error, error stack:\n-MPIR_Init_thread(565)..............: \n-MPID_Init(224).....................: channel initialization failed\n-MPIDI_CH3_Init(105)................: \n-MPID_nem_init(324).................: \n-MPID_nem_tcp_init(178).............: \n-MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n comm_create_group_idup\n 2\n ./comm\n fail\n Fatal error in PMPI_Init: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n contents\n 1\n ./datatype\n fail\n Fatal error in PMPI_Init: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n gaddress\n 1\n ./datatype\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n lbub\n 1\n ./datatype\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n localpack\n 1\n ./datatype\n fail\n Fatal error in PMPI_Init: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n simple-pack\n 1\n ./datatype\n fail\n Fatal error in PMPI_Init: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n simple-pack-external\n 1\n ./datatype\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n transpose-pack\n 1\n ./datatype\n fail\n Fatal error in PMPI_Init: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n slice-pack\n 1\n ./datatype\n fail\n Fatal error in PMPI_Init: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n struct-pack\n 1\n ./datatype\n fail\n Fatal error in PMPI_Init: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n structpack2\n 1\n ./datatype\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n typecommit\n 1\n ./datatype\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n typename\n 1\n ./datatype\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n typefree\n 1\n ./datatype\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n zeroparms\n 1\n ./datatype\n fail\n Fatal error in PMPI_Init: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n getpartelm\n 2\n ./datatype\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n-Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n-MPIR_Init_thread(565)..............: \n-MPID_Init(224).....................: channel initialization failed\n-MPIDI_CH3_Init(105)................: \n-MPID_nem_init(324).................: \n-MPID_nem_tcp_init(178).............: \n-MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n tresized\n 2\n ./datatype\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n tresized2\n 2\n ./datatype\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n-Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n-MPIR_Init_thread(565)..............: \n-MPID_Init(224).....................: channel initialization failed\n-MPIDI_CH3_Init(105)................: \n-MPID_nem_init(324).................: \n-MPID_nem_tcp_init(178).............: \n-MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n sendrecvt2\n 2\n ./datatype\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n-Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n-MPIR_Init_thread(565)..............: \n-MPID_Init(224).....................: channel initialization failed\n-MPIDI_CH3_Init(105)................: \n-MPID_nem_init(324).................: \n-MPID_nem_tcp_init(178).............: \n-MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n sendrecvt4\n 2\n ./datatype\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n-Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n-MPIR_Init_thread(565)..............: \n-MPID_Init(224).....................: channel initialization failed\n-MPIDI_CH3_Init(105)................: \n-MPID_nem_init(324).................: \n-MPID_nem_tcp_init(178).............: \n-MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n tmatchsize\n 1\n ./datatype\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n tfree\n 2\n ./datatype\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n typelb\n 1\n ./datatype\n fail\n Fatal error in PMPI_Init: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n contigstruct\n 1\n ./datatype\n fail\n Fatal error in PMPI_Init: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n struct-zero-count\n 1\n ./datatype\n fail\n Fatal error in PMPI_Init: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n blockindexed-zero-count\n 1\n ./datatype\n fail\n Fatal error in PMPI_Init: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n blockindexed-misc\n 1\n ./datatype\n fail\n Fatal error in PMPI_Init: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n indexed-misc\n 1\n ./datatype\n fail\n Fatal error in PMPI_Init: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n subarray-pack\n 1\n ./datatype\n fail\n Fatal error in PMPI_Init: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n subarray\n 2\n ./datatype\n fail\n Fatal error in PMPI_Init: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n Fatal error in PMPI_Init: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n darray-pack\n 1\n ./datatype\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n darray-pack\n 2\n ./datatype\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n darray-pack\n 2\n ./datatype\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n+Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n+MPIR_Init_thread(565)..............: \n+MPID_Init(224).....................: channel initialization failed\n+MPIDI_CH3_Init(105)................: \n+MPID_nem_init(324).................: \n+MPID_nem_tcp_init(178).............: \n+MPID_nem_tcp_get_business_card(425): \n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n darray-cyclic\n 2\n ./datatype\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n pairtype-size-extent\n 1\n ./datatype\n fail\n Fatal error in PMPI_Init: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n simple-commit\n 1\n ./datatype\n fail\n Fatal error in PMPI_Init: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n simple-size-extent\n 1\n ./datatype\n fail\n Fatal error in PMPI_Init: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n struct-no-real-types\n 1\n ./datatype\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n struct-empty-el\n 1\n ./datatype\n fail\n Fatal error in PMPI_Init: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n contig-zero-count\n 1\n ./datatype\n fail\n Fatal error in PMPI_Init: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n simple-resized\n 1\n ./datatype\n fail\n Fatal error in PMPI_Init: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n unusual-noncontigs\n 1\n ./datatype\n fail\n Fatal error in PMPI_Init: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n hindexed-zeros\n 1\n ./datatype\n fail\n Fatal error in PMPI_Init: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n lots-of-types\n 1\n ./datatype\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n get-elements-pairtype\n 1\n ./datatype\n fail\n Fatal error in PMPI_Init: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n unpack\n 1\n ./datatype\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n struct-ezhov\n 1\n ./datatype\n fail\n Fatal error in PMPI_Init: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n zeroblks\n 1\n ./datatype\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n struct-derived-zeros\n 1\n ./datatype\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n struct-verydeep\n 1\n ./datatype\n fail\n Fatal error in PMPI_Init: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n get-elements\n 1\n ./datatype\n fail\n Fatal error in PMPI_Init: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n hindexed_block\n 1\n ./datatype\n fail\n Fatal error in PMPI_Init: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n hindexed_block_contents\n 1\n ./datatype\n fail\n Fatal error in PMPI_Init: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n vecblklen\n 1\n ./datatype\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n hvecblklen\n 1\n ./datatype\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n longdouble\n 1\n ./datatype\n fail\n Fatal error in PMPI_Init: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n dataalign\n 2\n ./datatype\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n-Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n-MPIR_Init_thread(565)..............: \n-MPID_Init(224).....................: channel initialization failed\n-MPIDI_CH3_Init(105)................: \n-MPID_nem_init(324).................: \n-MPID_nem_tcp_init(178).............: \n-MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n cxx-types\n 1\n ./datatype\n fail\n Fatal error in PMPI_Init: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n adderr\n 1\n ./errhan\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n commcall\n 2\n ./errhan\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n errfatal\n 1\n ./errhan\n pass\n \n@@ -4617,245 +4316,237 @@\n Fatal error in PMPI_Init: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n predef_eh\n 2\n ./errhan\n fail\n Fatal error in PMPI_Init: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n-Fatal error in PMPI_Init: Other MPI error, error stack:\n-MPIR_Init_thread(565)..............: \n-MPID_Init(224).....................: channel initialization failed\n-MPIDI_CH3_Init(105)................: \n-MPID_nem_init(324).................: \n-MPID_nem_tcp_init(178).............: \n-MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n errstring2\n 1\n ./errhan\n fail\n Fatal error in PMPI_Init: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n dynamic_errcode_predefined_errclass\n 1\n ./errhan\n fail\n Fatal error in PMPI_Init: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n @groupcreate\n \n ./group\n fail\n- Failed to build @groupcreate; make[2]: Entering directory '/build/1st/mpi-testsuite-3.2+dfsg/build-mpich/group'\n+ Failed to build @groupcreate; make[2]: Entering directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-mpich/group'\n make[2]: *** No rule to make target '@groupcreate'. Stop.\n-make[2]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-mpich/group'\n+make[2]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-mpich/group'\n \n \n @grouptest\n \n ./group\n fail\n- Failed to build @grouptest; make[2]: Entering directory '/build/1st/mpi-testsuite-3.2+dfsg/build-mpich/group'\n+ Failed to build @grouptest; make[2]: Entering directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-mpich/group'\n make[2]: *** No rule to make target '@grouptest'. Stop.\n-make[2]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-mpich/group'\n+make[2]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-mpich/group'\n \n \n @grouptest2\n \n ./group\n fail\n- Failed to build @grouptest2; make[2]: Entering directory '/build/1st/mpi-testsuite-3.2+dfsg/build-mpich/group'\n+ Failed to build @grouptest2; make[2]: Entering directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-mpich/group'\n make[2]: *** No rule to make target '@grouptest2'. Stop.\n-make[2]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-mpich/group'\n+make[2]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-mpich/group'\n \n \n @groupnullincl\n \n ./group\n fail\n- Failed to build @groupnullincl; make[2]: Entering directory '/build/1st/mpi-testsuite-3.2+dfsg/build-mpich/group'\n+ Failed to build @groupnullincl; make[2]: Entering directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-mpich/group'\n make[2]: *** No rule to make target '@groupnullincl'. Stop.\n-make[2]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-mpich/group'\n+make[2]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-mpich/group'\n \n \n @gtranks\n \n ./group\n fail\n- Failed to build @gtranks; make[2]: Entering directory '/build/1st/mpi-testsuite-3.2+dfsg/build-mpich/group'\n+ Failed to build @gtranks; make[2]: Entering directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-mpich/group'\n make[2]: *** No rule to make target '@gtranks'. Stop.\n-make[2]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-mpich/group'\n+make[2]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-mpich/group'\n \n \n @gtranksperf\n \n ./group\n fail\n- Failed to build @gtranksperf; make[2]: Entering directory '/build/1st/mpi-testsuite-3.2+dfsg/build-mpich/group'\n+ Failed to build @gtranksperf; make[2]: Entering directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-mpich/group'\n make[2]: *** No rule to make target '@gtranksperf'. Stop.\n-make[2]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-mpich/group'\n+make[2]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-mpich/group'\n \n \n infodup\n 1\n ./info\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n infodel\n 1\n ./info\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n infovallen\n 1\n ./info\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n infoorder\n 1\n ./info\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n infomany\n 1\n ./info\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n infomany2\n 1\n ./info\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n infotest\n 1\n ./info\n fail\n Fatal error in PMPI_Init: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n infoget\n 1\n ./info\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n infoenv\n 1\n ./info\n fail\n Fatal error in PMPI_Init: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n exitst1\n 2\n ./init\n pass\n \n@@ -4873,10428 +4564,9914 @@\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n timeout\n 2\n ./init\n fail\n Fatal error in PMPI_Init: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n version\n 1\n ./init\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n finalized\n 1\n ./init\n fail\n- Fatal error in PMPI_Init: Other MPI error, error stack:\n-MPIR_Init_thread(565)..............: \n-MPID_Init(224).....................: channel initialization failed\n-MPIDI_CH3_Init(105)................: \n-MPID_nem_init(324).................: \n-MPID_nem_tcp_init(178).............: \n-MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+ \n \n \n attrself\n 1\n ./init\n fail\n Fatal error in PMPI_Init: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n library_version\n 1\n ./init\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n mpi_t_str\n 1\n ./mpi_t\n fail\n Fatal error in PMPI_Init: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n mpit_vars\n 1\n ./mpi_t\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n cvarwrite\n 1\n ./mpi_t\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n sendrecv1\n 2\n ./pt2pt\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n sendrecv2\n 2\n ./pt2pt\n fail\n Fatal error in PMPI_Init: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n Fatal error in PMPI_Init: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n sendrecv3\n 2\n ./pt2pt\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n sendflood\n 2\n ./pt2pt\n fail\n Fatal error in PMPI_Init: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n sendself\n 1\n ./pt2pt\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n sendall\n 2\n ./pt2pt\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n anyall\n 2\n ./pt2pt\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n-Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n-MPIR_Init_thread(565)..............: \n-MPID_Init(224).....................: channel initialization failed\n-MPIDI_CH3_Init(105)................: \n-MPID_nem_init(324).................: \n-MPID_nem_tcp_init(178).............: \n-MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n eagerdt\n 2\n ./pt2pt\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n pingping\n 2\n ./pt2pt\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n-Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n-MPIR_Init_thread(565)..............: \n-MPID_Init(224).....................: channel initialization failed\n-MPIDI_CH3_Init(105)................: \n-MPID_nem_init(324).................: \n-MPID_nem_tcp_init(178).............: \n-MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n bottom\n 2\n ./pt2pt\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n bsend1\n 1\n ./pt2pt\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n bsend2\n 1\n ./pt2pt\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n bsend3\n 1\n ./pt2pt\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n bsend4\n 1\n ./pt2pt\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n bsend5\n 2\n ./pt2pt\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n bsendalign\n 2\n ./pt2pt\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n+Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n+MPIR_Init_thread(565)..............: \n+MPID_Init(224).....................: channel initialization failed\n+MPIDI_CH3_Init(105)................: \n+MPID_nem_init(324).................: \n+MPID_nem_tcp_init(178).............: \n+MPID_nem_tcp_get_business_card(425): \n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n bsendpending\n 2\n ./pt2pt\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n isendself\n 1\n ./pt2pt\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n issendselfcancel\n 1\n ./pt2pt\n fail\n Fatal error in PMPI_Init: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n isendirecv\n 2\n ./pt2pt\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n+Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n+MPIR_Init_thread(565)..............: \n+MPID_Init(224).....................: channel initialization failed\n+MPIDI_CH3_Init(105)................: \n+MPID_nem_init(324).................: \n+MPID_nem_tcp_init(178).............: \n+MPID_nem_tcp_get_business_card(425): \n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n bsendfrag\n 2\n ./pt2pt\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n icsend\n 2\n ./pt2pt\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n+Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n+MPIR_Init_thread(565)..............: \n+MPID_Init(224).....................: channel initialization failed\n+MPIDI_CH3_Init(105)................: \n+MPID_nem_init(324).................: \n+MPID_nem_tcp_init(178).............: \n+MPID_nem_tcp_get_business_card(425): \n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n rqstatus\n 2\n ./pt2pt\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n rqfreeb\n 2\n ./pt2pt\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n-Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n-MPIR_Init_thread(565)..............: \n-MPID_Init(224).....................: channel initialization failed\n-MPIDI_CH3_Init(105)................: \n-MPID_nem_init(324).................: \n-MPID_nem_tcp_init(178).............: \n-MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n greq1\n 1\n ./pt2pt\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n probe-unexp\n 2\n ./pt2pt\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n probenull\n 1\n ./pt2pt\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n scancel\n 2\n ./pt2pt\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n scancel2\n 2\n ./pt2pt\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n pscancel\n 2\n ./pt2pt\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n+Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n+MPIR_Init_thread(565)..............: \n+MPID_Init(224).....................: channel initialization failed\n+MPIDI_CH3_Init(105)................: \n+MPID_nem_init(324).................: \n+MPID_nem_tcp_init(178).............: \n+MPID_nem_tcp_get_business_card(425): \n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n rcancel\n 2\n ./pt2pt\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n cancelrecv\n 2\n ./pt2pt\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n scancel_unmatch\n 2\n ./pt2pt\n fail\n Fatal error in PMPI_Init: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n Fatal error in PMPI_Init: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n cancelanysrc\n 2\n ./pt2pt\n fail\n Fatal error in PMPI_Init: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n-Fatal error in PMPI_Init: Other MPI error, error stack:\n-MPIR_Init_thread(565)..............: \n-MPID_Init(224).....................: channel initialization failed\n-MPIDI_CH3_Init(105)................: \n-MPID_nem_init(324).................: \n-MPID_nem_tcp_init(178).............: \n-MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n isendselfprobe\n 1\n ./pt2pt\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n inactivereq\n 1\n ./pt2pt\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n waittestnull\n 1\n ./pt2pt\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n waitany-null\n 1\n ./pt2pt\n fail\n Fatal error in PMPI_Init: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n mprobe\n 2\n ./pt2pt\n fail\n Fatal error in PMPI_Init: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n-Fatal error in PMPI_Init: Other MPI error, error stack:\n-MPIR_Init_thread(565)..............: \n-MPID_Init(224).....................: channel initialization failed\n-MPIDI_CH3_Init(105)................: \n-MPID_nem_init(324).................: \n-MPID_nem_tcp_init(178).............: \n-MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n big_count_status\n 1\n ./pt2pt\n fail\n Fatal error in PMPI_Init: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n winname\n 2\n ./rma\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n-Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n-MPIR_Init_thread(565)..............: \n-MPID_Init(224).....................: channel initialization failed\n-MPIDI_CH3_Init(105)................: \n-MPID_nem_init(324).................: \n-MPID_nem_tcp_init(178).............: \n-MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n allocmem\n 2\n ./rma\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n-Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n-MPIR_Init_thread(565)..............: \n-MPID_Init(224).....................: channel initialization failed\n-MPIDI_CH3_Init(105)................: \n-MPID_nem_init(324).................: \n-MPID_nem_tcp_init(178).............: \n-MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n putfidx\n 2\n ./rma\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n accfence1\n 2\n ./rma\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n-Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n-MPIR_Init_thread(565)..............: \n-MPID_Init(224).....................: channel initialization failed\n-MPIDI_CH3_Init(105)................: \n-MPID_nem_init(324).................: \n-MPID_nem_tcp_init(178).............: \n-MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n adlb_mimic1\n 2\n ./rma\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n accfence2\n 2\n ./rma\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n+Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n+MPIR_Init_thread(565)..............: \n+MPID_Init(224).....................: channel initialization failed\n+MPIDI_CH3_Init(105)................: \n+MPID_nem_init(324).................: \n+MPID_nem_tcp_init(178).............: \n+MPID_nem_tcp_get_business_card(425): \n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n putpscw1\n 2\n ./rma\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n-Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n-MPIR_Init_thread(565)..............: \n-MPID_Init(224).....................: channel initialization failed\n-MPIDI_CH3_Init(105)................: \n-MPID_nem_init(324).................: \n-MPID_nem_tcp_init(178).............: \n-MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n accpscw1\n 2\n ./rma\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n getgroup\n 2\n ./rma\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n transpose1\n 2\n ./rma\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n+Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n+MPIR_Init_thread(565)..............: \n+MPID_Init(224).....................: channel initialization failed\n+MPIDI_CH3_Init(105)................: \n+MPID_nem_init(324).................: \n+MPID_nem_tcp_init(178).............: \n+MPID_nem_tcp_get_business_card(425): \n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n transpose2\n 2\n ./rma\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n transpose3\n 2\n ./rma\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n-Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n-MPIR_Init_thread(565)..............: \n-MPID_Init(224).....................: channel initialization failed\n-MPIDI_CH3_Init(105)................: \n-MPID_nem_init(324).................: \n-MPID_nem_tcp_init(178).............: \n-MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n transpose3_shm\n 2\n ./rma\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n transpose5\n 2\n ./rma\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n transpose6\n 1\n ./rma\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n transpose7\n 2\n ./rma\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n test1\n 2\n ./rma\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n-Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n-MPIR_Init_thread(565)..............: \n-MPID_Init(224).....................: channel initialization failed\n-MPIDI_CH3_Init(105)................: \n-MPID_nem_init(324).................: \n-MPID_nem_tcp_init(178).............: \n-MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n test2\n 2\n ./rma\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n+Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n+MPIR_Init_thread(565)..............: \n+MPID_Init(224).....................: channel initialization failed\n+MPIDI_CH3_Init(105)................: \n+MPID_nem_init(324).................: \n+MPID_nem_tcp_init(178).............: \n+MPID_nem_tcp_get_business_card(425): \n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n test2_shm\n 2\n ./rma\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n+Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n+MPIR_Init_thread(565)..............: \n+MPID_Init(224).....................: channel initialization failed\n+MPIDI_CH3_Init(105)................: \n+MPID_nem_init(324).................: \n+MPID_nem_tcp_init(178).............: \n+MPID_nem_tcp_get_business_card(425): \n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n test3\n 2\n ./rma\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n test3_shm\n 2\n ./rma\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n+Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n+MPIR_Init_thread(565)..............: \n+MPID_Init(224).....................: channel initialization failed\n+MPIDI_CH3_Init(105)................: \n+MPID_nem_init(324).................: \n+MPID_nem_tcp_init(178).............: \n+MPID_nem_tcp_get_business_card(425): \n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n test4\n 2\n ./rma\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n test5\n 2\n ./rma\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n lockcontention\n 2\n ./rma\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n-Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n-MPIR_Init_thread(565)..............: \n-MPID_Init(224).....................: channel initialization failed\n-MPIDI_CH3_Init(105)................: \n-MPID_nem_init(324).................: \n-MPID_nem_tcp_init(178).............: \n-MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n lockcontention2\n 2\n ./rma\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n lockcontention2\n 2\n ./rma\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n-Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n-MPIR_Init_thread(565)..............: \n-MPID_Init(224).....................: channel initialization failed\n-MPIDI_CH3_Init(105)................: \n-MPID_nem_init(324).................: \n-MPID_nem_tcp_init(178).............: \n-MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n lockcontention3\n 2\n ./rma\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n+Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n+MPIR_Init_thread(565)..............: \n+MPID_Init(224).....................: channel initialization failed\n+MPIDI_CH3_Init(105)................: \n+MPID_nem_init(324).................: \n+MPID_nem_tcp_init(178).............: \n+MPID_nem_tcp_get_business_card(425): \n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n lockopts\n 2\n ./rma\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n lock_dt\n 2\n ./rma\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n lock_dt_flush\n 2\n ./rma\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n+Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n+MPIR_Init_thread(565)..............: \n+MPID_Init(224).....................: channel initialization failed\n+MPIDI_CH3_Init(105)................: \n+MPID_nem_init(324).................: \n+MPID_nem_tcp_init(178).............: \n+MPID_nem_tcp_get_business_card(425): \n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n lock_dt_flushlocal\n 2\n ./rma\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n lockall_dt\n 2\n ./rma\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n lockall_dt_flush\n 2\n ./rma\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n-Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n-MPIR_Init_thread(565)..............: \n-MPID_Init(224).....................: channel initialization failed\n-MPIDI_CH3_Init(105)................: \n-MPID_nem_init(324).................: \n-MPID_nem_tcp_init(178).............: \n-MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n lockall_dt_flushall\n 2\n ./rma\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n lockall_dt_flushlocal\n 2\n ./rma\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n+Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n+MPIR_Init_thread(565)..............: \n+MPID_Init(224).....................: channel initialization failed\n+MPIDI_CH3_Init(105)................: \n+MPID_nem_init(324).................: \n+MPID_nem_tcp_init(178).............: \n+MPID_nem_tcp_get_business_card(425): \n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n lockall_dt_flushlocalall\n 2\n ./rma\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n-Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n-MPIR_Init_thread(565)..............: \n-MPID_Init(224).....................: channel initialization failed\n-MPIDI_CH3_Init(105)................: \n-MPID_nem_init(324).................: \n-MPID_nem_tcp_init(178).............: \n-MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n lock_contention_dt\n 2\n ./rma\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n-Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n-MPIR_Init_thread(565)..............: \n-MPID_Init(224).....................: channel initialization failed\n-MPIDI_CH3_Init(105)................: \n-MPID_nem_init(324).................: \n-MPID_nem_tcp_init(178).............: \n-MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n transpose4\n 2\n ./rma\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n-Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n-MPIR_Init_thread(565)..............: \n-MPID_Init(224).....................: channel initialization failed\n-MPIDI_CH3_Init(105)................: \n-MPID_nem_init(324).................: \n-MPID_nem_tcp_init(178).............: \n-MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n fetchandadd\n 2\n ./rma\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n fetchandadd_tree\n 2\n ./rma\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n+Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n+MPIR_Init_thread(565)..............: \n+MPID_Init(224).....................: channel initialization failed\n+MPIDI_CH3_Init(105)................: \n+MPID_nem_init(324).................: \n+MPID_nem_tcp_init(178).............: \n+MPID_nem_tcp_get_business_card(425): \n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n wintest\n 2\n ./rma\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n wintest_shm\n 2\n ./rma\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n-Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n-MPIR_Init_thread(565)..............: \n-MPID_Init(224).....................: channel initialization failed\n-MPIDI_CH3_Init(105)................: \n-MPID_nem_init(324).................: \n-MPID_nem_tcp_init(178).............: \n-MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n contig_displ\n 1\n ./rma\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n test1_am\n 2\n ./rma\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n test2_am\n 2\n ./rma\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n-Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n-MPIR_Init_thread(565)..............: \n-MPID_Init(224).....................: channel initialization failed\n-MPIDI_CH3_Init(105)................: \n-MPID_nem_init(324).................: \n-MPID_nem_tcp_init(178).............: \n-MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n test2_am_shm\n 2\n ./rma\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n-Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n-MPIR_Init_thread(565)..............: \n-MPID_Init(224).....................: channel initialization failed\n-MPIDI_CH3_Init(105)................: \n-MPID_nem_init(324).................: \n-MPID_nem_tcp_init(178).............: \n-MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n test3_am\n 2\n ./rma\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n test3_am_shm\n 2\n ./rma\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n test4_am\n 2\n ./rma\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n test5_am\n 2\n ./rma\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n-Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n-MPIR_Init_thread(565)..............: \n-MPID_Init(224).....................: channel initialization failed\n-MPIDI_CH3_Init(105)................: \n-MPID_nem_init(324).................: \n-MPID_nem_tcp_init(178).............: \n-MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n fetchandadd_am\n 2\n ./rma\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n fetchandadd_tree_am\n 2\n ./rma\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n-Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n-MPIR_Init_thread(565)..............: \n-MPID_Init(224).....................: channel initialization failed\n-MPIDI_CH3_Init(105)................: \n-MPID_nem_init(324).................: \n-MPID_nem_tcp_init(178).............: \n-MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n accfence2_am\n 2\n ./rma\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n+Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n+MPIR_Init_thread(565)..............: \n+MPID_Init(224).....................: channel initialization failed\n+MPIDI_CH3_Init(105)................: \n+MPID_nem_init(324).................: \n+MPID_nem_tcp_init(178).............: \n+MPID_nem_tcp_get_business_card(425): \n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n test1_dt\n 2\n ./rma\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n+Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n+MPIR_Init_thread(565)..............: \n+MPID_Init(224).....................: channel initialization failed\n+MPIDI_CH3_Init(105)................: \n+MPID_nem_init(324).................: \n+MPID_nem_tcp_init(178).............: \n+MPID_nem_tcp_get_business_card(425): \n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n nullpscw\n 2\n ./rma\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n nullpscw_shm\n 2\n ./rma\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n attrorderwin\n 1\n ./rma\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n wincall\n 2\n ./rma\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n-Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n-MPIR_Init_thread(565)..............: \n-MPID_Init(224).....................: channel initialization failed\n-MPIDI_CH3_Init(105)................: \n-MPID_nem_init(324).................: \n-MPID_nem_tcp_init(178).............: \n-MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n baseattrwin\n 1\n ./rma\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n fkeyvalwin\n 1\n ./rma\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n selfrma\n 1\n ./rma\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n mixedsync\n 2\n ./rma\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n epochtest\n 2\n ./rma\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n+Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n+MPIR_Init_thread(565)..............: \n+MPID_Init(224).....................: channel initialization failed\n+MPIDI_CH3_Init(105)................: \n+MPID_nem_init(324).................: \n+MPID_nem_tcp_init(178).............: \n+MPID_nem_tcp_get_business_card(425): \n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n locknull\n 2\n ./rma\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n rmanull\n 2\n ./rma\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n-Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n-MPIR_Init_thread(565)..............: \n-MPID_Init(224).....................: channel initialization failed\n-MPIDI_CH3_Init(105)................: \n-MPID_nem_init(324).................: \n-MPID_nem_tcp_init(178).............: \n-MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n rmazero\n 2\n ./rma\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n strided_acc_indexed\n 2\n ./rma\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n-Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n-MPIR_Init_thread(565)..............: \n-MPID_Init(224).....................: channel initialization failed\n-MPIDI_CH3_Init(105)................: \n-MPID_nem_init(324).................: \n-MPID_nem_tcp_init(178).............: \n-MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n strided_acc_onelock\n 2\n ./rma\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n strided_acc_subarray\n 2\n ./rma\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n-Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n-MPIR_Init_thread(565)..............: \n-MPID_Init(224).....................: channel initialization failed\n-MPIDI_CH3_Init(105)................: \n-MPID_nem_init(324).................: \n-MPID_nem_tcp_init(178).............: \n-MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n strided_get_indexed\n 2\n ./rma\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n-Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n-MPIR_Init_thread(565)..............: \n-MPID_Init(224).....................: channel initialization failed\n-MPIDI_CH3_Init(105)................: \n-MPID_nem_init(324).................: \n-MPID_nem_tcp_init(178).............: \n-MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n strided_putget_indexed\n 2\n ./rma\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n strided_putget_indexed_shared\n 2\n ./rma\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n strided_getacc_indexed\n 2\n ./rma\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n strided_getacc_indexed_shared\n 2\n ./rma\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n-Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n-MPIR_Init_thread(565)..............: \n-MPID_Init(224).....................: channel initialization failed\n-MPIDI_CH3_Init(105)................: \n-MPID_nem_init(324).................: \n-MPID_nem_tcp_init(178).............: \n-MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n window_creation\n 2\n ./rma\n fail\n Fatal error in PMPI_Init: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n+Fatal error in PMPI_Init: Other MPI error, error stack:\n+MPIR_Init_thread(565)..............: \n+MPID_Init(224).....................: channel initialization failed\n+MPIDI_CH3_Init(105)................: \n+MPID_nem_init(324).................: \n+MPID_nem_tcp_init(178).............: \n+MPID_nem_tcp_get_business_card(425): \n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n contention_put\n 2\n ./rma\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n contention_putget\n 2\n ./rma\n fail\n Fatal error in PMPI_Init: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n-Fatal error in PMPI_Init: Other MPI error, error stack:\n-MPIR_Init_thread(565)..............: \n-MPID_Init(224).....................: channel initialization failed\n-MPIDI_CH3_Init(105)................: \n-MPID_nem_init(324).................: \n-MPID_nem_tcp_init(178).............: \n-MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n put_base\n 2\n ./rma\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n-Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n-MPIR_Init_thread(565)..............: \n-MPID_Init(224).....................: channel initialization failed\n-MPIDI_CH3_Init(105)................: \n-MPID_nem_init(324).................: \n-MPID_nem_tcp_init(178).............: \n-MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n put_bottom\n 2\n ./rma\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n-Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n-MPIR_Init_thread(565)..............: \n-MPID_Init(224).....................: channel initialization failed\n-MPIDI_CH3_Init(105)................: \n-MPID_nem_init(324).................: \n-MPID_nem_tcp_init(178).............: \n-MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n+ \n+ \n+ win_flavors\n+ 2\n+ ./rma\n+ fail\n+ \n \n \n win_flavors\n 2\n ./rma\n fail\n Fatal error in PMPI_Init: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n- win_flavors\n+ manyrma2\n 2\n ./rma\n fail\n Fatal error in PMPI_Init: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n Fatal error in PMPI_Init: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n- manyrma2\n+ manyrma2_shm\n 2\n ./rma\n fail\n Fatal error in PMPI_Init: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n- \n- \n- manyrma2_shm\n- 2\n- ./rma\n- fail\n- Fatal error in PMPI_Init: Other MPI error, error stack:\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n+Fatal error in PMPI_Init: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n manyrma3\n 2\n ./rma\n fail\n Fatal error in PMPI_Init: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n-Fatal error in PMPI_Init: Other MPI error, error stack:\n-MPIR_Init_thread(565)..............: \n-MPID_Init(224).....................: channel initialization failed\n-MPIDI_CH3_Init(105)................: \n-MPID_nem_init(324).................: \n-MPID_nem_tcp_init(178).............: \n-MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n win_shared\n 2\n ./rma\n fail\n Fatal error in PMPI_Init: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n-Fatal error in PMPI_Init: Other MPI error, error stack:\n-MPIR_Init_thread(565)..............: \n-MPID_Init(224).....................: channel initialization failed\n-MPIDI_CH3_Init(105)................: \n-MPID_nem_init(324).................: \n-MPID_nem_tcp_init(178).............: \n-MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n win_shared_create_allocshm\n 2\n ./rma\n fail\n Fatal error in PMPI_Init: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n+Fatal error in PMPI_Init: Other MPI error, error stack:\n+MPIR_Init_thread(565)..............: \n+MPID_Init(224).....................: channel initialization failed\n+MPIDI_CH3_Init(105)................: \n+MPID_nem_init(324).................: \n+MPID_nem_tcp_init(178).............: \n+MPID_nem_tcp_get_business_card(425): \n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n win_shared_create_no_allocshm\n 2\n ./rma\n fail\n Fatal error in PMPI_Init: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n Fatal error in PMPI_Init: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n win_shared_noncontig\n 2\n ./rma\n fail\n Fatal error in PMPI_Init: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n-Fatal error in PMPI_Init: Other MPI error, error stack:\n-MPIR_Init_thread(565)..............: \n-MPID_Init(224).....................: channel initialization failed\n-MPIDI_CH3_Init(105)................: \n-MPID_nem_init(324).................: \n-MPID_nem_tcp_init(178).............: \n-MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n win_shared_noncontig_put\n 2\n ./rma\n fail\n Fatal error in PMPI_Init: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n Fatal error in PMPI_Init: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n win_zero\n 2\n ./rma\n fail\n Fatal error in PMPI_Init: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n-Fatal error in PMPI_Init: Other MPI error, error stack:\n-MPIR_Init_thread(565)..............: \n-MPID_Init(224).....................: channel initialization failed\n-MPIDI_CH3_Init(105)................: \n-MPID_nem_init(324).................: \n-MPID_nem_tcp_init(178).............: \n-MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n win_dynamic_acc\n 2\n ./rma\n fail\n Fatal error in PMPI_Init: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n Fatal error in PMPI_Init: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n get_acc_local\n 1\n ./rma\n fail\n Fatal error in PMPI_Init: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n linked_list\n 2\n ./rma\n fail\n Fatal error in PMPI_Init: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n Fatal error in PMPI_Init: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n linked_list_fop\n 2\n ./rma\n fail\n Fatal error in PMPI_Init: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n compare_and_swap\n 2\n ./rma\n fail\n Fatal error in PMPI_Init: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n Fatal error in PMPI_Init: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n fetch_and_op_char\n 2\n ./rma\n fail\n Fatal error in PMPI_Init: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n fetch_and_op_short\n 2\n ./rma\n fail\n Fatal error in PMPI_Init: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n- \n- \n- fetch_and_op_int\n- 2\n- ./rma\n- fail\n- Fatal error in PMPI_Init: Other MPI error, error stack:\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n+Fatal error in PMPI_Init: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n- fetch_and_op_long\n+ fetch_and_op_int\n 2\n ./rma\n fail\n Fatal error in PMPI_Init: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n Fatal error in PMPI_Init: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n- \n- \n- fetch_and_op_double\n- 2\n- ./rma\n- fail\n- Fatal error in PMPI_Init: Other MPI error, error stack:\n-MPIR_Init_thread(565)..............: \n-MPID_Init(224).....................: channel initialization failed\n-MPIDI_CH3_Init(105)................: \n-MPID_nem_init(324).................: \n-MPID_nem_tcp_init(178).............: \n-MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n- fetch_and_op_long_double\n+ fetch_and_op_long\n 2\n ./rma\n fail\n Fatal error in PMPI_Init: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n- get_accumulate_double\n+ fetch_and_op_double\n 2\n ./rma\n fail\n Fatal error in PMPI_Init: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n- get_accumulate_double_derived\n+ fetch_and_op_long_double\n 2\n ./rma\n fail\n Fatal error in PMPI_Init: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n- get_accumulate_int\n+ get_accumulate_double\n 2\n ./rma\n fail\n Fatal error in PMPI_Init: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n Fatal error in PMPI_Init: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n- get_accumulate_int_derived\n+ get_accumulate_double_derived\n 2\n ./rma\n fail\n Fatal error in PMPI_Init: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n- get_accumulate_long\n+ get_accumulate_int\n 2\n ./rma\n fail\n Fatal error in PMPI_Init: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n- get_accumulate_long_derived\n+ get_accumulate_int_derived\n 2\n ./rma\n fail\n Fatal error in PMPI_Init: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n- get_accumulate_short\n+ get_accumulate_long\n 2\n ./rma\n fail\n Fatal error in PMPI_Init: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n+Fatal error in PMPI_Init: Other MPI error, error stack:\n+MPIR_Init_thread(565)..............: \n+MPID_Init(224).....................: channel initialization failed\n+MPIDI_CH3_Init(105)................: \n+MPID_nem_init(324).................: \n+MPID_nem_tcp_init(178).............: \n+MPID_nem_tcp_get_business_card(425): \n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n- get_accumulate_short_derived\n+ get_accumulate_long_derived\n 2\n ./rma\n fail\n Fatal error in PMPI_Init: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n Fatal error in PMPI_Init: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n- flush\n+ get_accumulate_short\n 2\n ./rma\n fail\n Fatal error in PMPI_Init: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n- reqops\n+ get_accumulate_short_derived\n 2\n ./rma\n fail\n Fatal error in PMPI_Init: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n- req_example\n+ flush\n 2\n ./rma\n fail\n Fatal error in PMPI_Init: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n+Fatal error in PMPI_Init: Other MPI error, error stack:\n+MPIR_Init_thread(565)..............: \n+MPID_Init(224).....................: channel initialization failed\n+MPIDI_CH3_Init(105)................: \n+MPID_nem_init(324).................: \n+MPID_nem_tcp_init(178).............: \n+MPID_nem_tcp_get_business_card(425): \n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n- req_example_shm\n+ reqops\n 2\n ./rma\n fail\n Fatal error in PMPI_Init: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n Fatal error in PMPI_Init: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n- rput_local_comp\n+ req_example\n 2\n ./rma\n fail\n Fatal error in PMPI_Init: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n Fatal error in PMPI_Init: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n- racc_local_comp\n+ req_example_shm\n 2\n ./rma\n fail\n Fatal error in PMPI_Init: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n- win_info\n+ rput_local_comp\n+ 2\n+ ./rma\n+ fail\n+ Fatal error in PMPI_Init: Other MPI error, error stack:\n+MPIR_Init_thread(565)..............: \n+MPID_Init(224).....................: channel initialization failed\n+MPIDI_CH3_Init(105)................: \n+MPID_nem_init(324).................: \n+MPID_nem_tcp_init(178).............: \n+MPID_nem_tcp_get_business_card(425): \n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n+ \n+ \n+ racc_local_comp\n 2\n ./rma\n fail\n Fatal error in PMPI_Init: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n Fatal error in PMPI_Init: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n- linked_list_lockall\n+ win_info\n 2\n ./rma\n fail\n Fatal error in PMPI_Init: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n-Fatal error in PMPI_Init: Other MPI error, error stack:\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n+ \n+ \n+ linked_list_lockall\n+ 2\n+ ./rma\n+ fail\n+ Fatal error in PMPI_Init: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n pscw_ordering\n 2\n ./rma\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n pscw_ordering_shm\n 2\n ./rma\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n+Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n+MPIR_Init_thread(565)..............: \n+MPID_Init(224).....................: channel initialization failed\n+MPIDI_CH3_Init(105)................: \n+MPID_nem_init(324).................: \n+MPID_nem_tcp_init(178).............: \n+MPID_nem_tcp_get_business_card(425): \n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n linked_list_bench_lock_all\n 2\n ./rma\n fail\n Fatal error in PMPI_Init: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n+Fatal error in PMPI_Init: Other MPI error, error stack:\n+MPIR_Init_thread(565)..............: \n+MPID_Init(224).....................: channel initialization failed\n+MPIDI_CH3_Init(105)................: \n+MPID_nem_init(324).................: \n+MPID_nem_tcp_init(178).............: \n+MPID_nem_tcp_get_business_card(425): \n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n linked_list_bench_lock_excl\n 2\n ./rma\n fail\n Fatal error in PMPI_Init: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n Fatal error in PMPI_Init: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n linked_list_bench_lock_shr\n 2\n ./rma\n fail\n Fatal error in PMPI_Init: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n+Fatal error in PMPI_Init: Other MPI error, error stack:\n+MPIR_Init_thread(565)..............: \n+MPID_Init(224).....................: channel initialization failed\n+MPIDI_CH3_Init(105)................: \n+MPID_nem_init(324).................: \n+MPID_nem_tcp_init(178).............: \n+MPID_nem_tcp_get_business_card(425): \n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n linked_list_bench_lock_shr_nocheck\n 2\n ./rma\n fail\n Fatal error in PMPI_Init: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n mutex_bench\n 2\n ./rma\n fail\n Fatal error in PMPI_Init: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n+Fatal error in PMPI_Init: Other MPI error, error stack:\n+MPIR_Init_thread(565)..............: \n+MPID_Init(224).....................: channel initialization failed\n+MPIDI_CH3_Init(105)................: \n+MPID_nem_init(324).................: \n+MPID_nem_tcp_init(178).............: \n+MPID_nem_tcp_get_business_card(425): \n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n mutex_bench_shared\n 2\n ./rma\n fail\n Fatal error in PMPI_Init: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n-Fatal error in PMPI_Init: Other MPI error, error stack:\n-MPIR_Init_thread(565)..............: \n-MPID_Init(224).....................: channel initialization failed\n-MPIDI_CH3_Init(105)................: \n-MPID_nem_init(324).................: \n-MPID_nem_tcp_init(178).............: \n-MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n mutex_bench_shm\n 2\n ./rma\n fail\n Fatal error in PMPI_Init: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n mutex_bench_shm_ordered\n 2\n ./rma\n fail\n Fatal error in PMPI_Init: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n rma-contig\n 2\n ./rma\n fail\n Fatal error in PMPI_Init: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n Fatal error in PMPI_Init: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n badrma\n 2\n ./rma\n fail\n Fatal error in PMPI_Init: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n Fatal error in PMPI_Init: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n acc-loc\n 2\n ./rma\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n-Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n-MPIR_Init_thread(565)..............: \n-MPID_Init(224).....................: channel initialization failed\n-MPIDI_CH3_Init(105)................: \n-MPID_nem_init(324).................: \n-MPID_nem_tcp_init(178).............: \n-MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n fence_shm\n 2\n ./rma\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n win_shared_zerobyte\n 2\n ./rma\n fail\n Fatal error in PMPI_Init: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n Fatal error in PMPI_Init: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n win_shared_put_flush_get\n 2\n ./rma\n fail\n Fatal error in PMPI_Init: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n-Fatal error in PMPI_Init: Other MPI error, error stack:\n-MPIR_Init_thread(565)..............: \n-MPID_Init(224).....................: channel initialization failed\n-MPIDI_CH3_Init(105)................: \n-MPID_nem_init(324).................: \n-MPID_nem_tcp_init(178).............: \n-MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n get-struct\n 2\n ./rma\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n+Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n+MPIR_Init_thread(565)..............: \n+MPID_Init(224).....................: channel initialization failed\n+MPIDI_CH3_Init(105)................: \n+MPID_nem_init(324).................: \n+MPID_nem_tcp_init(178).............: \n+MPID_nem_tcp_get_business_card(425): \n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n at_complete\n 2\n ./rma\n fail\n Fatal error in PMPI_Init: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n-Fatal error in PMPI_Init: Other MPI error, error stack:\n-MPIR_Init_thread(565)..............: \n-MPID_Init(224).....................: channel initialization failed\n-MPIDI_CH3_Init(105)................: \n-MPID_nem_init(324).................: \n-MPID_nem_tcp_init(178).............: \n-MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n atomic_rmw_fop\n 2\n ./rma\n fail\n Fatal error in PMPI_Init: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n-Fatal error in PMPI_Init: Other MPI error, error stack:\n-MPIR_Init_thread(565)..............: \n-MPID_Init(224).....................: channel initialization failed\n-MPIDI_CH3_Init(105)................: \n-MPID_nem_init(324).................: \n-MPID_nem_tcp_init(178).............: \n-MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n atomic_rmw_cas\n 2\n ./rma\n fail\n Fatal error in PMPI_Init: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n atomic_rmw_gacc\n 2\n ./rma\n fail\n Fatal error in PMPI_Init: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n atomic_get\n 2\n ./rma\n fail\n Fatal error in PMPI_Init: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n acc-pairtype\n 2\n ./rma\n fail\n Fatal error in PMPI_Init: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n-Fatal error in PMPI_Init: Other MPI error, error stack:\n-MPIR_Init_thread(565)..............: \n-MPID_Init(224).....................: channel initialization failed\n-MPIDI_CH3_Init(105)................: \n-MPID_nem_init(324).................: \n-MPID_nem_tcp_init(178).............: \n-MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n manyget\n 2\n ./rma\n fail\n Fatal error in PMPI_Init: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n Fatal error in PMPI_Init: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n derived-acc-flush_local\n 2\n ./rma\n fail\n Fatal error in PMPI_Init: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n-Fatal error in PMPI_Init: Other MPI error, error stack:\n-MPIR_Init_thread(565)..............: \n-MPID_Init(224).....................: channel initialization failed\n-MPIDI_CH3_Init(105)................: \n-MPID_nem_init(324).................: \n-MPID_nem_tcp_init(178).............: \n-MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n large-acc-flush_local\n 2\n ./rma\n fail\n Fatal error in PMPI_Init: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n+Fatal error in PMPI_Init: Other MPI error, error stack:\n+MPIR_Init_thread(565)..............: \n+MPID_Init(224).....................: channel initialization failed\n+MPIDI_CH3_Init(105)................: \n+MPID_nem_init(324).................: \n+MPID_nem_tcp_init(178).............: \n+MPID_nem_tcp_get_business_card(425): \n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n large-small-acc\n 2\n ./rma\n fail\n Fatal error in PMPI_Init: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n-Fatal error in PMPI_Init: Other MPI error, error stack:\n-MPIR_Init_thread(565)..............: \n-MPID_Init(224).....................: channel initialization failed\n-MPIDI_CH3_Init(105)................: \n-MPID_nem_init(324).................: \n-MPID_nem_tcp_init(178).............: \n-MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n win_shared_put_flush_load\n 2\n ./rma\n fail\n Fatal error in PMPI_Init: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n-Fatal error in PMPI_Init: Other MPI error, error stack:\n-MPIR_Init_thread(565)..............: \n-MPID_Init(224).....................: channel initialization failed\n-MPIDI_CH3_Init(105)................: \n-MPID_nem_init(324).................: \n-MPID_nem_tcp_init(178).............: \n-MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n win_shared_acc_flush_load\n 2\n ./rma\n fail\n Fatal error in PMPI_Init: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n Fatal error in PMPI_Init: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n win_shared_gacc_flush_load\n 2\n ./rma\n fail\n Fatal error in PMPI_Init: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n+Fatal error in PMPI_Init: Other MPI error, error stack:\n+MPIR_Init_thread(565)..............: \n+MPID_Init(224).....................: channel initialization failed\n+MPIDI_CH3_Init(105)................: \n+MPID_nem_init(324).................: \n+MPID_nem_tcp_init(178).............: \n+MPID_nem_tcp_get_business_card(425): \n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n win_shared_fop_flush_load\n 2\n ./rma\n fail\n Fatal error in PMPI_Init: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n-Fatal error in PMPI_Init: Other MPI error, error stack:\n-MPIR_Init_thread(565)..............: \n-MPID_Init(224).....................: channel initialization failed\n-MPIDI_CH3_Init(105)................: \n-MPID_nem_init(324).................: \n-MPID_nem_tcp_init(178).............: \n-MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n win_shared_cas_flush_load\n 2\n ./rma\n fail\n Fatal error in PMPI_Init: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n put_flush_get\n 2\n ./rma\n fail\n Fatal error in PMPI_Init: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n Fatal error in PMPI_Init: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n acc_flush_get\n 2\n ./rma\n fail\n Fatal error in PMPI_Init: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n-Fatal error in PMPI_Init: Other MPI error, error stack:\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n+ \n+ \n+ gacc_flush_get\n+ 2\n+ ./rma\n+ fail\n+ Fatal error in PMPI_Init: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n- gacc_flush_get\n+ fop_flush_get\n 2\n ./rma\n fail\n Fatal error in PMPI_Init: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n Fatal error in PMPI_Init: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n- fop_flush_get\n+ cas_flush_get\n 2\n ./rma\n fail\n Fatal error in PMPI_Init: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n- \n- \n- cas_flush_get\n- 2\n- ./rma\n- fail\n- Fatal error in PMPI_Init: Other MPI error, error stack:\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n+Fatal error in PMPI_Init: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n namepub\n 2\n ./spawn\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n spawn1\n 1\n ./spawn\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n spawn2\n 1\n ./spawn\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n spawninfo1\n 1\n ./spawn\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n spawnminfo1\n 1\n ./spawn\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n spawnintra\n 1\n ./spawn\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n spawnintra\n 2\n ./spawn\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n spawnargv\n 1\n ./spawn\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n spawnmanyarg\n 1\n ./spawn\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n spawnmult2\n 2\n ./spawn\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n+Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n+MPIR_Init_thread(565)..............: \n+MPID_Init(224).....................: channel initialization failed\n+MPIDI_CH3_Init(105)................: \n+MPID_nem_init(324).................: \n+MPID_nem_tcp_init(178).............: \n+MPID_nem_tcp_get_business_card(425): \n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n spaconacc\n 1\n ./spawn\n fail\n Fatal error in PMPI_Init: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n spaconacc2\n 1\n ./spawn\n fail\n Fatal error in PMPI_Init: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n selfconacc\n 2\n ./spawn\n fail\n Fatal error in PMPI_Init: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n spaiccreate\n 2\n ./spawn\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n-Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n-MPIR_Init_thread(565)..............: \n-MPID_Init(224).....................: channel initialization failed\n-MPIDI_CH3_Init(105)................: \n-MPID_nem_init(324).................: \n-MPID_nem_tcp_init(178).............: \n-MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n taskmaster\n 1\n ./spawn\n fail\n Fatal error in PMPI_Init: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n taskmaster\n 2\n ./spawn\n fail\n Fatal error in PMPI_Init: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n+Fatal error in PMPI_Init: Other MPI error, error stack:\n+MPIR_Init_thread(565)..............: \n+MPID_Init(224).....................: channel initialization failed\n+MPIDI_CH3_Init(105)................: \n+MPID_nem_init(324).................: \n+MPID_nem_tcp_init(178).............: \n+MPID_nem_tcp_get_business_card(425): \n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n join\n 2\n ./spawn\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n disconnect_reconnect\n 2\n ./spawn\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n disconnect_reconnect2\n 2\n ./spawn\n fail\n Fatal error in PMPI_Init: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n+Fatal error in PMPI_Init: Other MPI error, error stack:\n+MPIR_Init_thread(565)..............: \n+MPID_Init(224).....................: channel initialization failed\n+MPIDI_CH3_Init(105)................: \n+MPID_nem_init(324).................: \n+MPID_nem_tcp_init(178).............: \n+MPID_nem_tcp_get_business_card(425): \n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n disconnect_reconnect3\n 2\n ./spawn\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n-Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n-MPIR_Init_thread(565)..............: \n-MPID_Init(224).....................: channel initialization failed\n-MPIDI_CH3_Init(105)................: \n-MPID_nem_init(324).................: \n-MPID_nem_tcp_init(178).............: \n-MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n multiple_ports\n 2\n ./spawn\n fail\n Fatal error in PMPI_Init: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n-Fatal error in PMPI_Init: Other MPI error, error stack:\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n+ \n+ \n+ multiple_ports2\n+ 2\n+ ./spawn\n+ fail\n+ Fatal error in PMPI_Init: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n- multiple_ports2\n+ disconnect\n 2\n ./spawn\n fail\n- Fatal error in PMPI_Init: Other MPI error, error stack:\n+ Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n-Fatal error in PMPI_Init: Other MPI error, error stack:\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n+Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n- disconnect\n+ disconnect2\n 2\n ./spawn\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n- \n- \n- disconnect2\n- 2\n- ./spawn\n- fail\n- Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n+Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n disconnect3\n 2\n ./spawn\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n+Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n+MPIR_Init_thread(565)..............: \n+MPID_Init(224).....................: channel initialization failed\n+MPIDI_CH3_Init(105)................: \n+MPID_nem_init(324).................: \n+MPID_nem_tcp_init(178).............: \n+MPID_nem_tcp_get_business_card(425): \n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n concurrent_spawns\n 1\n ./spawn\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n pgroup_connect_test\n 2\n ./spawn\n fail\n Fatal error in PMPI_Init: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n+Fatal error in PMPI_Init: Other MPI error, error stack:\n+MPIR_Init_thread(565)..............: \n+MPID_Init(224).....................: channel initialization failed\n+MPIDI_CH3_Init(105)................: \n+MPID_nem_init(324).................: \n+MPID_nem_tcp_init(178).............: \n+MPID_nem_tcp_get_business_card(425): \n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n pgroup_intercomm_test\n 2\n ./spawn\n fail\n Fatal error in PMPI_Init: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n+Fatal error in PMPI_Init: Other MPI error, error stack:\n+MPIR_Init_thread(565)..............: \n+MPID_Init(224).....................: channel initialization failed\n+MPIDI_CH3_Init(105)................: \n+MPID_nem_init(324).................: \n+MPID_nem_tcp_init(178).............: \n+MPID_nem_tcp_get_business_card(425): \n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n spawn-rootargs\n 2\n ./spawn\n fail\n Fatal error in PMPI_Init: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n cartmap1\n 2\n ./topo\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n cartzero\n 2\n ./topo\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n cartshift1\n 2\n ./topo\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n cartsuball\n 2\n ./topo\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n-Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n-MPIR_Init_thread(565)..............: \n-MPID_Init(224).....................: channel initialization failed\n-MPIDI_CH3_Init(105)................: \n-MPID_nem_init(324).................: \n-MPID_nem_tcp_init(178).............: \n-MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n cartcreates\n 2\n ./topo\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n-Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n-MPIR_Init_thread(565)..............: \n-MPID_Init(224).....................: channel initialization failed\n-MPIDI_CH3_Init(105)................: \n-MPID_nem_init(324).................: \n-MPID_nem_tcp_init(178).............: \n-MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n dims1\n 2\n ./topo\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n-Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n-MPIR_Init_thread(565)..............: \n-MPID_Init(224).....................: channel initialization failed\n-MPIDI_CH3_Init(105)................: \n-MPID_nem_init(324).................: \n-MPID_nem_tcp_init(178).............: \n-MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n dims2\n 1\n ./topo\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n dims3\n 1\n ./topo\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n dims4\n 1\n ./topo\n fail\n Fatal error in PMPI_Init: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n graphmap1\n 2\n ./topo\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n+Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n+MPIR_Init_thread(565)..............: \n+MPID_Init(224).....................: channel initialization failed\n+MPIDI_CH3_Init(105)................: \n+MPID_nem_init(324).................: \n+MPID_nem_tcp_init(178).............: \n+MPID_nem_tcp_get_business_card(425): \n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n topotest\n 2\n ./topo\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n topodup\n 2\n ./topo\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n graphcr\n 2\n ./topo\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n+Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n+MPIR_Init_thread(565)..............: \n+MPID_Init(224).....................: channel initialization failed\n+MPIDI_CH3_Init(105)................: \n+MPID_nem_init(324).................: \n+MPID_nem_tcp_init(178).............: \n+MPID_nem_tcp_get_business_card(425): \n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n graphcr2\n 2\n ./topo\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n-Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n-MPIR_Init_thread(565)..............: \n-MPID_Init(224).....................: channel initialization failed\n-MPIDI_CH3_Init(105)................: \n-MPID_nem_init(324).................: \n-MPID_nem_tcp_init(178).............: \n-MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n distgraph1\n 2\n ./topo\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n+Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n+MPIR_Init_thread(565)..............: \n+MPID_Init(224).....................: channel initialization failed\n+MPIDI_CH3_Init(105)................: \n+MPID_nem_init(324).................: \n+MPID_nem_tcp_init(178).............: \n+MPID_nem_tcp_get_business_card(425): \n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n dgraph_unwgt\n 2\n ./topo\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n-Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n-MPIR_Init_thread(565)..............: \n-MPID_Init(224).....................: channel initialization failed\n-MPIDI_CH3_Init(105)................: \n-MPID_nem_init(324).................: \n-MPID_nem_tcp_init(178).............: \n-MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n neighb_coll\n 2\n ./topo\n fail\n Fatal error in PMPI_Init: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n rdwrord\n 2\n ./io\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n rdwrzero\n 2\n ./io\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n-Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n-MPIR_Init_thread(565)..............: \n-MPID_Init(224).....................: channel initialization failed\n-MPIDI_CH3_Init(105)................: \n-MPID_nem_init(324).................: \n-MPID_nem_tcp_init(178).............: \n-MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n getextent\n 2\n ./io\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n-Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n-MPIR_Init_thread(565)..............: \n-MPID_Init(224).....................: channel initialization failed\n-MPIDI_CH3_Init(105)................: \n-MPID_nem_init(324).................: \n-MPID_nem_tcp_init(178).............: \n-MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n setinfo\n 2\n ./io\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n-Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n-MPIR_Init_thread(565)..............: \n-MPID_Init(224).....................: channel initialization failed\n-MPIDI_CH3_Init(105)................: \n-MPID_nem_init(324).................: \n-MPID_nem_tcp_init(178).............: \n-MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n setviewcur\n 2\n ./io\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n+Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n+MPIR_Init_thread(565)..............: \n+MPID_Init(224).....................: channel initialization failed\n+MPIDI_CH3_Init(105)................: \n+MPID_nem_init(324).................: \n+MPID_nem_tcp_init(178).............: \n+MPID_nem_tcp_get_business_card(425): \n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n i_noncontig\n 2\n ./io\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n async\n 2\n ./io\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n async_any\n 2\n ./io\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n+Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n+MPIR_Init_thread(565)..............: \n+MPID_Init(224).....................: channel initialization failed\n+MPIDI_CH3_Init(105)................: \n+MPID_nem_init(324).................: \n+MPID_nem_tcp_init(178).............: \n+MPID_nem_tcp_get_business_card(425): \n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n userioerr\n 1\n ./io\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n resized\n 1\n ./io\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n resized2\n 1\n ./io\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n hindexed_io\n 1\n ./io\n fail\n Fatal error in PMPI_Init: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n simple_collective\n 1\n ./io\n fail\n Fatal error in PMPI_Init: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n attrmpi1f\n 1\n ./f77/attr\n fail\n Fatal error in PMPI_Init: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n baseattrf\n 1\n ./f77/attr\n fail\n Fatal error in PMPI_Init: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n baseattr2f\n 1\n ./f77/attr\n fail\n Fatal error in PMPI_Init: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n commattrf\n 1\n ./f77/attr\n fail\n Fatal error in PMPI_Init: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n commattr2f\n 1\n ./f77/attr\n fail\n Fatal error in PMPI_Init: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n commattr3f\n 1\n ./f77/attr\n fail\n Fatal error in PMPI_Init: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n commattr4f\n 1\n ./f77/attr\n fail\n Fatal error in PMPI_Init: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n typeattrf\n 1\n ./f77/attr\n fail\n Fatal error in PMPI_Init: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n typeattr2f\n 1\n ./f77/attr\n fail\n Fatal error in PMPI_Init: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n typeattr3f\n 1\n ./f77/attr\n fail\n Fatal error in PMPI_Init: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n uallreducef\n 2\n ./f77/coll\n fail\n Fatal error in PMPI_Init: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n exscanf\n 2\n ./f77/coll\n fail\n Fatal error in PMPI_Init: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n alltoallwf\n 2\n ./f77/coll\n fail\n Fatal error in PMPI_Init: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n Fatal error in PMPI_Init: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n alltoallvf\n 2\n ./f77/coll\n fail\n Fatal error in PMPI_Init: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n Fatal error in PMPI_Init: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n inplacef\n 2\n ./f77/coll\n fail\n Fatal error in PMPI_Init: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n-Fatal error in PMPI_Init: Other MPI error, error stack:\n-MPIR_Init_thread(565)..............: \n-MPID_Init(224).....................: channel initialization failed\n-MPIDI_CH3_Init(105)................: \n-MPID_nem_init(324).................: \n-MPID_nem_tcp_init(178).............: \n-MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n reducelocalf\n 2\n ./f77/coll\n fail\n Fatal error in PMPI_Init: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n redscatf\n 2\n ./f77/coll\n fail\n Fatal error in PMPI_Init: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n split_typef\n 2\n ./f77/coll\n fail\n Fatal error in PMPI_Init: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n nonblockingf\n 2\n ./f77/coll\n fail\n Fatal error in PMPI_Init: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n+Fatal error in PMPI_Init: Other MPI error, error stack:\n+MPIR_Init_thread(565)..............: \n+MPID_Init(224).....................: channel initialization failed\n+MPIDI_CH3_Init(105)................: \n+MPID_nem_init(324).................: \n+MPID_nem_tcp_init(178).............: \n+MPID_nem_tcp_get_business_card(425): \n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n vw_inplacef\n 2\n ./f77/coll\n fail\n Fatal error in PMPI_Init: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n-Fatal error in PMPI_Init: Other MPI error, error stack:\n-MPIR_Init_thread(565)..............: \n-MPID_Init(224).....................: channel initialization failed\n-MPIDI_CH3_Init(105)................: \n-MPID_nem_init(324).................: \n-MPID_nem_tcp_init(178).............: \n-MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n red_scat_blockf\n 2\n ./f77/coll\n fail\n Fatal error in PMPI_Init: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n-Fatal error in PMPI_Init: Other MPI error, error stack:\n-MPIR_Init_thread(565)..............: \n-MPID_Init(224).....................: channel initialization failed\n-MPIDI_CH3_Init(105)................: \n-MPID_nem_init(324).................: \n-MPID_nem_tcp_init(178).............: \n-MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n nonblocking_inpf\n 2\n ./f77/coll\n fail\n Fatal error in PMPI_Init: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n+Fatal error in PMPI_Init: Other MPI error, error stack:\n+MPIR_Init_thread(565)..............: \n+MPID_Init(224).....................: channel initialization failed\n+MPIDI_CH3_Init(105)................: \n+MPID_nem_init(324).................: \n+MPID_nem_tcp_init(178).............: \n+MPID_nem_tcp_get_business_card(425): \n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n typenamef\n 1\n ./f77/datatype\n fail\n Fatal error in PMPI_Init: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n typename3f\n 1\n ./f77/datatype\n fail\n Fatal error in PMPI_Init: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n typesnamef\n 1\n ./f77/datatype\n fail\n Fatal error in PMPI_Init: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n typecntsf\n 1\n ./f77/datatype\n fail\n Fatal error in PMPI_Init: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n typem2f\n 1\n ./f77/datatype\n fail\n Fatal error in PMPI_Init: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n typesubf\n 1\n ./f77/datatype\n fail\n Fatal error in PMPI_Init: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n packef\n 1\n ./f77/datatype\n fail\n Fatal error in PMPI_Init: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n gaddressf\n 1\n ./f77/datatype\n fail\n Fatal error in PMPI_Init: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n allctypesf\n 1\n ./f77/datatype\n fail\n Fatal error in PMPI_Init: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n hindex1f\n 1\n ./f77/datatype\n fail\n Fatal error in PMPI_Init: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n hindexed_blockf\n 1\n ./f77/datatype\n fail\n Fatal error in PMPI_Init: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n bottom\n 2\n ./f77/datatype\n fail\n Fatal error in PMPI_Init: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n-Fatal error in PMPI_Init: Other MPI error, error stack:\n-MPIR_Init_thread(565)..............: \n-MPID_Init(224).....................: channel initialization failed\n-MPIDI_CH3_Init(105)................: \n-MPID_nem_init(324).................: \n-MPID_nem_tcp_init(178).............: \n-MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n statusesf\n 1\n ./f77/pt2pt\n fail\n Fatal error in PMPI_Init: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n greqf\n 1\n ./f77/pt2pt\n fail\n Fatal error in PMPI_Init: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n allpairf\n 2\n ./f77/pt2pt\n fail\n Fatal error in PMPI_Init: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n-Fatal error in PMPI_Init: Other MPI error, error stack:\n-MPIR_Init_thread(565)..............: \n-MPID_Init(224).....................: channel initialization failed\n-MPIDI_CH3_Init(105)................: \n-MPID_nem_init(324).................: \n-MPID_nem_tcp_init(178).............: \n-MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n mprobef\n 2\n ./f77/pt2pt\n fail\n Fatal error in PMPI_Init: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n infotestf\n 1\n ./f77/info\n fail\n Fatal error in PMPI_Init: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n infotest2f\n 1\n ./f77/info\n fail\n Fatal error in PMPI_Init: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n namepubf\n 2\n ./f77/spawn\n fail\n Fatal error in PMPI_Init: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n+Fatal error in PMPI_Init: Other MPI error, error stack:\n+MPIR_Init_thread(565)..............: \n+MPID_Init(224).....................: channel initialization failed\n+MPIDI_CH3_Init(105)................: \n+MPID_nem_init(324).................: \n+MPID_nem_tcp_init(178).............: \n+MPID_nem_tcp_get_business_card(425): \n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n spawnf\n 1\n ./f77/spawn\n fail\n Fatal error in PMPI_Init: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n spawnargvf\n 1\n ./f77/spawn\n fail\n Fatal error in PMPI_Init: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n connaccf\n 2\n ./f77/spawn\n fail\n Fatal error in PMPI_Init: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n-Fatal error in PMPI_Init: Other MPI error, error stack:\n-MPIR_Init_thread(565)..............: \n-MPID_Init(224).....................: channel initialization failed\n-MPIDI_CH3_Init(105)................: \n-MPID_nem_init(324).................: \n-MPID_nem_tcp_init(178).............: \n-MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n spawnmultf\n 1\n ./f77/spawn\n fail\n Fatal error in PMPI_Init: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n spawnmult2f\n 2\n ./f77/spawn\n fail\n Fatal error in PMPI_Init: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n-Fatal error in PMPI_Init: Other MPI error, error stack:\n-MPIR_Init_thread(565)..............: \n-MPID_Init(224).....................: channel initialization failed\n-MPIDI_CH3_Init(105)................: \n-MPID_nem_init(324).................: \n-MPID_nem_tcp_init(178).............: \n-MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n iwriteatf\n 2\n ./f77/io\n fail\n Fatal error in PMPI_Init: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n-Fatal error in PMPI_Init: Other MPI error, error stack:\n-MPIR_Init_thread(565)..............: \n-MPID_Init(224).....................: channel initialization failed\n-MPIDI_CH3_Init(105)................: \n-MPID_nem_init(324).................: \n-MPID_nem_tcp_init(178).............: \n-MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n iwritef\n 2\n ./f77/io\n fail\n Fatal error in PMPI_Init: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n+Fatal error in PMPI_Init: Other MPI error, error stack:\n+MPIR_Init_thread(565)..............: \n+MPID_Init(224).....................: channel initialization failed\n+MPIDI_CH3_Init(105)................: \n+MPID_nem_init(324).................: \n+MPID_nem_tcp_init(178).............: \n+MPID_nem_tcp_get_business_card(425): \n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n iwriteshf\n 2\n ./f77/io\n fail\n Fatal error in PMPI_Init: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n Fatal error in PMPI_Init: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n writef\n 2\n ./f77/io\n fail\n Fatal error in PMPI_Init: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n writeatf\n 2\n ./f77/io\n fail\n Fatal error in PMPI_Init: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n+Fatal error in PMPI_Init: Other MPI error, error stack:\n+MPIR_Init_thread(565)..............: \n+MPID_Init(224).....................: channel initialization failed\n+MPIDI_CH3_Init(105)................: \n+MPID_nem_init(324).................: \n+MPID_nem_tcp_init(178).............: \n+MPID_nem_tcp_get_business_card(425): \n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n writeallf\n 2\n ./f77/io\n fail\n Fatal error in PMPI_Init: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n writeshf\n 2\n ./f77/io\n fail\n Fatal error in PMPI_Init: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n writeordf\n 2\n ./f77/io\n fail\n Fatal error in PMPI_Init: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n-Fatal error in PMPI_Init: Other MPI error, error stack:\n-MPIR_Init_thread(565)..............: \n-MPID_Init(224).....................: channel initialization failed\n-MPIDI_CH3_Init(105)................: \n-MPID_nem_init(324).................: \n-MPID_nem_tcp_init(178).............: \n-MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n writeatallf\n 2\n ./f77/io\n fail\n Fatal error in PMPI_Init: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n writeatallbef\n 2\n ./f77/io\n fail\n Fatal error in PMPI_Init: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n-Fatal error in PMPI_Init: Other MPI error, error stack:\n-MPIR_Init_thread(565)..............: \n-MPID_Init(224).....................: channel initialization failed\n-MPIDI_CH3_Init(105)................: \n-MPID_nem_init(324).................: \n-MPID_nem_tcp_init(178).............: \n-MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n writeallbef\n 2\n ./f77/io\n fail\n Fatal error in PMPI_Init: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n-Fatal error in PMPI_Init: Other MPI error, error stack:\n-MPIR_Init_thread(565)..............: \n-MPID_Init(224).....................: channel initialization failed\n-MPIDI_CH3_Init(105)................: \n-MPID_nem_init(324).................: \n-MPID_nem_tcp_init(178).............: \n-MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n writeordbef\n 2\n ./f77/io\n fail\n Fatal error in PMPI_Init: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n-Fatal error in PMPI_Init: Other MPI error, error stack:\n-MPIR_Init_thread(565)..............: \n-MPID_Init(224).....................: channel initialization failed\n-MPIDI_CH3_Init(105)................: \n-MPID_nem_init(324).................: \n-MPID_nem_tcp_init(178).............: \n-MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n fileerrf\n 1\n ./f77/io\n fail\n Fatal error in PMPI_Init: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n fileinfof\n 2\n ./f77/io\n fail\n Fatal error in PMPI_Init: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n-Fatal error in PMPI_Init: Other MPI error, error stack:\n-MPIR_Init_thread(565)..............: \n-MPID_Init(224).....................: channel initialization failed\n-MPIDI_CH3_Init(105)................: \n-MPID_nem_init(324).................: \n-MPID_nem_tcp_init(178).............: \n-MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n shpositionf\n 2\n ./f77/io\n fail\n Fatal error in PMPI_Init: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n-Fatal error in PMPI_Init: Other MPI error, error stack:\n-MPIR_Init_thread(565)..............: \n-MPID_Init(224).....................: channel initialization failed\n-MPIDI_CH3_Init(105)................: \n-MPID_nem_init(324).................: \n-MPID_nem_tcp_init(178).............: \n-MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n atomicityf\n 2\n ./f77/io\n fail\n Fatal error in PMPI_Init: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n-Fatal error in PMPI_Init: Other MPI error, error stack:\n-MPIR_Init_thread(565)..............: \n-MPID_Init(224).....................: channel initialization failed\n-MPIDI_CH3_Init(105)................: \n-MPID_nem_init(324).................: \n-MPID_nem_tcp_init(178).............: \n-MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n miscfilef\n 2\n ./f77/io\n fail\n Fatal error in PMPI_Init: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n-Fatal error in PMPI_Init: Other MPI error, error stack:\n-MPIR_Init_thread(565)..............: \n-MPID_Init(224).....................: channel initialization failed\n-MPIDI_CH3_Init(105)................: \n-MPID_nem_init(324).................: \n-MPID_nem_tcp_init(178).............: \n-MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n setviewcurf\n 2\n ./f77/io\n fail\n Fatal error in PMPI_Init: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n c2f2ciof\n 1\n ./f77/io\n fail\n Fatal error in PMPI_Init: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n c2fmultio\n 1\n ./f77/io\n fail\n Fatal error in PMPI_Init: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n winscale1f\n 2\n ./f77/rma\n fail\n Fatal error in PMPI_Init: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n-Fatal error in PMPI_Init: Other MPI error, error stack:\n-MPIR_Init_thread(565)..............: \n-MPID_Init(224).....................: channel initialization failed\n-MPIDI_CH3_Init(105)................: \n-MPID_nem_init(324).................: \n-MPID_nem_tcp_init(178).............: \n-MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n winfencef\n 2\n ./f77/rma\n fail\n Fatal error in PMPI_Init: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n-Fatal error in PMPI_Init: Other MPI error, error stack:\n-MPIR_Init_thread(565)..............: \n-MPID_Init(224).....................: channel initialization failed\n-MPIDI_CH3_Init(105)................: \n-MPID_nem_init(324).................: \n-MPID_nem_tcp_init(178).............: \n-MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n wingetf\n 2\n ./f77/rma\n fail\n Fatal error in PMPI_Init: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n-Fatal error in PMPI_Init: Other MPI error, error stack:\n-MPIR_Init_thread(565)..............: \n-MPID_Init(224).....................: channel initialization failed\n-MPIDI_CH3_Init(105)................: \n-MPID_nem_init(324).................: \n-MPID_nem_tcp_init(178).............: \n-MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n winscale2f\n 2\n ./f77/rma\n fail\n Fatal error in PMPI_Init: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n winerrf\n 1\n ./f77/rma\n fail\n Fatal error in PMPI_Init: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n winnamef\n 1\n ./f77/rma\n fail\n Fatal error in PMPI_Init: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n wingroupf\n 2\n ./f77/rma\n fail\n Fatal error in PMPI_Init: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n-Fatal error in PMPI_Init: Other MPI error, error stack:\n-MPIR_Init_thread(565)..............: \n-MPID_Init(224).....................: channel initialization failed\n-MPIDI_CH3_Init(105)................: \n-MPID_nem_init(324).................: \n-MPID_nem_tcp_init(178).............: \n-MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n winaccf\n 2\n ./f77/rma\n fail\n Fatal error in PMPI_Init: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n c2f2cwinf\n 1\n ./f77/rma\n fail\n Fatal error in PMPI_Init: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n baseattrwinf\n 1\n ./f77/rma\n fail\n Fatal error in PMPI_Init: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n winattrf\n 1\n ./f77/rma\n fail\n Fatal error in PMPI_Init: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n winattr2f\n 1\n ./f77/rma\n fail\n Fatal error in PMPI_Init: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n baseenvf\n 1\n ./f77/init\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n commnamef\n 2\n ./f77/comm\n fail\n Fatal error in PMPI_Init: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n+Fatal error in PMPI_Init: Other MPI error, error stack:\n+MPIR_Init_thread(565)..............: \n+MPID_Init(224).....................: channel initialization failed\n+MPIDI_CH3_Init(105)................: \n+MPID_nem_init(324).................: \n+MPID_nem_tcp_init(178).............: \n+MPID_nem_tcp_get_business_card(425): \n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n commerrf\n 2\n ./f77/comm\n fail\n Fatal error in PMPI_Init: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n-Fatal error in PMPI_Init: Other MPI error, error stack:\n-MPIR_Init_thread(565)..............: \n-MPID_Init(224).....................: channel initialization failed\n-MPIDI_CH3_Init(105)................: \n-MPID_nem_init(324).................: \n-MPID_nem_tcp_init(178).............: \n-MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n c2f2cf\n 1\n ./f77/ext\n fail\n Fatal error in PMPI_Init: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n c2fmult\n 1\n ./f77/ext\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n ctypesinf\n 1\n ./f77/ext\n fail\n Fatal error in PMPI_Init: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n allocmemf\n 1\n ./f77/ext\n fail\n Fatal error in PMPI_Init: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n cartcrf\n 2\n ./f77/topo\n fail\n Fatal error in PMPI_Init: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n-Fatal error in PMPI_Init: Other MPI error, error stack:\n-MPIR_Init_thread(565)..............: \n-MPID_Init(224).....................: channel initialization failed\n-MPIDI_CH3_Init(105)................: \n-MPID_nem_init(324).................: \n-MPID_nem_tcp_init(178).............: \n-MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n dgraph_wgtf\n 2\n ./f77/topo\n fail\n Fatal error in PMPI_Init: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n-Fatal error in PMPI_Init: Other MPI error, error stack:\n-MPIR_Init_thread(565)..............: \n-MPID_Init(224).....................: channel initialization failed\n-MPIDI_CH3_Init(105)................: \n-MPID_nem_init(324).................: \n-MPID_nem_tcp_init(178).............: \n-MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n dgraph_unwgtf\n 2\n ./f77/topo\n fail\n Fatal error in PMPI_Init: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n-Fatal error in PMPI_Init: Other MPI error, error stack:\n-MPIR_Init_thread(565)..............: \n-MPID_Init(224).....................: channel initialization failed\n-MPIDI_CH3_Init(105)................: \n-MPID_nem_init(324).................: \n-MPID_nem_tcp_init(178).............: \n-MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n profile1f\n 2\n ./f77/profile\n fail\n Fatal error in PMPI_Init: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n-Fatal error in PMPI_Init: Other MPI error, error stack:\n-MPIR_Init_thread(565)..............: \n-MPID_Init(224).....................: channel initialization failed\n-MPIDI_CH3_Init(105)................: \n-MPID_nem_init(324).................: \n-MPID_nem_tcp_init(178).............: \n-MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n attrtx\n 2\n ./cxx/attr\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n+Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n+MPIR_Init_thread(565)..............: \n+MPID_Init(224).....................: channel initialization failed\n+MPIDI_CH3_Init(105)................: \n+MPID_nem_init(324).................: \n+MPID_nem_tcp_init(178).............: \n+MPID_nem_tcp_get_business_card(425): \n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n attricx\n 2\n ./cxx/attr\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n-Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n-MPIR_Init_thread(565)..............: \n-MPID_Init(224).....................: channel initialization failed\n-MPIDI_CH3_Init(105)................: \n-MPID_nem_init(324).................: \n-MPID_nem_tcp_init(178).............: \n-MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n baseattrcommx\n 1\n ./cxx/attr\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n fkeyvalcommx\n 1\n ./cxx/attr\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n fkeyvaltypex\n 1\n ./cxx/attr\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n bsend1cxx\n 2\n ./cxx/pt2pt\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n-Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n-MPIR_Init_thread(565)..............: \n-MPID_Init(224).....................: channel initialization failed\n-MPIDI_CH3_Init(105)................: \n-MPID_nem_init(324).................: \n-MPID_nem_tcp_init(178).............: \n-MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n sendrecvx\n 2\n ./cxx/pt2pt\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n+Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n+MPIR_Init_thread(565)..............: \n+MPID_Init(224).....................: channel initialization failed\n+MPIDI_CH3_Init(105)................: \n+MPID_nem_init(324).................: \n+MPID_nem_tcp_init(178).............: \n+MPID_nem_tcp_get_business_card(425): \n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n commname2\n 2\n ./cxx/comm\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n arcomplex\n 2\n ./cxx/coll\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n-Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n-MPIR_Init_thread(565)..............: \n-MPID_Init(224).....................: channel initialization failed\n-MPIDI_CH3_Init(105)................: \n-MPID_nem_init(324).................: \n-MPID_nem_tcp_init(178).............: \n-MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n uallredx\n 2\n ./cxx/coll\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n-Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n-MPIR_Init_thread(565)..............: \n-MPID_Init(224).....................: channel initialization failed\n-MPIDI_CH3_Init(105)................: \n-MPID_nem_init(324).................: \n-MPID_nem_tcp_init(178).............: \n-MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n uallreduce\n 2\n ./cxx/coll\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n ureduce\n 2\n ./cxx/coll\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n-Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n-MPIR_Init_thread(565)..............: \n-MPID_Init(224).....................: channel initialization failed\n-MPIDI_CH3_Init(105)................: \n-MPID_nem_init(324).................: \n-MPID_nem_tcp_init(178).............: \n-MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n ureducelocal\n 2\n ./cxx/coll\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n uscan\n 2\n ./cxx/coll\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n uexscan\n 2\n ./cxx/coll\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n-Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n-MPIR_Init_thread(565)..............: \n-MPID_Init(224).....................: channel initialization failed\n-MPIDI_CH3_Init(105)................: \n-MPID_nem_init(324).................: \n-MPID_nem_tcp_init(178).............: \n-MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n alltoallw2x\n 2\n ./cxx/coll\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n icbcastx\n 2\n ./cxx/coll\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n-Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n-MPIR_Init_thread(565)..............: \n-MPID_Init(224).....................: channel initialization failed\n-MPIDI_CH3_Init(105)................: \n-MPID_nem_init(324).................: \n-MPID_nem_tcp_init(178).............: \n-MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n icbcastx\n 2\n ./cxx/coll\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n-Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n-MPIR_Init_thread(565)..............: \n-MPID_Init(224).....................: channel initialization failed\n-MPIDI_CH3_Init(105)................: \n-MPID_nem_init(324).................: \n-MPID_nem_tcp_init(178).............: \n-MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n icallreducex\n 2\n ./cxx/coll\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n icreducex\n 2\n ./cxx/coll\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n-Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n-MPIR_Init_thread(565)..............: \n-MPID_Init(224).....................: channel initialization failed\n-MPIDI_CH3_Init(105)................: \n-MPID_nem_init(324).................: \n-MPID_nem_tcp_init(178).............: \n-MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n icscatterx\n 2\n ./cxx/coll\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n icgatherx\n 2\n ./cxx/coll\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n icallgatherx\n 2\n ./cxx/coll\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n-Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n-MPIR_Init_thread(565)..............: \n-MPID_Init(224).....................: channel initialization failed\n-MPIDI_CH3_Init(105)................: \n-MPID_nem_init(324).................: \n-MPID_nem_tcp_init(178).............: \n-MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n icbarrierx\n 2\n ./cxx/coll\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n icallgathervx\n 2\n ./cxx/coll\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n icgathervx\n 2\n ./cxx/coll\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n+Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n+MPIR_Init_thread(565)..............: \n+MPID_Init(224).....................: channel initialization failed\n+MPIDI_CH3_Init(105)................: \n+MPID_nem_init(324).................: \n+MPID_nem_tcp_init(178).............: \n+MPID_nem_tcp_get_business_card(425): \n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n icscattervx\n 2\n ./cxx/coll\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n icalltoallx\n 2\n ./cxx/coll\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n+Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n+MPIR_Init_thread(565)..............: \n+MPID_Init(224).....................: channel initialization failed\n+MPIDI_CH3_Init(105)................: \n+MPID_nem_init(324).................: \n+MPID_nem_tcp_init(178).............: \n+MPID_nem_tcp_get_business_card(425): \n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n reduceboolx\n 2\n ./cxx/coll\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n+Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n+MPIR_Init_thread(565)..............: \n+MPID_Init(224).....................: channel initialization failed\n+MPIDI_CH3_Init(105)................: \n+MPID_nem_init(324).................: \n+MPID_nem_tcp_init(178).............: \n+MPID_nem_tcp_get_business_card(425): \n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n redscatblk\n 2\n ./cxx/coll\n fail\n Fatal error in PMPI_Init: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n-Fatal error in PMPI_Init: Other MPI error, error stack:\n-MPIR_Init_thread(565)..............: \n-MPID_Init(224).....................: channel initialization failed\n-MPIDI_CH3_Init(105)................: \n-MPID_nem_init(324).................: \n-MPID_nem_tcp_init(178).............: \n-MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n commcallx\n 2\n ./cxx/errhan\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n baseenv\n 1\n ./cxx/init\n fail\n Fatal error in PMPI_Init: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n initstatx\n 1\n ./cxx/init\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n initstat2x\n 1\n ./cxx/init\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n infodupx\n 1\n ./cxx/info\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n infodelx\n 1\n ./cxx/info\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n infovallenx\n 1\n ./cxx/info\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n infoorderx\n 1\n ./cxx/info\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n packsizex\n 1\n ./cxx/datatype\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n typecntsx\n 1\n ./cxx/datatype\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n typenamex\n 1\n ./cxx/datatype\n fail\n Fatal error in PMPI_Init: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n typemiscx\n 1\n ./cxx/datatype\n fail\n Fatal error in PMPI_Init: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n iwriteatx\n 2\n ./cxx/io\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n-Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n-MPIR_Init_thread(565)..............: \n-MPID_Init(224).....................: channel initialization failed\n-MPIDI_CH3_Init(105)................: \n-MPID_nem_init(324).................: \n-MPID_nem_tcp_init(178).............: \n-MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n iwritex\n 2\n ./cxx/io\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n-Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n-MPIR_Init_thread(565)..............: \n-MPID_Init(224).....................: channel initialization failed\n-MPIDI_CH3_Init(105)................: \n-MPID_nem_init(324).................: \n-MPID_nem_tcp_init(178).............: \n-MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n iwriteshx\n 2\n ./cxx/io\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n writex\n 2\n ./cxx/io\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n-Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n-MPIR_Init_thread(565)..............: \n-MPID_Init(224).....................: channel initialization failed\n-MPIDI_CH3_Init(105)................: \n-MPID_nem_init(324).................: \n-MPID_nem_tcp_init(178).............: \n-MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n writeatx\n 2\n ./cxx/io\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n writeallx\n 2\n ./cxx/io\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n-Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n-MPIR_Init_thread(565)..............: \n-MPID_Init(224).....................: channel initialization failed\n-MPIDI_CH3_Init(105)................: \n-MPID_nem_init(324).................: \n-MPID_nem_tcp_init(178).............: \n-MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n writeshx\n 2\n ./cxx/io\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n writeordx\n 2\n ./cxx/io\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n writeatallx\n 2\n ./cxx/io\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n writeatallbex\n 2\n ./cxx/io\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n writeallbex\n 2\n ./cxx/io\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n- \n- \n- writeordbex\n- 2\n- ./cxx/io\n- fail\n- Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n+Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n- iwriteatnosx\n+ writeordbex\n 2\n ./cxx/io\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n-Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n-MPIR_Init_thread(565)..............: \n-MPID_Init(224).....................: channel initialization failed\n-MPIDI_CH3_Init(105)................: \n-MPID_nem_init(324).................: \n-MPID_nem_tcp_init(178).............: \n-MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n- iwritenosx\n+ iwriteatnosx\n 2\n ./cxx/io\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n- iwriteshnosx\n+ iwritenosx\n 2\n ./cxx/io\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n- writenosx\n+ iwriteshnosx\n 2\n ./cxx/io\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n- writeatnosx\n+ writenosx\n 2\n ./cxx/io\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n- writeallnosx\n+ writeatnosx\n 2\n ./cxx/io\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n-Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n+ \n+ \n+ writeallnosx\n+ 2\n+ ./cxx/io\n+ fail\n+ Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n writeshnosx\n 2\n ./cxx/io\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n writeordnosx\n 2\n ./cxx/io\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n writeatallnosx\n 2\n ./cxx/io\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n writeatallbenosx\n 2\n ./cxx/io\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n writeallbenosx\n 2\n ./cxx/io\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n writeordbenosx\n 2\n ./cxx/io\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n+Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n+MPIR_Init_thread(565)..............: \n+MPID_Init(224).....................: channel initialization failed\n+MPIDI_CH3_Init(105)................: \n+MPID_nem_init(324).................: \n+MPID_nem_tcp_init(178).............: \n+MPID_nem_tcp_get_business_card(425): \n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n fileerrx\n 1\n ./cxx/io\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n fileinfox\n 2\n ./cxx/io\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n filemiscx\n 2\n ./cxx/io\n fail\n Fatal error in PMPI_Init: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n-Fatal error in PMPI_Init: Other MPI error, error stack:\n-MPIR_Init_thread(565)..............: \n-MPID_Init(224).....................: channel initialization failed\n-MPIDI_CH3_Init(105)................: \n-MPID_nem_init(324).................: \n-MPID_nem_tcp_init(178).............: \n-MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n shpositionx\n 2\n ./cxx/io\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n seekavail\n 1\n ./cxx/io\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n namepubx\n 2\n ./cxx/spawn\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n spawnintrax\n 1\n ./cxx/spawn\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n spawnintrax\n 2\n ./cxx/spawn\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n-Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n-MPIR_Init_thread(565)..............: \n-MPID_Init(224).....................: channel initialization failed\n-MPIDI_CH3_Init(105)................: \n-MPID_nem_init(324).................: \n-MPID_nem_tcp_init(178).............: \n-MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n spawnargvx\n 1\n ./cxx/spawn\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n selfconaccx\n 2\n ./cxx/spawn\n fail\n Fatal error in PMPI_Init: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n-Fatal error in PMPI_Init: Other MPI error, error stack:\n-MPIR_Init_thread(565)..............: \n-MPID_Init(224).....................: channel initialization failed\n-MPIDI_CH3_Init(105)................: \n-MPID_nem_init(324).................: \n-MPID_nem_tcp_init(178).............: \n-MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n winnamex\n 1\n ./cxx/rma\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n wincallx\n 1\n ./cxx/rma\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n getgroupx\n 2\n ./cxx/rma\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n-Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n-MPIR_Init_thread(565)..............: \n-MPID_Init(224).....................: channel initialization failed\n-MPIDI_CH3_Init(105)................: \n-MPID_nem_init(324).................: \n-MPID_nem_tcp_init(178).............: \n-MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n winfencex\n 2\n ./cxx/rma\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n-Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n-MPIR_Init_thread(565)..............: \n-MPID_Init(224).....................: channel initialization failed\n-MPIDI_CH3_Init(105)................: \n-MPID_nem_init(324).................: \n-MPID_nem_tcp_init(178).............: \n-MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n winscale1x\n 2\n ./cxx/rma\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n winscale2x\n 2\n ./cxx/rma\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n-Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n-MPIR_Init_thread(565)..............: \n-MPID_Init(224).....................: channel initialization failed\n-MPIDI_CH3_Init(105)................: \n-MPID_nem_init(324).................: \n-MPID_nem_tcp_init(178).............: \n-MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n fkeyvalwinx\n 1\n ./cxx/rma\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n attrmpi1f90\n 1\n ./f90/attr\n fail\n Fatal error in PMPI_Init: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n baseattrf90\n 1\n ./f90/attr\n fail\n Fatal error in PMPI_Init: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n baseattr2f90\n 1\n ./f90/attr\n fail\n Fatal error in PMPI_Init: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n commattrf90\n 1\n ./f90/attr\n fail\n Fatal error in PMPI_Init: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n commattr2f90\n 1\n ./f90/attr\n fail\n Fatal error in PMPI_Init: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n commattr3f90\n 1\n ./f90/attr\n fail\n Fatal error in PMPI_Init: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n commattr4f90\n 1\n ./f90/attr\n fail\n Fatal error in PMPI_Init: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n typeattrf90\n 1\n ./f90/attr\n fail\n Fatal error in PMPI_Init: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n typeattr2f90\n 1\n ./f90/attr\n fail\n Fatal error in PMPI_Init: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n typeattr3f90\n 1\n ./f90/attr\n fail\n Fatal error in PMPI_Init: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n fandcattrf90\n 1\n ./f90/attr\n fail\n Fatal error in PMPI_Init: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n baseattr3f90\n 1\n ./f90/attr\n fail\n Fatal error in PMPI_Init: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n attrlangf90\n 1\n ./f90/attr\n fail\n Fatal error in PMPI_Init: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n uallreducef90\n 2\n ./f90/coll\n fail\n Fatal error in PMPI_Init: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n-Fatal error in PMPI_Init: Other MPI error, error stack:\n-MPIR_Init_thread(565)..............: \n-MPID_Init(224).....................: channel initialization failed\n-MPIDI_CH3_Init(105)................: \n-MPID_nem_init(324).................: \n-MPID_nem_tcp_init(178).............: \n-MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n exscanf90\n 2\n ./f90/coll\n fail\n Fatal error in PMPI_Init: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n-[mpiexec@profitbricks-build5-amd64] control_cb (pm/pmiserv/pmiserv_cb.c:208): assert (!closed) failed\n-[mpiexec@profitbricks-build5-amd64] HYDT_dmxu_poll_wait_for_event (tools/demux/demux_poll.c:77): callback returned error status\n-[mpiexec@profitbricks-build5-amd64] HYDT_bscu_wait_for_completion (tools/bootstrap/utils/bscu_wait.c:60): error waiting for event\n-[mpiexec@profitbricks-build5-amd64] HYDT_bsci_wait_for_completion (tools/bootstrap/src/bsci_wait.c:22): launcher returned error waiting for completion\n- \n- \n- alltoallwf90\n- 2\n- ./f90/coll\n- fail\n- Fatal error in PMPI_Init: Other MPI error, error stack:\n-MPIR_Init_thread(565)..............: \n-MPID_Init(224).....................: channel initialization failed\n-MPIDI_CH3_Init(105)................: \n-MPID_nem_init(324).................: \n-MPID_nem_tcp_init(178).............: \n-MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n Fatal error in PMPI_Init: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n- alltoallvf90\n+ alltoallwf90\n 2\n ./f90/coll\n fail\n Fatal error in PMPI_Init: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n- inplacef90\n+ alltoallvf90\n 2\n ./f90/coll\n fail\n Fatal error in PMPI_Init: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n Fatal error in PMPI_Init: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n- reducelocalf90\n+ inplacef90\n 2\n ./f90/coll\n fail\n Fatal error in PMPI_Init: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n-Fatal error in PMPI_Init: Other MPI error, error stack:\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n+ \n+ \n+ reducelocalf90\n+ 2\n+ ./f90/coll\n+ fail\n+ Fatal error in PMPI_Init: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n redscatf90\n 2\n ./f90/coll\n fail\n Fatal error in PMPI_Init: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n split_typef90\n 2\n ./f90/coll\n fail\n Fatal error in PMPI_Init: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n-Fatal error in PMPI_Init: Other MPI error, error stack:\n-MPIR_Init_thread(565)..............: \n-MPID_Init(224).....................: channel initialization failed\n-MPIDI_CH3_Init(105)................: \n-MPID_nem_init(324).................: \n-MPID_nem_tcp_init(178).............: \n-MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n nonblockingf90\n 2\n ./f90/coll\n fail\n Fatal error in PMPI_Init: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n+Fatal error in PMPI_Init: Other MPI error, error stack:\n+MPIR_Init_thread(565)..............: \n+MPID_Init(224).....................: channel initialization failed\n+MPIDI_CH3_Init(105)................: \n+MPID_nem_init(324).................: \n+MPID_nem_tcp_init(178).............: \n+MPID_nem_tcp_get_business_card(425): \n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n vw_inplacef90\n 2\n ./f90/coll\n fail\n Fatal error in PMPI_Init: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n-Fatal error in PMPI_Init: Other MPI error, error stack:\n-MPIR_Init_thread(565)..............: \n-MPID_Init(224).....................: channel initialization failed\n-MPIDI_CH3_Init(105)................: \n-MPID_nem_init(324).................: \n-MPID_nem_tcp_init(178).............: \n-MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n red_scat_blockf90\n 2\n ./f90/coll\n fail\n Fatal error in PMPI_Init: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n Fatal error in PMPI_Init: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n nonblocking_inpf90\n 2\n ./f90/coll\n fail\n Fatal error in PMPI_Init: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n commnamef90\n 2\n ./f90/comm\n fail\n Fatal error in PMPI_Init: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n Fatal error in PMPI_Init: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n commerrf90\n 2\n ./f90/comm\n fail\n Fatal error in PMPI_Init: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n-Fatal error in PMPI_Init: Other MPI error, error stack:\n-MPIR_Init_thread(565)..............: \n-MPID_Init(224).....................: channel initialization failed\n-MPIDI_CH3_Init(105)................: \n-MPID_nem_init(324).................: \n-MPID_nem_tcp_init(178).............: \n-MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n c2f2cf90\n 1\n ./f90/ext\n fail\n Fatal error in PMPI_Init: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n c2f90mult\n 1\n ./f90/ext\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n ctypesinf90\n 1\n ./f90/ext\n fail\n Fatal error in PMPI_Init: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n allocmemf90\n 1\n ./f90/ext\n fail\n Fatal error in PMPI_Init: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n infotestf90\n 1\n ./f90/info\n fail\n Fatal error in PMPI_Init: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n infotest2f90\n 1\n ./f90/info\n fail\n Fatal error in PMPI_Init: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n baseenvf90\n 1\n ./f90/init\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n iwriteatf90\n 2\n ./f90/io\n fail\n Fatal error in PMPI_Init: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n+Fatal error in PMPI_Init: Other MPI error, error stack:\n+MPIR_Init_thread(565)..............: \n+MPID_Init(224).....................: channel initialization failed\n+MPIDI_CH3_Init(105)................: \n+MPID_nem_init(324).................: \n+MPID_nem_tcp_init(178).............: \n+MPID_nem_tcp_get_business_card(425): \n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n iwritef90\n 2\n ./f90/io\n fail\n Fatal error in PMPI_Init: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n Fatal error in PMPI_Init: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n iwriteshf90\n 2\n ./f90/io\n fail\n Fatal error in PMPI_Init: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n writef90\n 2\n ./f90/io\n fail\n Fatal error in PMPI_Init: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n-Fatal error in PMPI_Init: Other MPI error, error stack:\n-MPIR_Init_thread(565)..............: \n-MPID_Init(224).....................: channel initialization failed\n-MPIDI_CH3_Init(105)................: \n-MPID_nem_init(324).................: \n-MPID_nem_tcp_init(178).............: \n-MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n writeatf90\n 2\n ./f90/io\n fail\n Fatal error in PMPI_Init: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n+Fatal error in PMPI_Init: Other MPI error, error stack:\n+MPIR_Init_thread(565)..............: \n+MPID_Init(224).....................: channel initialization failed\n+MPIDI_CH3_Init(105)................: \n+MPID_nem_init(324).................: \n+MPID_nem_tcp_init(178).............: \n+MPID_nem_tcp_get_business_card(425): \n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n writeallf90\n 2\n ./f90/io\n fail\n Fatal error in PMPI_Init: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n writeshf90\n 2\n ./f90/io\n fail\n Fatal error in PMPI_Init: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n writeordf90\n 2\n ./f90/io\n fail\n Fatal error in PMPI_Init: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n writeatallf90\n 2\n ./f90/io\n fail\n Fatal error in PMPI_Init: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n writeatallbef90\n 2\n ./f90/io\n fail\n Fatal error in PMPI_Init: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n-Fatal error in PMPI_Init: Other MPI error, error stack:\n-MPIR_Init_thread(565)..............: \n-MPID_Init(224).....................: channel initialization failed\n-MPIDI_CH3_Init(105)................: \n-MPID_nem_init(324).................: \n-MPID_nem_tcp_init(178).............: \n-MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n writeallbef90\n 2\n ./f90/io\n fail\n Fatal error in PMPI_Init: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n writeordbef90\n 2\n ./f90/io\n fail\n Fatal error in PMPI_Init: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n+Fatal error in PMPI_Init: Other MPI error, error stack:\n+MPIR_Init_thread(565)..............: \n+MPID_Init(224).....................: channel initialization failed\n+MPIDI_CH3_Init(105)................: \n+MPID_nem_init(324).................: \n+MPID_nem_tcp_init(178).............: \n+MPID_nem_tcp_get_business_card(425): \n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n fileerrf90\n 1\n ./f90/io\n fail\n Fatal error in PMPI_Init: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n fileinfof90\n 2\n ./f90/io\n fail\n Fatal error in PMPI_Init: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n shpositionf90\n 2\n ./f90/io\n fail\n Fatal error in PMPI_Init: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n atomicityf90\n 2\n ./f90/io\n fail\n Fatal error in PMPI_Init: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n Fatal error in PMPI_Init: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n miscfilef90\n 2\n ./f90/io\n fail\n Fatal error in PMPI_Init: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n-Fatal error in PMPI_Init: Other MPI error, error stack:\n-MPIR_Init_thread(565)..............: \n-MPID_Init(224).....................: channel initialization failed\n-MPIDI_CH3_Init(105)................: \n-MPID_nem_init(324).................: \n-MPID_nem_tcp_init(178).............: \n-MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n setviewcurf90\n 2\n ./f90/io\n fail\n Fatal error in PMPI_Init: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n c2f2ciof90\n 1\n ./f90/io\n fail\n Fatal error in PMPI_Init: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n c2f90multio\n 1\n ./f90/io\n fail\n Fatal error in PMPI_Init: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n sizeof2\n 1\n ./f90/misc\n fail\n Fatal error in PMPI_Init: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n statusesf90\n 1\n ./f90/pt2pt\n fail\n Fatal error in PMPI_Init: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n greqf90\n 1\n ./f90/pt2pt\n fail\n Fatal error in PMPI_Init: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n allpairf90\n 2\n ./f90/pt2pt\n fail\n Fatal error in PMPI_Init: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n mprobef90\n 2\n ./f90/pt2pt\n fail\n Fatal error in PMPI_Init: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n-Fatal error in PMPI_Init: Other MPI error, error stack:\n-MPIR_Init_thread(565)..............: \n-MPID_Init(224).....................: channel initialization failed\n-MPIDI_CH3_Init(105)................: \n-MPID_nem_init(324).................: \n-MPID_nem_tcp_init(178).............: \n-MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n typenamef90\n 1\n ./f90/datatype\n fail\n Fatal error in PMPI_Init: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n typename3f90\n 1\n ./f90/datatype\n fail\n Fatal error in PMPI_Init: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n typesnamef90\n 1\n ./f90/datatype\n fail\n Fatal error in PMPI_Init: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n typecntsf90\n 1\n ./f90/datatype\n fail\n Fatal error in PMPI_Init: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n typem2f90\n 1\n ./f90/datatype\n fail\n Fatal error in PMPI_Init: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n typesubf90\n 1\n ./f90/datatype\n fail\n Fatal error in PMPI_Init: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n packef90\n 1\n ./f90/datatype\n fail\n Fatal error in PMPI_Init: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n gaddressf90\n 1\n ./f90/datatype\n fail\n Fatal error in PMPI_Init: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n allctypesf90\n 1\n ./f90/datatype\n fail\n Fatal error in PMPI_Init: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n hindex1f90\n 1\n ./f90/datatype\n fail\n Fatal error in PMPI_Init: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n hindexed_blockf90\n 1\n ./f90/datatype\n fail\n Fatal error in PMPI_Init: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n bottom\n 2\n ./f90/datatype\n fail\n Fatal error in PMPI_Init: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n structf\n 2\n ./f90/datatype\n fail\n Fatal error in PMPI_Init: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n indtype\n 2\n ./f90/datatype\n fail\n Fatal error in PMPI_Init: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n createf90\n 1\n ./f90/datatype\n fail\n Fatal error in PMPI_Init: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n sizeof\n 1\n ./f90/datatype\n fail\n Fatal error in PMPI_Init: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n kinds\n 2\n ./f90/datatype\n fail\n Fatal error in PMPI_Init: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n trf90\n 1\n ./f90/datatype\n fail\n Fatal error in PMPI_Init: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n get_elem_d\n 2\n ./f90/datatype\n fail\n Fatal error in PMPI_Init: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n get_elem_u\n 2\n ./f90/datatype\n fail\n Fatal error in PMPI_Init: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n createf90types\n 1\n ./f90/f90types\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n createf90types\n 1\n ./f90/f90types\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n winscale1f90\n 2\n ./f90/rma\n fail\n Fatal error in PMPI_Init: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n-Fatal error in PMPI_Init: Other MPI error, error stack:\n-MPIR_Init_thread(565)..............: \n-MPID_Init(224).....................: channel initialization failed\n-MPIDI_CH3_Init(105)................: \n-MPID_nem_init(324).................: \n-MPID_nem_tcp_init(178).............: \n-MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n winfencef90\n 2\n ./f90/rma\n fail\n Fatal error in PMPI_Init: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n wingetf90\n 2\n ./f90/rma\n fail\n Fatal error in PMPI_Init: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n winscale2f90\n 2\n ./f90/rma\n fail\n Fatal error in PMPI_Init: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n Fatal error in PMPI_Init: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n winerrf90\n 1\n ./f90/rma\n fail\n Fatal error in PMPI_Init: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n winnamef90\n 1\n ./f90/rma\n fail\n Fatal error in PMPI_Init: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n wingroupf90\n 2\n ./f90/rma\n fail\n Fatal error in PMPI_Init: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n winaccf90\n 2\n ./f90/rma\n fail\n Fatal error in PMPI_Init: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n c2f2cwinf90\n 1\n ./f90/rma\n fail\n Fatal error in PMPI_Init: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n baseattrwinf90\n 1\n ./f90/rma\n fail\n Fatal error in PMPI_Init: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n winattrf90\n 1\n ./f90/rma\n fail\n Fatal error in PMPI_Init: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n winattr2f90\n 1\n ./f90/rma\n fail\n Fatal error in PMPI_Init: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n namepubf90\n 2\n ./f90/spawn\n fail\n Fatal error in PMPI_Init: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n spawnf90\n 1\n ./f90/spawn\n fail\n Fatal error in PMPI_Init: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n spawnargvf90\n 1\n ./f90/spawn\n fail\n Fatal error in PMPI_Init: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n connaccf90\n 2\n ./f90/spawn\n fail\n Fatal error in PMPI_Init: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n Fatal error in PMPI_Init: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n spawnmultf90\n 1\n ./f90/spawn\n fail\n Fatal error in PMPI_Init: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n spawnmult2f90\n 2\n ./f90/spawn\n fail\n Fatal error in PMPI_Init: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n-Fatal error in PMPI_Init: Other MPI error, error stack:\n-MPIR_Init_thread(565)..............: \n-MPID_Init(224).....................: channel initialization failed\n-MPIDI_CH3_Init(105)................: \n-MPID_nem_init(324).................: \n-MPID_nem_tcp_init(178).............: \n-MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n spawnargvf03\n 1\n ./f90/spawn\n fail\n Fatal error in PMPI_Init: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n spawnmultf03\n 1\n ./f90/spawn\n fail\n Fatal error in PMPI_Init: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n wtimef90\n 1\n ./f90/timer\n fail\n Fatal error in PMPI_Init: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n cartcrf90\n 2\n ./f90/topo\n fail\n Fatal error in PMPI_Init: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n-Fatal error in PMPI_Init: Other MPI error, error stack:\n-MPIR_Init_thread(565)..............: \n-MPID_Init(224).....................: channel initialization failed\n-MPIDI_CH3_Init(105)................: \n-MPID_nem_init(324).................: \n-MPID_nem_tcp_init(178).............: \n-MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n dgraph_wgtf90\n 2\n ./f90/topo\n fail\n Fatal error in PMPI_Init: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n dgraph_unwgtf90\n 2\n ./f90/topo\n fail\n Fatal error in PMPI_Init: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n-Fatal error in PMPI_Init: Other MPI error, error stack:\n-MPIR_Init_thread(565)..............: \n-MPID_Init(224).....................: channel initialization failed\n-MPIDI_CH3_Init(105)................: \n-MPID_nem_init(324).................: \n-MPID_nem_tcp_init(178).............: \n-MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n profile1f90\n 2\n ./f90/profile\n fail\n Fatal error in PMPI_Init: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n keyvalmis\n 1\n ./errors/attr\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n noalias\n 2\n ./errors/coll\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n nb_noalias\n 2\n ./errors/coll\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n noalias2\n 2\n ./errors/coll\n fail\n Fatal error in PMPI_Init: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n-Fatal error in PMPI_Init: Other MPI error, error stack:\n-MPIR_Init_thread(565)..............: \n-MPID_Init(224).....................: channel initialization failed\n-MPIDI_CH3_Init(105)................: \n-MPID_nem_init(324).................: \n-MPID_nem_tcp_init(178).............: \n-MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n noalias3\n 2\n ./errors/coll\n fail\n Fatal error in PMPI_Init: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n rerr\n 2\n ./errors/coll\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n nb_rerr\n 2\n ./errors/coll\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n+Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n+MPIR_Init_thread(565)..............: \n+MPID_Init(224).....................: channel initialization failed\n+MPIDI_CH3_Init(105)................: \n+MPID_nem_init(324).................: \n+MPID_nem_tcp_init(178).............: \n+MPID_nem_tcp_get_business_card(425): \n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n reduce_local\n 1\n ./errors/coll\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n bcastlength\n 2\n ./errors/coll\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n-Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n-MPIR_Init_thread(565)..............: \n-MPID_Init(224).....................: channel initialization failed\n-MPIDI_CH3_Init(105)................: \n-MPID_nem_init(324).................: \n-MPID_nem_tcp_init(178).............: \n-MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n ibcastlength\n 2\n ./errors/coll\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n+Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n+MPIR_Init_thread(565)..............: \n+MPID_Init(224).....................: channel initialization failed\n+MPIDI_CH3_Init(105)................: \n+MPID_nem_init(324).................: \n+MPID_nem_tcp_init(178).............: \n+MPID_nem_tcp_get_business_card(425): \n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n cfree\n 2\n ./errors/comm\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n ccreate1\n 2\n ./errors/comm\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n userdup\n 2\n ./errors/comm\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n manysplit\n 2\n ./errors/comm\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n+Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n+MPIR_Init_thread(565)..............: \n+MPID_Init(224).....................: channel initialization failed\n+MPIDI_CH3_Init(105)................: \n+MPID_nem_init(324).................: \n+MPID_nem_tcp_init(178).............: \n+MPID_nem_tcp_get_business_card(425): \n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n gerr\n 1\n ./errors/group\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n proberank\n 1\n ./errors/pt2pt\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n truncmsg1\n 2\n ./errors/pt2pt\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n truncmsg2\n 2\n ./errors/pt2pt\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n errinstatts\n 2\n ./errors/pt2pt\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n errinstatta\n 2\n ./errors/pt2pt\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n-Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n-MPIR_Init_thread(565)..............: \n-MPID_Init(224).....................: channel initialization failed\n-MPIDI_CH3_Init(105)................: \n-MPID_nem_init(324).................: \n-MPID_nem_tcp_init(178).............: \n-MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n errinstatws\n 2\n ./errors/pt2pt\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n errinstatwa\n 2\n ./errors/pt2pt\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n-Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n-MPIR_Init_thread(565)..............: \n-MPID_Init(224).....................: channel initialization failed\n-MPIDI_CH3_Init(105)................: \n-MPID_nem_init(324).................: \n-MPID_nem_tcp_init(178).............: \n-MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n cartsmall\n 2\n ./errors/topo\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n winerr\n 2\n ./errors/rma\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n+Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n+MPIR_Init_thread(565)..............: \n+MPID_Init(224).....................: channel initialization failed\n+MPIDI_CH3_Init(105)................: \n+MPID_nem_init(324).................: \n+MPID_nem_tcp_init(178).............: \n+MPID_nem_tcp_get_business_card(425): \n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n winerr2\n 2\n ./errors/rma\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n cas_type_check\n 2\n ./errors/rma\n fail\n Fatal error in PMPI_Init: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n-Fatal error in PMPI_Init: Other MPI error, error stack:\n-MPIR_Init_thread(565)..............: \n-MPID_Init(224).....................: channel initialization failed\n-MPIDI_CH3_Init(105)................: \n-MPID_nem_init(324).................: \n-MPID_nem_tcp_init(178).............: \n-MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n win_sync_unlock\n 2\n ./errors/rma\n fail\n Fatal error in PMPI_Init: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n-Fatal error in PMPI_Init: Other MPI error, error stack:\n-MPIR_Init_thread(565)..............: \n-MPID_Init(224).....................: channel initialization failed\n-MPIDI_CH3_Init(105)................: \n-MPID_nem_init(324).................: \n-MPID_nem_tcp_init(178).............: \n-MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n win_sync_free_pt\n 2\n ./errors/rma\n fail\n Fatal error in PMPI_Init: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n win_sync_free_at\n 2\n ./errors/rma\n fail\n Fatal error in PMPI_Init: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n Fatal error in PMPI_Init: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n win_sync_complete\n 2\n ./errors/rma\n fail\n Fatal error in PMPI_Init: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n win_sync_lock_at\n 2\n ./errors/rma\n fail\n Fatal error in PMPI_Init: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n win_sync_lock_pt\n 2\n ./errors/rma\n fail\n Fatal error in PMPI_Init: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n-Fatal error in PMPI_Init: Other MPI error, error stack:\n-MPIR_Init_thread(565)..............: \n-MPID_Init(224).....................: channel initialization failed\n-MPIDI_CH3_Init(105)................: \n-MPID_nem_init(324).................: \n-MPID_nem_tcp_init(178).............: \n-MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n win_sync_lock_fence\n 2\n ./errors/rma\n fail\n Fatal error in PMPI_Init: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n win_sync_nested\n 2\n ./errors/rma\n fail\n Fatal error in PMPI_Init: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n Fatal error in PMPI_Init: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n win_sync_op\n 2\n ./errors/rma\n fail\n Fatal error in PMPI_Init: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n badport\n 2\n ./errors/spawn\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n-Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n-MPIR_Init_thread(565)..............: \n-MPID_Init(224).....................: channel initialization failed\n-MPIDI_CH3_Init(105)................: \n-MPID_nem_init(324).................: \n-MPID_nem_tcp_init(178).............: \n-MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n unpub\n 1\n ./errors/spawn\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n lookup_name\n 1\n ./errors/spawn\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n fileerrret\n 1\n ./errors/io\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n openerr\n 1\n ./errors/io\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n file_errhdl\n 1\n ./errors/io\n fail\n Fatal error in PMPI_Init: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n uerrhandf\n 1\n ./errors/f77/io\n fail\n Fatal error in PMPI_Init: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n errgetx\n 1\n ./errors/cxx/errhan\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n errsetx\n 1\n ./errors/cxx/errhan\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n throwtest\n 1\n ./errors/cxx/errhan\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n commerrx\n 2\n ./errors/cxx/errhan\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n fileerrretx\n 1\n ./errors/cxx/io\n fail\n Fatal error in PMPI_Init: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n errgetfilex\n 1\n ./errors/cxx/io\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n throwtestfilex\n 1\n ./errors/cxx/io\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n uerrhandf90\n 1\n ./errors/f90/io\n fail\n Fatal error in PMPI_Init: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n threads\n 2\n ./threads/pt2pt\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n+Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n+MPIR_Init_thread(565)..............: \n+MPID_Init(224).....................: channel initialization failed\n+MPIDI_CH3_Init(105)................: \n+MPID_nem_init(324).................: \n+MPID_nem_tcp_init(178).............: \n+MPID_nem_tcp_get_business_card(425): \n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n threaded_sr\n 2\n ./threads/pt2pt\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n alltoall\n 2\n ./threads/pt2pt\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n-Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n-MPIR_Init_thread(565)..............: \n-MPID_Init(224).....................: channel initialization failed\n-MPIDI_CH3_Init(105)................: \n-MPID_nem_init(324).................: \n-MPID_nem_tcp_init(178).............: \n-MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n sendselfth\n 1\n ./threads/pt2pt\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n multisend\n 2\n ./threads/pt2pt\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n multisend2\n 2\n ./threads/pt2pt\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n+Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n+MPIR_Init_thread(565)..............: \n+MPID_Init(224).....................: channel initialization failed\n+MPIDI_CH3_Init(105)................: \n+MPID_nem_init(324).................: \n+MPID_nem_tcp_init(178).............: \n+MPID_nem_tcp_get_business_card(425): \n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n multisend3\n 2\n ./threads/pt2pt\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n multisend4\n 2\n ./threads/pt2pt\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n-Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n-MPIR_Init_thread(565)..............: \n-MPID_Init(224).....................: channel initialization failed\n-MPIDI_CH3_Init(105)................: \n-MPID_nem_init(324).................: \n-MPID_nem_tcp_init(178).............: \n-MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n greq_wait\n 1\n ./threads/pt2pt\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n greq_test\n 1\n ./threads/pt2pt\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n ibsend\n 2\n ./threads/pt2pt\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n+Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n+MPIR_Init_thread(565)..............: \n+MPID_Init(224).....................: channel initialization failed\n+MPIDI_CH3_Init(105)................: \n+MPID_nem_init(324).................: \n+MPID_nem_tcp_init(178).............: \n+MPID_nem_tcp_get_business_card(425): \n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n ctxdup\n 2\n ./threads/comm\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n dup_leak_test\n 2\n ./threads/comm\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n comm_dup_deadlock\n 2\n ./threads/comm\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n-Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n-MPIR_Init_thread(565)..............: \n-MPID_Init(224).....................: channel initialization failed\n-MPIDI_CH3_Init(105)................: \n-MPID_nem_init(324).................: \n-MPID_nem_tcp_init(178).............: \n-MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n comm_create_threads\n 2\n ./threads/comm\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n comm_create_group_threads\n 2\n ./threads/comm\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n comm_create_group_threads2\n 2\n ./threads/comm\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n idup_deadlock\n 2\n ./threads/comm\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n+Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n+MPIR_Init_thread(565)..............: \n+MPID_Init(224).....................: channel initialization failed\n+MPIDI_CH3_Init(105)................: \n+MPID_nem_init(324).................: \n+MPID_nem_tcp_init(178).............: \n+MPID_nem_tcp_get_business_card(425): \n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n comm_idup\n 2\n ./threads/comm\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n ctxidup\n 2\n ./threads/comm\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n-Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n-MPIR_Init_thread(565)..............: \n-MPID_Init(224).....................: channel initialization failed\n-MPIDI_CH3_Init(105)................: \n-MPID_nem_init(324).................: \n-MPID_nem_tcp_init(178).............: \n-MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n idup_nb\n 2\n ./threads/comm\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n idup_comm_gen\n 2\n ./threads/comm\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n+Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n+MPIR_Init_thread(565)..............: \n+MPID_Init(224).....................: channel initialization failed\n+MPIDI_CH3_Init(105)................: \n+MPID_nem_init(324).................: \n+MPID_nem_tcp_init(178).............: \n+MPID_nem_tcp_get_business_card(425): \n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n initth\n 1\n ./threads/init\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n initth\n 2\n ./threads/init\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n mpit_threading\n 1\n ./threads/mpi_t\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n multispawn\n 1\n ./threads/spawn\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n th_taskmaster\n 1\n ./threads/spawn\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n th_taskmaster\n 2\n ./threads/spawn\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n+Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n+MPIR_Init_thread(565)..............: \n+MPID_Init(224).....................: channel initialization failed\n+MPIDI_CH3_Init(105)................: \n+MPID_nem_init(324).................: \n+MPID_nem_tcp_init(178).............: \n+MPID_nem_tcp_get_business_card(425): \n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n multirma\n 2\n ./threads/rma\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n multiget\n 2\n ./threads/rma\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n allred\n 2\n ./threads/coll\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n iallred\n 2\n ./threads/coll\n fail\n Fatal error in PMPI_Init_thread: Other MPI error, error stack:\n MPIR_Init_thread(565)..............: \n MPID_Init(224).....................: channel initialization failed\n MPIDI_CH3_Init(105)................: \n MPID_nem_init(324).................: \n MPID_nem_tcp_init(178).............: \n MPID_nem_tcp_get_business_card(425): \n-MPID_nem_tcp_init(384).............: gethostbyname failed, profitbricks-build5-amd64 (errno 2)\n+MPID_nem_tcp_init(384).............: gethostbyname failed, i-capture-the-hostname (errno 2)\n \n \n" } ] }, { "source1": "./usr/share/mpi-testsuite/results/openmpi/summary.junit.xml", "source2": "./usr/share/mpi-testsuite/results/openmpi/summary.junit.xml", "unified_diff": null, "details": [ { "source1": "./usr/share/mpi-testsuite/results/openmpi/summary.junit.xml", "source2": "./usr/share/mpi-testsuite/results/openmpi/summary.junit.xml", "unified_diff": "@@ -1,10 +1,10 @@\n \n \n- \n+ \n \n \n \n \n \n \n \n@@ -23,77 +23,77 @@\n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n@@ -125,331 +125,331 @@\n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n@@ -1386,126 +1386,126 @@\n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n@@ -1544,21 +1544,21 @@\n \n \n \n \n \n \n \n \n \n@@ -1571,15 +1571,15 @@\n \n \n \n \n \n \n@@ -1596,32 +1596,32 @@\n \n \n \n \n \n \n \n \n \n \n \n@@ -1630,15 +1630,15 @@\n \n \n \n \n \n \n@@ -1656,15 +1656,15 @@\n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n@@ -1809,15 +1809,15 @@\n \n \n \n \n \n \n \n \n \n \n \n \n \n message="requires MPI version 3.1">\n@@ -1919,15 +1919,15 @@\n \n \n \n \n \n \n \n \n \n \n \n \n@@ -2012,29 +2012,29 @@\n \n \n \n \n \n \n \n \n \n@@ -2054,72 +2054,76 @@\n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n@@ -2409,15 +2430,15 @@\n \n \n \n \n \n \n \n \n \n \n \n@@ -2517,15 +2538,15 @@\n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n@@ -2605,15 +2626,15 @@\n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n Returned \"Unreachable\" (-12) instead of \"Success\" (0)\n ## --------------------------------------------------------------------------\n-## [profitbricks-build5-amd64:32597] [[47742,2],1] ORTE_ERROR_LOG: Unreachable in file dpm/dpm.c at line 433\n-## [profitbricks-build5-amd64:32597] *** An error occurred in MPI_Init_thread\n-## [profitbricks-build5-amd64:32597] *** reported by process [3128819714,1]\n-## [profitbricks-build5-amd64:32597] *** on a NULL communicator\n-## [profitbricks-build5-amd64:32597] *** Unknown error\n-## [profitbricks-build5-amd64:32597] *** MPI_ERRORS_ARE_FATAL (processes in this communicator will now abort,\n-## [profitbricks-build5-amd64:32597] *** and potentially your MPI job)\n-## [profitbricks-build5-amd64:32594] [[47742,2],0] ORTE_ERROR_LOG: Unreachable in file dpm/dpm.c at line 433\n-## [profitbricks-build5-amd64:32477] [[47742,1],0] ORTE_ERROR_LOG: Unreachable in file dpm/dpm.c at line 433\n-## [profitbricks-build5-amd64:32477] *** An error occurred in MPI_Comm_spawn\n-## [profitbricks-build5-amd64:32477] *** reported by process [3128819713,0]\n-## [profitbricks-build5-amd64:32477] *** on communicator MPI_COMM_WORLD\n-## [profitbricks-build5-amd64:32477] *** MPI_ERR_INTERN: internal error\n-## [profitbricks-build5-amd64:32477] *** MPI_ERRORS_ARE_FATAL (processes in this communicator will now abort,\n-## [profitbricks-build5-amd64:32477] *** and potentially your MPI job)\n-## [profitbricks-build5-amd64:32428] 2 more processes have sent help message help-mca-bml-r2.txt / unreachable proc\n-## [profitbricks-build5-amd64:32428] Set MCA parameter \"orte_base_help_aggregate\" to 0 to see all help / error messages\n-## [profitbricks-build5-amd64:32428] 1 more process has sent help message help-mpi-runtime.txt / mpi_init:startup:internal-failure\n-## [profitbricks-build5-amd64:32428] 1 more process has sent help message help-mpi-errors.txt / mpi_errors_are_fatal unknown handle\n+## [i-capture-the-hostname:46135] *** An error occurred in MPI_Init_thread\n+## [i-capture-the-hostname:46135] *** reported by process [961019906,0]\n+## [i-capture-the-hostname:46135] *** on a NULL communicator\n+## [i-capture-the-hostname:46135] *** Unknown error\n+## [i-capture-the-hostname:46135] *** MPI_ERRORS_ARE_FATAL (processes in this communicator will now abort,\n+## [i-capture-the-hostname:46135] *** and potentially your MPI job)\n+## [i-capture-the-hostname:46105] 2 more processes have sent help message help-mca-bml-r2.txt / unreachable proc\n+## [i-capture-the-hostname:46105] Set MCA parameter \"orte_base_help_aggregate\" to 0 to see all help / error messages\n+## [i-capture-the-hostname:46105] 1 more process has sent help message help-mpi-runtime.txt / mpi_init:startup:internal-failure\n+## [i-capture-the-hostname:46105] 1 more process has sent help message help-mpi-errors.txt / mpi_errors_are_fatal unknown handle\n ]]> \n \n \n \n Returned \"Unreachable\" (-12) instead of \"Success\" (0)\n ## --------------------------------------------------------------------------\n-## [profitbricks-build5-amd64:33075] [[17614,2],1] ORTE_ERROR_LOG: Unreachable in file dpm/dpm.c at line 433\n-## [profitbricks-build5-amd64:33069] [[17614,2],0] ORTE_ERROR_LOG: Unreachable in file dpm/dpm.c at line 433\n-## [profitbricks-build5-amd64:32796] 2 more processes have sent help message help-mca-bml-r2.txt / unreachable proc\n-## [profitbricks-build5-amd64:32796] Set MCA parameter \"orte_base_help_aggregate\" to 0 to see all help / error messages\n-## [profitbricks-build5-amd64:32796] 1 more process has sent help message help-mpi-runtime.txt / mpi_init:startup:internal-failure\n+## [i-capture-the-hostname:46183] *** An error occurred in MPI_Init_thread\n+## [i-capture-the-hostname:46183] *** reported by process [958267394,0]\n+## [i-capture-the-hostname:46183] *** on a NULL communicator\n+## [i-capture-the-hostname:46183] *** Unknown error\n+## [i-capture-the-hostname:46183] *** MPI_ERRORS_ARE_FATAL (processes in this communicator will now abort,\n+## [i-capture-the-hostname:46183] *** and potentially your MPI job)\n+## [i-capture-the-hostname:46164] [[14622,1],0] ORTE_ERROR_LOG: Unreachable in file dpm/dpm.c at line 433\n+## [i-capture-the-hostname:46159] 2 more processes have sent help message help-mca-bml-r2.txt / unreachable proc\n+## [i-capture-the-hostname:46159] Set MCA parameter \"orte_base_help_aggregate\" to 0 to see all help / error messages\n ]]> \n \n \n \n \n \n \n \n \n \n \n \n Returned \"Unreachable\" (-12) instead of \"Success\" (0)\n ## --------------------------------------------------------------------------\n-## [profitbricks-build5-amd64:34436] [[17123,1],0] ORTE_ERROR_LOG: Unreachable in file dpm/dpm.c at line 433\n-## [profitbricks-build5-amd64:34586] [[17123,2],0] ORTE_ERROR_LOG: Unreachable in file dpm/dpm.c at line 433\n-## [profitbricks-build5-amd64:34588] [[17123,2],1] ORTE_ERROR_LOG: Unreachable in file dpm/dpm.c at line 433\n-## [profitbricks-build5-amd64:34586] *** An error occurred in MPI_Init_thread\n-## [profitbricks-build5-amd64:34586] *** reported by process [1122172930,0]\n-## [profitbricks-build5-amd64:34586] *** on a NULL communicator\n-## [profitbricks-build5-amd64:34586] *** Unknown error\n-## [profitbricks-build5-amd64:34586] *** MPI_ERRORS_ARE_FATAL (processes in this communicator will now abort,\n-## [profitbricks-build5-amd64:34586] *** and potentially your MPI job)\n-## [profitbricks-build5-amd64:34353] 2 more processes have sent help message help-mca-bml-r2.txt / unreachable proc\n-## [profitbricks-build5-amd64:34353] Set MCA parameter \"orte_base_help_aggregate\" to 0 to see all help / error messages\n-## [profitbricks-build5-amd64:34353] 1 more process has sent help message help-mpi-runtime.txt / mpi_init:startup:internal-failure\n-## [profitbricks-build5-amd64:34353] 1 more process has sent help message help-mpi-errors.txt / mpi_errors_are_fatal unknown handle\n+## [i-capture-the-hostname:46344] *** An error occurred in MPI_Init_thread\n+## [i-capture-the-hostname:46344] *** reported by process [968556546,1]\n+## [i-capture-the-hostname:46344] *** on a NULL communicator\n+## [i-capture-the-hostname:46344] *** Unknown error\n+## [i-capture-the-hostname:46344] *** MPI_ERRORS_ARE_FATAL (processes in this communicator will now abort,\n+## [i-capture-the-hostname:46344] *** and potentially your MPI job)\n+## [i-capture-the-hostname:46314] 2 more processes have sent help message help-mca-bml-r2.txt / unreachable proc\n+## [i-capture-the-hostname:46314] Set MCA parameter \"orte_base_help_aggregate\" to 0 to see all help / error messages\n+## [i-capture-the-hostname:46314] 1 more process has sent help message help-mpi-runtime.txt / mpi_init:startup:internal-failure\n+## [i-capture-the-hostname:46314] 1 more process has sent help message help-mpi-errors.txt / mpi_errors_are_fatal unknown handle\n ]]> \n \n \n \n Returned \"Unreachable\" (-12) instead of \"Success\" (0)\n ## --------------------------------------------------------------------------\n-## [profitbricks-build5-amd64:34974] [[17183,2],1] ORTE_ERROR_LOG: Unreachable in file dpm/dpm.c at line 433\n-## [profitbricks-build5-amd64:34850] [[17183,1],1] ORTE_ERROR_LOG: Unreachable in file dpm/dpm.c at line 433\n-## [profitbricks-build5-amd64:34849] [[17183,1],0] ORTE_ERROR_LOG: Unreachable in file dpm/dpm.c at line 433\n-## [profitbricks-build5-amd64:34966] [[17183,2],0] ORTE_ERROR_LOG: Unreachable in file dpm/dpm.c at line 433\n-## [profitbricks-build5-amd64:34974] *** An error occurred in MPI_Init_thread\n-## [profitbricks-build5-amd64:34974] *** reported by process [1126105090,1]\n-## [profitbricks-build5-amd64:34974] *** on a NULL communicator\n-## [profitbricks-build5-amd64:34974] *** Unknown error\n-## [profitbricks-build5-amd64:34974] *** MPI_ERRORS_ARE_FATAL (processes in this communicator will now abort,\n-## [profitbricks-build5-amd64:34974] *** and potentially your MPI job)\n-## [profitbricks-build5-amd64:34765] 3 more processes have sent help message help-mca-bml-r2.txt / unreachable proc\n-## [profitbricks-build5-amd64:34765] Set MCA parameter \"orte_base_help_aggregate\" to 0 to see all help / error messages\n-## [profitbricks-build5-amd64:34765] 1 more process has sent help message help-mpi-errors.txt / mpi_errors_are_fatal\n-## [profitbricks-build5-amd64:34765] 1 more process has sent help message help-mpi-runtime.txt / mpi_init:startup:internal-failure\n-## [profitbricks-build5-amd64:34765] 1 more process has sent help message help-mpi-errors.txt / mpi_errors_are_fatal unknown handle\n+## [i-capture-the-hostname:46398] [[14412,2],0] ORTE_ERROR_LOG: Unreachable in file dpm/dpm.c at line 433\n+## [i-capture-the-hostname:46399] *** An error occurred in MPI_Init_thread\n+## [i-capture-the-hostname:46399] *** reported by process [944504834,1]\n+## [i-capture-the-hostname:46399] *** on a NULL communicator\n+## [i-capture-the-hostname:46399] *** Unknown error\n+## [i-capture-the-hostname:46399] *** MPI_ERRORS_ARE_FATAL (processes in this communicator will now abort,\n+## [i-capture-the-hostname:46399] *** and potentially your MPI job)\n+## [i-capture-the-hostname:46365] 3 more processes have sent help message help-mca-bml-r2.txt / unreachable proc\n+## [i-capture-the-hostname:46365] Set MCA parameter \"orte_base_help_aggregate\" to 0 to see all help / error messages\n+## [i-capture-the-hostname:46365] 1 more process has sent help message help-mpi-errors.txt / mpi_errors_are_fatal\n+## [i-capture-the-hostname:46365] 1 more process has sent help message help-mpi-runtime.txt / mpi_init:startup:internal-failure\n+## [i-capture-the-hostname:46365] 1 more process has sent help message help-mpi-errors.txt / mpi_errors_are_fatal unknown handle\n ]]> \n \n \n \n Returned \"Unreachable\" (-12) instead of \"Success\" (0)\n ## --------------------------------------------------------------------------\n-## [profitbricks-build5-amd64:35244] [[19493,2],0] ORTE_ERROR_LOG: Unreachable in file dpm/dpm.c at line 433\n-## [profitbricks-build5-amd64:35244] *** An error occurred in MPI_Init_thread\n-## [profitbricks-build5-amd64:35244] *** reported by process [1277493250,0]\n-## [profitbricks-build5-amd64:35244] *** on a NULL communicator\n-## [profitbricks-build5-amd64:35244] *** Unknown error\n-## [profitbricks-build5-amd64:35244] *** MPI_ERRORS_ARE_FATAL (processes in this communicator will now abort,\n-## [profitbricks-build5-amd64:35244] *** and potentially your MPI job)\n-## [profitbricks-build5-amd64:35095] [[19493,1],0] ORTE_ERROR_LOG: Unreachable in file dpm/dpm.c at line 433\n-## [profitbricks-build5-amd64:35095] *** An error occurred in MPI_Comm_spawn\n-## [profitbricks-build5-amd64:35095] *** reported by process [1277493249,0]\n-## [profitbricks-build5-amd64:35095] *** on communicator MPI_COMM_WORLD\n-## [profitbricks-build5-amd64:35095] *** MPI_ERR_INTERN: internal error\n-## [profitbricks-build5-amd64:35095] *** MPI_ERRORS_ARE_FATAL (processes in this communicator will now abort,\n-## [profitbricks-build5-amd64:35095] *** and potentially your MPI job)\n-## [profitbricks-build5-amd64:35246] [[19493,2],1] ORTE_ERROR_LOG: Unreachable in file dpm/dpm.c at line 433\n-## [profitbricks-build5-amd64:35063] 2 more processes have sent help message help-mca-bml-r2.txt / unreachable proc\n-## [profitbricks-build5-amd64:35063] Set MCA parameter \"orte_base_help_aggregate\" to 0 to see all help / error messages\n-## [profitbricks-build5-amd64:35063] 1 more process has sent help message help-mpi-runtime.txt / mpi_init:startup:internal-failure\n+## [i-capture-the-hostname:46444] *** An error occurred in MPI_Init_thread\n+## [i-capture-the-hostname:46444] *** reported by process [939655170,0]\n+## [i-capture-the-hostname:46444] *** on a NULL communicator\n+## [i-capture-the-hostname:46444] *** Unknown error\n+## [i-capture-the-hostname:46444] *** MPI_ERRORS_ARE_FATAL (processes in this communicator will now abort,\n+## [i-capture-the-hostname:46444] *** and potentially your MPI job)\n+## [i-capture-the-hostname:46419] 2 more processes have sent help message help-mca-bml-r2.txt / unreachable proc\n+## [i-capture-the-hostname:46419] Set MCA parameter \"orte_base_help_aggregate\" to 0 to see all help / error messages\n+## [i-capture-the-hostname:46419] 1 more process has sent help message help-mpi-runtime.txt / mpi_init:startup:internal-failure\n+## [i-capture-the-hostname:46419] 1 more process has sent help message help-mpi-errors.txt / mpi_errors_are_fatal unknown handle\n ]]> \n \n \n \n Returned \"Unreachable\" (-12) instead of \"Success\" (0)\n ## --------------------------------------------------------------------------\n-## [profitbricks-build5-amd64:35871] [[20005,2],1] ORTE_ERROR_LOG: Unreachable in file dpm/dpm.c at line 433\n-## [profitbricks-build5-amd64:35866] [[20005,2],0] ORTE_ERROR_LOG: Unreachable in file dpm/dpm.c at line 433\n-## [profitbricks-build5-amd64:35871] *** An error occurred in MPI_Init_thread\n-## [profitbricks-build5-amd64:35871] *** reported by process [1311047682,1]\n-## [profitbricks-build5-amd64:35871] *** on a NULL communicator\n-## [profitbricks-build5-amd64:35871] *** Unknown error\n-## [profitbricks-build5-amd64:35871] *** MPI_ERRORS_ARE_FATAL (processes in this communicator will now abort,\n-## [profitbricks-build5-amd64:35871] *** and potentially your MPI job)\n-## [profitbricks-build5-amd64:35672] [[20005,1],0] ORTE_ERROR_LOG: Unreachable in file dpm/dpm.c at line 433\n-## [profitbricks-build5-amd64:35672] *** An error occurred in MPI_Comm_spawn\n-## [profitbricks-build5-amd64:35672] *** reported by process [1311047681,0]\n-## [profitbricks-build5-amd64:35672] *** on communicator MPI_COMM_WORLD\n-## [profitbricks-build5-amd64:35672] *** MPI_ERR_INTERN: internal error\n-## [profitbricks-build5-amd64:35672] *** MPI_ERRORS_ARE_FATAL (processes in this communicator will now abort,\n-## [profitbricks-build5-amd64:35672] *** and potentially your MPI job)\n-## [profitbricks-build5-amd64:35575] 2 more processes have sent help message help-mca-bml-r2.txt / unreachable proc\n-## [profitbricks-build5-amd64:35575] Set MCA parameter \"orte_base_help_aggregate\" to 0 to see all help / error messages\n-## [profitbricks-build5-amd64:35575] 1 more process has sent help message help-mpi-runtime.txt / mpi_init:startup:internal-failure\n-## [profitbricks-build5-amd64:35575] 1 more process has sent help message help-mpi-errors.txt / mpi_errors_are_fatal unknown handle\n+## [i-capture-the-hostname:46575] *** An error occurred in MPI_Init_thread\n+## [i-capture-the-hostname:46575] *** reported by process [952238082,1]\n+## [i-capture-the-hostname:46575] *** on a NULL communicator\n+## [i-capture-the-hostname:46575] *** Unknown error\n+## [i-capture-the-hostname:46575] *** MPI_ERRORS_ARE_FATAL (processes in this communicator will now abort,\n+## [i-capture-the-hostname:46575] *** and potentially your MPI job)\n+## [i-capture-the-hostname:46483] 2 more processes have sent help message help-mca-bml-r2.txt / unreachable proc\n+## [i-capture-the-hostname:46483] Set MCA parameter \"orte_base_help_aggregate\" to 0 to see all help / error messages\n+## [i-capture-the-hostname:46483] 1 more process has sent help message help-mpi-runtime.txt / mpi_init:startup:internal-failure\n+## [i-capture-the-hostname:46483] 1 more process has sent help message help-mpi-errors.txt / mpi_errors_are_fatal unknown handle\n ]]> \n \n \n \n Returned \"Unreachable\" (-12) instead of \"Success\" (0)\n-## --------------------------------------------------------------------------\n-## [profitbricks-build5-amd64:36278] *** An error occurred in MPI_Comm_spawn_multiple\n-## [profitbricks-build5-amd64:36278] *** reported by process [1240268801,0]\n-## [profitbricks-build5-amd64:36278] *** on communicator MPI_COMM_WORLD\n-## [profitbricks-build5-amd64:36278] *** MPI_ERR_INTERN: internal error\n-## [profitbricks-build5-amd64:36278] *** MPI_ERRORS_ARE_FATAL (processes in this communicator will now abort,\n-## [profitbricks-build5-amd64:36278] *** and potentially your MPI job)\n-## [profitbricks-build5-amd64:36283] [[18925,1],1] ORTE_ERROR_LOG: Unreachable in file dpm/dpm.c at line 433\n-## [profitbricks-build5-amd64:36365] [[18925,2],0] ORTE_ERROR_LOG: Unreachable in file dpm/dpm.c at line 433\n-## [profitbricks-build5-amd64:36366] *** An error occurred in MPI_Init_thread\n-## [profitbricks-build5-amd64:36366] *** reported by process [1240268802,1]\n-## [profitbricks-build5-amd64:36366] *** on a NULL communicator\n-## [profitbricks-build5-amd64:36366] *** Unknown error\n-## [profitbricks-build5-amd64:36366] *** MPI_ERRORS_ARE_FATAL (processes in this communicator will now abort,\n-## [profitbricks-build5-amd64:36366] *** and potentially your MPI job)\n-## [profitbricks-build5-amd64:36159] 3 more processes have sent help message help-mca-bml-r2.txt / unreachable proc\n-## [profitbricks-build5-amd64:36159] Set MCA parameter \"orte_base_help_aggregate\" to 0 to see all help / error messages\n-## [profitbricks-build5-amd64:36159] 1 more process has sent help message help-mpi-runtime.txt / mpi_init:startup:internal-failure\n-## [profitbricks-build5-amd64:36159] 1 more process has sent help message help-mpi-errors.txt / mpi_errors_are_fatal\n-## [profitbricks-build5-amd64:36159] 1 more process has sent help message help-mpi-errors.txt / mpi_errors_are_fatal unknown handle\n+## [i-capture-the-hostname:46869] [[15263,1],0] ORTE_ERROR_LOG: Unreachable in file dpm/dpm.c at line 433\n+## [i-capture-the-hostname:46869] *** An error occurred in MPI_Comm_spawn_multiple\n+## [i-capture-the-hostname:46869] *** reported by process [1000275969,0]\n+## [i-capture-the-hostname:46869] *** on communicator MPI_COMM_WORLD\n+## [i-capture-the-hostname:46869] *** MPI_ERR_INTERN: internal error\n+## [i-capture-the-hostname:46869] *** MPI_ERRORS_ARE_FATAL (processes in this communicator will now abort,\n+## [i-capture-the-hostname:46869] *** and potentially your MPI job)\n+## [i-capture-the-hostname:46798] 3 more processes have sent help message help-mca-bml-r2.txt / unreachable proc\n+## [i-capture-the-hostname:46798] Set MCA parameter \"orte_base_help_aggregate\" to 0 to see all help / error messages\n ]]> \n \n \n \n Returned \"Unreachable\" (-12) instead of \"Success\" (0)\n ## --------------------------------------------------------------------------\n-## [profitbricks-build5-amd64:36718] *** An error occurred in MPI_Init\n-## [profitbricks-build5-amd64:36718] *** reported by process [1241972738,0]\n-## [profitbricks-build5-amd64:36718] *** on a NULL communicator\n-## [profitbricks-build5-amd64:36718] *** Unknown error\n-## [profitbricks-build5-amd64:36718] *** MPI_ERRORS_ARE_FATAL (processes in this communicator will now abort,\n-## [profitbricks-build5-amd64:36718] *** and potentially your MPI job)\n-## [profitbricks-build5-amd64:36565] 1 more process has sent help message help-mca-bml-r2.txt / unreachable proc\n-## [profitbricks-build5-amd64:36565] Set MCA parameter \"orte_base_help_aggregate\" to 0 to see all help / error messages\n+## [i-capture-the-hostname:47022] *** An error occurred in MPI_Init\n+## [i-capture-the-hostname:47022] *** reported by process [987037698,0]\n+## [i-capture-the-hostname:47022] *** on a NULL communicator\n+## [i-capture-the-hostname:47022] *** Unknown error\n+## [i-capture-the-hostname:47022] *** MPI_ERRORS_ARE_FATAL (processes in this communicator will now abort,\n+## [i-capture-the-hostname:47022] *** and potentially your MPI job)\n+## [i-capture-the-hostname:46980] 1 more process has sent help message help-mca-bml-r2.txt / unreachable proc\n+## [i-capture-the-hostname:46980] Set MCA parameter \"orte_base_help_aggregate\" to 0 to see all help / error messages\n ]]> \n \n \n \n Returned \"Unreachable\" (-12) instead of \"Success\" (0)\n ## --------------------------------------------------------------------------\n-## [profitbricks-build5-amd64:36984] [[19251,2],0] ORTE_ERROR_LOG: Unreachable in file dpm/dpm.c at line 433\n-## [profitbricks-build5-amd64:36984] *** An error occurred in MPI_Init\n-## [profitbricks-build5-amd64:36984] *** reported by process [1261633538,0]\n-## [profitbricks-build5-amd64:36984] *** on a NULL communicator\n-## [profitbricks-build5-amd64:36984] *** Unknown error\n-## [profitbricks-build5-amd64:36984] *** MPI_ERRORS_ARE_FATAL (processes in this communicator will now abort,\n-## [profitbricks-build5-amd64:36984] *** and potentially your MPI job)\n-## [profitbricks-build5-amd64:36874] [[19251,1],0] ORTE_ERROR_LOG: Unreachable in file dpm/dpm.c at line 433\n-## [profitbricks-build5-amd64:36874] *** An error occurred in MPI_Comm_spawn\n-## [profitbricks-build5-amd64:36874] *** reported by process [1261633537,0]\n-## [profitbricks-build5-amd64:36874] *** on communicator MPI_COMM_SELF\n-## [profitbricks-build5-amd64:36874] *** MPI_ERR_INTERN: internal error\n-## [profitbricks-build5-amd64:36874] *** MPI_ERRORS_ARE_FATAL (processes in this communicator will now abort,\n-## [profitbricks-build5-amd64:36874] *** and potentially your MPI job)\n-## [profitbricks-build5-amd64:36833] 1 more process has sent help message help-mca-bml-r2.txt / unreachable proc\n-## [profitbricks-build5-amd64:36833] Set MCA parameter \"orte_base_help_aggregate\" to 0 to see all help / error messages\n+## [i-capture-the-hostname:47063] [[14994,2],0] ORTE_ERROR_LOG: Unreachable in file dpm/dpm.c at line 433\n+## [i-capture-the-hostname:47043] 1 more process has sent help message help-mca-bml-r2.txt / unreachable proc\n+## [i-capture-the-hostname:47043] Set MCA parameter \"orte_base_help_aggregate\" to 0 to see all help / error messages\n ]]> \n \n \n \n \n Returned \"Unreachable\" (-12) instead of \"Success\" (0)\n ## --------------------------------------------------------------------------\n-## [profitbricks-build5-amd64:37597] *** An error occurred in MPI_Init_thread\n-## [profitbricks-build5-amd64:37597] *** reported by process [1452736514,1]\n-## [profitbricks-build5-amd64:37597] *** on a NULL communicator\n-## [profitbricks-build5-amd64:37597] *** Unknown error\n-## [profitbricks-build5-amd64:37597] *** MPI_ERRORS_ARE_FATAL (processes in this communicator will now abort,\n-## [profitbricks-build5-amd64:37597] *** and potentially your MPI job)\n-## [profitbricks-build5-amd64:37596] [[22167,2],0] ORTE_ERROR_LOG: Unreachable in file dpm/dpm.c at line 433\n-## [profitbricks-build5-amd64:37501] [[22167,1],0] ORTE_ERROR_LOG: Unreachable in file dpm/dpm.c at line 433\n-## [profitbricks-build5-amd64:37501] *** An error occurred in MPI_Comm_spawn\n-## [profitbricks-build5-amd64:37501] *** reported by process [1452736513,0]\n-## [profitbricks-build5-amd64:37501] *** on communicator MPI_COMM_SELF\n-## [profitbricks-build5-amd64:37501] *** MPI_ERR_INTERN: internal error\n-## [profitbricks-build5-amd64:37501] *** MPI_ERRORS_ARE_FATAL (processes in this communicator will now abort,\n-## [profitbricks-build5-amd64:37501] *** and potentially your MPI job)\n-## [profitbricks-build5-amd64:37445] 2 more processes have sent help message help-mca-bml-r2.txt / unreachable proc\n-## [profitbricks-build5-amd64:37445] Set MCA parameter \"orte_base_help_aggregate\" to 0 to see all help / error messages\n-## [profitbricks-build5-amd64:37445] 1 more process has sent help message help-mpi-runtime.txt / mpi_init:startup:internal-failure\n-## [profitbricks-build5-amd64:37445] 1 more process has sent help message help-mpi-errors.txt / mpi_errors_are_fatal unknown handle\n+## [i-capture-the-hostname:47194] *** An error occurred in MPI_Comm_spawn\n+## [i-capture-the-hostname:47194] *** reported by process [890634241,0]\n+## [i-capture-the-hostname:47194] *** on communicator MPI_COMM_SELF\n+## [i-capture-the-hostname:47194] *** MPI_ERR_INTERN: internal error\n+## [i-capture-the-hostname:47194] *** MPI_ERRORS_ARE_FATAL (processes in this communicator will now abort,\n+## [i-capture-the-hostname:47194] *** and potentially your MPI job)\n+## [i-capture-the-hostname:47303] *** An error occurred in MPI_Init_thread\n+## [i-capture-the-hostname:47303] *** reported by process [890634242,0]\n+## [i-capture-the-hostname:47303] *** on a NULL communicator\n+## [i-capture-the-hostname:47303] *** Unknown error\n+## [i-capture-the-hostname:47303] *** MPI_ERRORS_ARE_FATAL (processes in this communicator will now abort,\n+## [i-capture-the-hostname:47303] *** and potentially your MPI job)\n+## [i-capture-the-hostname:47175] 2 more processes have sent help message help-mca-bml-r2.txt / unreachable proc\n+## [i-capture-the-hostname:47175] Set MCA parameter \"orte_base_help_aggregate\" to 0 to see all help / error messages\n+## [i-capture-the-hostname:47175] 1 more process has sent help message help-mpi-runtime.txt / mpi_init:startup:internal-failure\n+## [i-capture-the-hostname:47175] 1 more process has sent help message help-mpi-errors.txt / mpi_errors_are_fatal unknown handle\n ]]> \n \n \n \n Returned \"Unreachable\" (-12) instead of \"Success\" (0)\n ## --------------------------------------------------------------------------\n-## [profitbricks-build5-amd64:37968] *** An error occurred in MPI_Init\n-## [profitbricks-build5-amd64:37968] *** reported by process [1472069634,0]\n-## [profitbricks-build5-amd64:37968] *** on a NULL communicator\n-## [profitbricks-build5-amd64:37968] *** Unknown error\n-## [profitbricks-build5-amd64:37968] *** MPI_ERRORS_ARE_FATAL (processes in this communicator will now abort,\n-## [profitbricks-build5-amd64:37968] *** and potentially your MPI job)\n-## [profitbricks-build5-amd64:37840] *** An error occurred in MPI_Comm_spawn\n-## [profitbricks-build5-amd64:37840] *** reported by process [1472069633,0]\n-## [profitbricks-build5-amd64:37840] *** on communicator MPI_COMM_WORLD\n-## [profitbricks-build5-amd64:37840] *** MPI_ERR_INTERN: internal error\n-## [profitbricks-build5-amd64:37840] *** MPI_ERRORS_ARE_FATAL (processes in this communicator will now abort,\n-## [profitbricks-build5-amd64:37840] *** and potentially your MPI job)\n-## [profitbricks-build5-amd64:37840] [[22462,1],0] ORTE_ERROR_LOG: Unreachable in file dpm/dpm.c at line 433\n-## [profitbricks-build5-amd64:37740] 1 more process has sent help message help-mca-bml-r2.txt / unreachable proc\n-## [profitbricks-build5-amd64:37740] Set MCA parameter \"orte_base_help_aggregate\" to 0 to see all help / error messages\n+## [i-capture-the-hostname:47531] *** An error occurred in MPI_Comm_spawn\n+## [i-capture-the-hostname:47531] *** reported by process [874971137,0]\n+## [i-capture-the-hostname:47531] *** on communicator MPI_COMM_WORLD\n+## [i-capture-the-hostname:47531] *** MPI_ERR_INTERN: internal error\n+## [i-capture-the-hostname:47531] *** MPI_ERRORS_ARE_FATAL (processes in this communicator will now abort,\n+## [i-capture-the-hostname:47531] *** and potentially your MPI job)\n+## [i-capture-the-hostname:47649] *** An error occurred in MPI_Init\n+## [i-capture-the-hostname:47649] *** reported by process [874971138,0]\n+## [i-capture-the-hostname:47649] *** on a NULL communicator\n+## [i-capture-the-hostname:47649] *** Unknown error\n+## [i-capture-the-hostname:47649] *** MPI_ERRORS_ARE_FATAL (processes in this communicator will now abort,\n+## [i-capture-the-hostname:47649] *** and potentially your MPI job)\n+## [i-capture-the-hostname:47478] 1 more process has sent help message help-mca-bml-r2.txt / unreachable proc\n+## [i-capture-the-hostname:47478] Set MCA parameter \"orte_base_help_aggregate\" to 0 to see all help / error messages\n ]]> \n \n \n \n Returned \"Unreachable\" (-12) instead of \"Success\" (0)\n-## --------------------------------------------------------------------------\n-## [profitbricks-build5-amd64:38240] [[20979,2],0] ORTE_ERROR_LOG: Unreachable in file dpm/dpm.c at line 433\n-## [profitbricks-build5-amd64:38177] 2 more processes have sent help message help-mca-bml-r2.txt / unreachable proc\n-## [profitbricks-build5-amd64:38177] Set MCA parameter \"orte_base_help_aggregate\" to 0 to see all help / error messages\n-## [profitbricks-build5-amd64:38177] 1 more process has sent help message help-mpi-errors.txt / mpi_errors_are_fatal\n+## [i-capture-the-hostname:48848] [[12569,1],0] ORTE_ERROR_LOG: Unreachable in file dpm/dpm.c at line 433\n+## [i-capture-the-hostname:48848] *** An error occurred in MPI_Comm_spawn\n+## [i-capture-the-hostname:48848] *** reported by process [823721985,0]\n+## [i-capture-the-hostname:48848] *** on communicator MPI_COMM_WORLD\n+## [i-capture-the-hostname:48848] *** MPI_ERR_INTERN: internal error\n+## [i-capture-the-hostname:48848] *** MPI_ERRORS_ARE_FATAL (processes in this communicator will now abort,\n+## [i-capture-the-hostname:48848] *** and potentially your MPI job)\n+## [i-capture-the-hostname:48854] [[12569,1],1] ORTE_ERROR_LOG: Unreachable in file dpm/dpm.c at line 433\n+## [i-capture-the-hostname:48200] 2 more processes have sent help message help-mca-bml-r2.txt / unreachable proc\n+## [i-capture-the-hostname:48200] Set MCA parameter \"orte_base_help_aggregate\" to 0 to see all help / error messages\n+## [i-capture-the-hostname:48200] 1 more process has sent help message help-mpi-errors.txt / mpi_errors_are_fatal\n+## [warn] Epoll MOD(1) on fd 29 failed. Old events were 6; read change was 0 (none); write change was 2 (del); close change was 0 (none): Bad file descriptor\n ]]> \n \n \n \n \n \n Returned \"Unreachable\" (-12) instead of \"Success\" (0)\n ## --------------------------------------------------------------------------\n-## [profitbricks-build5-amd64:38468] [[21189,1],0] ORTE_ERROR_LOG: Unreachable in file dpm/dpm.c at line 433\n-## [profitbricks-build5-amd64:38615] *** An error occurred in MPI_Init_thread\n-## [profitbricks-build5-amd64:38615] *** reported by process [1388642306,2]\n-## [profitbricks-build5-amd64:38615] *** on a NULL communicator\n-## [profitbricks-build5-amd64:38615] *** Unknown error\n-## [profitbricks-build5-amd64:38615] *** MPI_ERRORS_ARE_FATAL (processes in this communicator will now abort,\n-## [profitbricks-build5-amd64:38615] *** and potentially your MPI job)\n-## [profitbricks-build5-amd64:38423] 4 more processes have sent help message help-mca-bml-r2.txt / unreachable proc\n-## [profitbricks-build5-amd64:38423] Set MCA parameter \"orte_base_help_aggregate\" to 0 to see all help / error messages\n-## [profitbricks-build5-amd64:38423] 1 more process has sent help message help-mpi-errors.txt / mpi_errors_are_fatal\n+## [i-capture-the-hostname:51043] [[19253,1],1] ORTE_ERROR_LOG: Unreachable in file dpm/dpm.c at line 433\n+## [i-capture-the-hostname:51043] *** An error occurred in MPI_Comm_spawn\n+## [i-capture-the-hostname:51043] *** reported by process [1261764609,1]\n+## [i-capture-the-hostname:51043] *** on communicator MPI_COMM_WORLD\n+## [i-capture-the-hostname:51043] *** MPI_ERR_INTERN: internal error\n+## [i-capture-the-hostname:51043] *** MPI_ERRORS_ARE_FATAL (processes in this communicator will now abort,\n+## [i-capture-the-hostname:51043] *** and potentially your MPI job)\n+## [i-capture-the-hostname:51399] *** An error occurred in MPI_Init_thread\n+## [i-capture-the-hostname:51399] *** reported by process [1261764610,0]\n+## [i-capture-the-hostname:51399] *** on a NULL communicator\n+## [i-capture-the-hostname:51399] *** Unknown error\n+## [i-capture-the-hostname:51399] *** MPI_ERRORS_ARE_FATAL (processes in this communicator will now abort,\n+## [i-capture-the-hostname:51399] *** and potentially your MPI job)\n+## [i-capture-the-hostname:51036] [[19253,1],0] ORTE_ERROR_LOG: Unreachable in file dpm/dpm.c at line 433\n+## [i-capture-the-hostname:51417] [[19253,2],2] ORTE_ERROR_LOG: Unreachable in file dpm/dpm.c at line 433\n+## [i-capture-the-hostname:50788] 4 more processes have sent help message help-mca-bml-r2.txt / unreachable proc\n+## [i-capture-the-hostname:50788] Set MCA parameter \"orte_base_help_aggregate\" to 0 to see all help / error messages\n+## [i-capture-the-hostname:50788] 1 more process has sent help message help-mpi-runtime.txt / mpi_init:startup:internal-failure\n+## [i-capture-the-hostname:50788] 1 more process has sent help message help-mpi-errors.txt / mpi_errors_are_fatal\n ]]> \n \n \n \n Returned \"Unreachable\" (-12) instead of \"Success\" (0)\n ## --------------------------------------------------------------------------\n-## [profitbricks-build5-amd64:39148] [[23641,1],0] ORTE_ERROR_LOG: Unreachable in file dpm/dpm.c at line 433\n-## [profitbricks-build5-amd64:39271] [[23641,2],2] ORTE_ERROR_LOG: Unreachable in file dpm/dpm.c at line 433\n-## [profitbricks-build5-amd64:39270] [[23641,2],1] ORTE_ERROR_LOG: Unreachable in file dpm/dpm.c at line 433\n-## [profitbricks-build5-amd64:39051] 4 more processes have sent help message help-mca-bml-r2.txt / unreachable proc\n-## [profitbricks-build5-amd64:39051] Set MCA parameter \"orte_base_help_aggregate\" to 0 to see all help / error messages\n-## [profitbricks-build5-amd64:39051] 1 more process has sent help message help-mpi-errors.txt / mpi_errors_are_fatal\n-## [profitbricks-build5-amd64:39051] 1 more process has sent help message help-mpi-runtime.txt / mpi_init:startup:internal-failure\n+## [i-capture-the-hostname:52119] *** An error occurred in MPI_Comm_spawn\n+## [i-capture-the-hostname:52119] *** reported by process [1194131457,1]\n+## [i-capture-the-hostname:52119] *** on communicator MPI_COMM_WORLD\n+## [i-capture-the-hostname:52119] *** MPI_ERR_INTERN: internal error\n+## [i-capture-the-hostname:52119] *** MPI_ERRORS_ARE_FATAL (processes in this communicator will now abort,\n+## [i-capture-the-hostname:52119] *** and potentially your MPI job)\n+## [i-capture-the-hostname:52518] *** An error occurred in MPI_Init\n+## [i-capture-the-hostname:52518] *** reported by process [1194131458,1]\n+## [i-capture-the-hostname:52518] *** on a NULL communicator\n+## [i-capture-the-hostname:52518] *** Unknown error\n+## [i-capture-the-hostname:52518] *** MPI_ERRORS_ARE_FATAL (processes in this communicator will now abort,\n+## [i-capture-the-hostname:52518] *** and potentially your MPI job)\n+## [i-capture-the-hostname:51836] 4 more processes have sent help message help-mca-bml-r2.txt / unreachable proc\n+## [i-capture-the-hostname:51836] Set MCA parameter \"orte_base_help_aggregate\" to 0 to see all help / error messages\n ]]> \n \n \n \n Returned \"Unreachable\" (-12) instead of \"Success\" (0)\n ## --------------------------------------------------------------------------\n-## [profitbricks-build5-amd64:39672] *** An error occurred in MPI_Init_thread\n-## [profitbricks-build5-amd64:39672] *** reported by process [1592983554,2]\n-## [profitbricks-build5-amd64:39672] *** on a NULL communicator\n-## [profitbricks-build5-amd64:39672] *** Unknown error\n-## [profitbricks-build5-amd64:39672] *** MPI_ERRORS_ARE_FATAL (processes in this communicator will now abort,\n-## [profitbricks-build5-amd64:39672] *** and potentially your MPI job)\n-## [profitbricks-build5-amd64:39457] 5 more processes have sent help message help-mca-bml-r2.txt / unreachable proc\n-## [profitbricks-build5-amd64:39457] Set MCA parameter \"orte_base_help_aggregate\" to 0 to see all help / error messages\n-## [profitbricks-build5-amd64:39457] 1 more process has sent help message help-mpi-errors.txt / mpi_errors_are_fatal\n-## [profitbricks-build5-amd64:39457] 3 more processes have sent help message help-mpi-runtime.txt / mpi_init:startup:internal-failure\n-## [profitbricks-build5-amd64:39457] 3 more processes have sent help message help-mpi-errors.txt / mpi_errors_are_fatal unknown handle\n+## [i-capture-the-hostname:53436] *** An error occurred in MPI_Comm_spawn\n+## [i-capture-the-hostname:53436] *** reported by process [1108213761,1]\n+## [i-capture-the-hostname:53436] *** on communicator MPI_COMM_WORLD\n+## [i-capture-the-hostname:53436] *** MPI_ERR_INTERN: internal error\n+## [i-capture-the-hostname:53436] *** MPI_ERRORS_ARE_FATAL (processes in this communicator will now abort,\n+## [i-capture-the-hostname:53436] *** and potentially your MPI job)\n+## [i-capture-the-hostname:53921] *** An error occurred in MPI_Init_thread\n+## [i-capture-the-hostname:53921] *** reported by process [1108213762,2]\n+## [i-capture-the-hostname:53921] *** on a NULL communicator\n+## [i-capture-the-hostname:53921] *** Unknown error\n+## [i-capture-the-hostname:53921] *** MPI_ERRORS_ARE_FATAL (processes in this communicator will now abort,\n+## [i-capture-the-hostname:53921] *** and potentially your MPI job)\n+## [i-capture-the-hostname:53428] [[16910,1],0] ORTE_ERROR_LOG: Unreachable in file dpm/dpm.c at line 433\n+## [i-capture-the-hostname:53436] [[16910,1],1] ORTE_ERROR_LOG: Unreachable in file dpm/dpm.c at line 433\n+## [i-capture-the-hostname:53930] [[16910,2],3] ORTE_ERROR_LOG: Unreachable in file dpm/dpm.c at line 433\n+## [i-capture-the-hostname:53911] [[16910,2],1] ORTE_ERROR_LOG: Unreachable in file dpm/dpm.c at line 433\n+## [i-capture-the-hostname:53087] 5 more processes have sent help message help-mca-bml-r2.txt / unreachable proc\n+## [i-capture-the-hostname:53087] Set MCA parameter \"orte_base_help_aggregate\" to 0 to see all help / error messages\n+## [i-capture-the-hostname:53087] 3 more processes have sent help message help-mpi-runtime.txt / mpi_init:startup:internal-failure\n+## [i-capture-the-hostname:53087] 1 more process has sent help message help-mpi-errors.txt / mpi_errors_are_fatal\n+## [i-capture-the-hostname:53087] 3 more processes have sent help message help-mpi-errors.txt / mpi_errors_are_fatal unknown handle\n ]]> \n \n \n \n \n \n \n \n \n \n \n \n Returned \"Unreachable\" (-12) instead of \"Success\" (0)\n ## --------------------------------------------------------------------------\n-## [profitbricks-build5-amd64:40873] *** An error occurred in MPI_Init_thread\n-## [profitbricks-build5-amd64:40873] *** reported by process [1524695042,2]\n-## [profitbricks-build5-amd64:40873] *** on a NULL communicator\n-## [profitbricks-build5-amd64:40873] *** Unknown error\n-## [profitbricks-build5-amd64:40873] *** MPI_ERRORS_ARE_FATAL (processes in this communicator will now abort,\n-## [profitbricks-build5-amd64:40873] *** and potentially your MPI job)\n-## [profitbricks-build5-amd64:40499] 4 more processes have sent help message help-mca-bml-r2.txt / unreachable proc\n-## [profitbricks-build5-amd64:40499] Set MCA parameter \"orte_base_help_aggregate\" to 0 to see all help / error messages\n+## [i-capture-the-hostname:55766] *** An error occurred in MPI_Init_thread\n+## [i-capture-the-hostname:55766] *** reported by process [1429667842,2]\n+## [i-capture-the-hostname:55766] *** on a NULL communicator\n+## [i-capture-the-hostname:55766] *** Unknown error\n+## [i-capture-the-hostname:55766] *** MPI_ERRORS_ARE_FATAL (processes in this communicator will now abort,\n+## [i-capture-the-hostname:55766] *** and potentially your MPI job)\n+## [i-capture-the-hostname:55398] 4 more processes have sent help message help-mca-bml-r2.txt / unreachable proc\n+## [i-capture-the-hostname:55398] Set MCA parameter \"orte_base_help_aggregate\" to 0 to see all help / error messages\n+## [i-capture-the-hostname:55398] 1 more process has sent help message help-mpi-errors.txt / mpi_errors_are_fatal\n+## [i-capture-the-hostname:55398] 2 more processes have sent help message help-mpi-runtime.txt / mpi_init:startup:internal-failure\n+## [i-capture-the-hostname:55398] 2 more processes have sent help message help-mpi-errors.txt / mpi_errors_are_fatal unknown handle\n ]]> \n \n \n \n Returned \"Unreachable\" (-12) instead of \"Success\" (0)\n ## --------------------------------------------------------------------------\n-## [profitbricks-build5-amd64:41301] [[25777,2],0] ORTE_ERROR_LOG: Unreachable in file dpm/dpm.c at line 433\n-## [profitbricks-build5-amd64:41301] *** An error occurred in MPI_Init_thread\n-## [profitbricks-build5-amd64:41301] *** reported by process [1689321474,0]\n-## [profitbricks-build5-amd64:41301] *** on a NULL communicator\n-## [profitbricks-build5-amd64:41301] *** Unknown error\n-## [profitbricks-build5-amd64:41301] *** MPI_ERRORS_ARE_FATAL (processes in this communicator will now abort,\n-## [profitbricks-build5-amd64:41301] *** and potentially your MPI job)\n-## [profitbricks-build5-amd64:41182] [[25777,1],0] ORTE_ERROR_LOG: Unreachable in file dpm/dpm.c at line 433\n-## [profitbricks-build5-amd64:41307] [[25777,2],1] ORTE_ERROR_LOG: Unreachable in file dpm/dpm.c at line 433\n-## [profitbricks-build5-amd64:41059] 4 more processes have sent help message help-mca-bml-r2.txt / unreachable proc\n-## [profitbricks-build5-amd64:41059] Set MCA parameter \"orte_base_help_aggregate\" to 0 to see all help / error messages\n-## [profitbricks-build5-amd64:41059] 1 more process has sent help message help-mpi-errors.txt / mpi_errors_are_fatal\n-## [profitbricks-build5-amd64:41059] 2 more processes have sent help message help-mpi-runtime.txt / mpi_init:startup:internal-failure\n-## [profitbricks-build5-amd64:41059] 1 more process has sent help message help-mpi-errors.txt / mpi_errors_are_fatal unknown handle\n+## [i-capture-the-hostname:56076] [[22521,1],0] ORTE_ERROR_LOG: Unreachable in file dpm/dpm.c at line 433\n+## [i-capture-the-hostname:56085] [[22521,1],1] ORTE_ERROR_LOG: Unreachable in file dpm/dpm.c at line 433\n+## [i-capture-the-hostname:56224] [[22521,2],0] ORTE_ERROR_LOG: Unreachable in file dpm/dpm.c at line 433\n+## [i-capture-the-hostname:56233] [[22521,2],2] ORTE_ERROR_LOG: Unreachable in file dpm/dpm.c at line 433\n+## [i-capture-the-hostname:56229] [[22521,2],1] ORTE_ERROR_LOG: Unreachable in file dpm/dpm.c at line 433\n+## [i-capture-the-hostname:56085] *** An error occurred in MPI_Comm_spawn\n+## [i-capture-the-hostname:56085] *** reported by process [1475936257,1]\n+## [i-capture-the-hostname:56085] *** on communicator MPI_COMM_WORLD\n+## [i-capture-the-hostname:56085] *** MPI_ERR_INTERN: internal error\n+## [i-capture-the-hostname:56085] *** MPI_ERRORS_ARE_FATAL (processes in this communicator will now abort,\n+## [i-capture-the-hostname:56085] *** and potentially your MPI job)\n+## [i-capture-the-hostname:56224] *** An error occurred in MPI_Init_thread\n+## [i-capture-the-hostname:56224] *** reported by process [1475936258,0]\n+## [i-capture-the-hostname:56224] *** on a NULL communicator\n+## [i-capture-the-hostname:56224] *** Unknown error\n+## [i-capture-the-hostname:56224] *** MPI_ERRORS_ARE_FATAL (processes in this communicator will now abort,\n+## [i-capture-the-hostname:56224] *** and potentially your MPI job)\n+## [i-capture-the-hostname:55976] 4 more processes have sent help message help-mca-bml-r2.txt / unreachable proc\n+## [i-capture-the-hostname:55976] Set MCA parameter \"orte_base_help_aggregate\" to 0 to see all help / error messages\n+## [i-capture-the-hostname:55976] 2 more processes have sent help message help-mpi-runtime.txt / mpi_init:startup:internal-failure\n+## [i-capture-the-hostname:55976] 1 more process has sent help message help-mpi-errors.txt / mpi_errors_are_fatal\n+## [i-capture-the-hostname:55976] 2 more processes have sent help message help-mpi-errors.txt / mpi_errors_are_fatal unknown handle\n ]]> \n \n \n \n Returned \"Unreachable\" (-12) instead of \"Success\" (0)\n ## --------------------------------------------------------------------------\n-## [profitbricks-build5-amd64:41692] *** An error occurred in MPI_Comm_spawn\n-## [profitbricks-build5-amd64:41692] *** reported by process [1721171969,1]\n-## [profitbricks-build5-amd64:41692] *** on communicator MPI_COMM_WORLD\n-## [profitbricks-build5-amd64:41692] *** MPI_ERR_INTERN: internal error\n-## [profitbricks-build5-amd64:41692] *** MPI_ERRORS_ARE_FATAL (processes in this communicator will now abort,\n-## [profitbricks-build5-amd64:41692] *** and potentially your MPI job)\n-## [profitbricks-build5-amd64:41811] [[26263,2],2] ORTE_ERROR_LOG: Unreachable in file dpm/dpm.c at line 433\n-## [profitbricks-build5-amd64:41810] *** An error occurred in MPI_Init_thread\n-## [profitbricks-build5-amd64:41810] *** reported by process [1721171970,1]\n-## [profitbricks-build5-amd64:41810] *** on a NULL communicator\n-## [profitbricks-build5-amd64:41810] *** Unknown error\n-## [profitbricks-build5-amd64:41810] *** MPI_ERRORS_ARE_FATAL (processes in this communicator will now abort,\n-## [profitbricks-build5-amd64:41810] *** and potentially your MPI job)\n-## [profitbricks-build5-amd64:41541] 4 more processes have sent help message help-mca-bml-r2.txt / unreachable proc\n-## [profitbricks-build5-amd64:41541] Set MCA parameter \"orte_base_help_aggregate\" to 0 to see all help / error messages\n-## [profitbricks-build5-amd64:41541] 2 more processes have sent help message help-mpi-runtime.txt / mpi_init:startup:internal-failure\n-## [profitbricks-build5-amd64:41541] 2 more processes have sent help message help-mpi-errors.txt / mpi_errors_are_fatal unknown handle\n-## [warn] Epoll MOD(1) on fd 31 failed. Old events were 6; read change was 0 (none); write change was 2 (del); close change was 0 (none): Bad file descriptor\n+## [i-capture-the-hostname:56698] [[20933,2],2] ORTE_ERROR_LOG: Unreachable in file dpm/dpm.c at line 433\n+## [i-capture-the-hostname:56695] [[20933,2],0] ORTE_ERROR_LOG: Unreachable in file dpm/dpm.c at line 433\n+## [i-capture-the-hostname:56697] [[20933,2],1] ORTE_ERROR_LOG: Unreachable in file dpm/dpm.c at line 433\n+## [i-capture-the-hostname:56468] 4 more processes have sent help message help-mca-bml-r2.txt / unreachable proc\n+## [i-capture-the-hostname:56468] Set MCA parameter \"orte_base_help_aggregate\" to 0 to see all help / error messages\n+## [i-capture-the-hostname:56468] 1 more process has sent help message help-mpi-errors.txt / mpi_errors_are_fatal\n+## [i-capture-the-hostname:56468] 2 more processes have sent help message help-mpi-runtime.txt / mpi_init:startup:internal-failure\n ]]> \n \n \n \n Returned \"Unreachable\" (-12) instead of \"Success\" (0)\n+## --------------------------------------------------------------------------\n+## [i-capture-the-hostname:56968] [[21366,1],0] ORTE_ERROR_LOG: Unreachable in file dpm/dpm.c at line 433\n+## [i-capture-the-hostname:56968] *** An error occurred in MPI_Comm_spawn\n+## [i-capture-the-hostname:56968] *** reported by process [1400242177,0]\n+## [i-capture-the-hostname:56968] *** on communicator MPI_COMM_WORLD\n+## [i-capture-the-hostname:56968] *** MPI_ERR_INTERN: internal error\n+## [i-capture-the-hostname:56968] *** MPI_ERRORS_ARE_FATAL (processes in this communicator will now abort,\n+## [i-capture-the-hostname:56968] *** and potentially your MPI job)\n+## [i-capture-the-hostname:57123] *** An error occurred in MPI_Init_thread\n+## [i-capture-the-hostname:57123] *** reported by process [1400242178,0]\n+## [i-capture-the-hostname:57123] *** on a NULL communicator\n+## [i-capture-the-hostname:57123] *** Unknown error\n+## [i-capture-the-hostname:57123] *** MPI_ERRORS_ARE_FATAL (processes in this communicator will now abort,\n+## [i-capture-the-hostname:57123] *** and potentially your MPI job)\n+## [i-capture-the-hostname:56871] 1 more process has sent help message help-mca-bml-r2.txt / unreachable proc\n+## [i-capture-the-hostname:56871] Set MCA parameter \"orte_base_help_aggregate\" to 0 to see all help / error messages\n ]]> \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n@@ -999049,92 +999051,92 @@\n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n@@ -999142,35 +999144,35 @@\n \n \n \n \n \n \n \n \n \n \n \n \n@@ -999264,71 +999266,71 @@\n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n@@ -999340,25 +999342,26 @@\n \n \n \n \n \n \n \n \n \n Returned \"Unreachable\" (-12) instead of \"Success\" (0)\n ## --------------------------------------------------------------------------\n-## [profitbricks-build5-amd64:05143] *** An error occurred in MPI_Comm_spawn\n-## [profitbricks-build5-amd64:05143] *** reported by process [3610050561,0]\n-## [profitbricks-build5-amd64:05143] *** on communicator MPI_COMM_WORLD\n-## [profitbricks-build5-amd64:05143] *** MPI_ERR_INTERN: internal error\n-## [profitbricks-build5-amd64:05143] *** MPI_ERRORS_ARE_FATAL (processes in this communicator will now abort,\n-## [profitbricks-build5-amd64:05143] *** and potentially your MPI job)\n-## [profitbricks-build5-amd64:05162] *** An error occurred in MPI_Init\n-## [profitbricks-build5-amd64:05162] *** reported by process [3610050562,0]\n-## [profitbricks-build5-amd64:05162] *** on a NULL communicator\n-## [profitbricks-build5-amd64:05162] *** Unknown error\n-## [profitbricks-build5-amd64:05162] *** MPI_ERRORS_ARE_FATAL (processes in this communicator will now abort,\n-## [profitbricks-build5-amd64:05162] *** and potentially your MPI job)\n-## [profitbricks-build5-amd64:05119] 2 more processes have sent help message help-mca-bml-r2.txt / unreachable proc\n-## [profitbricks-build5-amd64:05119] Set MCA parameter \"orte_base_help_aggregate\" to 0 to see all help / error messages\n-## [profitbricks-build5-amd64:05119] 1 more process has sent help message help-mpi-runtime.txt / mpi_init:startup:internal-failure\n-## [profitbricks-build5-amd64:05119] 1 more process has sent help message help-mpi-errors.txt / mpi_errors_are_fatal unknown handle\n+## [i-capture-the-hostname:05460] *** An error occurred in MPI_Comm_spawn\n+## [i-capture-the-hostname:05460] *** reported by process [2556690433,0]\n+## [i-capture-the-hostname:05460] *** on communicator MPI_COMM_WORLD\n+## [i-capture-the-hostname:05460] *** MPI_ERR_INTERN: internal error\n+## [i-capture-the-hostname:05460] *** MPI_ERRORS_ARE_FATAL (processes in this communicator will now abort,\n+## [i-capture-the-hostname:05460] *** and potentially your MPI job)\n+## [i-capture-the-hostname:05501] *** An error occurred in MPI_Init\n+## [i-capture-the-hostname:05501] *** reported by process [2556690434,0]\n+## [i-capture-the-hostname:05501] *** on a NULL communicator\n+## [i-capture-the-hostname:05501] *** Unknown error\n+## [i-capture-the-hostname:05501] *** MPI_ERRORS_ARE_FATAL (processes in this communicator will now abort,\n+## [i-capture-the-hostname:05501] *** and potentially your MPI job)\n+## [i-capture-the-hostname:05429] 2 more processes have sent help message help-mca-bml-r2.txt / unreachable proc\n+## [i-capture-the-hostname:05429] Set MCA parameter \"orte_base_help_aggregate\" to 0 to see all help / error messages\n+## [i-capture-the-hostname:05429] 1 more process has sent help message help-mpi-runtime.txt / mpi_init:startup:internal-failure\n+## [i-capture-the-hostname:05429] 1 more process has sent help message help-mpi-errors.txt / mpi_errors_are_fatal unknown handle\n ]]> \n \n \n \n Returned \"Unreachable\" (-12) instead of \"Success\" (0)\n ## --------------------------------------------------------------------------\n-## [profitbricks-build5-amd64:05418] [[53436,2],1] ORTE_ERROR_LOG: Unreachable in file dpm/dpm.c at line 433\n-## [profitbricks-build5-amd64:05418] *** An error occurred in MPI_Init\n-## [profitbricks-build5-amd64:05418] *** reported by process [3501981698,1]\n-## [profitbricks-build5-amd64:05418] *** on a NULL communicator\n-## [profitbricks-build5-amd64:05418] *** Unknown error\n-## [profitbricks-build5-amd64:05418] *** MPI_ERRORS_ARE_FATAL (processes in this communicator will now abort,\n-## [profitbricks-build5-amd64:05418] *** and potentially your MPI job)\n-## [profitbricks-build5-amd64:05258] *** An error occurred in MPI_Comm_spawn\n-## [profitbricks-build5-amd64:05258] *** reported by process [3501981697,0]\n-## [profitbricks-build5-amd64:05258] *** on communicator MPI_COMM_WORLD\n-## [profitbricks-build5-amd64:05258] *** MPI_ERR_INTERN: internal error\n-## [profitbricks-build5-amd64:05258] *** MPI_ERRORS_ARE_FATAL (processes in this communicator will now abort,\n-## [profitbricks-build5-amd64:05258] *** and potentially your MPI job)\n-## [profitbricks-build5-amd64:05258] [[53436,1],0] ORTE_ERROR_LOG: Unreachable in file dpm/dpm.c at line 433\n-## [profitbricks-build5-amd64:05414] [[53436,2],0] ORTE_ERROR_LOG: Unreachable in file dpm/dpm.c at line 433\n-## [profitbricks-build5-amd64:05230] 2 more processes have sent help message help-mca-bml-r2.txt / unreachable proc\n-## [profitbricks-build5-amd64:05230] Set MCA parameter \"orte_base_help_aggregate\" to 0 to see all help / error messages\n-## [profitbricks-build5-amd64:05230] 1 more process has sent help message help-mpi-runtime.txt / mpi_init:startup:internal-failure\n+## [i-capture-the-hostname:05632] *** An error occurred in MPI_Init\n+## [i-capture-the-hostname:05632] *** reported by process [2565668866,0]\n+## [i-capture-the-hostname:05632] *** on a NULL communicator\n+## [i-capture-the-hostname:05632] *** Unknown error\n+## [i-capture-the-hostname:05632] *** MPI_ERRORS_ARE_FATAL (processes in this communicator will now abort,\n+## [i-capture-the-hostname:05632] *** and potentially your MPI job)\n+## [i-capture-the-hostname:05564] 2 more processes have sent help message help-mca-bml-r2.txt / unreachable proc\n+## [i-capture-the-hostname:05564] Set MCA parameter \"orte_base_help_aggregate\" to 0 to see all help / error messages\n+## [i-capture-the-hostname:05564] 1 more process has sent help message help-mpi-runtime.txt / mpi_init:startup:internal-failure\n+## [i-capture-the-hostname:05564] 1 more process has sent help message help-mpi-errors.txt / mpi_errors_are_fatal unknown handle\n ]]> \n \n \n \n \n Returned \"Unreachable\" (-12) instead of \"Success\" (0)\n ## --------------------------------------------------------------------------\n-## [profitbricks-build5-amd64:06382] [[54051,2],0] ORTE_ERROR_LOG: Unreachable in file dpm/dpm.c at line 433\n-## [profitbricks-build5-amd64:06386] [[54051,2],1] ORTE_ERROR_LOG: Unreachable in file dpm/dpm.c at line 433\n-## [profitbricks-build5-amd64:06382] *** An error occurred in MPI_Init\n-## [profitbricks-build5-amd64:06382] *** reported by process [3542286338,0]\n-## [profitbricks-build5-amd64:06382] *** on a NULL communicator\n-## [profitbricks-build5-amd64:06382] *** Unknown error\n-## [profitbricks-build5-amd64:06382] *** MPI_ERRORS_ARE_FATAL (processes in this communicator will now abort,\n-## [profitbricks-build5-amd64:06382] *** and potentially your MPI job)\n-## [profitbricks-build5-amd64:06192] [[54051,1],0] ORTE_ERROR_LOG: Unreachable in file dpm/dpm.c at line 433\n-## [profitbricks-build5-amd64:06192] *** An error occurred in MPI_Comm_spawn_multiple\n-## [profitbricks-build5-amd64:06192] *** reported by process [3542286337,0]\n-## [profitbricks-build5-amd64:06192] *** on communicator MPI_COMM_WORLD\n-## [profitbricks-build5-amd64:06192] *** MPI_ERR_INTERN: internal error\n-## [profitbricks-build5-amd64:06192] *** MPI_ERRORS_ARE_FATAL (processes in this communicator will now abort,\n-## [profitbricks-build5-amd64:06192] *** and potentially your MPI job)\n-## [profitbricks-build5-amd64:06129] 2 more processes have sent help message help-mca-bml-r2.txt / unreachable proc\n-## [profitbricks-build5-amd64:06129] Set MCA parameter \"orte_base_help_aggregate\" to 0 to see all help / error messages\n-## [profitbricks-build5-amd64:06129] 1 more process has sent help message help-mpi-runtime.txt / mpi_init:startup:internal-failure\n-## [profitbricks-build5-amd64:06129] 1 more process has sent help message help-mpi-errors.txt / mpi_errors_are_fatal unknown handle\n+## [i-capture-the-hostname:05924] *** An error occurred in MPI_Comm_spawn_multiple\n+## [i-capture-the-hostname:05924] *** reported by process [2589261825,0]\n+## [i-capture-the-hostname:05924] *** on communicator MPI_COMM_WORLD\n+## [i-capture-the-hostname:05924] *** MPI_ERR_INTERN: internal error\n+## [i-capture-the-hostname:05924] *** MPI_ERRORS_ARE_FATAL (processes in this communicator will now abort,\n+## [i-capture-the-hostname:05924] *** and potentially your MPI job)\n+## [i-capture-the-hostname:05984] *** An error occurred in MPI_Init\n+## [i-capture-the-hostname:05984] *** reported by process [2589261826,0]\n+## [i-capture-the-hostname:05984] *** on a NULL communicator\n+## [i-capture-the-hostname:05984] *** Unknown error\n+## [i-capture-the-hostname:05984] *** MPI_ERRORS_ARE_FATAL (processes in this communicator will now abort,\n+## [i-capture-the-hostname:05984] *** and potentially your MPI job)\n+## [i-capture-the-hostname:05892] 2 more processes have sent help message help-mca-bml-r2.txt / unreachable proc\n+## [i-capture-the-hostname:05892] Set MCA parameter \"orte_base_help_aggregate\" to 0 to see all help / error messages\n+## [i-capture-the-hostname:05892] 1 more process has sent help message help-mpi-runtime.txt / mpi_init:startup:internal-failure\n+## [i-capture-the-hostname:05892] 1 more process has sent help message help-mpi-errors.txt / mpi_errors_are_fatal unknown handle\n ## [warn] Epoll MOD(1) on fd 29 failed. Old events were 6; read change was 0 (none); write change was 2 (del); close change was 0 (none): Bad file descriptor\n ]]> \n \n \n \n Returned \"Unreachable\" (-12) instead of \"Success\" (0)\n ## --------------------------------------------------------------------------\n-## [profitbricks-build5-amd64:06724] *** An error occurred in MPI_Init\n-## [profitbricks-build5-amd64:06724] *** reported by process [3713138690,1]\n-## [profitbricks-build5-amd64:06724] *** on a NULL communicator\n-## [profitbricks-build5-amd64:06724] *** Unknown error\n-## [profitbricks-build5-amd64:06724] *** MPI_ERRORS_ARE_FATAL (processes in this communicator will now abort,\n-## [profitbricks-build5-amd64:06724] *** and potentially your MPI job)\n-## [profitbricks-build5-amd64:06613] *** An error occurred in MPI_Comm_spawn_multiple\n-## [profitbricks-build5-amd64:06613] *** reported by process [3713138689,0]\n-## [profitbricks-build5-amd64:06613] *** on communicator MPI_COMM_WORLD\n-## [profitbricks-build5-amd64:06613] *** MPI_ERR_INTERN: internal error\n-## [profitbricks-build5-amd64:06613] *** MPI_ERRORS_ARE_FATAL (processes in this communicator will now abort,\n-## [profitbricks-build5-amd64:06613] *** and potentially your MPI job)\n-## [profitbricks-build5-amd64:06615] [[56658,1],1] ORTE_ERROR_LOG: Unreachable in file dpm/dpm.c at line 433\n-## [profitbricks-build5-amd64:06613] [[56658,1],0] ORTE_ERROR_LOG: Unreachable in file dpm/dpm.c at line 433\n-## [profitbricks-build5-amd64:06528] 3 more processes have sent help message help-mca-bml-r2.txt / unreachable proc\n-## [profitbricks-build5-amd64:06528] Set MCA parameter \"orte_base_help_aggregate\" to 0 to see all help / error messages\n-## [profitbricks-build5-amd64:06528] 1 more process has sent help message help-mpi-runtime.txt / mpi_init:startup:internal-failure\n-## [profitbricks-build5-amd64:06528] 1 more process has sent help message help-mpi-errors.txt / mpi_errors_are_fatal unknown handle\n-## [profitbricks-build5-amd64:06528] 1 more process has sent help message help-mpi-errors.txt / mpi_errors_are_fatal\n+## [i-capture-the-hostname:06230] [[39594,1],1] ORTE_ERROR_LOG: Unreachable in file dpm/dpm.c at line 433\n+## [i-capture-the-hostname:06352] [[39594,2],0] ORTE_ERROR_LOG: Unreachable in file dpm/dpm.c at line 433\n+## [i-capture-the-hostname:06224] [[39594,1],0] ORTE_ERROR_LOG: Unreachable in file dpm/dpm.c at line 433\n+## [i-capture-the-hostname:06230] *** An error occurred in MPI_Comm_spawn_multiple\n+## [i-capture-the-hostname:06230] *** reported by process [2594832385,1]\n+## [i-capture-the-hostname:06230] *** on communicator MPI_COMM_WORLD\n+## [i-capture-the-hostname:06230] *** MPI_ERR_INTERN: internal error\n+## [i-capture-the-hostname:06230] *** MPI_ERRORS_ARE_FATAL (processes in this communicator will now abort,\n+## [i-capture-the-hostname:06230] *** and potentially your MPI job)\n+## [i-capture-the-hostname:06358] [[39594,2],1] ORTE_ERROR_LOG: Unreachable in file dpm/dpm.c at line 433\n+## [i-capture-the-hostname:06352] *** An error occurred in MPI_Init\n+## [i-capture-the-hostname:06352] *** reported by process [2594832386,0]\n+## [i-capture-the-hostname:06352] *** on a NULL communicator\n+## [i-capture-the-hostname:06352] *** Unknown error\n+## [i-capture-the-hostname:06352] *** MPI_ERRORS_ARE_FATAL (processes in this communicator will now abort,\n+## [i-capture-the-hostname:06352] *** and potentially your MPI job)\n+## [i-capture-the-hostname:06139] 3 more processes have sent help message help-mca-bml-r2.txt / unreachable proc\n+## [i-capture-the-hostname:06139] Set MCA parameter \"orte_base_help_aggregate\" to 0 to see all help / error messages\n+## [i-capture-the-hostname:06139] 1 more process has sent help message help-mpi-runtime.txt / mpi_init:startup:internal-failure\n+## [i-capture-the-hostname:06139] 1 more process has sent help message help-mpi-errors.txt / mpi_errors_are_fatal\n+## [i-capture-the-hostname:06139] 1 more process has sent help message help-mpi-errors.txt / mpi_errors_are_fatal unknown handle\n ]]> \n \n \n- \n+ \n+ \n+ \n+ \n \n \n \n \n \n \n \n \n \n- \n- \n- \n+ \n+ \n+ \n+ \n+ \n+ \n+ \n+ \n+ \n+ \n+ \n+ \n \n \n \n- \n+ \n+ \n+ \n+ \n \n \n \n \n \n \n \n \n \n \n \n \n message="requires MPI version 3.1">\n@@ -999694,15 +999797,15 @@\n \n \n \n \n \n@@ -999727,99 +999830,99 @@\n \n \n \n \n \n \n \n \n \n \n \n datatype, &size );\n ## ^\n ## In file included from ../../../cxx/util/mtest.cxx:7:\n ## /usr/lib/x86_64-linux-gnu/openmpi/include/mpi.h:1758:20: note: declared here\n@@ -999851,34 +999954,34 @@\n ## ../../../cxx/util/mtest.cxx:848:50: error: 'NULL_COPY_FN' is not a member of 'MPI::Win'\n ## mem_keyval = MPI::Win::Create_keyval( MPI::Win::NULL_COPY_FN,\n ## ^~~~~~~~~~~~\n ## ../../../cxx/util/mtest.cxx:849:22: error: 'NULL_DELETE_FN' is not a member of 'MPI::Win'\n ## MPI::Win::NULL_DELETE_FN, 0 );\n ## ^~~~~~~~~~~~~~\n ## make[3]: *** [Makefile:416: mtest.o] Error 1\n-## make[3]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/util'\n+## make[3]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/util'\n ## make[2]: *** [Makefile:695: ../../cxx/util/mtest.o] Error 2\n-## make[2]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/attr'\n+## make[2]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/attr'\n ]]> \n \n \n \n datatype, &size );\n ## ^\n ## In file included from ../../../cxx/util/mtest.cxx:7:\n ## /usr/lib/x86_64-linux-gnu/openmpi/include/mpi.h:1758:20: note: declared here\n@@ -999910,34 +1000013,34 @@\n ## ../../../cxx/util/mtest.cxx:848:50: error: 'NULL_COPY_FN' is not a member of 'MPI::Win'\n ## mem_keyval = MPI::Win::Create_keyval( MPI::Win::NULL_COPY_FN,\n ## ^~~~~~~~~~~~\n ## ../../../cxx/util/mtest.cxx:849:22: error: 'NULL_DELETE_FN' is not a member of 'MPI::Win'\n ## MPI::Win::NULL_DELETE_FN, 0 );\n ## ^~~~~~~~~~~~~~\n ## make[3]: *** [Makefile:416: mtest.o] Error 1\n-## make[3]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/util'\n+## make[3]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/util'\n ## make[2]: *** [Makefile:695: ../../cxx/util/mtest.o] Error 2\n-## make[2]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/attr'\n+## make[2]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/attr'\n ]]> \n \n \n \n datatype, &size );\n ## ^\n ## In file included from ../../../cxx/util/mtest.cxx:7:\n ## /usr/lib/x86_64-linux-gnu/openmpi/include/mpi.h:1758:20: note: declared here\n@@ -999969,34 +1000072,34 @@\n ## ../../../cxx/util/mtest.cxx:848:50: error: 'NULL_COPY_FN' is not a member of 'MPI::Win'\n ## mem_keyval = MPI::Win::Create_keyval( MPI::Win::NULL_COPY_FN,\n ## ^~~~~~~~~~~~\n ## ../../../cxx/util/mtest.cxx:849:22: error: 'NULL_DELETE_FN' is not a member of 'MPI::Win'\n ## MPI::Win::NULL_DELETE_FN, 0 );\n ## ^~~~~~~~~~~~~~\n ## make[3]: *** [Makefile:416: mtest.o] Error 1\n-## make[3]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/util'\n+## make[3]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/util'\n ## make[2]: *** [Makefile:695: ../../cxx/util/mtest.o] Error 2\n-## make[2]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/attr'\n+## make[2]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/attr'\n ]]> \n \n \n \n datatype, &size );\n ## ^\n ## In file included from ../../../cxx/util/mtest.cxx:7:\n ## /usr/lib/x86_64-linux-gnu/openmpi/include/mpi.h:1758:20: note: declared here\n@@ -1000028,32 +1000131,32 @@\n ## ../../../cxx/util/mtest.cxx:848:50: error: 'NULL_COPY_FN' is not a member of 'MPI::Win'\n ## mem_keyval = MPI::Win::Create_keyval( MPI::Win::NULL_COPY_FN,\n ## ^~~~~~~~~~~~\n ## ../../../cxx/util/mtest.cxx:849:22: error: 'NULL_DELETE_FN' is not a member of 'MPI::Win'\n ## MPI::Win::NULL_DELETE_FN, 0 );\n ## ^~~~~~~~~~~~~~\n ## make[3]: *** [Makefile:416: mtest.o] Error 1\n-## make[3]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/util'\n+## make[3]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/util'\n ## make[2]: *** [Makefile:695: ../../cxx/util/mtest.o] Error 2\n-## make[2]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/attr'\n+## make[2]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/attr'\n ]]> \n \n \n \n \n \n \n \n datatype, &size );\n ## ^\n ## In file included from ../../../cxx/util/mtest.cxx:7:\n ## /usr/lib/x86_64-linux-gnu/openmpi/include/mpi.h:1758:20: note: declared here\n@@ -1000155,34 +1000258,34 @@\n ## ../../../cxx/util/mtest.cxx:848:50: error: 'NULL_COPY_FN' is not a member of 'MPI::Win'\n ## mem_keyval = MPI::Win::Create_keyval( MPI::Win::NULL_COPY_FN,\n ## ^~~~~~~~~~~~\n ## ../../../cxx/util/mtest.cxx:849:22: error: 'NULL_DELETE_FN' is not a member of 'MPI::Win'\n ## MPI::Win::NULL_DELETE_FN, 0 );\n ## ^~~~~~~~~~~~~~\n ## make[3]: *** [Makefile:416: mtest.o] Error 1\n-## make[3]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/util'\n+## make[3]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/util'\n ## make[2]: *** [Makefile:650: ../../cxx/util/mtest.o] Error 2\n-## make[2]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/pt2pt'\n+## make[2]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/pt2pt'\n ]]> \n \n \n \n datatype, &size );\n ## ^\n ## In file included from ../../../cxx/util/mtest.cxx:7:\n ## /usr/lib/x86_64-linux-gnu/openmpi/include/mpi.h:1758:20: note: declared here\n@@ -1000214,34 +1000317,34 @@\n ## ../../../cxx/util/mtest.cxx:848:50: error: 'NULL_COPY_FN' is not a member of 'MPI::Win'\n ## mem_keyval = MPI::Win::Create_keyval( MPI::Win::NULL_COPY_FN,\n ## ^~~~~~~~~~~~\n ## ../../../cxx/util/mtest.cxx:849:22: error: 'NULL_DELETE_FN' is not a member of 'MPI::Win'\n ## MPI::Win::NULL_DELETE_FN, 0 );\n ## ^~~~~~~~~~~~~~\n ## make[3]: *** [Makefile:416: mtest.o] Error 1\n-## make[3]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/util'\n+## make[3]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/util'\n ## make[2]: *** [Makefile:650: ../../cxx/util/mtest.o] Error 2\n-## make[2]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/pt2pt'\n+## make[2]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/pt2pt'\n ]]> \n \n \n \n datatype, &size );\n ## ^\n ## In file included from ../../../cxx/util/mtest.cxx:7:\n ## /usr/lib/x86_64-linux-gnu/openmpi/include/mpi.h:1758:20: note: declared here\n@@ -1000273,34 +1000376,34 @@\n ## ../../../cxx/util/mtest.cxx:848:50: error: 'NULL_COPY_FN' is not a member of 'MPI::Win'\n ## mem_keyval = MPI::Win::Create_keyval( MPI::Win::NULL_COPY_FN,\n ## ^~~~~~~~~~~~\n ## ../../../cxx/util/mtest.cxx:849:22: error: 'NULL_DELETE_FN' is not a member of 'MPI::Win'\n ## MPI::Win::NULL_DELETE_FN, 0 );\n ## ^~~~~~~~~~~~~~\n ## make[3]: *** [Makefile:416: mtest.o] Error 1\n-## make[3]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/util'\n+## make[3]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/util'\n ## make[2]: *** [Makefile:640: ../../cxx/util/mtest.o] Error 2\n-## make[2]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/comm'\n+## make[2]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/comm'\n ]]> \n \n \n \n datatype, &size );\n ## ^\n ## In file included from ../../../cxx/util/mtest.cxx:7:\n ## /usr/lib/x86_64-linux-gnu/openmpi/include/mpi.h:1758:20: note: declared here\n@@ -1000332,34 +1000435,34 @@\n ## ../../../cxx/util/mtest.cxx:848:50: error: 'NULL_COPY_FN' is not a member of 'MPI::Win'\n ## mem_keyval = MPI::Win::Create_keyval( MPI::Win::NULL_COPY_FN,\n ## ^~~~~~~~~~~~\n ## ../../../cxx/util/mtest.cxx:849:22: error: 'NULL_DELETE_FN' is not a member of 'MPI::Win'\n ## MPI::Win::NULL_DELETE_FN, 0 );\n ## ^~~~~~~~~~~~~~\n ## make[3]: *** [Makefile:416: mtest.o] Error 1\n-## make[3]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/util'\n+## make[3]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/util'\n ## make[2]: *** [Makefile:865: ../../cxx/util/mtest.o] Error 2\n-## make[2]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/coll'\n+## make[2]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/coll'\n ]]> \n \n \n \n datatype, &size );\n ## ^\n ## In file included from ../../../cxx/util/mtest.cxx:7:\n ## /usr/lib/x86_64-linux-gnu/openmpi/include/mpi.h:1758:20: note: declared here\n@@ -1000391,34 +1000494,34 @@\n ## ../../../cxx/util/mtest.cxx:848:50: error: 'NULL_COPY_FN' is not a member of 'MPI::Win'\n ## mem_keyval = MPI::Win::Create_keyval( MPI::Win::NULL_COPY_FN,\n ## ^~~~~~~~~~~~\n ## ../../../cxx/util/mtest.cxx:849:22: error: 'NULL_DELETE_FN' is not a member of 'MPI::Win'\n ## MPI::Win::NULL_DELETE_FN, 0 );\n ## ^~~~~~~~~~~~~~\n ## make[3]: *** [Makefile:416: mtest.o] Error 1\n-## make[3]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/util'\n+## make[3]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/util'\n ## make[2]: *** [Makefile:865: ../../cxx/util/mtest.o] Error 2\n-## make[2]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/coll'\n+## make[2]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/coll'\n ]]> \n \n \n \n datatype, &size );\n ## ^\n ## In file included from ../../../cxx/util/mtest.cxx:7:\n ## /usr/lib/x86_64-linux-gnu/openmpi/include/mpi.h:1758:20: note: declared here\n@@ -1000450,34 +1000553,34 @@\n ## ../../../cxx/util/mtest.cxx:848:50: error: 'NULL_COPY_FN' is not a member of 'MPI::Win'\n ## mem_keyval = MPI::Win::Create_keyval( MPI::Win::NULL_COPY_FN,\n ## ^~~~~~~~~~~~\n ## ../../../cxx/util/mtest.cxx:849:22: error: 'NULL_DELETE_FN' is not a member of 'MPI::Win'\n ## MPI::Win::NULL_DELETE_FN, 0 );\n ## ^~~~~~~~~~~~~~\n ## make[3]: *** [Makefile:416: mtest.o] Error 1\n-## make[3]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/util'\n+## make[3]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/util'\n ## make[2]: *** [Makefile:865: ../../cxx/util/mtest.o] Error 2\n-## make[2]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/coll'\n+## make[2]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/coll'\n ]]> \n \n \n \n datatype, &size );\n ## ^\n ## In file included from ../../../cxx/util/mtest.cxx:7:\n ## /usr/lib/x86_64-linux-gnu/openmpi/include/mpi.h:1758:20: note: declared here\n@@ -1000509,34 +1000612,34 @@\n ## ../../../cxx/util/mtest.cxx:848:50: error: 'NULL_COPY_FN' is not a member of 'MPI::Win'\n ## mem_keyval = MPI::Win::Create_keyval( MPI::Win::NULL_COPY_FN,\n ## ^~~~~~~~~~~~\n ## ../../../cxx/util/mtest.cxx:849:22: error: 'NULL_DELETE_FN' is not a member of 'MPI::Win'\n ## MPI::Win::NULL_DELETE_FN, 0 );\n ## ^~~~~~~~~~~~~~\n ## make[3]: *** [Makefile:416: mtest.o] Error 1\n-## make[3]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/util'\n+## make[3]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/util'\n ## make[2]: *** [Makefile:865: ../../cxx/util/mtest.o] Error 2\n-## make[2]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/coll'\n+## make[2]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/coll'\n ]]> \n \n \n \n datatype, &size );\n ## ^\n ## In file included from ../../../cxx/util/mtest.cxx:7:\n ## /usr/lib/x86_64-linux-gnu/openmpi/include/mpi.h:1758:20: note: declared here\n@@ -1000568,34 +1000671,34 @@\n ## ../../../cxx/util/mtest.cxx:848:50: error: 'NULL_COPY_FN' is not a member of 'MPI::Win'\n ## mem_keyval = MPI::Win::Create_keyval( MPI::Win::NULL_COPY_FN,\n ## ^~~~~~~~~~~~\n ## ../../../cxx/util/mtest.cxx:849:22: error: 'NULL_DELETE_FN' is not a member of 'MPI::Win'\n ## MPI::Win::NULL_DELETE_FN, 0 );\n ## ^~~~~~~~~~~~~~\n ## make[3]: *** [Makefile:416: mtest.o] Error 1\n-## make[3]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/util'\n+## make[3]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/util'\n ## make[2]: *** [Makefile:865: ../../cxx/util/mtest.o] Error 2\n-## make[2]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/coll'\n+## make[2]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/coll'\n ]]> \n \n \n \n datatype, &size );\n ## ^\n ## In file included from ../../../cxx/util/mtest.cxx:7:\n ## /usr/lib/x86_64-linux-gnu/openmpi/include/mpi.h:1758:20: note: declared here\n@@ -1000627,34 +1000730,34 @@\n ## ../../../cxx/util/mtest.cxx:848:50: error: 'NULL_COPY_FN' is not a member of 'MPI::Win'\n ## mem_keyval = MPI::Win::Create_keyval( MPI::Win::NULL_COPY_FN,\n ## ^~~~~~~~~~~~\n ## ../../../cxx/util/mtest.cxx:849:22: error: 'NULL_DELETE_FN' is not a member of 'MPI::Win'\n ## MPI::Win::NULL_DELETE_FN, 0 );\n ## ^~~~~~~~~~~~~~\n ## make[3]: *** [Makefile:416: mtest.o] Error 1\n-## make[3]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/util'\n+## make[3]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/util'\n ## make[2]: *** [Makefile:865: ../../cxx/util/mtest.o] Error 2\n-## make[2]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/coll'\n+## make[2]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/coll'\n ]]> \n \n \n \n datatype, &size );\n ## ^\n ## In file included from ../../../cxx/util/mtest.cxx:7:\n ## /usr/lib/x86_64-linux-gnu/openmpi/include/mpi.h:1758:20: note: declared here\n@@ -1000686,34 +1000789,34 @@\n ## ../../../cxx/util/mtest.cxx:848:50: error: 'NULL_COPY_FN' is not a member of 'MPI::Win'\n ## mem_keyval = MPI::Win::Create_keyval( MPI::Win::NULL_COPY_FN,\n ## ^~~~~~~~~~~~\n ## ../../../cxx/util/mtest.cxx:849:22: error: 'NULL_DELETE_FN' is not a member of 'MPI::Win'\n ## MPI::Win::NULL_DELETE_FN, 0 );\n ## ^~~~~~~~~~~~~~\n ## make[3]: *** [Makefile:416: mtest.o] Error 1\n-## make[3]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/util'\n+## make[3]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/util'\n ## make[2]: *** [Makefile:865: ../../cxx/util/mtest.o] Error 2\n-## make[2]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/coll'\n+## make[2]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/coll'\n ]]> \n \n \n \n datatype, &size );\n ## ^\n ## In file included from ../../../cxx/util/mtest.cxx:7:\n ## /usr/lib/x86_64-linux-gnu/openmpi/include/mpi.h:1758:20: note: declared here\n@@ -1000745,34 +1000848,34 @@\n ## ../../../cxx/util/mtest.cxx:848:50: error: 'NULL_COPY_FN' is not a member of 'MPI::Win'\n ## mem_keyval = MPI::Win::Create_keyval( MPI::Win::NULL_COPY_FN,\n ## ^~~~~~~~~~~~\n ## ../../../cxx/util/mtest.cxx:849:22: error: 'NULL_DELETE_FN' is not a member of 'MPI::Win'\n ## MPI::Win::NULL_DELETE_FN, 0 );\n ## ^~~~~~~~~~~~~~\n ## make[3]: *** [Makefile:416: mtest.o] Error 1\n-## make[3]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/util'\n+## make[3]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/util'\n ## make[2]: *** [Makefile:865: ../../cxx/util/mtest.o] Error 2\n-## make[2]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/coll'\n+## make[2]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/coll'\n ]]> \n \n \n \n datatype, &size );\n ## ^\n ## In file included from ../../../cxx/util/mtest.cxx:7:\n ## /usr/lib/x86_64-linux-gnu/openmpi/include/mpi.h:1758:20: note: declared here\n@@ -1000804,37 +1000907,37 @@\n ## ../../../cxx/util/mtest.cxx:848:50: error: 'NULL_COPY_FN' is not a member of 'MPI::Win'\n ## mem_keyval = MPI::Win::Create_keyval( MPI::Win::NULL_COPY_FN,\n ## ^~~~~~~~~~~~\n ## ../../../cxx/util/mtest.cxx:849:22: error: 'NULL_DELETE_FN' is not a member of 'MPI::Win'\n ## MPI::Win::NULL_DELETE_FN, 0 );\n ## ^~~~~~~~~~~~~~\n ## make[3]: *** [Makefile:416: mtest.o] Error 1\n-## make[3]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/util'\n+## make[3]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/util'\n ## make[2]: *** [Makefile:865: ../../cxx/util/mtest.o] Error 2\n-## make[2]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/coll'\n+## make[2]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/coll'\n ]]> \n \n \n \n datatype, &size );\n ## ^\n ## In file included from ../../../cxx/util/mtest.cxx:7:\n ## /usr/lib/x86_64-linux-gnu/openmpi/include/mpi.h:1758:20: note: declared here\n@@ -1000866,34 +1000969,34 @@\n ## ../../../cxx/util/mtest.cxx:848:50: error: 'NULL_COPY_FN' is not a member of 'MPI::Win'\n ## mem_keyval = MPI::Win::Create_keyval( MPI::Win::NULL_COPY_FN,\n ## ^~~~~~~~~~~~\n ## ../../../cxx/util/mtest.cxx:849:22: error: 'NULL_DELETE_FN' is not a member of 'MPI::Win'\n ## MPI::Win::NULL_DELETE_FN, 0 );\n ## ^~~~~~~~~~~~~~\n ## make[3]: *** [Makefile:416: mtest.o] Error 1\n-## make[3]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/util'\n+## make[3]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/util'\n ## make[2]: *** [Makefile:865: ../../cxx/util/mtest.o] Error 2\n-## make[2]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/coll'\n+## make[2]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/coll'\n ]]> \n \n \n \n datatype, &size );\n ## ^\n ## In file included from ../../../cxx/util/mtest.cxx:7:\n ## /usr/lib/x86_64-linux-gnu/openmpi/include/mpi.h:1758:20: note: declared here\n@@ -1000925,34 +1001028,34 @@\n ## ../../../cxx/util/mtest.cxx:848:50: error: 'NULL_COPY_FN' is not a member of 'MPI::Win'\n ## mem_keyval = MPI::Win::Create_keyval( MPI::Win::NULL_COPY_FN,\n ## ^~~~~~~~~~~~\n ## ../../../cxx/util/mtest.cxx:849:22: error: 'NULL_DELETE_FN' is not a member of 'MPI::Win'\n ## MPI::Win::NULL_DELETE_FN, 0 );\n ## ^~~~~~~~~~~~~~\n ## make[3]: *** [Makefile:416: mtest.o] Error 1\n-## make[3]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/util'\n+## make[3]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/util'\n ## make[2]: *** [Makefile:865: ../../cxx/util/mtest.o] Error 2\n-## make[2]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/coll'\n+## make[2]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/coll'\n ]]> \n \n \n \n datatype, &size );\n ## ^\n ## In file included from ../../../cxx/util/mtest.cxx:7:\n ## /usr/lib/x86_64-linux-gnu/openmpi/include/mpi.h:1758:20: note: declared here\n@@ -1000984,34 +1001087,34 @@\n ## ../../../cxx/util/mtest.cxx:848:50: error: 'NULL_COPY_FN' is not a member of 'MPI::Win'\n ## mem_keyval = MPI::Win::Create_keyval( MPI::Win::NULL_COPY_FN,\n ## ^~~~~~~~~~~~\n ## ../../../cxx/util/mtest.cxx:849:22: error: 'NULL_DELETE_FN' is not a member of 'MPI::Win'\n ## MPI::Win::NULL_DELETE_FN, 0 );\n ## ^~~~~~~~~~~~~~\n ## make[3]: *** [Makefile:416: mtest.o] Error 1\n-## make[3]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/util'\n+## make[3]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/util'\n ## make[2]: *** [Makefile:865: ../../cxx/util/mtest.o] Error 2\n-## make[2]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/coll'\n+## make[2]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/coll'\n ]]> \n \n \n \n datatype, &size );\n ## ^\n ## In file included from ../../../cxx/util/mtest.cxx:7:\n ## /usr/lib/x86_64-linux-gnu/openmpi/include/mpi.h:1758:20: note: declared here\n@@ -1001043,34 +1001146,34 @@\n ## ../../../cxx/util/mtest.cxx:848:50: error: 'NULL_COPY_FN' is not a member of 'MPI::Win'\n ## mem_keyval = MPI::Win::Create_keyval( MPI::Win::NULL_COPY_FN,\n ## ^~~~~~~~~~~~\n ## ../../../cxx/util/mtest.cxx:849:22: error: 'NULL_DELETE_FN' is not a member of 'MPI::Win'\n ## MPI::Win::NULL_DELETE_FN, 0 );\n ## ^~~~~~~~~~~~~~\n ## make[3]: *** [Makefile:416: mtest.o] Error 1\n-## make[3]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/util'\n+## make[3]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/util'\n ## make[2]: *** [Makefile:865: ../../cxx/util/mtest.o] Error 2\n-## make[2]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/coll'\n+## make[2]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/coll'\n ]]> \n \n \n \n datatype, &size );\n ## ^\n ## In file included from ../../../cxx/util/mtest.cxx:7:\n ## /usr/lib/x86_64-linux-gnu/openmpi/include/mpi.h:1758:20: note: declared here\n@@ -1001102,34 +1001205,34 @@\n ## ../../../cxx/util/mtest.cxx:848:50: error: 'NULL_COPY_FN' is not a member of 'MPI::Win'\n ## mem_keyval = MPI::Win::Create_keyval( MPI::Win::NULL_COPY_FN,\n ## ^~~~~~~~~~~~\n ## ../../../cxx/util/mtest.cxx:849:22: error: 'NULL_DELETE_FN' is not a member of 'MPI::Win'\n ## MPI::Win::NULL_DELETE_FN, 0 );\n ## ^~~~~~~~~~~~~~\n ## make[3]: *** [Makefile:416: mtest.o] Error 1\n-## make[3]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/util'\n+## make[3]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/util'\n ## make[2]: *** [Makefile:865: ../../cxx/util/mtest.o] Error 2\n-## make[2]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/coll'\n+## make[2]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/coll'\n ]]> \n \n \n \n datatype, &size );\n ## ^\n ## In file included from ../../../cxx/util/mtest.cxx:7:\n ## /usr/lib/x86_64-linux-gnu/openmpi/include/mpi.h:1758:20: note: declared here\n@@ -1001161,34 +1001264,34 @@\n ## ../../../cxx/util/mtest.cxx:848:50: error: 'NULL_COPY_FN' is not a member of 'MPI::Win'\n ## mem_keyval = MPI::Win::Create_keyval( MPI::Win::NULL_COPY_FN,\n ## ^~~~~~~~~~~~\n ## ../../../cxx/util/mtest.cxx:849:22: error: 'NULL_DELETE_FN' is not a member of 'MPI::Win'\n ## MPI::Win::NULL_DELETE_FN, 0 );\n ## ^~~~~~~~~~~~~~\n ## make[3]: *** [Makefile:416: mtest.o] Error 1\n-## make[3]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/util'\n+## make[3]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/util'\n ## make[2]: *** [Makefile:865: ../../cxx/util/mtest.o] Error 2\n-## make[2]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/coll'\n+## make[2]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/coll'\n ]]> \n \n \n \n datatype, &size );\n ## ^\n ## In file included from ../../../cxx/util/mtest.cxx:7:\n ## /usr/lib/x86_64-linux-gnu/openmpi/include/mpi.h:1758:20: note: declared here\n@@ -1001220,34 +1001323,34 @@\n ## ../../../cxx/util/mtest.cxx:848:50: error: 'NULL_COPY_FN' is not a member of 'MPI::Win'\n ## mem_keyval = MPI::Win::Create_keyval( MPI::Win::NULL_COPY_FN,\n ## ^~~~~~~~~~~~\n ## ../../../cxx/util/mtest.cxx:849:22: error: 'NULL_DELETE_FN' is not a member of 'MPI::Win'\n ## MPI::Win::NULL_DELETE_FN, 0 );\n ## ^~~~~~~~~~~~~~\n ## make[3]: *** [Makefile:416: mtest.o] Error 1\n-## make[3]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/util'\n+## make[3]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/util'\n ## make[2]: *** [Makefile:865: ../../cxx/util/mtest.o] Error 2\n-## make[2]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/coll'\n+## make[2]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/coll'\n ]]> \n \n \n \n datatype, &size );\n ## ^\n ## In file included from ../../../cxx/util/mtest.cxx:7:\n ## /usr/lib/x86_64-linux-gnu/openmpi/include/mpi.h:1758:20: note: declared here\n@@ -1001279,34 +1001382,34 @@\n ## ../../../cxx/util/mtest.cxx:848:50: error: 'NULL_COPY_FN' is not a member of 'MPI::Win'\n ## mem_keyval = MPI::Win::Create_keyval( MPI::Win::NULL_COPY_FN,\n ## ^~~~~~~~~~~~\n ## ../../../cxx/util/mtest.cxx:849:22: error: 'NULL_DELETE_FN' is not a member of 'MPI::Win'\n ## MPI::Win::NULL_DELETE_FN, 0 );\n ## ^~~~~~~~~~~~~~\n ## make[3]: *** [Makefile:416: mtest.o] Error 1\n-## make[3]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/util'\n+## make[3]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/util'\n ## make[2]: *** [Makefile:865: ../../cxx/util/mtest.o] Error 2\n-## make[2]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/coll'\n+## make[2]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/coll'\n ]]> \n \n \n \n datatype, &size );\n ## ^\n ## In file included from ../../../cxx/util/mtest.cxx:7:\n ## /usr/lib/x86_64-linux-gnu/openmpi/include/mpi.h:1758:20: note: declared here\n@@ -1001338,34 +1001441,34 @@\n ## ../../../cxx/util/mtest.cxx:848:50: error: 'NULL_COPY_FN' is not a member of 'MPI::Win'\n ## mem_keyval = MPI::Win::Create_keyval( MPI::Win::NULL_COPY_FN,\n ## ^~~~~~~~~~~~\n ## ../../../cxx/util/mtest.cxx:849:22: error: 'NULL_DELETE_FN' is not a member of 'MPI::Win'\n ## MPI::Win::NULL_DELETE_FN, 0 );\n ## ^~~~~~~~~~~~~~\n ## make[3]: *** [Makefile:416: mtest.o] Error 1\n-## make[3]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/util'\n+## make[3]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/util'\n ## make[2]: *** [Makefile:865: ../../cxx/util/mtest.o] Error 2\n-## make[2]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/coll'\n+## make[2]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/coll'\n ]]> \n \n \n \n datatype, &size );\n ## ^\n ## In file included from ../../../cxx/util/mtest.cxx:7:\n ## /usr/lib/x86_64-linux-gnu/openmpi/include/mpi.h:1758:20: note: declared here\n@@ -1001397,34 +1001500,34 @@\n ## ../../../cxx/util/mtest.cxx:848:50: error: 'NULL_COPY_FN' is not a member of 'MPI::Win'\n ## mem_keyval = MPI::Win::Create_keyval( MPI::Win::NULL_COPY_FN,\n ## ^~~~~~~~~~~~\n ## ../../../cxx/util/mtest.cxx:849:22: error: 'NULL_DELETE_FN' is not a member of 'MPI::Win'\n ## MPI::Win::NULL_DELETE_FN, 0 );\n ## ^~~~~~~~~~~~~~\n ## make[3]: *** [Makefile:416: mtest.o] Error 1\n-## make[3]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/util'\n+## make[3]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/util'\n ## make[2]: *** [Makefile:865: ../../cxx/util/mtest.o] Error 2\n-## make[2]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/coll'\n+## make[2]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/coll'\n ]]> \n \n \n \n datatype, &size );\n ## ^\n ## In file included from ../../../cxx/util/mtest.cxx:7:\n ## /usr/lib/x86_64-linux-gnu/openmpi/include/mpi.h:1758:20: note: declared here\n@@ -1001456,34 +1001559,34 @@\n ## ../../../cxx/util/mtest.cxx:848:50: error: 'NULL_COPY_FN' is not a member of 'MPI::Win'\n ## mem_keyval = MPI::Win::Create_keyval( MPI::Win::NULL_COPY_FN,\n ## ^~~~~~~~~~~~\n ## ../../../cxx/util/mtest.cxx:849:22: error: 'NULL_DELETE_FN' is not a member of 'MPI::Win'\n ## MPI::Win::NULL_DELETE_FN, 0 );\n ## ^~~~~~~~~~~~~~\n ## make[3]: *** [Makefile:416: mtest.o] Error 1\n-## make[3]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/util'\n+## make[3]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/util'\n ## make[2]: *** [Makefile:865: ../../cxx/util/mtest.o] Error 2\n-## make[2]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/coll'\n+## make[2]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/coll'\n ]]> \n \n \n \n datatype, &size );\n ## ^\n ## In file included from ../../../cxx/util/mtest.cxx:7:\n ## /usr/lib/x86_64-linux-gnu/openmpi/include/mpi.h:1758:20: note: declared here\n@@ -1001515,61 +1001618,61 @@\n ## ../../../cxx/util/mtest.cxx:848:50: error: 'NULL_COPY_FN' is not a member of 'MPI::Win'\n ## mem_keyval = MPI::Win::Create_keyval( MPI::Win::NULL_COPY_FN,\n ## ^~~~~~~~~~~~\n ## ../../../cxx/util/mtest.cxx:849:22: error: 'NULL_DELETE_FN' is not a member of 'MPI::Win'\n ## MPI::Win::NULL_DELETE_FN, 0 );\n ## ^~~~~~~~~~~~~~\n ## make[3]: *** [Makefile:416: mtest.o] Error 1\n-## make[3]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/util'\n+## make[3]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/util'\n ## make[2]: *** [Makefile:865: ../../cxx/util/mtest.o] Error 2\n-## make[2]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/coll'\n+## make[2]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/coll'\n ]]> \n \n \n \n \n \n \n \n datatype, &size );\n ## ^\n ## In file included from ../../../cxx/util/mtest.cxx:7:\n ## /usr/lib/x86_64-linux-gnu/openmpi/include/mpi.h:1758:20: note: declared here\n@@ -1001601,34 +1001704,34 @@\n ## ../../../cxx/util/mtest.cxx:848:50: error: 'NULL_COPY_FN' is not a member of 'MPI::Win'\n ## mem_keyval = MPI::Win::Create_keyval( MPI::Win::NULL_COPY_FN,\n ## ^~~~~~~~~~~~\n ## ../../../cxx/util/mtest.cxx:849:22: error: 'NULL_DELETE_FN' is not a member of 'MPI::Win'\n ## MPI::Win::NULL_DELETE_FN, 0 );\n ## ^~~~~~~~~~~~~~\n ## make[3]: *** [Makefile:416: mtest.o] Error 1\n-## make[3]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/util'\n+## make[3]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/util'\n ## make[2]: *** [Makefile:640: ../../cxx/util/mtest.o] Error 2\n-## make[2]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/errhan'\n+## make[2]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/errhan'\n ]]> \n \n \n \n datatype, &size );\n ## ^\n ## In file included from ../../../cxx/util/mtest.cxx:7:\n ## /usr/lib/x86_64-linux-gnu/openmpi/include/mpi.h:1758:20: note: declared here\n@@ -1001660,34 +1001763,34 @@\n ## ../../../cxx/util/mtest.cxx:848:50: error: 'NULL_COPY_FN' is not a member of 'MPI::Win'\n ## mem_keyval = MPI::Win::Create_keyval( MPI::Win::NULL_COPY_FN,\n ## ^~~~~~~~~~~~\n ## ../../../cxx/util/mtest.cxx:849:22: error: 'NULL_DELETE_FN' is not a member of 'MPI::Win'\n ## MPI::Win::NULL_DELETE_FN, 0 );\n ## ^~~~~~~~~~~~~~\n ## make[3]: *** [Makefile:416: mtest.o] Error 1\n-## make[3]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/util'\n+## make[3]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/util'\n ## make[2]: *** [Makefile:663: ../../cxx/util/mtest.o] Error 2\n-## make[2]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/init'\n+## make[2]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/init'\n ]]> \n \n \n \n datatype, &size );\n ## ^\n ## In file included from ../../../cxx/util/mtest.cxx:7:\n ## /usr/lib/x86_64-linux-gnu/openmpi/include/mpi.h:1758:20: note: declared here\n@@ -1001719,34 +1001822,34 @@\n ## ../../../cxx/util/mtest.cxx:848:50: error: 'NULL_COPY_FN' is not a member of 'MPI::Win'\n ## mem_keyval = MPI::Win::Create_keyval( MPI::Win::NULL_COPY_FN,\n ## ^~~~~~~~~~~~\n ## ../../../cxx/util/mtest.cxx:849:22: error: 'NULL_DELETE_FN' is not a member of 'MPI::Win'\n ## MPI::Win::NULL_DELETE_FN, 0 );\n ## ^~~~~~~~~~~~~~\n ## make[3]: *** [Makefile:416: mtest.o] Error 1\n-## make[3]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/util'\n+## make[3]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/util'\n ## make[2]: *** [Makefile:663: ../../cxx/util/mtest.o] Error 2\n-## make[2]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/init'\n+## make[2]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/init'\n ]]> \n \n \n \n datatype, &size );\n ## ^\n ## In file included from ../../../cxx/util/mtest.cxx:7:\n ## /usr/lib/x86_64-linux-gnu/openmpi/include/mpi.h:1758:20: note: declared here\n@@ -1001778,34 +1001881,34 @@\n ## ../../../cxx/util/mtest.cxx:848:50: error: 'NULL_COPY_FN' is not a member of 'MPI::Win'\n ## mem_keyval = MPI::Win::Create_keyval( MPI::Win::NULL_COPY_FN,\n ## ^~~~~~~~~~~~\n ## ../../../cxx/util/mtest.cxx:849:22: error: 'NULL_DELETE_FN' is not a member of 'MPI::Win'\n ## MPI::Win::NULL_DELETE_FN, 0 );\n ## ^~~~~~~~~~~~~~\n ## make[3]: *** [Makefile:416: mtest.o] Error 1\n-## make[3]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/util'\n+## make[3]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/util'\n ## make[2]: *** [Makefile:663: ../../cxx/util/mtest.o] Error 2\n-## make[2]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/init'\n+## make[2]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/init'\n ]]> \n \n \n \n datatype, &size );\n ## ^\n ## In file included from ../../../cxx/util/mtest.cxx:7:\n ## /usr/lib/x86_64-linux-gnu/openmpi/include/mpi.h:1758:20: note: declared here\n@@ -1001837,34 +1001940,34 @@\n ## ../../../cxx/util/mtest.cxx:848:50: error: 'NULL_COPY_FN' is not a member of 'MPI::Win'\n ## mem_keyval = MPI::Win::Create_keyval( MPI::Win::NULL_COPY_FN,\n ## ^~~~~~~~~~~~\n ## ../../../cxx/util/mtest.cxx:849:22: error: 'NULL_DELETE_FN' is not a member of 'MPI::Win'\n ## MPI::Win::NULL_DELETE_FN, 0 );\n ## ^~~~~~~~~~~~~~\n ## make[3]: *** [Makefile:416: mtest.o] Error 1\n-## make[3]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/util'\n+## make[3]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/util'\n ## make[2]: *** [Makefile:673: ../../cxx/util/mtest.o] Error 2\n-## make[2]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/info'\n+## make[2]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/info'\n ]]> \n \n \n \n datatype, &size );\n ## ^\n ## In file included from ../../../cxx/util/mtest.cxx:7:\n ## /usr/lib/x86_64-linux-gnu/openmpi/include/mpi.h:1758:20: note: declared here\n@@ -1001896,34 +1001999,34 @@\n ## ../../../cxx/util/mtest.cxx:848:50: error: 'NULL_COPY_FN' is not a member of 'MPI::Win'\n ## mem_keyval = MPI::Win::Create_keyval( MPI::Win::NULL_COPY_FN,\n ## ^~~~~~~~~~~~\n ## ../../../cxx/util/mtest.cxx:849:22: error: 'NULL_DELETE_FN' is not a member of 'MPI::Win'\n ## MPI::Win::NULL_DELETE_FN, 0 );\n ## ^~~~~~~~~~~~~~\n ## make[3]: *** [Makefile:416: mtest.o] Error 1\n-## make[3]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/util'\n+## make[3]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/util'\n ## make[2]: *** [Makefile:673: ../../cxx/util/mtest.o] Error 2\n-## make[2]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/info'\n+## make[2]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/info'\n ]]> \n \n \n \n datatype, &size );\n ## ^\n ## In file included from ../../../cxx/util/mtest.cxx:7:\n ## /usr/lib/x86_64-linux-gnu/openmpi/include/mpi.h:1758:20: note: declared here\n@@ -1001955,34 +1002058,34 @@\n ## ../../../cxx/util/mtest.cxx:848:50: error: 'NULL_COPY_FN' is not a member of 'MPI::Win'\n ## mem_keyval = MPI::Win::Create_keyval( MPI::Win::NULL_COPY_FN,\n ## ^~~~~~~~~~~~\n ## ../../../cxx/util/mtest.cxx:849:22: error: 'NULL_DELETE_FN' is not a member of 'MPI::Win'\n ## MPI::Win::NULL_DELETE_FN, 0 );\n ## ^~~~~~~~~~~~~~\n ## make[3]: *** [Makefile:416: mtest.o] Error 1\n-## make[3]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/util'\n+## make[3]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/util'\n ## make[2]: *** [Makefile:673: ../../cxx/util/mtest.o] Error 2\n-## make[2]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/info'\n+## make[2]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/info'\n ]]> \n \n \n \n datatype, &size );\n ## ^\n ## In file included from ../../../cxx/util/mtest.cxx:7:\n ## /usr/lib/x86_64-linux-gnu/openmpi/include/mpi.h:1758:20: note: declared here\n@@ -1002014,34 +1002117,34 @@\n ## ../../../cxx/util/mtest.cxx:848:50: error: 'NULL_COPY_FN' is not a member of 'MPI::Win'\n ## mem_keyval = MPI::Win::Create_keyval( MPI::Win::NULL_COPY_FN,\n ## ^~~~~~~~~~~~\n ## ../../../cxx/util/mtest.cxx:849:22: error: 'NULL_DELETE_FN' is not a member of 'MPI::Win'\n ## MPI::Win::NULL_DELETE_FN, 0 );\n ## ^~~~~~~~~~~~~~\n ## make[3]: *** [Makefile:416: mtest.o] Error 1\n-## make[3]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/util'\n+## make[3]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/util'\n ## make[2]: *** [Makefile:673: ../../cxx/util/mtest.o] Error 2\n-## make[2]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/info'\n+## make[2]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/info'\n ]]> \n \n \n \n datatype, &size );\n ## ^\n ## In file included from ../../../cxx/util/mtest.cxx:7:\n ## /usr/lib/x86_64-linux-gnu/openmpi/include/mpi.h:1758:20: note: declared here\n@@ -1002073,34 +1002176,34 @@\n ## ../../../cxx/util/mtest.cxx:848:50: error: 'NULL_COPY_FN' is not a member of 'MPI::Win'\n ## mem_keyval = MPI::Win::Create_keyval( MPI::Win::NULL_COPY_FN,\n ## ^~~~~~~~~~~~\n ## ../../../cxx/util/mtest.cxx:849:22: error: 'NULL_DELETE_FN' is not a member of 'MPI::Win'\n ## MPI::Win::NULL_DELETE_FN, 0 );\n ## ^~~~~~~~~~~~~~\n ## make[3]: *** [Makefile:416: mtest.o] Error 1\n-## make[3]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/util'\n+## make[3]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/util'\n ## make[2]: *** [Makefile:673: ../../cxx/util/mtest.o] Error 2\n-## make[2]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/datatype'\n+## make[2]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/datatype'\n ]]> \n \n \n \n datatype, &size );\n ## ^\n ## In file included from ../../../cxx/util/mtest.cxx:7:\n ## /usr/lib/x86_64-linux-gnu/openmpi/include/mpi.h:1758:20: note: declared here\n@@ -1002132,32 +1002235,32 @@\n ## ../../../cxx/util/mtest.cxx:848:50: error: 'NULL_COPY_FN' is not a member of 'MPI::Win'\n ## mem_keyval = MPI::Win::Create_keyval( MPI::Win::NULL_COPY_FN,\n ## ^~~~~~~~~~~~\n ## ../../../cxx/util/mtest.cxx:849:22: error: 'NULL_DELETE_FN' is not a member of 'MPI::Win'\n ## MPI::Win::NULL_DELETE_FN, 0 );\n ## ^~~~~~~~~~~~~~\n ## make[3]: *** [Makefile:416: mtest.o] Error 1\n-## make[3]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/util'\n+## make[3]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/util'\n ## make[2]: *** [Makefile:673: ../../cxx/util/mtest.o] Error 2\n-## make[2]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/datatype'\n+## make[2]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/datatype'\n ]]> \n \n \n \n \n \n \n \n \n \n \n \n datatype, &size );\n ## ^\n ## In file included from ../../../cxx/util/mtest.cxx:7:\n ## /usr/lib/x86_64-linux-gnu/openmpi/include/mpi.h:1758:20: note: declared here\n@@ -1002351,34 +1002454,34 @@\n ## ../../../cxx/util/mtest.cxx:848:50: error: 'NULL_COPY_FN' is not a member of 'MPI::Win'\n ## mem_keyval = MPI::Win::Create_keyval( MPI::Win::NULL_COPY_FN,\n ## ^~~~~~~~~~~~\n ## ../../../cxx/util/mtest.cxx:849:22: error: 'NULL_DELETE_FN' is not a member of 'MPI::Win'\n ## MPI::Win::NULL_DELETE_FN, 0 );\n ## ^~~~~~~~~~~~~~\n ## make[3]: *** [Makefile:416: mtest.o] Error 1\n-## make[3]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/util'\n+## make[3]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/util'\n ## make[2]: *** [Makefile:989: ../../cxx/util/mtest.o] Error 2\n-## make[2]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/io'\n+## make[2]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/io'\n ]]> \n \n \n \n datatype, &size );\n ## ^\n ## In file included from ../../../cxx/util/mtest.cxx:7:\n ## /usr/lib/x86_64-linux-gnu/openmpi/include/mpi.h:1758:20: note: declared here\n@@ -1002410,34 +1002513,34 @@\n ## ../../../cxx/util/mtest.cxx:848:50: error: 'NULL_COPY_FN' is not a member of 'MPI::Win'\n ## mem_keyval = MPI::Win::Create_keyval( MPI::Win::NULL_COPY_FN,\n ## ^~~~~~~~~~~~\n ## ../../../cxx/util/mtest.cxx:849:22: error: 'NULL_DELETE_FN' is not a member of 'MPI::Win'\n ## MPI::Win::NULL_DELETE_FN, 0 );\n ## ^~~~~~~~~~~~~~\n ## make[3]: *** [Makefile:416: mtest.o] Error 1\n-## make[3]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/util'\n+## make[3]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/util'\n ## make[2]: *** [Makefile:989: ../../cxx/util/mtest.o] Error 2\n-## make[2]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/io'\n+## make[2]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/io'\n ]]> \n \n \n \n datatype, &size );\n ## ^\n ## In file included from ../../../cxx/util/mtest.cxx:7:\n ## /usr/lib/x86_64-linux-gnu/openmpi/include/mpi.h:1758:20: note: declared here\n@@ -1002469,34 +1002572,34 @@\n ## ../../../cxx/util/mtest.cxx:848:50: error: 'NULL_COPY_FN' is not a member of 'MPI::Win'\n ## mem_keyval = MPI::Win::Create_keyval( MPI::Win::NULL_COPY_FN,\n ## ^~~~~~~~~~~~\n ## ../../../cxx/util/mtest.cxx:849:22: error: 'NULL_DELETE_FN' is not a member of 'MPI::Win'\n ## MPI::Win::NULL_DELETE_FN, 0 );\n ## ^~~~~~~~~~~~~~\n ## make[3]: *** [Makefile:416: mtest.o] Error 1\n-## make[3]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/util'\n+## make[3]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/util'\n ## make[2]: *** [Makefile:989: ../../cxx/util/mtest.o] Error 2\n-## make[2]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/io'\n+## make[2]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/io'\n ]]> \n \n \n \n datatype, &size );\n ## ^\n ## In file included from ../../../cxx/util/mtest.cxx:7:\n ## /usr/lib/x86_64-linux-gnu/openmpi/include/mpi.h:1758:20: note: declared here\n@@ -1002528,34 +1002631,34 @@\n ## ../../../cxx/util/mtest.cxx:848:50: error: 'NULL_COPY_FN' is not a member of 'MPI::Win'\n ## mem_keyval = MPI::Win::Create_keyval( MPI::Win::NULL_COPY_FN,\n ## ^~~~~~~~~~~~\n ## ../../../cxx/util/mtest.cxx:849:22: error: 'NULL_DELETE_FN' is not a member of 'MPI::Win'\n ## MPI::Win::NULL_DELETE_FN, 0 );\n ## ^~~~~~~~~~~~~~\n ## make[3]: *** [Makefile:416: mtest.o] Error 1\n-## make[3]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/util'\n+## make[3]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/util'\n ## make[2]: *** [Makefile:989: ../../cxx/util/mtest.o] Error 2\n-## make[2]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/io'\n+## make[2]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/io'\n ]]> \n \n \n \n datatype, &size );\n ## ^\n ## In file included from ../../../cxx/util/mtest.cxx:7:\n ## /usr/lib/x86_64-linux-gnu/openmpi/include/mpi.h:1758:20: note: declared here\n@@ -1002587,34 +1002690,34 @@\n ## ../../../cxx/util/mtest.cxx:848:50: error: 'NULL_COPY_FN' is not a member of 'MPI::Win'\n ## mem_keyval = MPI::Win::Create_keyval( MPI::Win::NULL_COPY_FN,\n ## ^~~~~~~~~~~~\n ## ../../../cxx/util/mtest.cxx:849:22: error: 'NULL_DELETE_FN' is not a member of 'MPI::Win'\n ## MPI::Win::NULL_DELETE_FN, 0 );\n ## ^~~~~~~~~~~~~~\n ## make[3]: *** [Makefile:416: mtest.o] Error 1\n-## make[3]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/util'\n+## make[3]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/util'\n ## make[2]: *** [Makefile:989: ../../cxx/util/mtest.o] Error 2\n-## make[2]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/io'\n+## make[2]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/io'\n ]]> \n \n \n \n datatype, &size );\n ## ^\n ## In file included from ../../../cxx/util/mtest.cxx:7:\n ## /usr/lib/x86_64-linux-gnu/openmpi/include/mpi.h:1758:20: note: declared here\n@@ -1002646,34 +1002749,34 @@\n ## ../../../cxx/util/mtest.cxx:848:50: error: 'NULL_COPY_FN' is not a member of 'MPI::Win'\n ## mem_keyval = MPI::Win::Create_keyval( MPI::Win::NULL_COPY_FN,\n ## ^~~~~~~~~~~~\n ## ../../../cxx/util/mtest.cxx:849:22: error: 'NULL_DELETE_FN' is not a member of 'MPI::Win'\n ## MPI::Win::NULL_DELETE_FN, 0 );\n ## ^~~~~~~~~~~~~~\n ## make[3]: *** [Makefile:416: mtest.o] Error 1\n-## make[3]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/util'\n+## make[3]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/util'\n ## make[2]: *** [Makefile:989: ../../cxx/util/mtest.o] Error 2\n-## make[2]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/io'\n+## make[2]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/io'\n ]]> \n \n \n \n datatype, &size );\n ## ^\n ## In file included from ../../../cxx/util/mtest.cxx:7:\n ## /usr/lib/x86_64-linux-gnu/openmpi/include/mpi.h:1758:20: note: declared here\n@@ -1002705,34 +1002808,34 @@\n ## ../../../cxx/util/mtest.cxx:848:50: error: 'NULL_COPY_FN' is not a member of 'MPI::Win'\n ## mem_keyval = MPI::Win::Create_keyval( MPI::Win::NULL_COPY_FN,\n ## ^~~~~~~~~~~~\n ## ../../../cxx/util/mtest.cxx:849:22: error: 'NULL_DELETE_FN' is not a member of 'MPI::Win'\n ## MPI::Win::NULL_DELETE_FN, 0 );\n ## ^~~~~~~~~~~~~~\n ## make[3]: *** [Makefile:416: mtest.o] Error 1\n-## make[3]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/util'\n+## make[3]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/util'\n ## make[2]: *** [Makefile:989: ../../cxx/util/mtest.o] Error 2\n-## make[2]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/io'\n+## make[2]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/io'\n ]]> \n \n \n \n datatype, &size );\n ## ^\n ## In file included from ../../../cxx/util/mtest.cxx:7:\n ## /usr/lib/x86_64-linux-gnu/openmpi/include/mpi.h:1758:20: note: declared here\n@@ -1002764,34 +1002867,34 @@\n ## ../../../cxx/util/mtest.cxx:848:50: error: 'NULL_COPY_FN' is not a member of 'MPI::Win'\n ## mem_keyval = MPI::Win::Create_keyval( MPI::Win::NULL_COPY_FN,\n ## ^~~~~~~~~~~~\n ## ../../../cxx/util/mtest.cxx:849:22: error: 'NULL_DELETE_FN' is not a member of 'MPI::Win'\n ## MPI::Win::NULL_DELETE_FN, 0 );\n ## ^~~~~~~~~~~~~~\n ## make[3]: *** [Makefile:416: mtest.o] Error 1\n-## make[3]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/util'\n+## make[3]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/util'\n ## make[2]: *** [Makefile:989: ../../cxx/util/mtest.o] Error 2\n-## make[2]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/io'\n+## make[2]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/io'\n ]]> \n \n \n \n datatype, &size );\n ## ^\n ## In file included from ../../../cxx/util/mtest.cxx:7:\n ## /usr/lib/x86_64-linux-gnu/openmpi/include/mpi.h:1758:20: note: declared here\n@@ -1002823,34 +1002926,34 @@\n ## ../../../cxx/util/mtest.cxx:848:50: error: 'NULL_COPY_FN' is not a member of 'MPI::Win'\n ## mem_keyval = MPI::Win::Create_keyval( MPI::Win::NULL_COPY_FN,\n ## ^~~~~~~~~~~~\n ## ../../../cxx/util/mtest.cxx:849:22: error: 'NULL_DELETE_FN' is not a member of 'MPI::Win'\n ## MPI::Win::NULL_DELETE_FN, 0 );\n ## ^~~~~~~~~~~~~~\n ## make[3]: *** [Makefile:416: mtest.o] Error 1\n-## make[3]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/util'\n+## make[3]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/util'\n ## make[2]: *** [Makefile:989: ../../cxx/util/mtest.o] Error 2\n-## make[2]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/io'\n+## make[2]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/io'\n ]]> \n \n \n \n datatype, &size );\n ## ^\n ## In file included from ../../../cxx/util/mtest.cxx:7:\n ## /usr/lib/x86_64-linux-gnu/openmpi/include/mpi.h:1758:20: note: declared here\n@@ -1002882,34 +1002985,34 @@\n ## ../../../cxx/util/mtest.cxx:848:50: error: 'NULL_COPY_FN' is not a member of 'MPI::Win'\n ## mem_keyval = MPI::Win::Create_keyval( MPI::Win::NULL_COPY_FN,\n ## ^~~~~~~~~~~~\n ## ../../../cxx/util/mtest.cxx:849:22: error: 'NULL_DELETE_FN' is not a member of 'MPI::Win'\n ## MPI::Win::NULL_DELETE_FN, 0 );\n ## ^~~~~~~~~~~~~~\n ## make[3]: *** [Makefile:416: mtest.o] Error 1\n-## make[3]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/util'\n+## make[3]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/util'\n ## make[2]: *** [Makefile:989: ../../cxx/util/mtest.o] Error 2\n-## make[2]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/io'\n+## make[2]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/io'\n ]]> \n \n \n \n datatype, &size );\n ## ^\n ## In file included from ../../../cxx/util/mtest.cxx:7:\n ## /usr/lib/x86_64-linux-gnu/openmpi/include/mpi.h:1758:20: note: declared here\n@@ -1002941,34 +1003044,34 @@\n ## ../../../cxx/util/mtest.cxx:848:50: error: 'NULL_COPY_FN' is not a member of 'MPI::Win'\n ## mem_keyval = MPI::Win::Create_keyval( MPI::Win::NULL_COPY_FN,\n ## ^~~~~~~~~~~~\n ## ../../../cxx/util/mtest.cxx:849:22: error: 'NULL_DELETE_FN' is not a member of 'MPI::Win'\n ## MPI::Win::NULL_DELETE_FN, 0 );\n ## ^~~~~~~~~~~~~~\n ## make[3]: *** [Makefile:416: mtest.o] Error 1\n-## make[3]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/util'\n+## make[3]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/util'\n ## make[2]: *** [Makefile:989: ../../cxx/util/mtest.o] Error 2\n-## make[2]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/io'\n+## make[2]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/io'\n ]]> \n \n \n \n datatype, &size );\n ## ^\n ## In file included from ../../../cxx/util/mtest.cxx:7:\n ## /usr/lib/x86_64-linux-gnu/openmpi/include/mpi.h:1758:20: note: declared here\n@@ -1003000,34 +1003103,34 @@\n ## ../../../cxx/util/mtest.cxx:848:50: error: 'NULL_COPY_FN' is not a member of 'MPI::Win'\n ## mem_keyval = MPI::Win::Create_keyval( MPI::Win::NULL_COPY_FN,\n ## ^~~~~~~~~~~~\n ## ../../../cxx/util/mtest.cxx:849:22: error: 'NULL_DELETE_FN' is not a member of 'MPI::Win'\n ## MPI::Win::NULL_DELETE_FN, 0 );\n ## ^~~~~~~~~~~~~~\n ## make[3]: *** [Makefile:416: mtest.o] Error 1\n-## make[3]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/util'\n+## make[3]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/util'\n ## make[2]: *** [Makefile:989: ../../cxx/util/mtest.o] Error 2\n-## make[2]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/io'\n+## make[2]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/io'\n ]]> \n \n \n \n datatype, &size );\n ## ^\n ## In file included from ../../../cxx/util/mtest.cxx:7:\n ## /usr/lib/x86_64-linux-gnu/openmpi/include/mpi.h:1758:20: note: declared here\n@@ -1003059,34 +1003162,34 @@\n ## ../../../cxx/util/mtest.cxx:848:50: error: 'NULL_COPY_FN' is not a member of 'MPI::Win'\n ## mem_keyval = MPI::Win::Create_keyval( MPI::Win::NULL_COPY_FN,\n ## ^~~~~~~~~~~~\n ## ../../../cxx/util/mtest.cxx:849:22: error: 'NULL_DELETE_FN' is not a member of 'MPI::Win'\n ## MPI::Win::NULL_DELETE_FN, 0 );\n ## ^~~~~~~~~~~~~~\n ## make[3]: *** [Makefile:416: mtest.o] Error 1\n-## make[3]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/util'\n+## make[3]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/util'\n ## make[2]: *** [Makefile:989: ../../cxx/util/mtest.o] Error 2\n-## make[2]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/io'\n+## make[2]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/io'\n ]]> \n \n \n \n datatype, &size );\n ## ^\n ## In file included from ../../../cxx/util/mtest.cxx:7:\n ## /usr/lib/x86_64-linux-gnu/openmpi/include/mpi.h:1758:20: note: declared here\n@@ -1003118,34 +1003221,34 @@\n ## ../../../cxx/util/mtest.cxx:848:50: error: 'NULL_COPY_FN' is not a member of 'MPI::Win'\n ## mem_keyval = MPI::Win::Create_keyval( MPI::Win::NULL_COPY_FN,\n ## ^~~~~~~~~~~~\n ## ../../../cxx/util/mtest.cxx:849:22: error: 'NULL_DELETE_FN' is not a member of 'MPI::Win'\n ## MPI::Win::NULL_DELETE_FN, 0 );\n ## ^~~~~~~~~~~~~~\n ## make[3]: *** [Makefile:416: mtest.o] Error 1\n-## make[3]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/util'\n+## make[3]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/util'\n ## make[2]: *** [Makefile:989: ../../cxx/util/mtest.o] Error 2\n-## make[2]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/io'\n+## make[2]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/io'\n ]]> \n \n \n \n datatype, &size );\n ## ^\n ## In file included from ../../../cxx/util/mtest.cxx:7:\n ## /usr/lib/x86_64-linux-gnu/openmpi/include/mpi.h:1758:20: note: declared here\n@@ -1003177,34 +1003280,34 @@\n ## ../../../cxx/util/mtest.cxx:848:50: error: 'NULL_COPY_FN' is not a member of 'MPI::Win'\n ## mem_keyval = MPI::Win::Create_keyval( MPI::Win::NULL_COPY_FN,\n ## ^~~~~~~~~~~~\n ## ../../../cxx/util/mtest.cxx:849:22: error: 'NULL_DELETE_FN' is not a member of 'MPI::Win'\n ## MPI::Win::NULL_DELETE_FN, 0 );\n ## ^~~~~~~~~~~~~~\n ## make[3]: *** [Makefile:416: mtest.o] Error 1\n-## make[3]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/util'\n+## make[3]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/util'\n ## make[2]: *** [Makefile:989: ../../cxx/util/mtest.o] Error 2\n-## make[2]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/io'\n+## make[2]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/io'\n ]]> \n \n \n \n datatype, &size );\n ## ^\n ## In file included from ../../../cxx/util/mtest.cxx:7:\n ## /usr/lib/x86_64-linux-gnu/openmpi/include/mpi.h:1758:20: note: declared here\n@@ -1003236,34 +1003339,34 @@\n ## ../../../cxx/util/mtest.cxx:848:50: error: 'NULL_COPY_FN' is not a member of 'MPI::Win'\n ## mem_keyval = MPI::Win::Create_keyval( MPI::Win::NULL_COPY_FN,\n ## ^~~~~~~~~~~~\n ## ../../../cxx/util/mtest.cxx:849:22: error: 'NULL_DELETE_FN' is not a member of 'MPI::Win'\n ## MPI::Win::NULL_DELETE_FN, 0 );\n ## ^~~~~~~~~~~~~~\n ## make[3]: *** [Makefile:416: mtest.o] Error 1\n-## make[3]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/util'\n+## make[3]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/util'\n ## make[2]: *** [Makefile:989: ../../cxx/util/mtest.o] Error 2\n-## make[2]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/io'\n+## make[2]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/io'\n ]]> \n \n \n \n datatype, &size );\n ## ^\n ## In file included from ../../../cxx/util/mtest.cxx:7:\n ## /usr/lib/x86_64-linux-gnu/openmpi/include/mpi.h:1758:20: note: declared here\n@@ -1003295,34 +1003398,34 @@\n ## ../../../cxx/util/mtest.cxx:848:50: error: 'NULL_COPY_FN' is not a member of 'MPI::Win'\n ## mem_keyval = MPI::Win::Create_keyval( MPI::Win::NULL_COPY_FN,\n ## ^~~~~~~~~~~~\n ## ../../../cxx/util/mtest.cxx:849:22: error: 'NULL_DELETE_FN' is not a member of 'MPI::Win'\n ## MPI::Win::NULL_DELETE_FN, 0 );\n ## ^~~~~~~~~~~~~~\n ## make[3]: *** [Makefile:416: mtest.o] Error 1\n-## make[3]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/util'\n+## make[3]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/util'\n ## make[2]: *** [Makefile:989: ../../cxx/util/mtest.o] Error 2\n-## make[2]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/io'\n+## make[2]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/io'\n ]]> \n \n \n \n datatype, &size );\n ## ^\n ## In file included from ../../../cxx/util/mtest.cxx:7:\n ## /usr/lib/x86_64-linux-gnu/openmpi/include/mpi.h:1758:20: note: declared here\n@@ -1003354,34 +1003457,34 @@\n ## ../../../cxx/util/mtest.cxx:848:50: error: 'NULL_COPY_FN' is not a member of 'MPI::Win'\n ## mem_keyval = MPI::Win::Create_keyval( MPI::Win::NULL_COPY_FN,\n ## ^~~~~~~~~~~~\n ## ../../../cxx/util/mtest.cxx:849:22: error: 'NULL_DELETE_FN' is not a member of 'MPI::Win'\n ## MPI::Win::NULL_DELETE_FN, 0 );\n ## ^~~~~~~~~~~~~~\n ## make[3]: *** [Makefile:416: mtest.o] Error 1\n-## make[3]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/util'\n+## make[3]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/util'\n ## make[2]: *** [Makefile:989: ../../cxx/util/mtest.o] Error 2\n-## make[2]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/io'\n+## make[2]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/io'\n ]]> \n \n \n \n datatype, &size );\n ## ^\n ## In file included from ../../../cxx/util/mtest.cxx:7:\n ## /usr/lib/x86_64-linux-gnu/openmpi/include/mpi.h:1758:20: note: declared here\n@@ -1003413,34 +1003516,34 @@\n ## ../../../cxx/util/mtest.cxx:848:50: error: 'NULL_COPY_FN' is not a member of 'MPI::Win'\n ## mem_keyval = MPI::Win::Create_keyval( MPI::Win::NULL_COPY_FN,\n ## ^~~~~~~~~~~~\n ## ../../../cxx/util/mtest.cxx:849:22: error: 'NULL_DELETE_FN' is not a member of 'MPI::Win'\n ## MPI::Win::NULL_DELETE_FN, 0 );\n ## ^~~~~~~~~~~~~~\n ## make[3]: *** [Makefile:416: mtest.o] Error 1\n-## make[3]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/util'\n+## make[3]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/util'\n ## make[2]: *** [Makefile:989: ../../cxx/util/mtest.o] Error 2\n-## make[2]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/io'\n+## make[2]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/io'\n ]]> \n \n \n \n datatype, &size );\n ## ^\n ## In file included from ../../../cxx/util/mtest.cxx:7:\n ## /usr/lib/x86_64-linux-gnu/openmpi/include/mpi.h:1758:20: note: declared here\n@@ -1003472,34 +1003575,34 @@\n ## ../../../cxx/util/mtest.cxx:848:50: error: 'NULL_COPY_FN' is not a member of 'MPI::Win'\n ## mem_keyval = MPI::Win::Create_keyval( MPI::Win::NULL_COPY_FN,\n ## ^~~~~~~~~~~~\n ## ../../../cxx/util/mtest.cxx:849:22: error: 'NULL_DELETE_FN' is not a member of 'MPI::Win'\n ## MPI::Win::NULL_DELETE_FN, 0 );\n ## ^~~~~~~~~~~~~~\n ## make[3]: *** [Makefile:416: mtest.o] Error 1\n-## make[3]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/util'\n+## make[3]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/util'\n ## make[2]: *** [Makefile:989: ../../cxx/util/mtest.o] Error 2\n-## make[2]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/io'\n+## make[2]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/io'\n ]]> \n \n \n \n datatype, &size );\n ## ^\n ## In file included from ../../../cxx/util/mtest.cxx:7:\n ## /usr/lib/x86_64-linux-gnu/openmpi/include/mpi.h:1758:20: note: declared here\n@@ -1003531,34 +1003634,34 @@\n ## ../../../cxx/util/mtest.cxx:848:50: error: 'NULL_COPY_FN' is not a member of 'MPI::Win'\n ## mem_keyval = MPI::Win::Create_keyval( MPI::Win::NULL_COPY_FN,\n ## ^~~~~~~~~~~~\n ## ../../../cxx/util/mtest.cxx:849:22: error: 'NULL_DELETE_FN' is not a member of 'MPI::Win'\n ## MPI::Win::NULL_DELETE_FN, 0 );\n ## ^~~~~~~~~~~~~~\n ## make[3]: *** [Makefile:416: mtest.o] Error 1\n-## make[3]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/util'\n+## make[3]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/util'\n ## make[2]: *** [Makefile:989: ../../cxx/util/mtest.o] Error 2\n-## make[2]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/io'\n+## make[2]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/io'\n ]]> \n \n \n \n datatype, &size );\n ## ^\n ## In file included from ../../../cxx/util/mtest.cxx:7:\n ## /usr/lib/x86_64-linux-gnu/openmpi/include/mpi.h:1758:20: note: declared here\n@@ -1003590,34 +1003693,34 @@\n ## ../../../cxx/util/mtest.cxx:848:50: error: 'NULL_COPY_FN' is not a member of 'MPI::Win'\n ## mem_keyval = MPI::Win::Create_keyval( MPI::Win::NULL_COPY_FN,\n ## ^~~~~~~~~~~~\n ## ../../../cxx/util/mtest.cxx:849:22: error: 'NULL_DELETE_FN' is not a member of 'MPI::Win'\n ## MPI::Win::NULL_DELETE_FN, 0 );\n ## ^~~~~~~~~~~~~~\n ## make[3]: *** [Makefile:416: mtest.o] Error 1\n-## make[3]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/util'\n+## make[3]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/util'\n ## make[2]: *** [Makefile:989: ../../cxx/util/mtest.o] Error 2\n-## make[2]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/io'\n+## make[2]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/io'\n ]]> \n \n \n \n datatype, &size );\n ## ^\n ## In file included from ../../../cxx/util/mtest.cxx:7:\n ## /usr/lib/x86_64-linux-gnu/openmpi/include/mpi.h:1758:20: note: declared here\n@@ -1003649,34 +1003752,34 @@\n ## ../../../cxx/util/mtest.cxx:848:50: error: 'NULL_COPY_FN' is not a member of 'MPI::Win'\n ## mem_keyval = MPI::Win::Create_keyval( MPI::Win::NULL_COPY_FN,\n ## ^~~~~~~~~~~~\n ## ../../../cxx/util/mtest.cxx:849:22: error: 'NULL_DELETE_FN' is not a member of 'MPI::Win'\n ## MPI::Win::NULL_DELETE_FN, 0 );\n ## ^~~~~~~~~~~~~~\n ## make[3]: *** [Makefile:416: mtest.o] Error 1\n-## make[3]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/util'\n+## make[3]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/util'\n ## make[2]: *** [Makefile:989: ../../cxx/util/mtest.o] Error 2\n-## make[2]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/io'\n+## make[2]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/io'\n ]]> \n \n \n \n datatype, &size );\n ## ^\n ## In file included from ../../../cxx/util/mtest.cxx:7:\n ## /usr/lib/x86_64-linux-gnu/openmpi/include/mpi.h:1758:20: note: declared here\n@@ -1003708,34 +1003811,34 @@\n ## ../../../cxx/util/mtest.cxx:848:50: error: 'NULL_COPY_FN' is not a member of 'MPI::Win'\n ## mem_keyval = MPI::Win::Create_keyval( MPI::Win::NULL_COPY_FN,\n ## ^~~~~~~~~~~~\n ## ../../../cxx/util/mtest.cxx:849:22: error: 'NULL_DELETE_FN' is not a member of 'MPI::Win'\n ## MPI::Win::NULL_DELETE_FN, 0 );\n ## ^~~~~~~~~~~~~~\n ## make[3]: *** [Makefile:416: mtest.o] Error 1\n-## make[3]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/util'\n+## make[3]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/util'\n ## make[2]: *** [Makefile:989: ../../cxx/util/mtest.o] Error 2\n-## make[2]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/io'\n+## make[2]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/io'\n ]]> \n \n \n \n datatype, &size );\n ## ^\n ## In file included from ../../../cxx/util/mtest.cxx:7:\n ## /usr/lib/x86_64-linux-gnu/openmpi/include/mpi.h:1758:20: note: declared here\n@@ -1003767,34 +1003870,34 @@\n ## ../../../cxx/util/mtest.cxx:848:50: error: 'NULL_COPY_FN' is not a member of 'MPI::Win'\n ## mem_keyval = MPI::Win::Create_keyval( MPI::Win::NULL_COPY_FN,\n ## ^~~~~~~~~~~~\n ## ../../../cxx/util/mtest.cxx:849:22: error: 'NULL_DELETE_FN' is not a member of 'MPI::Win'\n ## MPI::Win::NULL_DELETE_FN, 0 );\n ## ^~~~~~~~~~~~~~\n ## make[3]: *** [Makefile:416: mtest.o] Error 1\n-## make[3]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/util'\n+## make[3]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/util'\n ## make[2]: *** [Makefile:989: ../../cxx/util/mtest.o] Error 2\n-## make[2]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/io'\n+## make[2]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/io'\n ]]> \n \n \n \n datatype, &size );\n ## ^\n ## In file included from ../../../cxx/util/mtest.cxx:7:\n ## /usr/lib/x86_64-linux-gnu/openmpi/include/mpi.h:1758:20: note: declared here\n@@ -1003826,34 +1003929,34 @@\n ## ../../../cxx/util/mtest.cxx:848:50: error: 'NULL_COPY_FN' is not a member of 'MPI::Win'\n ## mem_keyval = MPI::Win::Create_keyval( MPI::Win::NULL_COPY_FN,\n ## ^~~~~~~~~~~~\n ## ../../../cxx/util/mtest.cxx:849:22: error: 'NULL_DELETE_FN' is not a member of 'MPI::Win'\n ## MPI::Win::NULL_DELETE_FN, 0 );\n ## ^~~~~~~~~~~~~~\n ## make[3]: *** [Makefile:416: mtest.o] Error 1\n-## make[3]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/util'\n+## make[3]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/util'\n ## make[2]: *** [Makefile:989: ../../cxx/util/mtest.o] Error 2\n-## make[2]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/io'\n+## make[2]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/io'\n ]]> \n \n \n \n datatype, &size );\n ## ^\n ## In file included from ../../../cxx/util/mtest.cxx:7:\n ## /usr/lib/x86_64-linux-gnu/openmpi/include/mpi.h:1758:20: note: declared here\n@@ -1003885,34 +1003988,34 @@\n ## ../../../cxx/util/mtest.cxx:848:50: error: 'NULL_COPY_FN' is not a member of 'MPI::Win'\n ## mem_keyval = MPI::Win::Create_keyval( MPI::Win::NULL_COPY_FN,\n ## ^~~~~~~~~~~~\n ## ../../../cxx/util/mtest.cxx:849:22: error: 'NULL_DELETE_FN' is not a member of 'MPI::Win'\n ## MPI::Win::NULL_DELETE_FN, 0 );\n ## ^~~~~~~~~~~~~~\n ## make[3]: *** [Makefile:416: mtest.o] Error 1\n-## make[3]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/util'\n+## make[3]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/util'\n ## make[2]: *** [Makefile:989: ../../cxx/util/mtest.o] Error 2\n-## make[2]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/io'\n+## make[2]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/io'\n ]]> \n \n \n \n datatype, &size );\n ## ^\n ## In file included from ../../../cxx/util/mtest.cxx:7:\n ## /usr/lib/x86_64-linux-gnu/openmpi/include/mpi.h:1758:20: note: declared here\n@@ -1003944,34 +1004047,34 @@\n ## ../../../cxx/util/mtest.cxx:848:50: error: 'NULL_COPY_FN' is not a member of 'MPI::Win'\n ## mem_keyval = MPI::Win::Create_keyval( MPI::Win::NULL_COPY_FN,\n ## ^~~~~~~~~~~~\n ## ../../../cxx/util/mtest.cxx:849:22: error: 'NULL_DELETE_FN' is not a member of 'MPI::Win'\n ## MPI::Win::NULL_DELETE_FN, 0 );\n ## ^~~~~~~~~~~~~~\n ## make[3]: *** [Makefile:416: mtest.o] Error 1\n-## make[3]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/util'\n+## make[3]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/util'\n ## make[2]: *** [Makefile:989: ../../cxx/util/mtest.o] Error 2\n-## make[2]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/io'\n+## make[2]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/io'\n ]]> \n \n \n \n datatype, &size );\n ## ^\n ## In file included from ../../../cxx/util/mtest.cxx:7:\n ## /usr/lib/x86_64-linux-gnu/openmpi/include/mpi.h:1758:20: note: declared here\n@@ -1004003,34 +1004106,34 @@\n ## ../../../cxx/util/mtest.cxx:848:50: error: 'NULL_COPY_FN' is not a member of 'MPI::Win'\n ## mem_keyval = MPI::Win::Create_keyval( MPI::Win::NULL_COPY_FN,\n ## ^~~~~~~~~~~~\n ## ../../../cxx/util/mtest.cxx:849:22: error: 'NULL_DELETE_FN' is not a member of 'MPI::Win'\n ## MPI::Win::NULL_DELETE_FN, 0 );\n ## ^~~~~~~~~~~~~~\n ## make[3]: *** [Makefile:416: mtest.o] Error 1\n-## make[3]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/util'\n+## make[3]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/util'\n ## make[2]: *** [Makefile:989: ../../cxx/util/mtest.o] Error 2\n-## make[2]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/io'\n+## make[2]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/io'\n ]]> \n \n \n \n datatype, &size );\n ## ^\n ## In file included from ../../../cxx/util/mtest.cxx:7:\n ## /usr/lib/x86_64-linux-gnu/openmpi/include/mpi.h:1758:20: note: declared here\n@@ -1004062,34 +1004165,34 @@\n ## ../../../cxx/util/mtest.cxx:848:50: error: 'NULL_COPY_FN' is not a member of 'MPI::Win'\n ## mem_keyval = MPI::Win::Create_keyval( MPI::Win::NULL_COPY_FN,\n ## ^~~~~~~~~~~~\n ## ../../../cxx/util/mtest.cxx:849:22: error: 'NULL_DELETE_FN' is not a member of 'MPI::Win'\n ## MPI::Win::NULL_DELETE_FN, 0 );\n ## ^~~~~~~~~~~~~~\n ## make[3]: *** [Makefile:416: mtest.o] Error 1\n-## make[3]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/util'\n+## make[3]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/util'\n ## make[2]: *** [Makefile:673: ../../cxx/util/mtest.o] Error 2\n-## make[2]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/spawn'\n+## make[2]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/spawn'\n ]]> \n \n \n \n datatype, &size );\n ## ^\n ## In file included from ../../../cxx/util/mtest.cxx:7:\n ## /usr/lib/x86_64-linux-gnu/openmpi/include/mpi.h:1758:20: note: declared here\n@@ -1004121,37 +1004224,37 @@\n ## ../../../cxx/util/mtest.cxx:848:50: error: 'NULL_COPY_FN' is not a member of 'MPI::Win'\n ## mem_keyval = MPI::Win::Create_keyval( MPI::Win::NULL_COPY_FN,\n ## ^~~~~~~~~~~~\n ## ../../../cxx/util/mtest.cxx:849:22: error: 'NULL_DELETE_FN' is not a member of 'MPI::Win'\n ## MPI::Win::NULL_DELETE_FN, 0 );\n ## ^~~~~~~~~~~~~~\n ## make[3]: *** [Makefile:416: mtest.o] Error 1\n-## make[3]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/util'\n+## make[3]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/util'\n ## make[2]: *** [Makefile:673: ../../cxx/util/mtest.o] Error 2\n-## make[2]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/spawn'\n+## make[2]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/spawn'\n ]]> \n \n \n \n datatype, &size );\n ## ^\n ## In file included from ../../../cxx/util/mtest.cxx:7:\n ## /usr/lib/x86_64-linux-gnu/openmpi/include/mpi.h:1758:20: note: declared here\n@@ -1004183,34 +1004286,34 @@\n ## ../../../cxx/util/mtest.cxx:848:50: error: 'NULL_COPY_FN' is not a member of 'MPI::Win'\n ## mem_keyval = MPI::Win::Create_keyval( MPI::Win::NULL_COPY_FN,\n ## ^~~~~~~~~~~~\n ## ../../../cxx/util/mtest.cxx:849:22: error: 'NULL_DELETE_FN' is not a member of 'MPI::Win'\n ## MPI::Win::NULL_DELETE_FN, 0 );\n ## ^~~~~~~~~~~~~~\n ## make[3]: *** [Makefile:416: mtest.o] Error 1\n-## make[3]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/util'\n+## make[3]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/util'\n ## make[2]: *** [Makefile:673: ../../cxx/util/mtest.o] Error 2\n-## make[2]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/spawn'\n+## make[2]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/spawn'\n ]]> \n \n \n \n datatype, &size );\n ## ^\n ## In file included from ../../../cxx/util/mtest.cxx:7:\n ## /usr/lib/x86_64-linux-gnu/openmpi/include/mpi.h:1758:20: note: declared here\n@@ -1004242,34 +1004345,34 @@\n ## ../../../cxx/util/mtest.cxx:848:50: error: 'NULL_COPY_FN' is not a member of 'MPI::Win'\n ## mem_keyval = MPI::Win::Create_keyval( MPI::Win::NULL_COPY_FN,\n ## ^~~~~~~~~~~~\n ## ../../../cxx/util/mtest.cxx:849:22: error: 'NULL_DELETE_FN' is not a member of 'MPI::Win'\n ## MPI::Win::NULL_DELETE_FN, 0 );\n ## ^~~~~~~~~~~~~~\n ## make[3]: *** [Makefile:416: mtest.o] Error 1\n-## make[3]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/util'\n+## make[3]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/util'\n ## make[2]: *** [Makefile:673: ../../cxx/util/mtest.o] Error 2\n-## make[2]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/spawn'\n+## make[2]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/spawn'\n ]]> \n \n \n \n datatype, &size );\n ## ^\n ## In file included from ../../../cxx/util/mtest.cxx:7:\n ## /usr/lib/x86_64-linux-gnu/openmpi/include/mpi.h:1758:20: note: declared here\n@@ -1004301,34 +1004404,34 @@\n ## ../../../cxx/util/mtest.cxx:848:50: error: 'NULL_COPY_FN' is not a member of 'MPI::Win'\n ## mem_keyval = MPI::Win::Create_keyval( MPI::Win::NULL_COPY_FN,\n ## ^~~~~~~~~~~~\n ## ../../../cxx/util/mtest.cxx:849:22: error: 'NULL_DELETE_FN' is not a member of 'MPI::Win'\n ## MPI::Win::NULL_DELETE_FN, 0 );\n ## ^~~~~~~~~~~~~~\n ## make[3]: *** [Makefile:416: mtest.o] Error 1\n-## make[3]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/util'\n+## make[3]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/util'\n ## make[2]: *** [Makefile:673: ../../cxx/util/mtest.o] Error 2\n-## make[2]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/spawn'\n+## make[2]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/spawn'\n ]]> \n \n \n \n datatype, &size );\n ## ^\n ## In file included from ../../../cxx/util/mtest.cxx:7:\n ## /usr/lib/x86_64-linux-gnu/openmpi/include/mpi.h:1758:20: note: declared here\n@@ -1004360,34 +1004463,34 @@\n ## ../../../cxx/util/mtest.cxx:848:50: error: 'NULL_COPY_FN' is not a member of 'MPI::Win'\n ## mem_keyval = MPI::Win::Create_keyval( MPI::Win::NULL_COPY_FN,\n ## ^~~~~~~~~~~~\n ## ../../../cxx/util/mtest.cxx:849:22: error: 'NULL_DELETE_FN' is not a member of 'MPI::Win'\n ## MPI::Win::NULL_DELETE_FN, 0 );\n ## ^~~~~~~~~~~~~~\n ## make[3]: *** [Makefile:416: mtest.o] Error 1\n-## make[3]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/util'\n+## make[3]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/util'\n ## make[2]: *** [Makefile:700: ../../cxx/util/mtest.o] Error 2\n-## make[2]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/rma'\n+## make[2]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/rma'\n ]]> \n \n \n \n datatype, &size );\n ## ^\n ## In file included from ../../../cxx/util/mtest.cxx:7:\n ## /usr/lib/x86_64-linux-gnu/openmpi/include/mpi.h:1758:20: note: declared here\n@@ -1004419,34 +1004522,34 @@\n ## ../../../cxx/util/mtest.cxx:848:50: error: 'NULL_COPY_FN' is not a member of 'MPI::Win'\n ## mem_keyval = MPI::Win::Create_keyval( MPI::Win::NULL_COPY_FN,\n ## ^~~~~~~~~~~~\n ## ../../../cxx/util/mtest.cxx:849:22: error: 'NULL_DELETE_FN' is not a member of 'MPI::Win'\n ## MPI::Win::NULL_DELETE_FN, 0 );\n ## ^~~~~~~~~~~~~~\n ## make[3]: *** [Makefile:416: mtest.o] Error 1\n-## make[3]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/util'\n+## make[3]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/util'\n ## make[2]: *** [Makefile:700: ../../cxx/util/mtest.o] Error 2\n-## make[2]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/rma'\n+## make[2]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/rma'\n ]]> \n \n \n \n datatype, &size );\n ## ^\n ## In file included from ../../../cxx/util/mtest.cxx:7:\n ## /usr/lib/x86_64-linux-gnu/openmpi/include/mpi.h:1758:20: note: declared here\n@@ -1004478,34 +1004581,34 @@\n ## ../../../cxx/util/mtest.cxx:848:50: error: 'NULL_COPY_FN' is not a member of 'MPI::Win'\n ## mem_keyval = MPI::Win::Create_keyval( MPI::Win::NULL_COPY_FN,\n ## ^~~~~~~~~~~~\n ## ../../../cxx/util/mtest.cxx:849:22: error: 'NULL_DELETE_FN' is not a member of 'MPI::Win'\n ## MPI::Win::NULL_DELETE_FN, 0 );\n ## ^~~~~~~~~~~~~~\n ## make[3]: *** [Makefile:416: mtest.o] Error 1\n-## make[3]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/util'\n+## make[3]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/util'\n ## make[2]: *** [Makefile:700: ../../cxx/util/mtest.o] Error 2\n-## make[2]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/rma'\n+## make[2]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/rma'\n ]]> \n \n \n \n datatype, &size );\n ## ^\n ## In file included from ../../../cxx/util/mtest.cxx:7:\n ## /usr/lib/x86_64-linux-gnu/openmpi/include/mpi.h:1758:20: note: declared here\n@@ -1004537,34 +1004640,34 @@\n ## ../../../cxx/util/mtest.cxx:848:50: error: 'NULL_COPY_FN' is not a member of 'MPI::Win'\n ## mem_keyval = MPI::Win::Create_keyval( MPI::Win::NULL_COPY_FN,\n ## ^~~~~~~~~~~~\n ## ../../../cxx/util/mtest.cxx:849:22: error: 'NULL_DELETE_FN' is not a member of 'MPI::Win'\n ## MPI::Win::NULL_DELETE_FN, 0 );\n ## ^~~~~~~~~~~~~~\n ## make[3]: *** [Makefile:416: mtest.o] Error 1\n-## make[3]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/util'\n+## make[3]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/util'\n ## make[2]: *** [Makefile:700: ../../cxx/util/mtest.o] Error 2\n-## make[2]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/rma'\n+## make[2]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/rma'\n ]]> \n \n \n \n datatype, &size );\n ## ^\n ## In file included from ../../../cxx/util/mtest.cxx:7:\n ## /usr/lib/x86_64-linux-gnu/openmpi/include/mpi.h:1758:20: note: declared here\n@@ -1004596,34 +1004699,34 @@\n ## ../../../cxx/util/mtest.cxx:848:50: error: 'NULL_COPY_FN' is not a member of 'MPI::Win'\n ## mem_keyval = MPI::Win::Create_keyval( MPI::Win::NULL_COPY_FN,\n ## ^~~~~~~~~~~~\n ## ../../../cxx/util/mtest.cxx:849:22: error: 'NULL_DELETE_FN' is not a member of 'MPI::Win'\n ## MPI::Win::NULL_DELETE_FN, 0 );\n ## ^~~~~~~~~~~~~~\n ## make[3]: *** [Makefile:416: mtest.o] Error 1\n-## make[3]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/util'\n+## make[3]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/util'\n ## make[2]: *** [Makefile:700: ../../cxx/util/mtest.o] Error 2\n-## make[2]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/rma'\n+## make[2]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/rma'\n ]]> \n \n \n \n datatype, &size );\n ## ^\n ## In file included from ../../../cxx/util/mtest.cxx:7:\n ## /usr/lib/x86_64-linux-gnu/openmpi/include/mpi.h:1758:20: note: declared here\n@@ -1004655,42 +1004758,42 @@\n ## ../../../cxx/util/mtest.cxx:848:50: error: 'NULL_COPY_FN' is not a member of 'MPI::Win'\n ## mem_keyval = MPI::Win::Create_keyval( MPI::Win::NULL_COPY_FN,\n ## ^~~~~~~~~~~~\n ## ../../../cxx/util/mtest.cxx:849:22: error: 'NULL_DELETE_FN' is not a member of 'MPI::Win'\n ## MPI::Win::NULL_DELETE_FN, 0 );\n ## ^~~~~~~~~~~~~~\n ## make[3]: *** [Makefile:416: mtest.o] Error 1\n-## make[3]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/util'\n+## make[3]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/util'\n ## make[2]: *** [Makefile:700: ../../cxx/util/mtest.o] Error 2\n-## make[2]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/rma'\n+## make[2]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/rma'\n ]]> \n \n \n \n \n \n \n \n \n \n \n@@ -1004708,130 +1004811,150 @@\n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n- \n+ \n+ \n+ \n+ \n \n \n \n \n \n \n \n \n \n \n- \n- \n- \n- \n-\n+ \n+ \n \n- \n- \n- \n- \n- \n-\n+ \n+ \n \n+ \n+ \n+ \n+ \n+ \n+ \n+ \n+ \n \n \n \n \n \n \n \n \n \n \n \n \n message="requires MPI version 3.1">\n@@ -1004946,19 +1005104,19 @@\n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n@@ -1005095,15 +1005253,15 @@\n \n \n \n \n \n \n \n \n \n \n \n \n \n \n@@ -1005201,15 +1005359,15 @@\n \n \n \n \n \n@@ -1005227,438 +1005385,424 @@\n \n \n Returned \"Unreachable\" (-12) instead of \"Success\" (0)\n ## --------------------------------------------------------------------------\n-## [profitbricks-build5-amd64:24549] *** An error occurred in MPI_Comm_spawn\n-## [profitbricks-build5-amd64:24549] *** reported by process [2607808513,0]\n-## [profitbricks-build5-amd64:24549] *** on communicator MPI_COMM_WORLD\n-## [profitbricks-build5-amd64:24549] *** MPI_ERR_INTERN: internal error\n-## [profitbricks-build5-amd64:24549] *** MPI_ERRORS_ARE_FATAL (processes in this communicator will now abort,\n-## [profitbricks-build5-amd64:24549] *** and potentially your MPI job)\n-## [profitbricks-build5-amd64:24613] *** An error occurred in MPI_Init\n-## [profitbricks-build5-amd64:24613] *** reported by process [2607808514,0]\n-## [profitbricks-build5-amd64:24613] *** on a NULL communicator\n-## [profitbricks-build5-amd64:24613] *** Unknown error\n-## [profitbricks-build5-amd64:24613] *** MPI_ERRORS_ARE_FATAL (processes in this communicator will now abort,\n-## [profitbricks-build5-amd64:24613] *** and potentially your MPI job)\n-## [profitbricks-build5-amd64:24482] 2 more processes have sent help message help-mca-bml-r2.txt / unreachable proc\n-## [profitbricks-build5-amd64:24482] Set MCA parameter \"orte_base_help_aggregate\" to 0 to see all help / error messages\n-## [profitbricks-build5-amd64:24482] 1 more process has sent help message help-mpi-runtime.txt / mpi_init:startup:internal-failure\n-## [profitbricks-build5-amd64:24482] 1 more process has sent help message help-mpi-errors.txt / mpi_errors_are_fatal unknown handle\n-## [warn] Epoll MOD(1) on fd 27 failed. Old events were 6; read change was 0 (none); write change was 2 (del); close change was 0 (none): Bad file descriptor\n+## [i-capture-the-hostname:07987] [[37874,2],0] ORTE_ERROR_LOG: Unreachable in file dpm/dpm.c at line 433\n+## [i-capture-the-hostname:07932] [[37874,1],0] ORTE_ERROR_LOG: Unreachable in file dpm/dpm.c at line 433\n+## [i-capture-the-hostname:07988] [[37874,2],1] ORTE_ERROR_LOG: Unreachable in file dpm/dpm.c at line 433\n+## [i-capture-the-hostname:07932] *** An error occurred in MPI_Comm_spawn\n+## [i-capture-the-hostname:07932] *** reported by process [2482110465,0]\n+## [i-capture-the-hostname:07932] *** on communicator MPI_COMM_WORLD\n+## [i-capture-the-hostname:07932] *** MPI_ERR_INTERN: internal error\n+## [i-capture-the-hostname:07932] *** MPI_ERRORS_ARE_FATAL (processes in this communicator will now abort,\n+## [i-capture-the-hostname:07932] *** and potentially your MPI job)\n+## [i-capture-the-hostname:07988] *** An error occurred in MPI_Init\n+## [i-capture-the-hostname:07988] *** reported by process [2482110466,1]\n+## [i-capture-the-hostname:07988] *** on a NULL communicator\n+## [i-capture-the-hostname:07988] *** Unknown error\n+## [i-capture-the-hostname:07988] *** MPI_ERRORS_ARE_FATAL (processes in this communicator will now abort,\n+## [i-capture-the-hostname:07988] *** and potentially your MPI job)\n+## [i-capture-the-hostname:07843] 2 more processes have sent help message help-mca-bml-r2.txt / unreachable proc\n+## [i-capture-the-hostname:07843] Set MCA parameter \"orte_base_help_aggregate\" to 0 to see all help / error messages\n+## [i-capture-the-hostname:07843] 1 more process has sent help message help-mpi-runtime.txt / mpi_init:startup:internal-failure\n+## [i-capture-the-hostname:07843] 1 more process has sent help message help-mpi-errors.txt / mpi_errors_are_fatal unknown handle\n ]]> \n \n \n \n Returned \"Unreachable\" (-12) instead of \"Success\" (0)\n ## --------------------------------------------------------------------------\n-## [profitbricks-build5-amd64:24799] [[42070,2],0] ORTE_ERROR_LOG: Unreachable in file dpm/dpm.c at line 433\n-## [profitbricks-build5-amd64:24799] *** An error occurred in MPI_Init\n-## [profitbricks-build5-amd64:24799] *** reported by process [2757099522,0]\n-## [profitbricks-build5-amd64:24799] *** on a NULL communicator\n-## [profitbricks-build5-amd64:24799] *** Unknown error\n-## [profitbricks-build5-amd64:24799] *** MPI_ERRORS_ARE_FATAL (processes in this communicator will now abort,\n-## [profitbricks-build5-amd64:24799] *** and potentially your MPI job)\n-## [profitbricks-build5-amd64:24752] [[42070,1],0] ORTE_ERROR_LOG: Unreachable in file dpm/dpm.c at line 433\n-## [profitbricks-build5-amd64:24800] [[42070,2],1] ORTE_ERROR_LOG: Unreachable in file dpm/dpm.c at line 433\n-## [profitbricks-build5-amd64:24752] *** An error occurred in MPI_Comm_spawn\n-## [profitbricks-build5-amd64:24752] *** reported by process [2757099521,0]\n-## [profitbricks-build5-amd64:24752] *** on communicator MPI_COMM_WORLD\n-## [profitbricks-build5-amd64:24752] *** MPI_ERR_INTERN: internal error\n-## [profitbricks-build5-amd64:24752] *** MPI_ERRORS_ARE_FATAL (processes in this communicator will now abort,\n-## [profitbricks-build5-amd64:24752] *** and potentially your MPI job)\n-## [profitbricks-build5-amd64:24708] 2 more processes have sent help message help-mca-bml-r2.txt / unreachable proc\n-## [profitbricks-build5-amd64:24708] Set MCA parameter \"orte_base_help_aggregate\" to 0 to see all help / error messages\n-## [profitbricks-build5-amd64:24708] 1 more process has sent help message help-mpi-runtime.txt / mpi_init:startup:internal-failure\n-## [profitbricks-build5-amd64:24708] 1 more process has sent help message help-mpi-errors.txt / mpi_errors_are_fatal unknown handle\n-## [warn] Epoll MOD(1) on fd 29 failed. Old events were 6; read change was 0 (none); write change was 2 (del); close change was 0 (none): Bad file descriptor\n+## [i-capture-the-hostname:08273] [[44378,2],0] ORTE_ERROR_LOG: Unreachable in file dpm/dpm.c at line 433\n+## [i-capture-the-hostname:08274] [[44378,2],1] ORTE_ERROR_LOG: Unreachable in file dpm/dpm.c at line 433\n+## [i-capture-the-hostname:08203] 2 more processes have sent help message help-mca-bml-r2.txt / unreachable proc\n+## [i-capture-the-hostname:08203] Set MCA parameter \"orte_base_help_aggregate\" to 0 to see all help / error messages\n+## [i-capture-the-hostname:08203] 1 more process has sent help message help-mpi-runtime.txt / mpi_init:startup:internal-failure\n ]]> \n \n \n \n \n Returned \"Unreachable\" (-12) instead of \"Success\" (0)\n ## --------------------------------------------------------------------------\n-## [profitbricks-build5-amd64:25217] [[42300,2],1] ORTE_ERROR_LOG: Unreachable in file dpm/dpm.c at line 433\n-## [profitbricks-build5-amd64:25214] [[42300,2],0] ORTE_ERROR_LOG: Unreachable in file dpm/dpm.c at line 433\n-## [profitbricks-build5-amd64:25217] *** An error occurred in MPI_Init\n-## [profitbricks-build5-amd64:25217] *** reported by process [2772172802,1]\n-## [profitbricks-build5-amd64:25217] *** on a NULL communicator\n-## [profitbricks-build5-amd64:25217] *** Unknown error\n-## [profitbricks-build5-amd64:25217] *** MPI_ERRORS_ARE_FATAL (processes in this communicator will now abort,\n-## [profitbricks-build5-amd64:25217] *** and potentially your MPI job)\n-## [profitbricks-build5-amd64:25070] 2 more processes have sent help message help-mca-bml-r2.txt / unreachable proc\n-## [profitbricks-build5-amd64:25070] Set MCA parameter \"orte_base_help_aggregate\" to 0 to see all help / error messages\n-## [profitbricks-build5-amd64:25070] 1 more process has sent help message help-mpi-runtime.txt / mpi_init:startup:internal-failure\n-## [profitbricks-build5-amd64:25070] 1 more process has sent help message help-mpi-errors.txt / mpi_errors_are_fatal unknown handle\n+## [i-capture-the-hostname:08387] [[44498,2],1] ORTE_ERROR_LOG: Unreachable in file dpm/dpm.c at line 433\n+## [i-capture-the-hostname:08386] [[44498,2],0] ORTE_ERROR_LOG: Unreachable in file dpm/dpm.c at line 433\n+## [i-capture-the-hostname:08323] 2 more processes have sent help message help-mca-bml-r2.txt / unreachable proc\n+## [i-capture-the-hostname:08323] Set MCA parameter \"orte_base_help_aggregate\" to 0 to see all help / error messages\n+## [i-capture-the-hostname:08323] 1 more process has sent help message help-mpi-runtime.txt / mpi_init:startup:internal-failure\n ]]> \n \n \n \n Returned \"Unreachable\" (-12) instead of \"Success\" (0)\n ## --------------------------------------------------------------------------\n-## [profitbricks-build5-amd64:25571] *** An error occurred in MPI_Init\n-## [profitbricks-build5-amd64:25571] *** reported by process [2818441218,0]\n-## [profitbricks-build5-amd64:25571] *** on a NULL communicator\n-## [profitbricks-build5-amd64:25571] *** Unknown error\n-## [profitbricks-build5-amd64:25571] *** MPI_ERRORS_ARE_FATAL (processes in this communicator will now abort,\n-## [profitbricks-build5-amd64:25571] *** and potentially your MPI job)\n-## [profitbricks-build5-amd64:25388] 3 more processes have sent help message help-mca-bml-r2.txt / unreachable proc\n-## [profitbricks-build5-amd64:25388] Set MCA parameter \"orte_base_help_aggregate\" to 0 to see all help / error messages\n-## [profitbricks-build5-amd64:25388] 1 more process has sent help message help-mpi-errors.txt / mpi_errors_are_fatal\n-## [profitbricks-build5-amd64:25388] 1 more process has sent help message help-mpi-runtime.txt / mpi_init:startup:internal-failure\n-## [profitbricks-build5-amd64:25388] 1 more process has sent help message help-mpi-errors.txt / mpi_errors_are_fatal unknown handle\n+## [i-capture-the-hostname:08524] *** An error occurred in MPI_Comm_spawn_multiple\n+## [i-capture-the-hostname:08524] *** reported by process [2890072065,1]\n+## [i-capture-the-hostname:08524] *** on communicator MPI_COMM_WORLD\n+## [i-capture-the-hostname:08524] *** MPI_ERR_INTERN: internal error\n+## [i-capture-the-hostname:08524] *** MPI_ERRORS_ARE_FATAL (processes in this communicator will now abort,\n+## [i-capture-the-hostname:08524] *** and potentially your MPI job)\n+## [i-capture-the-hostname:08519] [[44099,1],0] ORTE_ERROR_LOG: Unreachable in file dpm/dpm.c at line 433\n+## [i-capture-the-hostname:08590] *** An error occurred in MPI_Init\n+## [i-capture-the-hostname:08590] *** reported by process [2890072066,0]\n+## [i-capture-the-hostname:08590] *** on a NULL communicator\n+## [i-capture-the-hostname:08590] *** Unknown error\n+## [i-capture-the-hostname:08590] *** MPI_ERRORS_ARE_FATAL (processes in this communicator will now abort,\n+## [i-capture-the-hostname:08590] *** and potentially your MPI job)\n+## [i-capture-the-hostname:08466] 3 more processes have sent help message help-mca-bml-r2.txt / unreachable proc\n+## [i-capture-the-hostname:08466] Set MCA parameter \"orte_base_help_aggregate\" to 0 to see all help / error messages\n+## [i-capture-the-hostname:08466] 1 more process has sent help message help-mpi-runtime.txt / mpi_init:startup:internal-failure\n+## [i-capture-the-hostname:08466] 1 more process has sent help message help-mpi-errors.txt / mpi_errors_are_fatal\n+## [i-capture-the-hostname:08466] 1 more process has sent help message help-mpi-errors.txt / mpi_errors_are_fatal unknown handle\n ]]> \n \n \n \n Returned \"Unreachable\" (-12) instead of \"Success\" (0)\n ## --------------------------------------------------------------------------\n-## [profitbricks-build5-amd64:25812] *** An error occurred in MPI_Comm_spawn\n-## [profitbricks-build5-amd64:25812] *** reported by process [2695364609,0]\n-## [profitbricks-build5-amd64:25812] *** on communicator MPI_COMM_WORLD\n-## [profitbricks-build5-amd64:25812] *** MPI_ERR_INTERN: internal error\n-## [profitbricks-build5-amd64:25812] *** MPI_ERRORS_ARE_FATAL (processes in this communicator will now abort,\n-## [profitbricks-build5-amd64:25812] *** and potentially your MPI job)\n-## [profitbricks-build5-amd64:25937] [[41128,2],0] ORTE_ERROR_LOG: Unreachable in file dpm/dpm.c at line 433\n-## [profitbricks-build5-amd64:25937] *** An error occurred in MPI_Init\n-## [profitbricks-build5-amd64:25937] *** reported by process [2695364610,0]\n-## [profitbricks-build5-amd64:25937] *** on a NULL communicator\n-## [profitbricks-build5-amd64:25937] *** Unknown error\n-## [profitbricks-build5-amd64:25937] *** MPI_ERRORS_ARE_FATAL (processes in this communicator will now abort,\n-## [profitbricks-build5-amd64:25937] *** and potentially your MPI job)\n-## [profitbricks-build5-amd64:25722] 2 more processes have sent help message help-mca-bml-r2.txt / unreachable proc\n-## [profitbricks-build5-amd64:25722] Set MCA parameter \"orte_base_help_aggregate\" to 0 to see all help / error messages\n-## [profitbricks-build5-amd64:25722] 1 more process has sent help message help-mpi-runtime.txt / mpi_init:startup:internal-failure\n-## [profitbricks-build5-amd64:25722] 1 more process has sent help message help-mpi-errors.txt / mpi_errors_are_fatal unknown handle\n+## [i-capture-the-hostname:08792] [[44881,1],0] ORTE_ERROR_LOG: Unreachable in file dpm/dpm.c at line 433\n+## [i-capture-the-hostname:08900] [[44881,2],1] ORTE_ERROR_LOG: Unreachable in file dpm/dpm.c at line 433\n+## [i-capture-the-hostname:08792] *** An error occurred in MPI_Comm_spawn\n+## [i-capture-the-hostname:08792] *** reported by process [2941321217,0]\n+## [i-capture-the-hostname:08792] *** on communicator MPI_COMM_WORLD\n+## [i-capture-the-hostname:08792] *** MPI_ERR_INTERN: internal error\n+## [i-capture-the-hostname:08792] *** MPI_ERRORS_ARE_FATAL (processes in this communicator will now abort,\n+## [i-capture-the-hostname:08792] *** and potentially your MPI job)\n+## [i-capture-the-hostname:08894] *** An error occurred in MPI_Init\n+## [i-capture-the-hostname:08894] *** reported by process [2941321218,0]\n+## [i-capture-the-hostname:08894] *** on a NULL communicator\n+## [i-capture-the-hostname:08894] *** Unknown error\n+## [i-capture-the-hostname:08894] *** MPI_ERRORS_ARE_FATAL (processes in this communicator will now abort,\n+## [i-capture-the-hostname:08894] *** and potentially your MPI job)\n+## [i-capture-the-hostname:08704] 2 more processes have sent help message help-mca-bml-r2.txt / unreachable proc\n+## [i-capture-the-hostname:08704] Set MCA parameter \"orte_base_help_aggregate\" to 0 to see all help / error messages\n+## [i-capture-the-hostname:08704] 1 more process has sent help message help-mpi-runtime.txt / mpi_init:startup:internal-failure\n+## [i-capture-the-hostname:08704] 1 more process has sent help message help-mpi-errors.txt / mpi_errors_are_fatal unknown handle\n+## [warn] Epoll MOD(1) on fd 25 failed. Old events were 6; read change was 0 (none); write change was 2 (del); close change was 0 (none): Bad file descriptor\n ]]> \n \n \n \n Returned \"Unreachable\" (-12) instead of \"Success\" (0)\n ## --------------------------------------------------------------------------\n-## [profitbricks-build5-amd64:26209] [[41220,2],1] ORTE_ERROR_LOG: Unreachable in file dpm/dpm.c at line 433\n-## [profitbricks-build5-amd64:26203] [[41220,2],0] ORTE_ERROR_LOG: Unreachable in file dpm/dpm.c at line 433\n-## [profitbricks-build5-amd64:26209] *** An error occurred in MPI_Init\n-## [profitbricks-build5-amd64:26209] *** reported by process [2701393922,1]\n-## [profitbricks-build5-amd64:26209] *** on a NULL communicator\n-## [profitbricks-build5-amd64:26209] *** Unknown error\n-## [profitbricks-build5-amd64:26209] *** MPI_ERRORS_ARE_FATAL (processes in this communicator will now abort,\n-## [profitbricks-build5-amd64:26209] *** and potentially your MPI job)\n-## [profitbricks-build5-amd64:26070] 2 more processes have sent help message help-mca-bml-r2.txt / unreachable proc\n-## [profitbricks-build5-amd64:26070] Set MCA parameter \"orte_base_help_aggregate\" to 0 to see all help / error messages\n-## [profitbricks-build5-amd64:26070] 1 more process has sent help message help-mpi-runtime.txt / mpi_init:startup:internal-failure\n-## [profitbricks-build5-amd64:26070] 1 more process has sent help message help-mpi-errors.txt / mpi_errors_are_fatal unknown handle\n+## [i-capture-the-hostname:09025] *** An error occurred in MPI_Comm_spawn_multiple\n+## [i-capture-the-hostname:09025] *** reported by process [2924019713,0]\n+## [i-capture-the-hostname:09025] *** on communicator MPI_COMM_WORLD\n+## [i-capture-the-hostname:09025] *** MPI_ERR_INTERN: internal error\n+## [i-capture-the-hostname:09025] *** MPI_ERRORS_ARE_FATAL (processes in this communicator will now abort,\n+## [i-capture-the-hostname:09025] *** and potentially your MPI job)\n+## [i-capture-the-hostname:09053] *** An error occurred in MPI_Init\n+## [i-capture-the-hostname:09053] *** reported by process [2924019714,0]\n+## [i-capture-the-hostname:09053] *** on a NULL communicator\n+## [i-capture-the-hostname:09053] *** Unknown error\n+## [i-capture-the-hostname:09053] *** MPI_ERRORS_ARE_FATAL (processes in this communicator will now abort,\n+## [i-capture-the-hostname:09053] *** and potentially your MPI job)\n+## [i-capture-the-hostname:08984] 2 more processes have sent help message help-mca-bml-r2.txt / unreachable proc\n+## [i-capture-the-hostname:08984] Set MCA parameter \"orte_base_help_aggregate\" to 0 to see all help / error messages\n+## [i-capture-the-hostname:08984] 1 more process has sent help message help-mpi-runtime.txt / mpi_init:startup:internal-failure\n+## [i-capture-the-hostname:08984] 1 more process has sent help message help-mpi-errors.txt / mpi_errors_are_fatal unknown handle\n+## [warn] Epoll MOD(1) on fd 29 failed. Old events were 6; read change was 0 (none); write change was 2 (del); close change was 0 (none): Bad file descriptor\n ]]> \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n@@ -1005716,15 +1005860,15 @@\n \n \n \n@@ -1005732,15 +1005876,15 @@\n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n message="non-strict test, strict MPI mode requested">\n \n@@ -1005969,70 +1006113,83 @@\n \n \n \n \n message="non-strict test, strict MPI mode requested">\n \n \n- \n+ \n+ \n+ \n+ \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n@@ -1006041,129 +1006198,129 @@\n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n@@ -1006231,37 +1006388,37 @@\n \n \n \n \n \n \n datatype, &size );\n ## ^\n ## In file included from ../../../cxx/util/mtest.cxx:7:\n ## /usr/lib/x86_64-linux-gnu/openmpi/include/mpi.h:1758:20: note: declared here\n@@ -1006293,34 +1006450,34 @@\n ## ../../../cxx/util/mtest.cxx:848:50: error: 'NULL_COPY_FN' is not a member of 'MPI::Win'\n ## mem_keyval = MPI::Win::Create_keyval( MPI::Win::NULL_COPY_FN,\n ## ^~~~~~~~~~~~\n ## ../../../cxx/util/mtest.cxx:849:22: error: 'NULL_DELETE_FN' is not a member of 'MPI::Win'\n ## MPI::Win::NULL_DELETE_FN, 0 );\n ## ^~~~~~~~~~~~~~\n ## make[3]: *** [Makefile:416: mtest.o] Error 1\n-## make[3]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/util'\n+## make[3]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/util'\n ## make[2]: *** [Makefile:670: ../../../cxx/util/mtest.o] Error 2\n-## make[2]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/errors/cxx/errhan'\n+## make[2]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/errors/cxx/errhan'\n ]]> \n \n \n \n datatype, &size );\n ## ^\n ## In file included from ../../../cxx/util/mtest.cxx:7:\n ## /usr/lib/x86_64-linux-gnu/openmpi/include/mpi.h:1758:20: note: declared here\n@@ -1006352,34 +1006509,34 @@\n ## ../../../cxx/util/mtest.cxx:848:50: error: 'NULL_COPY_FN' is not a member of 'MPI::Win'\n ## mem_keyval = MPI::Win::Create_keyval( MPI::Win::NULL_COPY_FN,\n ## ^~~~~~~~~~~~\n ## ../../../cxx/util/mtest.cxx:849:22: error: 'NULL_DELETE_FN' is not a member of 'MPI::Win'\n ## MPI::Win::NULL_DELETE_FN, 0 );\n ## ^~~~~~~~~~~~~~\n ## make[3]: *** [Makefile:416: mtest.o] Error 1\n-## make[3]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/util'\n+## make[3]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/util'\n ## make[2]: *** [Makefile:670: ../../../cxx/util/mtest.o] Error 2\n-## make[2]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/errors/cxx/errhan'\n+## make[2]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/errors/cxx/errhan'\n ]]> \n \n \n \n datatype, &size );\n ## ^\n ## In file included from ../../../cxx/util/mtest.cxx:7:\n ## /usr/lib/x86_64-linux-gnu/openmpi/include/mpi.h:1758:20: note: declared here\n@@ -1006411,32 +1006568,32 @@\n ## ../../../cxx/util/mtest.cxx:848:50: error: 'NULL_COPY_FN' is not a member of 'MPI::Win'\n ## mem_keyval = MPI::Win::Create_keyval( MPI::Win::NULL_COPY_FN,\n ## ^~~~~~~~~~~~\n ## ../../../cxx/util/mtest.cxx:849:22: error: 'NULL_DELETE_FN' is not a member of 'MPI::Win'\n ## MPI::Win::NULL_DELETE_FN, 0 );\n ## ^~~~~~~~~~~~~~\n ## make[3]: *** [Makefile:416: mtest.o] Error 1\n-## make[3]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/util'\n+## make[3]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/util'\n ## make[2]: *** [Makefile:670: ../../../cxx/util/mtest.o] Error 2\n-## make[2]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/errors/cxx/errhan'\n+## make[2]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/errors/cxx/errhan'\n ]]> \n \n \n \n \n \n \n \n datatype, &size );\n ## ^\n ## In file included from ../../../cxx/util/mtest.cxx:7:\n ## /usr/lib/x86_64-linux-gnu/openmpi/include/mpi.h:1758:20: note: declared here\n@@ -1006504,34 +1006661,34 @@\n ## ../../../cxx/util/mtest.cxx:848:50: error: 'NULL_COPY_FN' is not a member of 'MPI::Win'\n ## mem_keyval = MPI::Win::Create_keyval( MPI::Win::NULL_COPY_FN,\n ## ^~~~~~~~~~~~\n ## ../../../cxx/util/mtest.cxx:849:22: error: 'NULL_DELETE_FN' is not a member of 'MPI::Win'\n ## MPI::Win::NULL_DELETE_FN, 0 );\n ## ^~~~~~~~~~~~~~\n ## make[3]: *** [Makefile:416: mtest.o] Error 1\n-## make[3]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/util'\n+## make[3]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/util'\n ## make[2]: *** [Makefile:661: ../../../cxx/util/mtest.o] Error 2\n-## make[2]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/errors/cxx/io'\n+## make[2]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/errors/cxx/io'\n ]]> \n \n \n \n datatype, &size );\n ## ^\n ## In file included from ../../../cxx/util/mtest.cxx:7:\n ## /usr/lib/x86_64-linux-gnu/openmpi/include/mpi.h:1758:20: note: declared here\n@@ -1006563,34 +1006720,34 @@\n ## ../../../cxx/util/mtest.cxx:848:50: error: 'NULL_COPY_FN' is not a member of 'MPI::Win'\n ## mem_keyval = MPI::Win::Create_keyval( MPI::Win::NULL_COPY_FN,\n ## ^~~~~~~~~~~~\n ## ../../../cxx/util/mtest.cxx:849:22: error: 'NULL_DELETE_FN' is not a member of 'MPI::Win'\n ## MPI::Win::NULL_DELETE_FN, 0 );\n ## ^~~~~~~~~~~~~~\n ## make[3]: *** [Makefile:416: mtest.o] Error 1\n-## make[3]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/util'\n+## make[3]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/util'\n ## make[2]: *** [Makefile:661: ../../../cxx/util/mtest.o] Error 2\n-## make[2]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/errors/cxx/io'\n+## make[2]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/errors/cxx/io'\n ]]> \n \n \n \n datatype, &size );\n ## ^\n ## In file included from ../../../cxx/util/mtest.cxx:7:\n ## /usr/lib/x86_64-linux-gnu/openmpi/include/mpi.h:1758:20: note: declared here\n@@ -1006622,27 +1006779,27 @@\n ## ../../../cxx/util/mtest.cxx:848:50: error: 'NULL_COPY_FN' is not a member of 'MPI::Win'\n ## mem_keyval = MPI::Win::Create_keyval( MPI::Win::NULL_COPY_FN,\n ## ^~~~~~~~~~~~\n ## ../../../cxx/util/mtest.cxx:849:22: error: 'NULL_DELETE_FN' is not a member of 'MPI::Win'\n ## MPI::Win::NULL_DELETE_FN, 0 );\n ## ^~~~~~~~~~~~~~\n ## make[3]: *** [Makefile:416: mtest.o] Error 1\n-## make[3]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/util'\n+## make[3]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/util'\n ## make[2]: *** [Makefile:661: ../../../cxx/util/mtest.o] Error 2\n-## make[2]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/errors/cxx/io'\n+## make[2]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/errors/cxx/io'\n ]]> \n \n \n \n \n \n" } ] }, { "source1": "./usr/share/mpi-testsuite/results/openmpi/summary.tap", "source2": "./usr/share/mpi-testsuite/results/openmpi/summary.tap", "unified_diff": "@@ -1,10 +1,10 @@\n TAP version 13\n # MPICH test suite results (TAP format)\n-# date 2020-01-04-22-31\n+# date 2018-12-03-15-03\n ok 1 - ./attr/attrt 2\n ok 2 - ./attr/attric 2\n ok 3 - ./attr/attrerr 1\n ok 4 - ./attr/attrend 1\n ok 5 - ./attr/attrend 2\n ok 6 - ./attr/attrend2 1\n ok 7 - ./attr/attrend2 2\n@@ -21,65 +21,65 @@\n ok 18 - ./attr/fkeyvalcomm 1\n ok 19 - ./attr/fkeyvaltype 1\n not ok 20 - ./attr/keyval_double_free 1\n ---\n Directory: ./attr\n File: keyval_double_free\n Num-procs: 1\n- Date: \"Sat Jan 4 22:31:47 2020\"\n+ Date: \"Mon Dec 3 15:03:39 2018\"\n ...\n ## Test output (expected 'No Errors'):\n-## [profitbricks-build5-amd64:49454] *** An error occurred in MPI_Keyval_free\n-## [profitbricks-build5-amd64:49454] *** reported by process [78446593,0]\n-## [profitbricks-build5-amd64:49454] *** on communicator MPI_COMM_WORLD\n-## [profitbricks-build5-amd64:49454] *** MPI_ERR_OTHER: known error not in list\n-## [profitbricks-build5-amd64:49454] *** MPI_ERRORS_ARE_FATAL (processes in this communicator will now abort,\n-## [profitbricks-build5-amd64:49454] *** and potentially your MPI job)\n+## [i-capture-the-hostname:19747] *** An error occurred in MPI_Keyval_free\n+## [i-capture-the-hostname:19747] *** reported by process [3226271745,0]\n+## [i-capture-the-hostname:19747] *** on communicator MPI_COMM_WORLD\n+## [i-capture-the-hostname:19747] *** MPI_ERR_OTHER: known error not in list\n+## [i-capture-the-hostname:19747] *** MPI_ERRORS_ARE_FATAL (processes in this communicator will now abort,\n+## [i-capture-the-hostname:19747] *** and potentially your MPI job)\n not ok 21 - ./attr/keyval_double_free_comm 1\n ---\n Directory: ./attr\n File: keyval_double_free_comm\n Num-procs: 1\n- Date: \"Sat Jan 4 22:31:49 2020\"\n+ Date: \"Mon Dec 3 15:03:41 2018\"\n ...\n ## Test output (expected 'No Errors'):\n-## [profitbricks-build5-amd64:50353] *** An error occurred in MPI_Comm_free_keyval\n-## [profitbricks-build5-amd64:50353] *** reported by process [9568257,0]\n-## [profitbricks-build5-amd64:50353] *** on communicator MPI_COMM_WORLD\n-## [profitbricks-build5-amd64:50353] *** MPI_ERR_OTHER: known error not in list\n-## [profitbricks-build5-amd64:50353] *** MPI_ERRORS_ARE_FATAL (processes in this communicator will now abort,\n-## [profitbricks-build5-amd64:50353] *** and potentially your MPI job)\n+## [i-capture-the-hostname:19933] *** An error occurred in MPI_Comm_free_keyval\n+## [i-capture-the-hostname:19933] *** reported by process [3230728193,0]\n+## [i-capture-the-hostname:19933] *** on communicator MPI_COMM_WORLD\n+## [i-capture-the-hostname:19933] *** MPI_ERR_OTHER: known error not in list\n+## [i-capture-the-hostname:19933] *** MPI_ERRORS_ARE_FATAL (processes in this communicator will now abort,\n+## [i-capture-the-hostname:19933] *** and potentially your MPI job)\n not ok 22 - ./attr/keyval_double_free_type 1\n ---\n Directory: ./attr\n File: keyval_double_free_type\n Num-procs: 1\n- Date: \"Sat Jan 4 22:31:50 2020\"\n+ Date: \"Mon Dec 3 15:03:42 2018\"\n ...\n ## Test output (expected 'No Errors'):\n-## [profitbricks-build5-amd64:51175] *** An error occurred in MPI_Type_free_keyval\n-## [profitbricks-build5-amd64:51175] *** reported by process [59441153,0]\n-## [profitbricks-build5-amd64:51175] *** on communicator MPI_COMM_WORLD\n-## [profitbricks-build5-amd64:51175] *** MPI_ERR_OTHER: known error not in list\n-## [profitbricks-build5-amd64:51175] *** MPI_ERRORS_ARE_FATAL (processes in this communicator will now abort,\n-## [profitbricks-build5-amd64:51175] *** and potentially your MPI job)\n+## [i-capture-the-hostname:20060] *** An error occurred in MPI_Type_free_keyval\n+## [i-capture-the-hostname:20060] *** reported by process [3273261057,0]\n+## [i-capture-the-hostname:20060] *** on communicator MPI_COMM_WORLD\n+## [i-capture-the-hostname:20060] *** MPI_ERR_OTHER: known error not in list\n+## [i-capture-the-hostname:20060] *** MPI_ERRORS_ARE_FATAL (processes in this communicator will now abort,\n+## [i-capture-the-hostname:20060] *** and potentially your MPI job)\n not ok 23 - ./attr/keyval_double_free_win 1\n ---\n Directory: ./attr\n File: keyval_double_free_win\n Num-procs: 1\n- Date: \"Sat Jan 4 22:31:52 2020\"\n+ Date: \"Mon Dec 3 15:03:44 2018\"\n ...\n ## Test output (expected 'No Errors'):\n-## [profitbricks-build5-amd64:52212] *** An error occurred in MPI_Win_free_keyval\n-## [profitbricks-build5-amd64:52212] *** reported by process [257163265,0]\n-## [profitbricks-build5-amd64:52212] *** on communicator MPI_COMM_WORLD\n-## [profitbricks-build5-amd64:52212] *** MPI_ERR_OTHER: known error not in list\n-## [profitbricks-build5-amd64:52212] *** MPI_ERRORS_ARE_FATAL (processes in this communicator will now abort,\n-## [profitbricks-build5-amd64:52212] *** and potentially your MPI job)\n+## [i-capture-the-hostname:20144] *** An error occurred in MPI_Win_free_keyval\n+## [i-capture-the-hostname:20144] *** reported by process [3285450753,0]\n+## [i-capture-the-hostname:20144] *** on communicator MPI_COMM_WORLD\n+## [i-capture-the-hostname:20144] *** MPI_ERR_OTHER: known error not in list\n+## [i-capture-the-hostname:20144] *** MPI_ERRORS_ARE_FATAL (processes in this communicator will now abort,\n+## [i-capture-the-hostname:20144] *** and potentially your MPI job)\n ok 24 - ./coll/allred 2\n ok 25 - ./coll/allred 2\n ok 26 - ./coll/allred 2\n ok 27 - ./coll/allredmany 2\n ok 28 - ./coll/allred2 2\n ok 29 - ./coll/allred3 2\n ok 30 - ./coll/allred4 2\n@@ -107,327 +107,327 @@\n ok 52 - ./coll/allgatherv2 2\n ok 53 - ./coll/allgatherv3 2\n not ok 54 - ./coll/allgatherv4 2\n ---\n Directory: ./coll\n File: allgatherv4\n Num-procs: 2\n- Date: \"Sat Jan 4 22:32:21 2020\"\n+ Date: \"Mon Dec 3 15:04:15 2018\"\n ...\n ## Test output (expected 'No Errors'):\n-## [profitbricks-build5-amd64:63127] *** Process received signal ***\n-## [profitbricks-build5-amd64:63127] Signal: Floating point exception (8)\n-## [profitbricks-build5-amd64:63127] Signal code: Integer divide-by-zero (1)\n-## [profitbricks-build5-amd64:63127] Failing at address: 0x55d2c4ae4b5f\n-## [profitbricks-build5-amd64:63127] [ 0] /lib/x86_64-linux-gnu/libpthread.so.0(+0x126b0)[0x7f8a11f5a6b0]\n-## [profitbricks-build5-amd64:63127] [ 1] ./allgatherv4(+0x3b5f)[0x55d2c4ae4b5f]\n-## [profitbricks-build5-amd64:63127] [ 2] ./allgatherv4(+0x3dce)[0x55d2c4ae4dce]\n-## [profitbricks-build5-amd64:63127] [ 3] ./allgatherv4(+0x3787)[0x55d2c4ae4787]\n-## [profitbricks-build5-amd64:63127] [ 4] /lib/x86_64-linux-gnu/libc.so.6(__libc_start_main+0xeb)[0x7f8a11da409b]\n-## [profitbricks-build5-amd64:63127] [ 5] ./allgatherv4(+0x386a)[0x55d2c4ae486a]\n-## [profitbricks-build5-amd64:63127] *** End of error message ***\n+## [i-capture-the-hostname:23554] *** Process received signal ***\n+## [i-capture-the-hostname:23554] Signal: Floating point exception (8)\n+## [i-capture-the-hostname:23554] Signal code: Integer divide-by-zero (1)\n+## [i-capture-the-hostname:23554] Failing at address: 0x556580c17b5f\n+## [i-capture-the-hostname:23554] [ 0] /lib/x86_64-linux-gnu/libpthread.so.0(+0x126b0)[0x7f904d4f56b0]\n+## [i-capture-the-hostname:23554] [ 1] ./allgatherv4(+0x3b5f)[0x556580c17b5f]\n+## [i-capture-the-hostname:23554] [ 2] ./allgatherv4(+0x3dce)[0x556580c17dce]\n+## [i-capture-the-hostname:23554] [ 3] ./allgatherv4(+0x3787)[0x556580c17787]\n+## [i-capture-the-hostname:23554] [ 4] /lib/x86_64-linux-gnu/libc.so.6(__libc_start_main+0xeb)[0x7f904d33f09b]\n+## [i-capture-the-hostname:23554] [ 5] ./allgatherv4(+0x386a)[0x556580c1786a]\n+## [i-capture-the-hostname:23554] *** End of error message ***\n ## --------------------------------------------------------------------------\n ## Primary job terminated normally, but 1 process returned\n ## a non-zero exit code. Per user-direction, the job has been aborted.\n ## --------------------------------------------------------------------------\n ## --------------------------------------------------------------------------\n-## mpiexec.openmpi noticed that process rank 0 with PID 0 on node profitbricks-build5-amd64 exited on signal 8 (Floating point exception).\n+## mpiexec.openmpi noticed that process rank 0 with PID 0 on node i-capture-the-hostname exited on signal 8 (Floating point exception).\n ## --------------------------------------------------------------------------\n ok 55 - ./coll/allgather_struct 2\n ok 56 - ./coll/bcasttest 2\n ok 57 - ./coll/bcasttest 2\n ok 58 - ./coll/bcasttest 2\n not ok 59 - ./coll/bcast_full 2\n ---\n Directory: ./coll\n File: bcast_full\n Num-procs: 2\n- Date: \"Sat Jan 4 22:37:26 2020\"\n+ Date: \"Mon Dec 3 15:09:20 2018\"\n ...\n ## Test output (expected 'No Errors'):\n+## Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n+## Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n ## Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n ## Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n ## Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n ## Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n-## Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n-## Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n-## Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n-## Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n-## Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n-## Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n-## Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n-## Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n-## Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n-## Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n-## Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n-## Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n-## Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n-## Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n-## Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n ## Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n ## Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n ## Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n ## Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n ## Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n ## Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n ## Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n ## Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n ## Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n ## Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n ## Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n ## Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n ## Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n ## Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n-## Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n-## Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n-## Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n-## Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n-## Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n ## Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n ## Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n ## Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n ## Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n-## Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n-## Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n+## Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n+## Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n ## Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n ## Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n ## Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n ## Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n ## Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n+## Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n+## Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n ## Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n+## Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n ## Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n+## Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n ## Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n ## Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n ## Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n ## Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n ## Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n ## Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n ## Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n ## Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n+## Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n+## Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n ## Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n ## Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n-## Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n-## Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n-## Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n-## Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n+## Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n ## Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n ## Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n ## Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n ## Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n ## Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n ## Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n ## Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n-## Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n-## Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n-## Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n-## Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n-## Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n-## Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n-## Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n-## Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n-## Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n-## Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n-## Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n-## Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n ## Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n ## Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n ## Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n ## Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n ## Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n ## Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n+## Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n+## Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n ## Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n ## Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n-## Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n-## Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n+## Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n ## Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n ## Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n+## Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n ## Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n ## Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n ## Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n ## Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n ## Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n ## Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n-## Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n-## Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n-## Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n ## Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n ## Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n ## Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n ## Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n ## Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n ## Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n ## Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n+## Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n+## Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n ## Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n ## Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n ## Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n ## Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n ## Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n+## Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n ## Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n+## Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n ## Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n+## Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n ## Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n ## Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n ## Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n ## Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n ## Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n ## Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n ## Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n+## Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n+## Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n+## Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n+## Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n+## Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n+## Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n ## Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n ## Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n ## Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n ## Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n ## Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n ## Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n ## Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n ## Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n ## Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n ## Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n-## Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n+## Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n ## Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n ## Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n ## Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n+## Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n+## Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n ## Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n+## Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n ## Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n+## Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n+## Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n ## Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n ## Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n+## Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n ## Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n ## Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n ## Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n ## Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n ## Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n ## Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n ## Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n ## Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n ## Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n-## Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n-## Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n-## Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n-## Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n ## Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n ## Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n ## Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n+## Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n ## Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n+## Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n ## Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n+## Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n+## Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n ## Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n ## Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n+## Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n ## Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n ## Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n ## Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n ## Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n+## Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n+## Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n+## Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n ## Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n ## Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n ## Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n ## Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n+## Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n+## Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n ## Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n+## Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n ## Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n ## Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n-## Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n-## Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n-## Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n-## Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n+## Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n ## Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n ## Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n ## Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n ## Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n ## Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n ## Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n ## Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n ## Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n ## Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n-## Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n-## Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n+## Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n+## Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n ## Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n ## Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n ## Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n ## Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n ## Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n ## Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n ## Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n ## Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n+## Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n ## Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n ## Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n ## Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n ## Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n ## Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n ## Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n-## Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n-## Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n ## Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n ## Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n ## Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n ## Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n+## Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n ## Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n ## Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n ## Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n+## Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n ## Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n+## Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n ## Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n+## Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n ## Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n+## Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n ## Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n ## Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n ## Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n ## Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n ## Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n ## Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n ## Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n+## Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n ## Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n ## Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n ## Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n-## Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n-## Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n-## Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n ## Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n ## Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n ## Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n ## Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n ## Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n+## Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n ## Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n ## Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n ## Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n-## Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n ## Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n ## Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n ## Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n ## Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n+## Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n ## Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n ## Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n-## Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n-## Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n-## Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n ## Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n ## Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n ## Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n ## Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n ## Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n ## Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n ## Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n ## Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n ## Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n ## Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n+## Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n+## Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n+## Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n ## Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n+## Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n+## Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n ## Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n ## Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n ## Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n ## Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n ## Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n ## Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n+## Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n ## Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n ## Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n ## Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n+## Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n+## Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n+## Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n+## Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n ## Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n ## Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n ## Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n ## Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n ## Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n ## Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n-## Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n+## Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n ## Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n ## Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n ## Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n-## Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n-## Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n-## Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n ## MTestTypeSubarrayCheckbuf error\n ## MTestTypeSubarrayCheckbuf error\n ## MTestTypeSubarrayCheckbuf error\n ## MTestTypeSubarrayCheckbuf error\n ## MTestTypeSubarrayCheckbuf error\n ## MTestTypeSubarrayCheckbuf error\n ## MTestTypeSubarrayCheckbuf error\n@@ -446,90 +446,90 @@\n ## MTestTypeSubarrayCheckbuf error\n ## MTestTypeSubarrayCheckbuf error\n ## MTestTypeSubarrayCheckbuf error\n ## MTestTypeSubarrayCheckbuf error\n ## MTestTypeSubarrayCheckbuf error\n ## MTestTypeSubarrayCheckbuf error\n ## MTestTypeSubarrayCheckbuf error\n-## Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n-## Error with communicator MPI_COMM_WORLD and datatype MPI_INT\n ## MTestTypeSubarrayCheckbuf error\n ## MTestTypeSubarrayCheckbuf error\n+## Error with communicator MPI_COMM_WORLD and datatype MPI_INT\n+## Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n ## --------------------------------------------------------------------------\n ## The user-provided time limit for job execution has been reached:\n ## \n ## Timeout: 300 seconds\n ## \n ## The job will now be aborted. Please check your code and/or\n ## adjust/remove the job execution time limit (as specified by --timeout\n ## command line option or MPIEXEC_TIMEOUT environment variable).\n ## --------------------------------------------------------------------------\n not ok 60 - ./coll/bcast_min_datatypes 2\n ---\n Directory: ./coll\n File: bcast_min_datatypes\n Num-procs: 2\n- Date: \"Sat Jan 4 22:57:26 2020\"\n+ Date: \"Mon Dec 3 15:29:23 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n-## Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n-## Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n-## Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n-## Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n-## Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n-## Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n-## Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n-## Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n-## Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n-## Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n-## Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n-## Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n-## Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n-## Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n-## Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n-## Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n-## Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n-## Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n-## Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n-## Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n-## Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n-## Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n-## Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n-## Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n-## Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n-## Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n-## Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n ## Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n+## Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n ## Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n ## Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n+## Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n ## Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n+## Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n ## Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n+## Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n ## Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n+## Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n ## Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n+## Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n ## Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n+## Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n+## Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n ## Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n+## Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n ## Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n+## Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n ## Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n ## Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n+## Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n+## Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n ## Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n+## Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n ## Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n+## Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n ## Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n ## Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n+## Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n+## Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n ## Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n+## Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n ## Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n+## Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n ## Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n+## Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n ## Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n+## Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n ## Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n+## Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n ## Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n ## Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n+## Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n ## Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n+## Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n ## Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n+## Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n ## Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n+## Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n+## Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n ## Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n+## Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n ## Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n ## Error with communicator MPI_COMM_WORLD and datatype MPI_INT\n ## Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n ## --------------------------------------------------------------------------\n ## The user-provided time limit for job execution has been reached:\n ## \n ## Timeout: 1200 seconds\n@@ -539,296 +539,296 @@\n ## command line option or MPIEXEC_TIMEOUT environment variable).\n ## --------------------------------------------------------------------------\n not ok 61 - ./coll/bcast_comm_world 2\n ---\n Directory: ./coll\n File: bcast_comm_world\n Num-procs: 2\n- Date: \"Sat Jan 4 23:17:27 2020\"\n+ Date: \"Mon Dec 3 15:49:24 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n+## Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n ## Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n+## Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n+## Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n+## Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n ## Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n ## Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n ## Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n ## Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n-## Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n-## Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n-## Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n-## Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n-## Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n-## Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n-## Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n-## Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n-## Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n-## Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n-## Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n-## Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n-## Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n-## Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n-## Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n-## Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n ## Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n ## Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n+## Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n+## Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n ## Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n ## Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n ## Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n ## Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n ## Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n ## Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n ## Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n ## Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n-## Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n-## Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n-## Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n-## Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n-## Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n-## Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n-## Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n-## Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n-## Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n-## Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n-## Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n-## Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n-## Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n-## Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n-## Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n-## Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n-## Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n-## Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n-## Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n-## Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n-## Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n-## Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n-## Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n-## Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n-## Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n-## Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n-## Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n-## Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n-## Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n-## Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n-## Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n-## Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n-## Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n-## Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n-## Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n-## Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n-## Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n-## Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n-## Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n-## Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n-## Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n-## Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n-## Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n-## Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n-## Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n-## Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n-## Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n-## Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n-## Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n-## Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n-## Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n-## Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n-## Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n-## Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n-## Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n-## Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n-## Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n-## Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n-## Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n-## Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n-## Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n-## Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n-## Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n-## Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n-## Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n-## Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n-## Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n-## Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n-## Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n-## Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n-## Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n-## Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n-## Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n-## Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n-## Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n-## Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n-## Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n-## Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n-## Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n-## Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n-## Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n-## Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n-## Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n-## Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n-## Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n-## Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n-## Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n-## Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n-## Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n-## Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n-## Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n-## Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n-## Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n-## Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n-## Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n-## Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n-## Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n-## Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n-## Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n-## Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n-## Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n-## Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n-## Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n-## Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n-## Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n-## Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n-## Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n-## Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n-## Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n-## Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n-## Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n-## Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n-## Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n-## Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n-## Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n ## Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n ## Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n+## Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n+## Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n ## Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n ## Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n ## Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n ## Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n+## Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n+## Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n+## Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n+## Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n ## Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n ## Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n ## Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n ## Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n+## Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n+## Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n+## Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n+## Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n ## Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n ## Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n ## Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n ## Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n+## Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n+## Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n ## Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n ## Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n+## Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n+## Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n+## Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n+## Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n ## Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n ## Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n+## Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n+## Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n ## Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n ## Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n ## Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n ## Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n+## Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n+## Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n+## Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n+## Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n ## Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n ## Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n+## Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n+## Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n ## Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n ## Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n+## Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n+## Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n ## Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n ## Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n ## Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n ## Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n+## Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n+## Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n+## Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n+## Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n ## Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n ## Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n+## Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n+## Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n ## Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n ## Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n ## Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n ## Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n+## Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n+## Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n+## Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n ## Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n ## Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n+## Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n ## Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n+## Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n+## Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n ## Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n ## Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n+## Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n ## Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n+## Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n ## Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n+## Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n+## Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n ## Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n+## Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n ## Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n+## Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n ## Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n+## Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n ## Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n+## Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n ## Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n ## Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n+## Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n ## Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n+## Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n+## Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n ## Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n+## Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n ## Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n+## Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n ## Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n+## Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n ## Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n+## Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n ## Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n+## Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n ## Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n+## Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n ## Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n ## Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n+## Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n+## Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n ## Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n+## Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n ## Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n+## Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n ## Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n ## Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n+## Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n+## Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n ## Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n+## Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n ## Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n ## Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n+## Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n ## Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n+## Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n+## Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n ## Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n+## Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n ## Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n ## Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n+## Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n ## Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n+## Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n ## Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n+## Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n ## Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n+## Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n ## Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n+## Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n+## Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n ## Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n+## Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n ## Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n ## Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n+## Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n+## Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n ## Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n ## Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n+## Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n ## Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n+## Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n ## Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n+## Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n+## Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n ## Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n+## Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n ## Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n ## Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n+## Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n ## Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n+## Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n ## Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n+## Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n+## Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n ## Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n ## Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n+## Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n ## Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n+## Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n ## Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n+## Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n ## Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n+## Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n ## Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n+## Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n ## Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n+## Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n+## Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n ## Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n+## Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n ## Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n ## Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n+## Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n ## Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n+## Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n ## Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n+## Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n ## Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n+## Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n+## Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n ## Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n+## Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n ## Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n+## Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n ## Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n ## Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n+## Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n ## Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n+## Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n+## Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n ## Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n+## Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n ## Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n ## Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n+## Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n ## Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n+## Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n ## Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n+## Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n ## Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n+## Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n ## Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n+## Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n+## Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n ## Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n+## Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n ## Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n ## Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n ## Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n+## Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n ## Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n ## Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n+## Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n ## Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n+## Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n+## Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n ## Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n+## Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n ## Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n ## Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n+## Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n ## Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n+## Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n+## Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n ## Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n+## Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n ## Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n ## Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n+## Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n ## Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n+## Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n+## Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n ## Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n+## Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n ## Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n ## Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n+## Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n ## Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n+## Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n+## Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n ## Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n-## Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n-## Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n+## Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n ## Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n ## MTestTypeSubarrayCheckbuf error\n ## MTestTypeSubarrayCheckbuf error\n ## MTestTypeSubarrayCheckbuf error\n ## MTestTypeSubarrayCheckbuf error\n ## MTestTypeSubarrayCheckbuf error\n ## MTestTypeSubarrayCheckbuf error\n@@ -850,16 +850,16 @@\n ## MTestTypeSubarrayCheckbuf error\n ## MTestTypeSubarrayCheckbuf error\n ## MTestTypeSubarrayCheckbuf error\n ## MTestTypeSubarrayCheckbuf error\n ## MTestTypeSubarrayCheckbuf error\n ## MTestTypeSubarrayCheckbuf error\n ## MTestTypeSubarrayCheckbuf error\n-## Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n ## Error with communicator MPI_COMM_WORLD and datatype MPI_INT\n+## Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n ## --------------------------------------------------------------------------\n ## The user-provided time limit for job execution has been reached:\n ## \n ## Timeout: 1200 seconds\n ## \n ## The job will now be aborted. Please check your code and/or\n ## adjust/remove the job execution time limit (as specified by --timeout\n@@ -955,338 +955,338 @@\n ok 149 - ./coll/uoplong 2\n ok 150 - ./coll/uoplong 2\n not ok 151 - ./coll/nonblocking 2\n ---\n Directory: ./coll\n File: nonblocking\n Num-procs: 2\n- Date: \"Sat Jan 4 23:19:42 2020\"\n+ Date: \"Mon Dec 3 15:50:52 2018\"\n ...\n ## Test output (expected 'No Errors'):\n-## [profitbricks-build5-amd64:30536] *** Process received signal ***\n-## [profitbricks-build5-amd64:30536] Signal: Segmentation fault (11)\n-## [profitbricks-build5-amd64:30536] Signal code: (128)\n-## [profitbricks-build5-amd64:30536] Failing at address: (nil)\n-## [profitbricks-build5-amd64:30536] [ 0] /lib/x86_64-linux-gnu/libpthread.so.0(+0x126b0)[0x7f22e27ca6b0]\n-## [profitbricks-build5-amd64:30536] [ 1] /usr/lib/x86_64-linux-gnu/libmpi.so.40(PMPI_Ialltoallw+0x117)[0x7f22e2843287]\n-## [profitbricks-build5-amd64:30536] [ 2] ./nonblocking(+0x3b85)[0x559a420b0b85]\n-## [profitbricks-build5-amd64:30536] [ 3] /lib/x86_64-linux-gnu/libc.so.6(__libc_start_main+0xeb)[0x7f22e261409b]\n-## [profitbricks-build5-amd64:30536] [ 4] ./nonblocking(+0x415a)[0x559a420b115a]\n-## [profitbricks-build5-amd64:30536] *** End of error message ***\n-## [profitbricks-build5-amd64:30537] *** Process received signal ***\n-## [profitbricks-build5-amd64:30537] Signal: Segmentation fault (11)\n-## [profitbricks-build5-amd64:30537] Signal code: (128)\n-## [profitbricks-build5-amd64:30537] Failing at address: (nil)\n+## [i-capture-the-hostname:37607] *** Process received signal ***\n+## [i-capture-the-hostname:37607] Signal: Segmentation fault (11)\n+## [i-capture-the-hostname:37607] Signal code: (128)\n+## [i-capture-the-hostname:37607] Failing at address: (nil)\n+## [i-capture-the-hostname:37609] *** Process received signal ***\n+## [i-capture-the-hostname:37609] Signal: Segmentation fault (11)\n+## [i-capture-the-hostname:37609] Signal code: (128)\n+## [i-capture-the-hostname:37609] Failing at address: (nil)\n+## [i-capture-the-hostname:37607] [ 0] /lib/x86_64-linux-gnu/libpthread.so.0(+0x126b0)[0x7f60f867d6b0]\n+## [i-capture-the-hostname:37607] [ 1] /usr/lib/x86_64-linux-gnu/libmpi.so.40(PMPI_Ialltoallw+0x117)[0x7f60f86f6287]\n+## [i-capture-the-hostname:37607] [ 2] ./nonblocking(+0x3b85)[0x55d5d13c4b85]\n+## [i-capture-the-hostname:37607] [ 3] /lib/x86_64-linux-gnu/libc.so.6(__libc_start_main+0xeb)[0x7f60f84c709b]\n+## [i-capture-the-hostname:37607] [ 4] ./nonblocking(+0x415a)[0x55d5d13c515a]\n+## [i-capture-the-hostname:37607] *** End of error message ***\n+## [i-capture-the-hostname:37609] [ 0] /lib/x86_64-linux-gnu/libpthread.so.0(+0x126b0)[0x7f6fb1d956b0]\n+## [i-capture-the-hostname:37609] [ 1] /usr/lib/x86_64-linux-gnu/libmpi.so.40(PMPI_Ialltoallw+0x117)[0x7f6fb1e0e287]\n+## [i-capture-the-hostname:37609] [ 2] ./nonblocking(+0x3b85)[0x55a43c14fb85]\n+## [i-capture-the-hostname:37609] [ 3] /lib/x86_64-linux-gnu/libc.so.6(__libc_start_main+0xeb)[0x7f6fb1bdf09b]\n+## [i-capture-the-hostname:37609] [ 4] ./nonblocking(+0x415a)[0x55a43c15015a]\n+## [i-capture-the-hostname:37609] *** End of error message ***\n ## --------------------------------------------------------------------------\n ## Primary job terminated normally, but 1 process returned\n ## a non-zero exit code. Per user-direction, the job has been aborted.\n ## --------------------------------------------------------------------------\n-## [profitbricks-build5-amd64:30537] [ 0] /lib/x86_64-linux-gnu/libpthread.so.0(+0x126b0)[0x7fb16074a6b0]\n-## [profitbricks-build5-amd64:30537] [ 1] /usr/lib/x86_64-linux-gnu/libmpi.so.40(PMPI_Ialltoallw+0x117)[0x7fb1607c3287]\n-## [profitbricks-build5-amd64:30537] [ 2] ./nonblocking(+0x3b85)[0x5602f3c53b85]\n-## [profitbricks-build5-amd64:30537] [ 3] /lib/x86_64-linux-gnu/libc.so.6(__libc_start_main+0xeb)[0x7fb16059409b]\n-## [profitbricks-build5-amd64:30537] [ 4] ./nonblocking(+0x415a)[0x5602f3c5415a]\n-## [profitbricks-build5-amd64:30537] *** End of error message ***\n ## --------------------------------------------------------------------------\n-## mpiexec.openmpi noticed that process rank 0 with PID 0 on node profitbricks-build5-amd64 exited on signal 11 (Segmentation fault).\n+## mpiexec.openmpi noticed that process rank 0 with PID 0 on node i-capture-the-hostname exited on signal 11 (Segmentation fault).\n ## --------------------------------------------------------------------------\n not ok 152 - ./coll/nonblocking 2\n ---\n Directory: ./coll\n File: nonblocking\n Num-procs: 2\n- Date: \"Sat Jan 4 23:19:44 2020\"\n+ Date: \"Mon Dec 3 15:50:55 2018\"\n ...\n ## Test output (expected 'No Errors'):\n-## [profitbricks-build5-amd64:30785] *** Process received signal ***\n-## [profitbricks-build5-amd64:30785] Signal: Segmentation fault (11)\n-## [profitbricks-build5-amd64:30785] Signal code: (128)\n-## [profitbricks-build5-amd64:30785] Failing at address: (nil)\n-## [profitbricks-build5-amd64:30783] *** Process received signal ***\n-## [profitbricks-build5-amd64:30783] Signal: Segmentation fault (11)\n-## [profitbricks-build5-amd64:30783] Signal code: (128)\n-## [profitbricks-build5-amd64:30783] Failing at address: (nil)\n-## [profitbricks-build5-amd64:30783] [ 0] /lib/x86_64-linux-gnu/libpthread.so.0(+0x126b0)[0x7fd0c2caa6b0]\n-## [profitbricks-build5-amd64:30783] [ 1] /usr/lib/x86_64-linux-gnu/libmpi.so.40(PMPI_Ialltoallw+0x117)[0x7fd0c2d23287]\n-## [profitbricks-build5-amd64:30783] [ 2] ./nonblocking(+0x3b85)[0x55e59ba0eb85]\n-## [profitbricks-build5-amd64:30783] [ 3] /lib/x86_64-linux-gnu/libc.so.6(__libc_start_main+0xeb)[0x7fd0c2af409b]\n-## [profitbricks-build5-amd64:30783] [ 4] ./nonblocking(+0x415a)[0x55e59ba0f15a]\n-## [profitbricks-build5-amd64:30783] *** End of error message ***\n+## [i-capture-the-hostname:37754] *** Process received signal ***\n+## [i-capture-the-hostname:37754] Signal: Segmentation fault (11)\n+## [i-capture-the-hostname:37754] Signal code: (128)\n+## [i-capture-the-hostname:37754] Failing at address: (nil)\n+## [i-capture-the-hostname:37755] *** Process received signal ***\n+## [i-capture-the-hostname:37755] Signal: Segmentation fault (11)\n+## [i-capture-the-hostname:37755] Signal code: (128)\n+## [i-capture-the-hostname:37755] Failing at address: (nil)\n+## [i-capture-the-hostname:37754] [ 0] /lib/x86_64-linux-gnu/libpthread.so.0(+0x126b0)[0x7fadeb1a56b0]\n+## [i-capture-the-hostname:37754] [ 1] /usr/lib/x86_64-linux-gnu/libmpi.so.40(PMPI_Ialltoallw+0x117)[0x7fadeb21e287]\n+## [i-capture-the-hostname:37754] [ 2] ./nonblocking(+0x3b85)[0x565461425b85]\n+## [i-capture-the-hostname:37754] [ 3] /lib/x86_64-linux-gnu/libc.so.6(__libc_start_main+0xeb)[0x7fadeafef09b]\n+## [i-capture-the-hostname:37754] [ 4] ./nonblocking(+0x415a)[0x56546142615a]\n+## [i-capture-the-hostname:37754] *** End of error message ***\n+## [i-capture-the-hostname:37755] [ 0] /lib/x86_64-linux-gnu/libpthread.so.0(+0x126b0)[0x7f05f0ef56b0]\n+## [i-capture-the-hostname:37755] [ 1] /usr/lib/x86_64-linux-gnu/libmpi.so.40(PMPI_Ialltoallw+0x117)[0x7f05f0f6e287]\n+## [i-capture-the-hostname:37755] [ 2] ./nonblocking(+0x3b85)[0x559ff9929b85]\n+## [i-capture-the-hostname:37755] [ 3] /lib/x86_64-linux-gnu/libc.so.6(__libc_start_main+0xeb)[0x7f05f0d3f09b]\n+## [i-capture-the-hostname:37755] [ 4] ./nonblocking(+0x415a)[0x559ff992a15a]\n+## [i-capture-the-hostname:37755] *** End of error message ***\n ## --------------------------------------------------------------------------\n ## Primary job terminated normally, but 1 process returned\n ## a non-zero exit code. Per user-direction, the job has been aborted.\n ## --------------------------------------------------------------------------\n-## [profitbricks-build5-amd64:30785] [ 0] /lib/x86_64-linux-gnu/libpthread.so.0(+0x126b0)[0x7fde328326b0]\n-## [profitbricks-build5-amd64:30785] [ 1] /usr/lib/x86_64-linux-gnu/libmpi.so.40(PMPI_Ialltoallw+0x117)[0x7fde328ab287]\n-## [profitbricks-build5-amd64:30785] [ 2] ./nonblocking(+0x3b85)[0x55df699ffb85]\n-## [profitbricks-build5-amd64:30785] [ 3] /lib/x86_64-linux-gnu/libc.so.6(__libc_start_main+0xeb)[0x7fde3267c09b]\n-## [profitbricks-build5-amd64:30785] [ 4] ./nonblocking(+0x415a)[0x55df69a0015a]\n-## [profitbricks-build5-amd64:30785] *** End of error message ***\n ## --------------------------------------------------------------------------\n-## mpiexec.openmpi noticed that process rank 0 with PID 0 on node profitbricks-build5-amd64 exited on signal 11 (Segmentation fault).\n+## mpiexec.openmpi noticed that process rank 0 with PID 0 on node i-capture-the-hostname exited on signal 11 (Segmentation fault).\n ## --------------------------------------------------------------------------\n not ok 153 - ./coll/nonblocking 2\n ---\n Directory: ./coll\n File: nonblocking\n Num-procs: 2\n- Date: \"Sat Jan 4 23:19:47 2020\"\n+ Date: \"Mon Dec 3 15:50:58 2018\"\n ...\n ## Test output (expected 'No Errors'):\n-## [profitbricks-build5-amd64:30846] *** Process received signal ***\n-## [profitbricks-build5-amd64:30846] Signal: Segmentation fault (11)\n-## [profitbricks-build5-amd64:30846] Signal code: (128)\n-## [profitbricks-build5-amd64:30846] Failing at address: (nil)\n-## [profitbricks-build5-amd64:30847] *** Process received signal ***\n-## [profitbricks-build5-amd64:30847] Signal: Segmentation fault (11)\n-## [profitbricks-build5-amd64:30847] Signal code: (128)\n-## [profitbricks-build5-amd64:30847] Failing at address: (nil)\n-## [profitbricks-build5-amd64:30846] [ 0] /lib/x86_64-linux-gnu/libpthread.so.0(+0x126b0)[0x7f4cb65826b0]\n-## [profitbricks-build5-amd64:30846] [ 1] /usr/lib/x86_64-linux-gnu/libmpi.so.40(PMPI_Ialltoallw+0x117)[0x7f4cb65fb287]\n-## [profitbricks-build5-amd64:30846] [ 2] ./nonblocking(+0x3b85)[0x55e34c213b85]\n-## [profitbricks-build5-amd64:30846] [ 3] /lib/x86_64-linux-gnu/libc.so.6(__libc_start_main+0xeb)[0x7f4cb63cc09b]\n-## [profitbricks-build5-amd64:30846] [ 4] ./nonblocking(+0x415a)[0x55e34c21415a]\n-## [profitbricks-build5-amd64:30846] *** End of error message ***\n-## [profitbricks-build5-amd64:30847] [ 0] /lib/x86_64-linux-gnu/libpthread.so.0(+0x126b0)[0x7f0c8d5da6b0]\n-## [profitbricks-build5-amd64:30847] [ 1] /usr/lib/x86_64-linux-gnu/libmpi.so.40(PMPI_Ialltoallw+0x117)[0x7f0c8d653287]\n-## [profitbricks-build5-amd64:30847] [ 2] ./nonblocking(+0x3b85)[0x55886d5e5b85]\n-## [profitbricks-build5-amd64:30847] [ 3] /lib/x86_64-linux-gnu/libc.so.6(__libc_start_main+0xeb)[0x7f0c8d42409b]\n-## [profitbricks-build5-amd64:30847] [ 4] ./nonblocking(+0x415a)[0x55886d5e615a]\n-## [profitbricks-build5-amd64:30847] *** End of error message ***\n+## [i-capture-the-hostname:37870] *** Process received signal ***\n+## [i-capture-the-hostname:37870] Signal: Segmentation fault (11)\n+## [i-capture-the-hostname:37870] Signal code: (128)\n+## [i-capture-the-hostname:37870] Failing at address: (nil)\n+## [i-capture-the-hostname:37871] *** Process received signal ***\n+## [i-capture-the-hostname:37871] Signal: Segmentation fault (11)\n+## [i-capture-the-hostname:37871] Signal code: (128)\n+## [i-capture-the-hostname:37871] Failing at address: (nil)\n+## [i-capture-the-hostname:37870] [ 0] /lib/x86_64-linux-gnu/libpthread.so.0(+0x126b0)[0x7f392c84d6b0]\n+## [i-capture-the-hostname:37870] [ 1] [i-capture-the-hostname:37871] [ 0] /lib/x86_64-linux-gnu/libpthread.so.0(+0x126b0)[0x7fb44a9956b0]\n+## [i-capture-the-hostname:37871] [ 1] /usr/lib/x86_64-linux-gnu/libmpi.so.40(PMPI_Ialltoallw+0x117)[0x7fb44aa0e287]\n+## [i-capture-the-hostname:37871] [ 2] ./nonblocking(+0x3b85)[0x55e08bbeab85]\n+## [i-capture-the-hostname:37871] [ 3] /usr/lib/x86_64-linux-gnu/libmpi.so.40(PMPI_Ialltoallw+0x117)[0x7f392c8c6287]\n+## [i-capture-the-hostname:37870] [ 2] ./nonblocking(+0x3b85)[0x555cf20f5b85]\n+## [i-capture-the-hostname:37870] [ 3] /lib/x86_64-linux-gnu/libc.so.6(__libc_start_main+0xeb)[0x7fb44a7df09b]\n+## [i-capture-the-hostname:37871] [ 4] ./nonblocking(+0x415a)[0x55e08bbeb15a]\n+## [i-capture-the-hostname:37871] *** End of error message ***\n+## /lib/x86_64-linux-gnu/libc.so.6(__libc_start_main+0xeb)[0x7f392c69709b]\n+## [i-capture-the-hostname:37870] [ 4] ./nonblocking(+0x415a)[0x555cf20f615a]\n+## [i-capture-the-hostname:37870] *** End of error message ***\n ## --------------------------------------------------------------------------\n ## Primary job terminated normally, but 1 process returned\n ## a non-zero exit code. Per user-direction, the job has been aborted.\n ## --------------------------------------------------------------------------\n ## --------------------------------------------------------------------------\n-## mpiexec.openmpi noticed that process rank 0 with PID 0 on node profitbricks-build5-amd64 exited on signal 11 (Segmentation fault).\n+## mpiexec.openmpi noticed that process rank 0 with PID 0 on node i-capture-the-hostname exited on signal 11 (Segmentation fault).\n ## --------------------------------------------------------------------------\n not ok 154 - ./coll/nonblocking2 1\n ---\n Directory: ./coll\n File: nonblocking2\n Num-procs: 1\n- Date: \"Sat Jan 4 23:19:50 2020\"\n+ Date: \"Mon Dec 3 15:51:00 2018\"\n ...\n ## Test output (expected 'No Errors'):\n-## [profitbricks-build5-amd64:30894] *** Process received signal ***\n-## [profitbricks-build5-amd64:30894] Signal: Segmentation fault (11)\n-## [profitbricks-build5-amd64:30894] Signal code: Address not mapped (1)\n-## [profitbricks-build5-amd64:30894] Failing at address: 0x7fc404afc450\n-## [profitbricks-build5-amd64:30894] [ 0] /lib/x86_64-linux-gnu/libpthread.so.0(+0x126b0)[0x7fc4632b26b0]\n-## [profitbricks-build5-amd64:30894] [ 1] /usr/lib/x86_64-linux-gnu/libmpi.so.40(PMPI_Ialltoallw+0x117)[0x7fc46332b287]\n-## [profitbricks-build5-amd64:30894] [ 2] ./nonblocking2(+0x4dc3)[0x55aa04aefdc3]\n-## [profitbricks-build5-amd64:30894] [ 3] /lib/x86_64-linux-gnu/libc.so.6(__libc_start_main+0xeb)[0x7fc4630fc09b]\n-## [profitbricks-build5-amd64:30894] [ 4] ./nonblocking2(+0x50ca)[0x55aa04af00ca]\n-## [profitbricks-build5-amd64:30894] *** End of error message ***\n+## [i-capture-the-hostname:37895] *** Process received signal ***\n+## [i-capture-the-hostname:37895] Signal: Segmentation fault (11)\n+## [i-capture-the-hostname:37895] Signal code: Address not mapped (1)\n+## [i-capture-the-hostname:37895] Failing at address: 0x7fe7ac466450\n+## [i-capture-the-hostname:37895] [ 0] /lib/x86_64-linux-gnu/libpthread.so.0(+0x126b0)[0x7fe73f7056b0]\n+## [i-capture-the-hostname:37895] [ 1] /usr/lib/x86_64-linux-gnu/libmpi.so.40(PMPI_Ialltoallw+0x117)[0x7fe73f77e287]\n+## [i-capture-the-hostname:37895] [ 2] ./nonblocking2(+0x4dc3)[0x558fac459dc3]\n+## [i-capture-the-hostname:37895] [ 3] /lib/x86_64-linux-gnu/libc.so.6(__libc_start_main+0xeb)[0x7fe73f54f09b]\n+## [i-capture-the-hostname:37895] [ 4] ./nonblocking2(+0x50ca)[0x558fac45a0ca]\n+## [i-capture-the-hostname:37895] *** End of error message ***\n ## --------------------------------------------------------------------------\n ## Primary job terminated normally, but 1 process returned\n ## a non-zero exit code. Per user-direction, the job has been aborted.\n ## --------------------------------------------------------------------------\n ## --------------------------------------------------------------------------\n-## mpiexec.openmpi noticed that process rank 0 with PID 0 on node profitbricks-build5-amd64 exited on signal 11 (Segmentation fault).\n+## mpiexec.openmpi noticed that process rank 0 with PID 0 on node i-capture-the-hostname exited on signal 11 (Segmentation fault).\n ## --------------------------------------------------------------------------\n not ok 155 - ./coll/nonblocking2 2\n ---\n Directory: ./coll\n File: nonblocking2\n Num-procs: 2\n- Date: \"Sat Jan 4 23:19:53 2020\"\n+ Date: \"Mon Dec 3 15:51:02 2018\"\n ...\n ## Test output (expected 'No Errors'):\n-## [profitbricks-build5-amd64:31326] *** Process received signal ***\n-## [profitbricks-build5-amd64:31326] Signal: Segmentation fault (11)\n-## [profitbricks-build5-amd64:31326] Signal code: Address not mapped (1)\n-## [profitbricks-build5-amd64:31326] Failing at address: (nil)\n-## [profitbricks-build5-amd64:31326] [ 0] /lib/x86_64-linux-gnu/libpthread.so.0(+0x126b0)[0x7f73b474a6b0]\n-## [profitbricks-build5-amd64:31326] *** End of error message ***\n+## [i-capture-the-hostname:37937] *** Process received signal ***\n+## [i-capture-the-hostname:37937] Signal: Segmentation fault (11)\n+## [i-capture-the-hostname:37937] Signal code: Address not mapped (1)\n+## [i-capture-the-hostname:37937] Failing at address: (nil)\n+## [i-capture-the-hostname:37937] [ 0] /lib/x86_64-linux-gnu/libpthread.so.0(+0x126b0)[0x7fb16f00d6b0]\n+## [i-capture-the-hostname:37937] *** End of error message ***\n ## --------------------------------------------------------------------------\n ## Primary job terminated normally, but 1 process returned\n ## a non-zero exit code. Per user-direction, the job has been aborted.\n ## --------------------------------------------------------------------------\n ## --------------------------------------------------------------------------\n-## mpiexec.openmpi noticed that process rank 0 with PID 0 on node profitbricks-build5-amd64 exited on signal 11 (Segmentation fault).\n+## mpiexec.openmpi noticed that process rank 0 with PID 0 on node i-capture-the-hostname exited on signal 11 (Segmentation fault).\n ## --------------------------------------------------------------------------\n not ok 156 - ./coll/nonblocking2 2\n ---\n Directory: ./coll\n File: nonblocking2\n Num-procs: 2\n- Date: \"Sat Jan 4 23:19:55 2020\"\n+ Date: \"Mon Dec 3 15:51:04 2018\"\n ...\n ## Test output (expected 'No Errors'):\n-## [profitbricks-build5-amd64:31879] *** Process received signal ***\n-## [profitbricks-build5-amd64:31879] Signal: Segmentation fault (11)\n-## [profitbricks-build5-amd64:31879] Signal code: Address not mapped (1)\n-## [profitbricks-build5-amd64:31879] Failing at address: (nil)\n-## [profitbricks-build5-amd64:31879] [ 0] /lib/x86_64-linux-gnu/libpthread.so.0(+0x126b0)[0x7f5e5d35a6b0]\n-## [profitbricks-build5-amd64:31879] *** End of error message ***\n+## [i-capture-the-hostname:38074] *** Process received signal ***\n+## [i-capture-the-hostname:38074] Signal: Segmentation fault (11)\n+## [i-capture-the-hostname:38074] Signal code: Address not mapped (1)\n+## [i-capture-the-hostname:38074] Failing at address: (nil)\n+## [i-capture-the-hostname:38074] [ 0] /lib/x86_64-linux-gnu/libpthread.so.0(+0x126b0)[0x7f28e79956b0]\n+## [i-capture-the-hostname:38074] *** End of error message ***\n ## --------------------------------------------------------------------------\n ## Primary job terminated normally, but 1 process returned\n ## a non-zero exit code. Per user-direction, the job has been aborted.\n ## --------------------------------------------------------------------------\n ## --------------------------------------------------------------------------\n-## mpiexec.openmpi noticed that process rank 0 with PID 0 on node profitbricks-build5-amd64 exited on signal 11 (Segmentation fault).\n+## mpiexec.openmpi noticed that process rank 0 with PID 0 on node i-capture-the-hostname exited on signal 11 (Segmentation fault).\n ## --------------------------------------------------------------------------\n not ok 157 - ./coll/nonblocking2 2\n ---\n Directory: ./coll\n File: nonblocking2\n Num-procs: 2\n- Date: \"Sat Jan 4 23:19:57 2020\"\n+ Date: \"Mon Dec 3 15:51:06 2018\"\n ...\n ## Test output (expected 'No Errors'):\n-## [profitbricks-build5-amd64:32249] *** Process received signal ***\n-## [profitbricks-build5-amd64:32249] Signal: Segmentation fault (11)\n-## [profitbricks-build5-amd64:32249] Signal code: Address not mapped (1)\n-## [profitbricks-build5-amd64:32249] Failing at address: (nil)\n-## [profitbricks-build5-amd64:32249] [ 0] /lib/x86_64-linux-gnu/libpthread.so.0(+0x126b0)[0x7f739c3ea6b0]\n-## [profitbricks-build5-amd64:32249] *** End of error message ***\n+## [i-capture-the-hostname:38185] *** Process received signal ***\n+## [i-capture-the-hostname:38185] Signal: Segmentation fault (11)\n+## [i-capture-the-hostname:38185] Signal code: Address not mapped (1)\n+## [i-capture-the-hostname:38185] Failing at address: (nil)\n+## [i-capture-the-hostname:38185] [ 0] /lib/x86_64-linux-gnu/libpthread.so.0(+0x126b0)[0x7fe1b2c656b0]\n+## [i-capture-the-hostname:38185] *** End of error message ***\n ## --------------------------------------------------------------------------\n ## Primary job terminated normally, but 1 process returned\n ## a non-zero exit code. Per user-direction, the job has been aborted.\n ## --------------------------------------------------------------------------\n ## --------------------------------------------------------------------------\n-## mpiexec.openmpi noticed that process rank 0 with PID 0 on node profitbricks-build5-amd64 exited on signal 11 (Segmentation fault).\n+## mpiexec.openmpi noticed that process rank 0 with PID 0 on node i-capture-the-hostname exited on signal 11 (Segmentation fault).\n ## --------------------------------------------------------------------------\n not ok 158 - ./coll/nonblocking3 1\n ---\n Directory: ./coll\n File: nonblocking3\n Num-procs: 1\n- Date: \"Sat Jan 4 23:20:00 2020\"\n+ Date: \"Mon Dec 3 15:51:09 2018\"\n ...\n ## Test output (expected 'No Errors'):\n-## [profitbricks-build5-amd64:32604] *** Process received signal ***\n-## [profitbricks-build5-amd64:32604] Signal: Segmentation fault (11)\n-## [profitbricks-build5-amd64:32604] Signal code: Address not mapped (1)\n-## [profitbricks-build5-amd64:32604] Failing at address: 0x4137d590\n-## [profitbricks-build5-amd64:32604] [ 0] /lib/x86_64-linux-gnu/libpthread.so.0(+0x126b0)[0x7fd8e51026b0]\n-## [profitbricks-build5-amd64:32604] [ 1] /usr/lib/x86_64-linux-gnu/libmpi.so.40(PMPI_Ialltoallw+0x117)[0x7fd8e517b287]\n-## [profitbricks-build5-amd64:32604] [ 2] ./nonblocking3(+0x5d84)[0x557941370d84]\n-## [profitbricks-build5-amd64:32604] [ 3] ./nonblocking3(+0x3820)[0x55794136e820]\n-## [profitbricks-build5-amd64:32604] [ 4] /lib/x86_64-linux-gnu/libc.so.6(__libc_start_main+0xeb)[0x7fd8e4f4c09b]\n-## [profitbricks-build5-amd64:32604] [ 5] ./nonblocking3(+0x50ba)[0x5579413700ba]\n-## [profitbricks-build5-amd64:32604] *** End of error message ***\n+## [i-capture-the-hostname:38277] *** Process received signal ***\n+## [i-capture-the-hostname:38277] Signal: Segmentation fault (11)\n+## [i-capture-the-hostname:38277] Signal code: Address not mapped (1)\n+## [i-capture-the-hostname:38277] Failing at address: 0x60fc6590\n+## [i-capture-the-hostname:38277] [ 0] /lib/x86_64-linux-gnu/libpthread.so.0(+0x126b0)[0x7fd76add56b0]\n+## [i-capture-the-hostname:38277] [ 1] /usr/lib/x86_64-linux-gnu/libmpi.so.40(PMPI_Ialltoallw+0x117)[0x7fd76ae4e287]\n+## [i-capture-the-hostname:38277] [ 2] ./nonblocking3(+0x5d84)[0x55ec60fb9d84]\n+## [i-capture-the-hostname:38277] [ 3] ./nonblocking3(+0x3820)[0x55ec60fb7820]\n+## [i-capture-the-hostname:38277] [ 4] /lib/x86_64-linux-gnu/libc.so.6(__libc_start_main+0xeb)[0x7fd76ac1f09b]\n+## [i-capture-the-hostname:38277] [ 5] ./nonblocking3(+0x50ba)[0x55ec60fb90ba]\n+## [i-capture-the-hostname:38277] *** End of error message ***\n ## --------------------------------------------------------------------------\n ## Primary job terminated normally, but 1 process returned\n ## a non-zero exit code. Per user-direction, the job has been aborted.\n ## --------------------------------------------------------------------------\n ## --------------------------------------------------------------------------\n-## mpiexec.openmpi noticed that process rank 0 with PID 0 on node profitbricks-build5-amd64 exited on signal 11 (Segmentation fault).\n+## mpiexec.openmpi noticed that process rank 0 with PID 0 on node i-capture-the-hostname exited on signal 11 (Segmentation fault).\n ## --------------------------------------------------------------------------\n not ok 159 - ./coll/nonblocking3 2\n ---\n Directory: ./coll\n File: nonblocking3\n Num-procs: 2\n- Date: \"Sat Jan 4 23:20:03 2020\"\n+ Date: \"Mon Dec 3 15:51:12 2018\"\n ...\n ## Test output (expected 'No Errors'):\n-## [profitbricks-build5-amd64:32946] *** Process received signal ***\n-## [profitbricks-build5-amd64:32946] Signal: Segmentation fault (11)\n-## [profitbricks-build5-amd64:32946] Signal code: (128)\n-## [profitbricks-build5-amd64:32946] Failing at address: (nil)\n-## [profitbricks-build5-amd64:32946] [ 0] /lib/x86_64-linux-gnu/libpthread.so.0(+0x126b0)[0x7f9c909ea6b0]\n-## [profitbricks-build5-amd64:32946] [ 1] /usr/lib/x86_64-linux-gnu/libmpi.so.40(PMPI_Ialltoallw+0x117)[0x7f9c90a63287]\n-## [profitbricks-build5-amd64:32946] [ 2] ./nonblocking3(+0x5d84)[0x55c0a8161d84]\n-## [profitbricks-build5-amd64:32946] [ 3] ./nonblocking3(+0x3820)[0x55c0a815f820]\n-## [profitbricks-build5-amd64:32946] [ 4] /lib/x86_64-linux-gnu/libc.so.6(__libc_start_main+0xeb)[0x7f9c9083409b]\n-## [profitbricks-build5-amd64:32946] [ 5] ./nonblocking3(+0x50ba)[0x55c0a81610ba]\n-## [profitbricks-build5-amd64:32946] *** End of error message ***\n-## [profitbricks-build5-amd64:32947] *** Process received signal ***\n-## [profitbricks-build5-amd64:32947] Signal: Segmentation fault (11)\n-## [profitbricks-build5-amd64:32947] Signal code: (128)\n-## [profitbricks-build5-amd64:32947] Failing at address: (nil)\n-## [profitbricks-build5-amd64:32947] [ 0] /lib/x86_64-linux-gnu/libpthread.so.0(+0x126b0)[0x7f54dea026b0]\n-## [profitbricks-build5-amd64:32947] [ 1] /usr/lib/x86_64-linux-gnu/libmpi.so.40(PMPI_Ialltoallw+0x117)[0x7f54dea7b287]\n-## [profitbricks-build5-amd64:32947] [ 2] ./nonblocking3(+0x5d84)[0x55b22304cd84]\n-## [profitbricks-build5-amd64:32947] [ 3] ./nonblocking3(+0x3820)[0x55b22304a820]\n-## [profitbricks-build5-amd64:32947] [ 4] /lib/x86_64-linux-gnu/libc.so.6(__libc_start_main+0xeb)[0x7f54de84c09b]\n-## [profitbricks-build5-amd64:32947] [ 5] ./nonblocking3(+0x50ba)[0x55b22304c0ba]\n-## [profitbricks-build5-amd64:32947] *** End of error message ***\n+## [i-capture-the-hostname:38299] *** Process received signal ***\n+## [i-capture-the-hostname:38299] Signal: Segmentation fault (11)\n+## [i-capture-the-hostname:38299] Signal code: (128)\n+## [i-capture-the-hostname:38299] Failing at address: (nil)\n+## [i-capture-the-hostname:38299] [ 0] /lib/x86_64-linux-gnu/libpthread.so.0(+0x126b0)[0x7fc1859cd6b0]\n+## [i-capture-the-hostname:38299] [ 1] /usr/lib/x86_64-linux-gnu/libmpi.so.40(PMPI_Ialltoallw+0x117)[0x7fc185a46287]\n+## [i-capture-the-hostname:38299] [ 2] ./nonblocking3(+0x5d84)[0x55b7dbc50d84]\n+## [i-capture-the-hostname:38299] [ 3] ./nonblocking3(+0x3820)[0x55b7dbc4e820]\n+## [i-capture-the-hostname:38299] [ 4] /lib/x86_64-linux-gnu/libc.so.6(__libc_start_main+0xeb)[0x7fc18581709b]\n+## [i-capture-the-hostname:38299] [ 5] ./nonblocking3(+0x50ba)[0x55b7dbc500ba]\n+## [i-capture-the-hostname:38299] *** End of error message ***\n+## [i-capture-the-hostname:38300] *** Process received signal ***\n+## [i-capture-the-hostname:38300] Signal: Segmentation fault (11)\n+## [i-capture-the-hostname:38300] Signal code: (128)\n+## [i-capture-the-hostname:38300] Failing at address: (nil)\n+## [i-capture-the-hostname:38300] [ 0] /lib/x86_64-linux-gnu/libpthread.so.0(+0x126b0)[0x7f60b80c56b0]\n+## [i-capture-the-hostname:38300] [ 1] /usr/lib/x86_64-linux-gnu/libmpi.so.40(PMPI_Ialltoallw+0x117)[0x7f60b813e287]\n+## [i-capture-the-hostname:38300] [ 2] ./nonblocking3(+0x5d84)[0x5600d2373d84]\n+## [i-capture-the-hostname:38300] [ 3] ./nonblocking3(+0x3820)[0x5600d2371820]\n+## [i-capture-the-hostname:38300] [ 4] /lib/x86_64-linux-gnu/libc.so.6(__libc_start_main+0xeb)[0x7f60b7f0f09b]\n+## [i-capture-the-hostname:38300] [ 5] ./nonblocking3(+0x50ba)[0x5600d23730ba]\n+## [i-capture-the-hostname:38300] *** End of error message ***\n ## --------------------------------------------------------------------------\n ## Primary job terminated normally, but 1 process returned\n ## a non-zero exit code. Per user-direction, the job has been aborted.\n ## --------------------------------------------------------------------------\n ## --------------------------------------------------------------------------\n-## mpiexec.openmpi noticed that process rank 0 with PID 0 on node profitbricks-build5-amd64 exited on signal 11 (Segmentation fault).\n+## mpiexec.openmpi noticed that process rank 0 with PID 0 on node i-capture-the-hostname exited on signal 11 (Segmentation fault).\n ## --------------------------------------------------------------------------\n not ok 160 - ./coll/nonblocking3 2\n ---\n Directory: ./coll\n File: nonblocking3\n Num-procs: 2\n- Date: \"Sat Jan 4 23:20:06 2020\"\n+ Date: \"Mon Dec 3 15:51:15 2018\"\n ...\n ## Test output (expected 'No Errors'):\n-## [profitbricks-build5-amd64:33277] *** Process received signal ***\n-## [profitbricks-build5-amd64:33277] Signal: Segmentation fault (11)\n-## [profitbricks-build5-amd64:33277] Signal code: (128)\n-## [profitbricks-build5-amd64:33277] Failing at address: (nil)\n-## [profitbricks-build5-amd64:33282] *** Process received signal ***\n-## [profitbricks-build5-amd64:33282] Signal: Segmentation fault (11)\n-## [profitbricks-build5-amd64:33282] Signal code: (128)\n-## [profitbricks-build5-amd64:33282] Failing at address: (nil)\n-## [profitbricks-build5-amd64:33277] [ 0] /lib/x86_64-linux-gnu/libpthread.so.0(+0x126b0)[0x7fa10ee726b0]\n-## [profitbricks-build5-amd64:33277] [ 1] /usr/lib/x86_64-linux-gnu/libmpi.so.40(PMPI_Ialltoallw+0x117)[0x7fa10eeeb287]\n-## [profitbricks-build5-amd64:33277] [ 2] ./nonblocking3(+0x5d84)[0x560e67595d84]\n-## [profitbricks-build5-amd64:33277] [ 3] ./nonblocking3(+0x3820)[0x560e67593820]\n-## [profitbricks-build5-amd64:33277] [ 4] /lib/x86_64-linux-gnu/libc.so.6(__libc_start_main+0xeb)[0x7fa10ecbc09b]\n-## [profitbricks-build5-amd64:33277] [ 5] ./nonblocking3(+0x50ba)[0x560e675950ba]\n-## [profitbricks-build5-amd64:33277] *** End of error message ***\n-## [profitbricks-build5-amd64:33282] [ 0] /lib/x86_64-linux-gnu/libpthread.so.0(+0x126b0)[0x7f57161326b0]\n-## [profitbricks-build5-amd64:33282] [ 1] /usr/lib/x86_64-linux-gnu/libmpi.so.40(PMPI_Ialltoallw+0x117)[0x7f57161ab287]\n-## [profitbricks-build5-amd64:33282] [ 2] ./nonblocking3(+0x5d84)[0x55f7faef6d84]\n-## [profitbricks-build5-amd64:33282] [ 3] ./nonblocking3(+0x3820)[0x55f7faef4820]\n-## [profitbricks-build5-amd64:33282] [ 4] /lib/x86_64-linux-gnu/libc.so.6(__libc_start_main+0xeb)[0x7f5715f7c09b]\n-## [profitbricks-build5-amd64:33282] [ 5] ./nonblocking3(+0x50ba)[0x55f7faef60ba]\n-## [profitbricks-build5-amd64:33282] *** End of error message ***\n+## [i-capture-the-hostname:38313] *** Process received signal ***\n+## [i-capture-the-hostname:38313] Signal: Segmentation fault (11)\n+## [i-capture-the-hostname:38313] Signal code: (128)\n+## [i-capture-the-hostname:38313] Failing at address: (nil)\n+## [i-capture-the-hostname:38314] *** Process received signal ***\n+## [i-capture-the-hostname:38314] Signal: Segmentation fault (11)\n+## [i-capture-the-hostname:38314] Signal code: (128)\n+## [i-capture-the-hostname:38314] Failing at address: (nil)\n+## [i-capture-the-hostname:38313] [ 0] /lib/x86_64-linux-gnu/libpthread.so.0(+0x126b0)[0x7fd4ef1cd6b0]\n+## [i-capture-the-hostname:38313] [ 1] /usr/lib/x86_64-linux-gnu/libmpi.so.40(PMPI_Ialltoallw+0x117)[0x7fd4ef246287]\n+## [i-capture-the-hostname:38313] [ 2] ./nonblocking3(+0x5d84)[0x55bbd8de9d84]\n+## [i-capture-the-hostname:38313] [ 3] ./nonblocking3(+0x3820)[0x55bbd8de7820]\n+## [i-capture-the-hostname:38313] [ 4] /lib/x86_64-linux-gnu/libc.so.6(__libc_start_main+0xeb)[0x7fd4ef01709b]\n+## [i-capture-the-hostname:38313] [ 5] ./nonblocking3(+0x50ba)[0x55bbd8de90ba]\n+## [i-capture-the-hostname:38313] *** End of error message ***\n+## [i-capture-the-hostname:38314] [ 0] /lib/x86_64-linux-gnu/libpthread.so.0(+0x126b0)[0x7f87432956b0]\n+## [i-capture-the-hostname:38314] [ 1] /usr/lib/x86_64-linux-gnu/libmpi.so.40(PMPI_Ialltoallw+0x117)[0x7f874330e287]\n+## [i-capture-the-hostname:38314] [ 2] ./nonblocking3(+0x5d84)[0x562cb341bd84]\n+## [i-capture-the-hostname:38314] [ 3] ./nonblocking3(+0x3820)[0x562cb3419820]\n+## [i-capture-the-hostname:38314] [ 4] /lib/x86_64-linux-gnu/libc.so.6(__libc_start_main+0xeb)[0x7f87430df09b]\n+## [i-capture-the-hostname:38314] [ 5] ./nonblocking3(+0x50ba)[0x562cb341b0ba]\n+## [i-capture-the-hostname:38314] *** End of error message ***\n ## --------------------------------------------------------------------------\n ## Primary job terminated normally, but 1 process returned\n ## a non-zero exit code. Per user-direction, the job has been aborted.\n ## --------------------------------------------------------------------------\n ## --------------------------------------------------------------------------\n-## mpiexec.openmpi noticed that process rank 0 with PID 0 on node profitbricks-build5-amd64 exited on signal 11 (Segmentation fault).\n+## mpiexec.openmpi noticed that process rank 0 with PID 0 on node i-capture-the-hostname exited on signal 11 (Segmentation fault).\n ## --------------------------------------------------------------------------\n not ok 161 - ./coll/nonblocking3 2\n ---\n Directory: ./coll\n File: nonblocking3\n Num-procs: 2\n- Date: \"Sat Jan 4 23:20:08 2020\"\n+ Date: \"Mon Dec 3 15:51:17 2018\"\n ...\n ## Test output (expected 'No Errors'):\n-## [profitbricks-build5-amd64:33779] *** Process received signal ***\n-## [profitbricks-build5-amd64:33779] Signal: Segmentation fault (11)\n-## [profitbricks-build5-amd64:33779] Signal code: (128)\n-## [profitbricks-build5-amd64:33779] Failing at address: (nil)\n-## [profitbricks-build5-amd64:33782] *** Process received signal ***\n-## [profitbricks-build5-amd64:33782] Signal: Segmentation fault (11)\n-## [profitbricks-build5-amd64:33782] Signal code: (128)\n-## [profitbricks-build5-amd64:33782] Failing at address: (nil)\n-## [profitbricks-build5-amd64:33779] [ 0] /lib/x86_64-linux-gnu/libpthread.so.0(+0x126b0)[0x7f994932a6b0]\n-## [profitbricks-build5-amd64:33779] [ 1] /usr/lib/x86_64-linux-gnu/libmpi.so.40(PMPI_Ialltoallw+0x117)[0x7f99493a3287]\n-## [profitbricks-build5-amd64:33779] [ 2] ./nonblocking3(+0x5d84)[0x56021c71bd84]\n-## [profitbricks-build5-amd64:33779] [ 3] ./nonblocking3(+0x3820)[0x56021c719820]\n-## [profitbricks-build5-amd64:33779] [ 4] /lib/x86_64-linux-gnu/libc.so.6(__libc_start_main+0xeb)[0x7f994917409b]\n-## [profitbricks-build5-amd64:33779] [ 5] ./nonblocking3(+0x50ba)[0x56021c71b0ba]\n-## [profitbricks-build5-amd64:33779] *** End of error message ***\n-## [profitbricks-build5-amd64:33782] [ 0] /lib/x86_64-linux-gnu/libpthread.so.0(+0x126b0)[0x7fe0b826a6b0]\n-## [profitbricks-build5-amd64:33782] [ 1] /usr/lib/x86_64-linux-gnu/libmpi.so.40(PMPI_Ialltoallw+0x117)[0x7fe0b82e3287]\n-## [profitbricks-build5-amd64:33782] [ 2] ./nonblocking3(+0x5d84)[0x55da0ae00d84]\n-## [profitbricks-build5-amd64:33782] [ 3] ./nonblocking3(+0x3820)[0x55da0adfe820]\n-## [profitbricks-build5-amd64:33782] [ 4] /lib/x86_64-linux-gnu/libc.so.6(__libc_start_main+0xeb)[0x7fe0b80b409b]\n-## [profitbricks-build5-amd64:33782] [ 5] ./nonblocking3(+0x50ba)[0x55da0ae000ba]\n-## [profitbricks-build5-amd64:33782] *** End of error message ***\n+## [i-capture-the-hostname:38338] *** Process received signal ***\n+## [i-capture-the-hostname:38338] Signal: Segmentation fault (11)\n+## [i-capture-the-hostname:38338] Signal code: (128)\n+## [i-capture-the-hostname:38338] Failing at address: (nil)\n+## [i-capture-the-hostname:38338] [ 0] /lib/x86_64-linux-gnu/libpthread.so.0(+0x126b0)[0x7f6100d556b0]\n+## [i-capture-the-hostname:38338] [ 1] /usr/lib/x86_64-linux-gnu/libmpi.so.40(PMPI_Ialltoallw+0x117)[0x7f6100dce287]\n+## [i-capture-the-hostname:38338] [ 2] ./nonblocking3(+0x5d84)[0x55867f79ed84]\n+## [i-capture-the-hostname:38338] [ 3] ./nonblocking3(+0x3820)[0x55867f79c820]\n+## [i-capture-the-hostname:38338] [ 4] /lib/x86_64-linux-gnu/libc.so.6(__libc_start_main+0xeb)[0x7f6100b9f09b]\n+## [i-capture-the-hostname:38338] [ 5] ./nonblocking3(+0x50ba)[0x55867f79e0ba]\n+## [i-capture-the-hostname:38338] *** End of error message ***\n+## [i-capture-the-hostname:38339] *** Process received signal ***\n+## [i-capture-the-hostname:38339] Signal: Segmentation fault (11)\n+## [i-capture-the-hostname:38339] Signal code: (128)\n+## [i-capture-the-hostname:38339] Failing at address: (nil)\n+## [i-capture-the-hostname:38339] [ 0] /lib/x86_64-linux-gnu/libpthread.so.0(+0x126b0)[0x7f0be281d6b0]\n+## [i-capture-the-hostname:38339] [ 1] /usr/lib/x86_64-linux-gnu/libmpi.so.40(PMPI_Ialltoallw+0x117)[0x7f0be2896287]\n+## [i-capture-the-hostname:38339] [ 2] ./nonblocking3(+0x5d84)[0x555ef259ed84]\n+## [i-capture-the-hostname:38339] [ 3] ./nonblocking3(+0x3820)[0x555ef259c820]\n+## [i-capture-the-hostname:38339] [ 4] /lib/x86_64-linux-gnu/libc.so.6(__libc_start_main+0xeb)[0x7f0be266709b]\n+## [i-capture-the-hostname:38339] [ 5] ./nonblocking3(+0x50ba)[0x555ef259e0ba]\n+## [i-capture-the-hostname:38339] *** End of error message ***\n ## --------------------------------------------------------------------------\n ## Primary job terminated normally, but 1 process returned\n ## a non-zero exit code. Per user-direction, the job has been aborted.\n ## --------------------------------------------------------------------------\n ## --------------------------------------------------------------------------\n-## mpiexec.openmpi noticed that process rank 0 with PID 0 on node profitbricks-build5-amd64 exited on signal 11 (Segmentation fault).\n+## mpiexec.openmpi noticed that process rank 0 with PID 0 on node i-capture-the-hostname exited on signal 11 (Segmentation fault).\n ## --------------------------------------------------------------------------\n ok 162 - ./coll/iallred 2\n ok 163 - ./coll/ibarrier 2\n ok 164 - ./coll/nballtoall1 2\n ok 165 - ./coll/nbcoll2 2\n ok 166 - ./coll/nbredscat 2\n ok 167 - ./coll/nbredscat 2\n@@ -1308,110 +1308,110 @@\n ok 183 - ./comm/dup 2\n ok 184 - ./comm/dupic 2\n not ok 185 - ./comm/commcreate1 2\n ---\n Directory: ./comm\n File: commcreate1\n Num-procs: 2\n- Date: \"Sun Jan 5 00:20:37 2020\"\n+ Date: \"Mon Dec 3 15:51:40 2018\"\n ...\n ## Test output (expected 'No Errors'):\n-## [profitbricks-build5-amd64:55812] *** An error occurred in MPI_Group_range_incl\n-## [profitbricks-build5-amd64:55812] *** reported by process [497090561,0]\n-## [profitbricks-build5-amd64:55812] *** on communicator MPI_COMM_WORLD\n-## [profitbricks-build5-amd64:55812] *** MPI_ERR_RANK: invalid rank\n-## [profitbricks-build5-amd64:55812] *** MPI_ERRORS_ARE_FATAL (processes in this communicator will now abort,\n-## [profitbricks-build5-amd64:55812] *** and potentially your MPI job)\n-## [profitbricks-build5-amd64:55667] 1 more process has sent help message help-mpi-errors.txt / mpi_errors_are_fatal\n-## [profitbricks-build5-amd64:55667] Set MCA parameter \"orte_base_help_aggregate\" to 0 to see all help / error messages\n+## [i-capture-the-hostname:39200] *** An error occurred in MPI_Group_range_incl\n+## [i-capture-the-hostname:39200] *** reported by process [340393985,0]\n+## [i-capture-the-hostname:39200] *** on communicator MPI_COMM_WORLD\n+## [i-capture-the-hostname:39200] *** MPI_ERR_RANK: invalid rank\n+## [i-capture-the-hostname:39200] *** MPI_ERRORS_ARE_FATAL (processes in this communicator will now abort,\n+## [i-capture-the-hostname:39200] *** and potentially your MPI job)\n+## [i-capture-the-hostname:39195] 1 more process has sent help message help-mpi-errors.txt / mpi_errors_are_fatal\n+## [i-capture-the-hostname:39195] Set MCA parameter \"orte_base_help_aggregate\" to 0 to see all help / error messages\n ok 186 - ./comm/commname 2\n ok 187 - ./comm/ic1 2\n not ok 188 - ./comm/ic2 2\n ---\n Directory: ./comm\n File: ic2\n Num-procs: 2\n- Date: \"Sun Jan 5 00:20:41 2020\"\n+ Date: \"Mon Dec 3 15:51:44 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## ERROR: this test requires at least 33 processes\n ## ERROR: this test requires at least 33 processes\n ## --------------------------------------------------------------------------\n ## MPI_ABORT was invoked on rank 0 in communicator MPI_COMM_WORLD\n ## with errorcode 1.\n ## \n ## NOTE: invoking MPI_ABORT causes Open MPI to kill all MPI processes.\n ## You may or may not see output from other processes, depending on\n ## exactly when Open MPI kills them.\n ## --------------------------------------------------------------------------\n-## [profitbricks-build5-amd64:57101] 1 more process has sent help message help-mpi-api.txt / mpi-abort\n-## [profitbricks-build5-amd64:57101] Set MCA parameter \"orte_base_help_aggregate\" to 0 to see all help / error messages\n+## [i-capture-the-hostname:39256] 1 more process has sent help message help-mpi-api.txt / mpi-abort\n+## [i-capture-the-hostname:39256] Set MCA parameter \"orte_base_help_aggregate\" to 0 to see all help / error messages\n ok 189 - ./comm/icgroup 2\n ok 190 - ./comm/icm 2\n not ok 191 - ./comm/icsplit 2\n ---\n Directory: ./comm\n File: icsplit\n Num-procs: 2\n- Date: \"Sun Jan 5 00:20:44 2020\"\n+ Date: \"Mon Dec 3 15:51:48 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## This test requires at least 4 processes\n ## This test requires at least 4 processes\n ## --------------------------------------------------------------------------\n ## MPI_ABORT was invoked on rank 1 in communicator MPI_COMM_WORLD\n ## with errorcode 1.\n ## \n ## NOTE: invoking MPI_ABORT causes Open MPI to kill all MPI processes.\n ## You may or may not see output from other processes, depending on\n ## exactly when Open MPI kills them.\n ## --------------------------------------------------------------------------\n-## [profitbricks-build5-amd64:58296] 1 more process has sent help message help-mpi-api.txt / mpi-abort\n-## [profitbricks-build5-amd64:58296] Set MCA parameter \"orte_base_help_aggregate\" to 0 to see all help / error messages\n+## [i-capture-the-hostname:39298] 1 more process has sent help message help-mpi-api.txt / mpi-abort\n+## [i-capture-the-hostname:39298] Set MCA parameter \"orte_base_help_aggregate\" to 0 to see all help / error messages\n not ok 192 - ./comm/iccreate 2\n ---\n Directory: ./comm\n File: iccreate\n Num-procs: 2\n- Date: \"Sun Jan 5 00:20:46 2020\"\n+ Date: \"Mon Dec 3 15:51:50 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## This test requires at least 4 processes\n ## This test requires at least 4 processes\n ## --------------------------------------------------------------------------\n ## MPI_ABORT was invoked on rank 1 in communicator MPI_COMM_WORLD\n ## with errorcode 1.\n ## \n ## NOTE: invoking MPI_ABORT causes Open MPI to kill all MPI processes.\n ## You may or may not see output from other processes, depending on\n ## exactly when Open MPI kills them.\n ## --------------------------------------------------------------------------\n-## [profitbricks-build5-amd64:58904] 1 more process has sent help message help-mpi-api.txt / mpi-abort\n-## [profitbricks-build5-amd64:58904] Set MCA parameter \"orte_base_help_aggregate\" to 0 to see all help / error messages\n+## [i-capture-the-hostname:39312] 1 more process has sent help message help-mpi-api.txt / mpi-abort\n+## [i-capture-the-hostname:39312] Set MCA parameter \"orte_base_help_aggregate\" to 0 to see all help / error messages\n ok 193 - ./comm/ctxalloc 2\n ok 194 - ./comm/ctxsplit 2\n ok 195 - ./comm/cmfree 2\n not ok 196 - ./comm/cmsplit 2\n ---\n Directory: ./comm\n File: cmsplit\n Num-procs: 2\n- Date: \"Sun Jan 5 00:21:00 2020\"\n+ Date: \"Mon Dec 3 15:52:03 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## This test requires at least four processes.This test requires at least four processes.--------------------------------------------------------------------------\n ## MPI_ABORT was invoked on rank 0 in communicator MPI_COMM_WORLD\n ## with errorcode 1.\n ## \n ## NOTE: invoking MPI_ABORT causes Open MPI to kill all MPI processes.\n ## You may or may not see output from other processes, depending on\n ## exactly when Open MPI kills them.\n ## --------------------------------------------------------------------------\n-## [profitbricks-build5-amd64:63276] 1 more process has sent help message help-mpi-api.txt / mpi-abort\n-## [profitbricks-build5-amd64:63276] Set MCA parameter \"orte_base_help_aggregate\" to 0 to see all help / error messages\n+## [i-capture-the-hostname:39425] 1 more process has sent help message help-mpi-api.txt / mpi-abort\n+## [i-capture-the-hostname:39425] Set MCA parameter \"orte_base_help_aggregate\" to 0 to see all help / error messages\n ok 197 - ./comm/cmsplit2 2\n ok 198 - ./comm/probe-intercomm 2\n ok 199 - ./comm/cmsplit_type 2\n ok 200 - ./comm/comm_create_group 2\n ok 201 - ./comm/comm_create_group 2\n ok 202 - ./comm/comm_group_half 2\n ok 203 - ./comm/comm_group_half 2\n@@ -1446,21 +1446,21 @@\n ok 232 - ./datatype/structpack2 1\n ok 233 - ./datatype/typecommit 1\n not ok 234 - ./datatype/typename 1\n ---\n Directory: ./datatype\n File: typename\n Num-procs: 1\n- Date: \"Sun Jan 5 00:21:32 2020\"\n+ Date: \"Mon Dec 3 15:52:37 2018\"\n ...\n ## Test output (expected 'No Errors'):\n+## Found 3 errors\n ## Expected MPI_C_FLOAT_COMPLEX but got MPI_C_COMPLEX\n ## Expected MPI_LONG_LONG but got MPI_LONG_LONG_INT\n ## MPI_INTEGER16 is not available\n-## Found 3 errors\n ok 235 - ./datatype/typefree 1\n ok 236 - ./datatype/zeroparms 1\n ok 237 - ./datatype/getpartelm 2\n ok 238 - ./datatype/tresized 2\n ok 239 - ./datatype/tresized2 2\n ok 240 - ./datatype/sendrecvt2 2\n ok 241 - ./datatype/sendrecvt4 2\n@@ -1469,15 +1469,15 @@\n ok 244 - ./datatype/typelb 1\n ok 245 - ./datatype/contigstruct 1\n not ok 246 - ./datatype/struct-zero-count 1\n ---\n Directory: ./datatype\n File: struct-zero-count\n Num-procs: 1\n- Date: \"Sun Jan 5 00:21:43 2020\"\n+ Date: \"Mon Dec 3 15:52:47 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Found 1 errors\n ok 247 - ./datatype/blockindexed-zero-count 1\n ok 248 - ./datatype/blockindexed-misc 1\n ok 249 - ./datatype/indexed-misc 1\n ok 250 - ./datatype/subarray-pack 1\n@@ -1490,41 +1490,41 @@\n ok 257 - ./datatype/simple-commit 1\n ok 258 - ./datatype/simple-size-extent 1\n not ok 259 - ./datatype/struct-no-real-types 1\n ---\n Directory: ./datatype\n File: struct-no-real-types\n Num-procs: 1\n- Date: \"Sun Jan 5 00:21:54 2020\"\n+ Date: \"Mon Dec 3 15:52:58 2018\"\n ...\n ## Test output (expected 'No Errors'):\n+## Found 1 errors\n ## error: extent is 9223372036854775798 but should be -10 in no_real_types_test()\n ## type map is { (LB,10) }, so UB is 0 and extent is ub-lb\n ## 1 errors in blockindexed test.\n-## Found 1 errors\n ok 260 - ./datatype/struct-empty-el 1\n not ok 261 - ./datatype/contig-zero-count 1\n ---\n Directory: ./datatype\n File: contig-zero-count\n Num-procs: 1\n- Date: \"Sun Jan 5 00:21:56 2020\"\n+ Date: \"Mon Dec 3 15:53:00 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Found 1 errors\n ok 262 - ./datatype/simple-resized 1\n ok 263 - ./datatype/unusual-noncontigs 1\n ok 264 - ./datatype/hindexed-zeros 1\n ok 265 - ./datatype/lots-of-types 1\n not ok 266 - ./datatype/get-elements-pairtype 1\n ---\n Directory: ./datatype\n File: get-elements-pairtype\n Num-procs: 1\n- Date: \"Sun Jan 5 00:22:00 2020\"\n+ Date: \"Mon Dec 3 15:53:05 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Found 1 errors\n ok 267 - ./datatype/unpack 1\n ok 268 - ./datatype/struct-ezhov 1\n ok 269 - ./datatype/zeroblks 1\n ok 270 - ./datatype/struct-derived-zeros 1\n@@ -1538,15 +1538,15 @@\n ok 278 - ./datatype/dataalign 2\n ok 279 - ./datatype/cxx-types 1\n not ok 280 - ./errhan/adderr 1\n ---\n Directory: ./errhan\n File: adderr\n Num-procs: 1\n- Date: \"Sun Jan 5 00:22:12 2020\"\n+ Date: \"Mon Dec 3 15:53:17 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Error class 0 is not a valid error code e 5d\n ## Error class 1 is not a valid error code e 63\n ## Error class 2 is not a valid error code e 69\n ## Error class 3 is not a valid error code e 6f\n ## Error class 4 is not a valid error code e 75\n@@ -1585,75 +1585,75 @@\n ok 285 - ./errhan/errstring2 1\n ok 286 - ./errhan/dynamic_errcode_predefined_errclass 1\n not ok 287 - ./group/@groupcreate \n ---\n Directory: ./group\n File: @groupcreate\n Num-procs: \n- Date: \"Sun Jan 5 00:22:19 2020\"\n+ Date: \"Mon Dec 3 15:53:23 2018\"\n ...\n ## Test output (expected 'No Errors'):\n-## Failed to build @groupcreate; make[2]: Entering directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/group'\n+## Failed to build @groupcreate; make[2]: Entering directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/group'\n ## make[2]: *** No rule to make target '@groupcreate'. Stop.\n-## make[2]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/group'\n+## make[2]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/group'\n not ok 288 - ./group/@grouptest \n ---\n Directory: ./group\n File: @grouptest\n Num-procs: \n- Date: \"Sun Jan 5 00:22:19 2020\"\n+ Date: \"Mon Dec 3 15:53:23 2018\"\n ...\n ## Test output (expected 'No Errors'):\n-## Failed to build @grouptest; make[2]: Entering directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/group'\n+## Failed to build @grouptest; make[2]: Entering directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/group'\n ## make[2]: *** No rule to make target '@grouptest'. Stop.\n-## make[2]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/group'\n+## make[2]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/group'\n not ok 289 - ./group/@grouptest2 \n ---\n Directory: ./group\n File: @grouptest2\n Num-procs: \n- Date: \"Sun Jan 5 00:22:19 2020\"\n+ Date: \"Mon Dec 3 15:53:23 2018\"\n ...\n ## Test output (expected 'No Errors'):\n-## Failed to build @grouptest2; make[2]: Entering directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/group'\n+## Failed to build @grouptest2; make[2]: Entering directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/group'\n ## make[2]: *** No rule to make target '@grouptest2'. Stop.\n-## make[2]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/group'\n+## make[2]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/group'\n not ok 290 - ./group/@groupnullincl \n ---\n Directory: ./group\n File: @groupnullincl\n Num-procs: \n- Date: \"Sun Jan 5 00:22:19 2020\"\n+ Date: \"Mon Dec 3 15:53:23 2018\"\n ...\n ## Test output (expected 'No Errors'):\n-## Failed to build @groupnullincl; make[2]: Entering directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/group'\n+## Failed to build @groupnullincl; make[2]: Entering directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/group'\n ## make[2]: *** No rule to make target '@groupnullincl'. Stop.\n-## make[2]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/group'\n+## make[2]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/group'\n not ok 291 - ./group/@gtranks \n ---\n Directory: ./group\n File: @gtranks\n Num-procs: \n- Date: \"Sun Jan 5 00:22:19 2020\"\n+ Date: \"Mon Dec 3 15:53:23 2018\"\n ...\n ## Test output (expected 'No Errors'):\n-## Failed to build @gtranks; make[2]: Entering directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/group'\n+## Failed to build @gtranks; make[2]: Entering directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/group'\n ## make[2]: *** No rule to make target '@gtranks'. Stop.\n-## make[2]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/group'\n+## make[2]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/group'\n not ok 292 - ./group/@gtranksperf \n ---\n Directory: ./group\n File: @gtranksperf\n Num-procs: \n- Date: \"Sun Jan 5 00:22:19 2020\"\n+ Date: \"Mon Dec 3 15:53:23 2018\"\n ...\n ## Test output (expected 'No Errors'):\n-## Failed to build @gtranksperf; make[2]: Entering directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/group'\n+## Failed to build @gtranksperf; make[2]: Entering directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/group'\n ## make[2]: *** No rule to make target '@gtranksperf'. Stop.\n-## make[2]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/group'\n+## make[2]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/group'\n ok 293 - ./info/infodup 1\n ok 294 - ./info/infodel 1\n ok 295 - ./info/infovallen 1\n ok 296 - ./info/infoorder 1\n ok 297 - ./info/infomany 1\n ok 298 - ./info/infomany2 1\n ok 299 - ./info/infotest 1\n@@ -1663,15 +1663,15 @@\n ok 303 - ./init/exitst2 2\n ok 304 - ./init/initstat 1\n not ok 305 - ./init/timeout 2\n ---\n Directory: ./init\n File: timeout\n Num-procs: 2\n- Date: \"Sun Jan 5 00:22:43 2020\"\n+ Date: \"Mon Dec 3 15:53:48 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## --------------------------------------------------------------------------\n ## The user-provided time limit for job execution has been reached:\n ## \n ## Timeout: 10 seconds\n ## \n@@ -1684,52 +1684,52 @@\n ok 308 - ./init/attrself 1\n ok 309 - ./init/library_version 1\n not ok 310 - ./mpi_t/mpi_t_str 1\n ---\n Directory: ./mpi_t\n File: mpi_t_str\n Num-procs: 1\n- Date: \"Sun Jan 5 00:22:48 2020\"\n+ Date: \"Mon Dec 3 15:53:53 2018\"\n ...\n ## Test output (expected 'No Errors'):\n+## found 1464 errors\n ## check failed: ((strlen(desc) + 1) == min(desc_len, STR_SZ)), line 88\n ## check failed: ((strlen(desc) + 1) == min(desc_len, STR_SZ)), line 88\n ## check failed: ((strlen(desc) + 1) == min(desc_len, STR_SZ)), line 88\n ## check failed: ((strlen(desc) + 1) == min(desc_len, STR_SZ)), line 88\n ## check failed: ((strlen(desc) + 1) == min(desc_len, STR_SZ)), line 88\n ## check failed: ((strlen(desc) + 1) == min(desc_len, STR_SZ)), line 88\n ## check failed: ((strlen(desc) + 1) == min(desc_len, STR_SZ)), line 88\n ## check failed: ((strlen(desc) + 1) == min(desc_len, STR_SZ)), line 88\n ## check failed: ((strlen(desc) + 1) == min(desc_len, STR_SZ)), line 88\n-## found 1464 errors\n not ok 311 - ./mpi_t/mpit_vars 1\n ---\n Directory: ./mpi_t\n File: mpit_vars\n Num-procs: 1\n- Date: \"Sun Jan 5 00:22:50 2020\"\n+ Date: \"Mon Dec 3 15:53:56 2018\"\n ...\n ## Test output (expected 'No Errors'):\n-## [profitbricks-build5-amd64:45361] *** Process received signal ***\n-## [profitbricks-build5-amd64:45361] Signal: Segmentation fault (11)\n-## [profitbricks-build5-amd64:45361] Signal code: Address not mapped (1)\n-## [profitbricks-build5-amd64:45361] Failing at address: (nil)\n-## [profitbricks-build5-amd64:45361] [ 0] /lib/x86_64-linux-gnu/libpthread.so.0(+0x126b0)[0x7fe71f54a6b0]\n-## [profitbricks-build5-amd64:45361] [ 1] /usr/lib/x86_64-linux-gnu/libmpi.so.40(PMPI_T_cvar_read+0x4e)[0x7fe71f5f485e]\n-## [profitbricks-build5-amd64:45361] [ 2] ./mpit_vars(+0x457f)[0x55dbdd7d757f]\n-## [profitbricks-build5-amd64:45361] [ 3] ./mpit_vars(+0x36b5)[0x55dbdd7d66b5]\n-## [profitbricks-build5-amd64:45361] [ 4] /lib/x86_64-linux-gnu/libc.so.6(__libc_start_main+0xeb)[0x7fe71f39409b]\n-## [profitbricks-build5-amd64:45361] [ 5] ./mpit_vars(+0x377a)[0x55dbdd7d677a]\n-## [profitbricks-build5-amd64:45361] *** End of error message ***\n+## [i-capture-the-hostname:49703] *** Process received signal ***\n+## [i-capture-the-hostname:49703] Signal: Segmentation fault (11)\n+## [i-capture-the-hostname:49703] Signal code: Address not mapped (1)\n+## [i-capture-the-hostname:49703] Failing at address: (nil)\n+## [i-capture-the-hostname:49703] [ 0] /lib/x86_64-linux-gnu/libpthread.so.0(+0x126b0)[0x7f0df58356b0]\n+## [i-capture-the-hostname:49703] [ 1] /usr/lib/x86_64-linux-gnu/libmpi.so.40(PMPI_T_cvar_read+0x4e)[0x7f0df58df85e]\n+## [i-capture-the-hostname:49703] [ 2] ./mpit_vars(+0x457f)[0x5590e91d957f]\n+## [i-capture-the-hostname:49703] [ 3] ./mpit_vars(+0x36b5)[0x5590e91d86b5]\n+## [i-capture-the-hostname:49703] [ 4] /lib/x86_64-linux-gnu/libc.so.6(__libc_start_main+0xeb)[0x7f0df567f09b]\n+## [i-capture-the-hostname:49703] [ 5] ./mpit_vars(+0x377a)[0x5590e91d877a]\n+## [i-capture-the-hostname:49703] *** End of error message ***\n ## --------------------------------------------------------------------------\n ## Primary job terminated normally, but 1 process returned\n ## a non-zero exit code. Per user-direction, the job has been aborted.\n ## --------------------------------------------------------------------------\n ## --------------------------------------------------------------------------\n-## mpiexec.openmpi noticed that process rank 0 with PID 0 on node profitbricks-build5-amd64 exited on signal 11 (Segmentation fault).\n+## mpiexec.openmpi noticed that process rank 0 with PID 0 on node i-capture-the-hostname exited on signal 11 (Segmentation fault).\n ## --------------------------------------------------------------------------\n ok 312 - ./mpi_t/cvarwrite 1\n ok 313 - ./mpi_t/getindex 1 # SKIP requires MPI version 3.1\n ok 313 - ./pt2pt/sendrecv1 2\n ok 314 - ./pt2pt/sendrecv2 2\n ok 315 - ./pt2pt/sendrecv3 2\n ok 316 - ./pt2pt/sendflood 2\n@@ -1757,15 +1757,15 @@\n ok 338 - ./pt2pt/probe-unexp 2\n ok 339 - ./pt2pt/probenull 1\n not ok 340 - ./pt2pt/scancel 2\n ---\n Directory: ./pt2pt\n File: scancel\n Num-procs: 2\n- Date: \"Sun Jan 5 00:27:30 2020\"\n+ Date: \"Mon Dec 3 15:58:34 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Failed to cancel an Isend request\n ## Failed to cancel an Ibsend request\n ## --------------------------------------------------------------------------\n ## The user-provided time limit for job execution has been reached:\n ## \n@@ -1777,15 +1777,15 @@\n ## --------------------------------------------------------------------------\n ok 341 - ./pt2pt/scancel2 2\n not ok 342 - ./pt2pt/pscancel 2\n ---\n Directory: ./pt2pt\n File: pscancel\n Num-procs: 2\n- Date: \"Sun Jan 5 00:30:32 2020\"\n+ Date: \"Mon Dec 3 16:01:36 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Failed to cancel a persistent send request\n ## Failed to cancel a persistent bsend request\n ## --------------------------------------------------------------------------\n ## The user-provided time limit for job execution has been reached:\n ## \n@@ -1797,15 +1797,15 @@\n ## --------------------------------------------------------------------------\n ok 343 - ./pt2pt/rcancel 2\n not ok 344 - ./pt2pt/cancelrecv 2\n ---\n Directory: ./pt2pt\n File: cancelrecv\n Num-procs: 2\n- Date: \"Sun Jan 5 00:33:34 2020\"\n+ Date: \"Mon Dec 3 16:04:37 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## --------------------------------------------------------------------------\n ## The user-provided time limit for job execution has been reached:\n ## \n ## Timeout: 180 seconds\n ## \n@@ -1814,15 +1814,15 @@\n ## command line option or MPIEXEC_TIMEOUT environment variable).\n ## --------------------------------------------------------------------------\n not ok 345 - ./pt2pt/scancel_unmatch 2\n ---\n Directory: ./pt2pt\n File: scancel_unmatch\n Num-procs: 2\n- Date: \"Sun Jan 5 00:33:35 2020\"\n+ Date: \"Mon Dec 3 16:04:38 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Failed to cancel send Found 1 errors\n ok 346 - ./pt2pt/cancelanysrc 2\n ok 347 - ./pt2pt/isendselfprobe 1\n ok 348 - ./pt2pt/inactivereq 1\n ok 349 - ./pt2pt/waittestnull 1\n@@ -1834,29 +1834,29 @@\n ok 355 - ./rma/putfidx 2\n ok 356 - ./rma/accfence1 2\n not ok 357 - ./rma/adlb_mimic1 2\n ---\n Directory: ./rma\n File: adlb_mimic1\n Num-procs: 2\n- Date: \"Sun Jan 5 00:34:00 2020\"\n+ Date: \"Mon Dec 3 16:05:03 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## This test requires at least 3 processes\n ## This test requires at least 3 processes\n ## --------------------------------------------------------------------------\n-## MPI_ABORT was invoked on rank 1 in communicator MPI_COMM_WORLD\n+## MPI_ABORT was invoked on rank 0 in communicator MPI_COMM_WORLD\n ## with errorcode 1.\n ## \n ## NOTE: invoking MPI_ABORT causes Open MPI to kill all MPI processes.\n ## You may or may not see output from other processes, depending on\n ## exactly when Open MPI kills them.\n ## --------------------------------------------------------------------------\n-## [profitbricks-build5-amd64:38388] 1 more process has sent help message help-mpi-api.txt / mpi-abort\n-## [profitbricks-build5-amd64:38388] Set MCA parameter \"orte_base_help_aggregate\" to 0 to see all help / error messages\n+## [i-capture-the-hostname:33131] 1 more process has sent help message help-mpi-api.txt / mpi-abort\n+## [i-capture-the-hostname:33131] Set MCA parameter \"orte_base_help_aggregate\" to 0 to see all help / error messages\n ok 358 - ./rma/accfence2 2\n ok 359 - ./rma/putpscw1 2\n ok 360 - ./rma/accpscw1 2\n ok 361 - ./rma/getgroup 2\n ok 362 - ./rma/transpose1 2\n ok 363 - ./rma/transpose2 2\n ok 364 - ./rma/transpose3 2\n@@ -1872,292 +1872,313 @@\n ok 374 - ./rma/test4 2\n ok 375 - ./rma/test5 2\n not ok 376 - ./rma/lockcontention 2\n ---\n Directory: ./rma\n File: lockcontention\n Num-procs: 2\n- Date: \"Sun Jan 5 00:34:41 2020\"\n+ Date: \"Mon Dec 3 16:05:47 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Run this program with 3 or more processes\n ## Run this program with 3 or more processes\n ## --------------------------------------------------------------------------\n ## MPI_ABORT was invoked on rank 1 in communicator MPI_COMM_WORLD\n ## with errorcode 1.\n ## \n ## NOTE: invoking MPI_ABORT causes Open MPI to kill all MPI processes.\n ## You may or may not see output from other processes, depending on\n ## exactly when Open MPI kills them.\n ## --------------------------------------------------------------------------\n-## [profitbricks-build5-amd64:51493] 1 more process has sent help message help-mpi-api.txt / mpi-abort\n-## [profitbricks-build5-amd64:51493] Set MCA parameter \"orte_base_help_aggregate\" to 0 to see all help / error messages\n+## [i-capture-the-hostname:36578] 1 more process has sent help message help-mpi-api.txt / mpi-abort\n+## [i-capture-the-hostname:36578] Set MCA parameter \"orte_base_help_aggregate\" to 0 to see all help / error messages\n ok 377 - ./rma/lockcontention2 2\n ok 378 - ./rma/lockcontention2 2\n not ok 379 - ./rma/lockcontention3 2\n ---\n Directory: ./rma\n File: lockcontention3\n Num-procs: 2\n- Date: \"Sun Jan 5 00:34:45 2020\"\n+ Date: \"Mon Dec 3 16:05:51 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## This test requires at least 3 processes\n ## This test requires at least 3 processes\n ## --------------------------------------------------------------------------\n-## MPI_ABORT was invoked on rank 1 in communicator MPI_COMM_WORLD\n+## MPI_ABORT was invoked on rank 0 in communicator MPI_COMM_WORLD\n ## with errorcode 1.\n ## \n ## NOTE: invoking MPI_ABORT causes Open MPI to kill all MPI processes.\n ## You may or may not see output from other processes, depending on\n ## exactly when Open MPI kills them.\n ## --------------------------------------------------------------------------\n-## [profitbricks-build5-amd64:52424] 1 more process has sent help message help-mpi-api.txt / mpi-abort\n-## [profitbricks-build5-amd64:52424] Set MCA parameter \"orte_base_help_aggregate\" to 0 to see all help / error messages\n+## [i-capture-the-hostname:37006] 1 more process has sent help message help-mpi-api.txt / mpi-abort\n+## [i-capture-the-hostname:37006] Set MCA parameter \"orte_base_help_aggregate\" to 0 to see all help / error messages\n ok 380 - ./rma/lockopts 2\n not ok 381 - ./rma/lock_dt 2\n ---\n Directory: ./rma\n File: lock_dt\n Num-procs: 2\n- Date: \"Sun Jan 5 00:37:46 2020\"\n+ Date: \"Mon Dec 3 16:08:54 2018\"\n ...\n ## Test output (expected 'No Errors'):\n+## [i-capture-the-hostname:37260] *** Process received signal ***\n+## [i-capture-the-hostname:37260] Signal: Segmentation fault (11)\n+## [i-capture-the-hostname:37260] Signal code: Address not mapped (1)\n+## [i-capture-the-hostname:37260] Failing at address: 0x55d831d1e0c0\n ## malloc(): corrupted top size\n-## [profitbricks-build5-amd64:53184] *** Process received signal ***\n-## [profitbricks-build5-amd64:53184] Signal: Aborted (6)\n-## [profitbricks-build5-amd64:53184] Signal code: (-6)\n+## [i-capture-the-hostname:37260] *** Process received signal ***\n+## [i-capture-the-hostname:37260] Signal: Aborted (6)\n+## [i-capture-the-hostname:37260] Signal code: (-6)\n ## --------------------------------------------------------------------------\n ## The user-provided time limit for job execution has been reached:\n ## \n ## Timeout: 180 seconds\n ## \n ## The job will now be aborted. Please check your code and/or\n ## adjust/remove the job execution time limit (as specified by --timeout\n ## command line option or MPIEXEC_TIMEOUT environment variable).\n ## --------------------------------------------------------------------------\n not ok 382 - ./rma/lock_dt_flush 2\n ---\n Directory: ./rma\n File: lock_dt_flush\n Num-procs: 2\n- Date: \"Sun Jan 5 00:37:48 2020\"\n+ Date: \"Mon Dec 3 16:08:56 2018\"\n ...\n ## Test output (expected 'No Errors'):\n-## [profitbricks-build5-amd64:51145] *** Process received signal ***\n-## [profitbricks-build5-amd64:51145] Signal: Segmentation fault (11)\n-## [profitbricks-build5-amd64:51145] Signal code: Address not mapped (1)\n-## [profitbricks-build5-amd64:51145] Failing at address: 0x55bcbd5cb3e0\n-## [profitbricks-build5-amd64:51145] [ 0] /lib/x86_64-linux-gnu/libpthread.so.0(+0x126b0)[0x7f55c0dba6b0]\n-## [profitbricks-build5-amd64:51145] [ 1] /lib/x86_64-linux-gnu/libc.so.6(+0x13ad67)[0x7f55c0d1ad67]\n-## [profitbricks-build5-amd64:51145] [ 2] /usr/lib/x86_64-linux-gnu/libopen-pal.so.40(opal_generic_simple_unpack+0x4e6)[0x7f55c0ab4596]\n-## [profitbricks-build5-amd64:51145] [ 3] /usr/lib/x86_64-linux-gnu/openmpi/lib/openmpi3/mca_pml_ob1.so(mca_pml_ob1_recv_request_progress_frag+0x1c2)[0x7f55be1e8b62]\n-## [profitbricks-build5-amd64:51145] [ 4] /usr/lib/x86_64-linux-gnu/openmpi/lib/openmpi3/mca_btl_vader.so(mca_btl_vader_poll_handle_frag+0x8f)[0x7f55be2246bf]\n-## [profitbricks-build5-amd64:51145] [ 5] /usr/lib/x86_64-linux-gnu/openmpi/lib/openmpi3/mca_btl_vader.so(+0x4b8a)[0x7f55be224b8a]\n-## [profitbricks-build5-amd64:51145] [ 6] /usr/lib/x86_64-linux-gnu/libopen-pal.so.40(opal_progress+0x2c)[0x7f55c0a98e0c]\n-## [profitbricks-build5-amd64:51145] [ 7] /usr/lib/x86_64-linux-gnu/libopen-pal.so.40(ompi_sync_wait_mt+0x115)[0x7f55c0a9f835]\n-## [profitbricks-build5-amd64:51145] [ 8] /usr/lib/x86_64-linux-gnu/libmpi.so.40(ompi_request_default_wait+0x1d9)[0x7f55c0e1a9b9]\n-## [profitbricks-build5-amd64:51145] [ 9] /usr/lib/x86_64-linux-gnu/libmpi.so.40(ompi_coll_base_barrier_intra_two_procs+0xea)[0x7f55c0e77c8a]\n-## [profitbricks-build5-amd64:51145] [10] /usr/lib/x86_64-linux-gnu/libmpi.so.40(MPI_Barrier+0xa8)[0x7f55c0e33978]\n-## [profitbricks-build5-amd64:51145] [11] ./lock_dt_flush(+0x386b)[0x55bcbbc8386b]\n-## [profitbricks-build5-amd64:51145] [12] /lib/x86_64-linux-gnu/libc.so.6(__libc_start_main+0xeb)[0x7f55c0c0409b]\n-## [profitbricks-build5-amd64:51145] [13] ./lock_dt_flush(+0x393a)[0x55bcbbc8393a]\n-## [profitbricks-build5-amd64:51145] *** End of error message ***\n+## [i-capture-the-hostname:50873] *** Process received signal ***\n+## [i-capture-the-hostname:50873] Signal: Segmentation fault (11)\n+## [i-capture-the-hostname:50873] Signal code: Address not mapped (1)\n+## [i-capture-the-hostname:50873] Failing at address: 0x562662628260\n+## [i-capture-the-hostname:50873] [ 0] /lib/x86_64-linux-gnu/libpthread.so.0(+0x126b0)[0x7f386cd7d6b0]\n+## [i-capture-the-hostname:50873] [ 1] /lib/x86_64-linux-gnu/libc.so.6(+0x13ad67)[0x7f386ccddd67]\n+## [i-capture-the-hostname:50873] [ 2] /usr/lib/x86_64-linux-gnu/libopen-pal.so.40(opal_generic_simple_unpack+0x4e6)[0x7f386ca77596]\n+## [i-capture-the-hostname:50873] [ 3] /usr/lib/x86_64-linux-gnu/openmpi/lib/openmpi3/mca_pml_ob1.so(mca_pml_ob1_recv_request_progress_frag+0x1c2)[0x7f386a1abb62]\n+## [i-capture-the-hostname:50873] [ 4] /usr/lib/x86_64-linux-gnu/openmpi/lib/openmpi3/mca_btl_vader.so(mca_btl_vader_poll_handle_frag+0x8f)[0x7f386a1e76bf]\n+## [i-capture-the-hostname:50873] [ 5] /usr/lib/x86_64-linux-gnu/openmpi/lib/openmpi3/mca_btl_vader.so(+0x4b8a)[0x7f386a1e7b8a]\n+## [i-capture-the-hostname:50873] [ 6] /usr/lib/x86_64-linux-gnu/libopen-pal.so.40(opal_progress+0x2c)[0x7f386ca5be0c]\n+## [i-capture-the-hostname:50873] [ 7] /usr/lib/x86_64-linux-gnu/libopen-pal.so.40(ompi_sync_wait_mt+0x115)[0x7f386ca62835]\n+## [i-capture-the-hostname:50873] [ 8] /usr/lib/x86_64-linux-gnu/libmpi.so.40(ompi_request_default_wait+0x1d9)[0x7f386cddd9b9]\n+## [i-capture-the-hostname:50873] [ 9] /usr/lib/x86_64-linux-gnu/libmpi.so.40(ompi_coll_base_barrier_intra_two_procs+0xea)[0x7f386ce3ac8a]\n+## [i-capture-the-hostname:50873] [10] /usr/lib/x86_64-linux-gnu/libmpi.so.40(MPI_Barrier+0xa8)[0x7f386cdf6978]\n+## [i-capture-the-hostname:50873] [11] ./lock_dt_flush(+0x386b)[0x5626603df86b]\n+## [i-capture-the-hostname:50873] [12] /lib/x86_64-linux-gnu/libc.so.6(__libc_start_main+0xeb)[0x7f386cbc709b]\n+## [i-capture-the-hostname:50873] [13] ./lock_dt_flush(+0x393a)[0x5626603df93a]\n+## [i-capture-the-hostname:50873] *** End of error message ***\n ## --------------------------------------------------------------------------\n ## Primary job terminated normally, but 1 process returned\n ## a non-zero exit code. Per user-direction, the job has been aborted.\n ## --------------------------------------------------------------------------\n ## --------------------------------------------------------------------------\n-## mpiexec.openmpi noticed that process rank 1 with PID 0 on node profitbricks-build5-amd64 exited on signal 11 (Segmentation fault).\n+## mpiexec.openmpi noticed that process rank 1 with PID 0 on node i-capture-the-hostname exited on signal 11 (Segmentation fault).\n ## --------------------------------------------------------------------------\n not ok 383 - ./rma/lock_dt_flushlocal 2\n ---\n Directory: ./rma\n File: lock_dt_flushlocal\n Num-procs: 2\n- Date: \"Sun Jan 5 00:40:49 2020\"\n+ Date: \"Mon Dec 3 16:08:58 2018\"\n ...\n ## Test output (expected 'No Errors'):\n-## [profitbricks-build5-amd64:51804] *** Process received signal ***\n-## [profitbricks-build5-amd64:51804] Signal: Segmentation fault (11)\n-## [profitbricks-build5-amd64:51804] Signal code: Address not mapped (1)\n-## [profitbricks-build5-amd64:51804] Failing at address: 0x560c5b483030\n-## malloc(): corrupted top size\n-## [profitbricks-build5-amd64:51804] *** Process received signal ***\n-## [profitbricks-build5-amd64:51804] Signal: Aborted (6)\n-## [profitbricks-build5-amd64:51804] Signal code: (-6)\n+## [i-capture-the-hostname:50905] *** Process received signal ***\n+## [i-capture-the-hostname:50905] Signal: Segmentation fault (11)\n+## [i-capture-the-hostname:50905] Signal code: Address not mapped (1)\n+## [i-capture-the-hostname:50905] Failing at address: 0x55ab28e87050\n+## [i-capture-the-hostname:50905] [ 0] /lib/x86_64-linux-gnu/libpthread.so.0(+0x126b0)[0x7f68e594d6b0]\n+## [i-capture-the-hostname:50905] [ 1] /lib/x86_64-linux-gnu/libc.so.6(+0x13ad67)[0x7f68e58add67]\n+## [i-capture-the-hostname:50905] [ 2] /usr/lib/x86_64-linux-gnu/libopen-pal.so.40(opal_generic_simple_unpack+0x4e6)[0x7f68e5647596]\n+## [i-capture-the-hostname:50905] [ 3] /usr/lib/x86_64-linux-gnu/openmpi/lib/openmpi3/mca_pml_ob1.so(mca_pml_ob1_recv_request_progress_frag+0x1c2)[0x7f68e2d7bb62]\n+## [i-capture-the-hostname:50905] [ 4] /usr/lib/x86_64-linux-gnu/openmpi/lib/openmpi3/mca_btl_vader.so(mca_btl_vader_poll_handle_frag+0x8f)[0x7f68e2db76bf]\n+## [i-capture-the-hostname:50905] [ 5] /usr/lib/x86_64-linux-gnu/openmpi/lib/openmpi3/mca_btl_vader.so(+0x4b8a)[0x7f68e2db7b8a]\n+## [i-capture-the-hostname:50905] [ 6] /usr/lib/x86_64-linux-gnu/libopen-pal.so.40(opal_progress+0x2c)[0x7f68e562be0c]\n+## [i-capture-the-hostname:50905] [ 7] /usr/lib/x86_64-linux-gnu/libopen-pal.so.40(ompi_sync_wait_mt+0x115)[0x7f68e5632835]\n+## [i-capture-the-hostname:50905] [ 8] /usr/lib/x86_64-linux-gnu/libmpi.so.40(ompi_request_default_wait+0x1d9)[0x7f68e59ad9b9]\n+## [i-capture-the-hostname:50905] [ 9] /usr/lib/x86_64-linux-gnu/libmpi.so.40(ompi_coll_base_barrier_intra_two_procs+0xea)[0x7f68e5a0ac8a]\n+## [i-capture-the-hostname:50905] [10] /usr/lib/x86_64-linux-gnu/libmpi.so.40(MPI_Barrier+0xa8)[0x7f68e59c6978]\n+## [i-capture-the-hostname:50905] [11] ./lock_dt_flushlocal(+0x38be)[0x55ab27eee8be]\n+## [i-capture-the-hostname:50905] [12] /lib/x86_64-linux-gnu/libc.so.6(__libc_start_main+0xeb)[0x7f68e579709b]\n+## [i-capture-the-hostname:50905] [13] ./lock_dt_flushlocal(+0x398a)[0x55ab27eee98a]\n+## [i-capture-the-hostname:50905] *** End of error message ***\n ## --------------------------------------------------------------------------\n-## The user-provided time limit for job execution has been reached:\n-## \n-## Timeout: 180 seconds\n-## \n-## The job will now be aborted. Please check your code and/or\n-## adjust/remove the job execution time limit (as specified by --timeout\n-## command line option or MPIEXEC_TIMEOUT environment variable).\n+## Primary job terminated normally, but 1 process returned\n+## a non-zero exit code. Per user-direction, the job has been aborted.\n+## --------------------------------------------------------------------------\n+## --------------------------------------------------------------------------\n+## mpiexec.openmpi noticed that process rank 1 with PID 0 on node i-capture-the-hostname exited on signal 11 (Segmentation fault).\n ## --------------------------------------------------------------------------\n not ok 384 - ./rma/lockall_dt 2\n ---\n Directory: ./rma\n File: lockall_dt\n Num-procs: 2\n- Date: \"Sun Jan 5 00:44:49 2020\"\n+ Date: \"Mon Dec 3 16:09:00 2018\"\n+ ...\n+## Test output (expected 'No Errors'):\n+## [i-capture-the-hostname:51015] *** Process received signal ***\n+## [i-capture-the-hostname:51015] Signal: Segmentation fault (11)\n+## [i-capture-the-hostname:51015] Signal code: Address not mapped (1)\n+## [i-capture-the-hostname:51015] Failing at address: 0x56121ca8c290\n+## [i-capture-the-hostname:51015] [ 0] /lib/x86_64-linux-gnu/libpthread.so.0(+0x126b0)[0x7f00c40656b0]\n+## [i-capture-the-hostname:51015] [ 1] /lib/x86_64-linux-gnu/libc.so.6(+0x13ad67)[0x7f00c3fc5d67]\n+## [i-capture-the-hostname:51015] [ 2] /usr/lib/x86_64-linux-gnu/libopen-pal.so.40(opal_generic_simple_unpack+0x4e6)[0x7f00c3d5f596]\n+## [i-capture-the-hostname:51015] [ 3] /usr/lib/x86_64-linux-gnu/openmpi/lib/openmpi3/mca_pml_ob1.so(mca_pml_ob1_recv_request_progress_frag+0x1c2)[0x7f00c1493b62]\n+## [i-capture-the-hostname:51015] [ 4] /usr/lib/x86_64-linux-gnu/openmpi/lib/openmpi3/mca_btl_vader.so(mca_btl_vader_poll_handle_frag+0x8f)[0x7f00c14cf6bf]\n+## [i-capture-the-hostname:51015] [ 5] /usr/lib/x86_64-linux-gnu/openmpi/lib/openmpi3/mca_btl_vader.so(+0x4b8a)[0x7f00c14cfb8a]\n+## [i-capture-the-hostname:51015] [ 6] /usr/lib/x86_64-linux-gnu/libopen-pal.so.40(opal_progress+0x2c)[0x7f00c3d43e0c]\n+## [i-capture-the-hostname:51015] [ 7] /usr/lib/x86_64-linux-gnu/libopen-pal.so.40(ompi_sync_wait_mt+0x115)[0x7f00c3d4a835]\n+## [i-capture-the-hostname:51015] [ 8] /usr/lib/x86_64-linux-gnu/libmpi.so.40(ompi_request_default_wait+0x1d9)[0x7f00c40c59b9]\n+## [i-capture-the-hostname:51015] [ 9] /usr/lib/x86_64-linux-gnu/libmpi.so.40(ompi_coll_base_barrier_intra_two_procs+0xea)[0x7f00c4122c8a]\n+## [i-capture-the-hostname:51015] [10] /usr/lib/x86_64-linux-gnu/libmpi.so.40(MPI_Barrier+0xa8)[0x7f00c40de978]\n+## [i-capture-the-hostname:51015] [11] ./lockall_dt(+0x368c)[0x56121c1cd68c]\n+## [i-capture-the-hostname:51015] [12] /lib/x86_64-linux-gnu/libc.so.6(__libc_start_main+0xeb)[0x7f00c3eaf09b]\n+## [i-capture-the-hostname:51015] [13] ./lockall_dt(+0x392a)[0x56121c1cd92a]\n+## [i-capture-the-hostname:51015] *** End of error message ***\n+## --------------------------------------------------------------------------\n+## Primary job terminated normally, but 1 process returned\n+## a non-zero exit code. Per user-direction, the job has been aborted.\n+## --------------------------------------------------------------------------\n+## --------------------------------------------------------------------------\n+## mpiexec.openmpi noticed that process rank 1 with PID 0 on node i-capture-the-hostname exited on signal 11 (Segmentation fault).\n+## --------------------------------------------------------------------------\n+not ok 385 - ./rma/lockall_dt_flush 2\n+ ---\n+ Directory: ./rma\n+ File: lockall_dt_flush\n+ Num-procs: 2\n+ Date: \"Mon Dec 3 16:13:01 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## malloc(): corrupted top size\n-## [profitbricks-build5-amd64:59257] *** Process received signal ***\n-## [profitbricks-build5-amd64:59257] Signal: Aborted (6)\n-## [profitbricks-build5-amd64:59257] Signal code: (-6)\n+## [i-capture-the-hostname:51104] *** Process received signal ***\n+## [i-capture-the-hostname:51104] Signal: Aborted (6)\n+## [i-capture-the-hostname:51104] Signal code: (-6)\n ## --------------------------------------------------------------------------\n ## The user-provided time limit for job execution has been reached:\n ## \n ## Timeout: 240 seconds\n ## \n ## The job will now be aborted. Please check your code and/or\n ## adjust/remove the job execution time limit (as specified by --timeout\n ## command line option or MPIEXEC_TIMEOUT environment variable).\n ## --------------------------------------------------------------------------\n-not ok 385 - ./rma/lockall_dt_flush 2\n+not ok 386 - ./rma/lockall_dt_flushall 2\n ---\n Directory: ./rma\n- File: lockall_dt_flush\n+ File: lockall_dt_flushall\n Num-procs: 2\n- Date: \"Sun Jan 5 00:44:52 2020\"\n+ Date: \"Mon Dec 3 16:13:03 2018\"\n ...\n ## Test output (expected 'No Errors'):\n-## [profitbricks-build5-amd64:41357] *** Process received signal ***\n-## [profitbricks-build5-amd64:41357] Signal: Segmentation fault (11)\n-## [profitbricks-build5-amd64:41357] Signal code: Address not mapped (1)\n-## [profitbricks-build5-amd64:41357] Failing at address: 0x5572fd3763e0\n-## [profitbricks-build5-amd64:41357] [ 0] /lib/x86_64-linux-gnu/libpthread.so.0(+0x126b0)[0x7f966dc826b0]\n-## [profitbricks-build5-amd64:41357] [ 1] /lib/x86_64-linux-gnu/libc.so.6(+0x13ad67)[0x7f966dbe2d67]\n-## [profitbricks-build5-amd64:41357] [ 2] /usr/lib/x86_64-linux-gnu/libopen-pal.so.40(opal_generic_simple_unpack+0x4e6)[0x7f966d97c596]\n-## [profitbricks-build5-amd64:41357] [ 3] /usr/lib/x86_64-linux-gnu/openmpi/lib/openmpi3/mca_pml_ob1.so(mca_pml_ob1_recv_request_progress_frag+0x1c2)[0x7f966b0b0b62]\n-## [profitbricks-build5-amd64:41357] [ 4] /usr/lib/x86_64-linux-gnu/openmpi/lib/openmpi3/mca_btl_vader.so(mca_btl_vader_poll_handle_frag+0x8f)[0x7f966b0ec6bf]\n-## [profitbricks-build5-amd64:41357] [ 5] /usr/lib/x86_64-linux-gnu/openmpi/lib/openmpi3/mca_btl_vader.so(+0x4b8a)[0x7f966b0ecb8a]\n-## [profitbricks-build5-amd64:41357] [ 6] /usr/lib/x86_64-linux-gnu/libopen-pal.so.40(opal_progress+0x2c)[0x7f966d960e0c]\n-## [profitbricks-build5-amd64:41357] [ 7] /usr/lib/x86_64-linux-gnu/libopen-pal.so.40(ompi_sync_wait_mt+0x115)[0x7f966d967835]\n-## [profitbricks-build5-amd64:41357] [ 8] /usr/lib/x86_64-linux-gnu/libmpi.so.40(ompi_request_default_wait+0x1d9)[0x7f966dce29b9]\n-## [profitbricks-build5-amd64:41357] [ 9] /usr/lib/x86_64-linux-gnu/libmpi.so.40(ompi_coll_base_barrier_intra_two_procs+0xea)[0x7f966dd3fc8a]\n-## [profitbricks-build5-amd64:41357] [10] /usr/lib/x86_64-linux-gnu/libmpi.so.40(MPI_Barrier+0xa8)[0x7f966dcfb978]\n-## [profitbricks-build5-amd64:41357] [11] ./lockall_dt_flush(+0x36a6)[0x5572fbda66a6]\n-## [profitbricks-build5-amd64:41357] [12] /lib/x86_64-linux-gnu/libc.so.6(__libc_start_main+0xeb)[0x7f966dacc09b]\n-## [profitbricks-build5-amd64:41357] [13] ./lockall_dt_flush(+0x396a)[0x5572fbda696a]\n-## [profitbricks-build5-amd64:41357] *** End of error message ***\n+## [i-capture-the-hostname:63672] *** Process received signal ***\n+## [i-capture-the-hostname:63672] Signal: Segmentation fault (11)\n+## [i-capture-the-hostname:63672] Signal code: Address not mapped (1)\n+## [i-capture-the-hostname:63672] Failing at address: 0x558ffefe7260\n+## [i-capture-the-hostname:63672] [ 0] /lib/x86_64-linux-gnu/libpthread.so.0(+0x126b0)[0x7f704fd9d6b0]\n+## [i-capture-the-hostname:63672] [ 1] /lib/x86_64-linux-gnu/libc.so.6(+0x13ad67)[0x7f704fcfdd67]\n+## [i-capture-the-hostname:63672] [ 2] /usr/lib/x86_64-linux-gnu/libopen-pal.so.40(opal_generic_simple_unpack+0x4e6)[0x7f704fa97596]\n+## [i-capture-the-hostname:63672] [ 3] /usr/lib/x86_64-linux-gnu/openmpi/lib/openmpi3/mca_pml_ob1.so(mca_pml_ob1_recv_request_progress_frag+0x1c2)[0x7f704d1cbb62]\n+## [i-capture-the-hostname:63672] [ 4] /usr/lib/x86_64-linux-gnu/openmpi/lib/openmpi3/mca_btl_vader.so(mca_btl_vader_poll_handle_frag+0x8f)[0x7f704d2076bf]\n+## [i-capture-the-hostname:63672] [ 5] /usr/lib/x86_64-linux-gnu/openmpi/lib/openmpi3/mca_btl_vader.so(+0x4b8a)[0x7f704d207b8a]\n+## [i-capture-the-hostname:63672] [ 6] /usr/lib/x86_64-linux-gnu/libopen-pal.so.40(opal_progress+0x2c)[0x7f704fa7be0c]\n+## [i-capture-the-hostname:63672] [ 7] /usr/lib/x86_64-linux-gnu/libopen-pal.so.40(ompi_sync_wait_mt+0x115)[0x7f704fa82835]\n+## [i-capture-the-hostname:63672] [ 8] /usr/lib/x86_64-linux-gnu/libmpi.so.40(ompi_request_default_wait+0x1d9)[0x7f704fdfd9b9]\n+## [i-capture-the-hostname:63672] [ 9] /usr/lib/x86_64-linux-gnu/libmpi.so.40(ompi_coll_base_barrier_intra_two_procs+0xea)[0x7f704fe5ac8a]\n+## [i-capture-the-hostname:63672] [10] /usr/lib/x86_64-linux-gnu/libmpi.so.40(MPI_Barrier+0xa8)[0x7f704fe16978]\n+## [i-capture-the-hostname:63672] [11] ./lockall_dt_flushall(+0x36a6)[0x558ffe64e6a6]\n+## [i-capture-the-hostname:63672] [12] /lib/x86_64-linux-gnu/libc.so.6(__libc_start_main+0xeb)[0x7f704fbe709b]\n+## [i-capture-the-hostname:63672] [13] ./lockall_dt_flushall(+0x396a)[0x558ffe64e96a]\n+## [i-capture-the-hostname:63672] *** End of error message ***\n ## --------------------------------------------------------------------------\n ## Primary job terminated normally, but 1 process returned\n ## a non-zero exit code. Per user-direction, the job has been aborted.\n ## --------------------------------------------------------------------------\n ## --------------------------------------------------------------------------\n-## mpiexec.openmpi noticed that process rank 1 with PID 0 on node profitbricks-build5-amd64 exited on signal 11 (Segmentation fault).\n+## mpiexec.openmpi noticed that process rank 1 with PID 0 on node i-capture-the-hostname exited on signal 11 (Segmentation fault).\n ## --------------------------------------------------------------------------\n-not ok 386 - ./rma/lockall_dt_flushall 2\n+not ok 387 - ./rma/lockall_dt_flushlocal 2\n ---\n Directory: ./rma\n- File: lockall_dt_flushall\n+ File: lockall_dt_flushlocal\n Num-procs: 2\n- Date: \"Sun Jan 5 00:48:53 2020\"\n+ Date: \"Mon Dec 3 16:17:04 2018\"\n ...\n ## Test output (expected 'No Errors'):\n+## [i-capture-the-hostname:63776] *** Process received signal ***\n+## [i-capture-the-hostname:63776] Signal: Segmentation fault (11)\n+## [i-capture-the-hostname:63776] Signal code: Address not mapped (1)\n+## [i-capture-the-hostname:63776] Failing at address: 0x55b1515b30e0\n ## malloc(): corrupted top size\n-## [profitbricks-build5-amd64:41800] *** Process received signal ***\n-## [profitbricks-build5-amd64:41800] Signal: Aborted (6)\n-## [profitbricks-build5-amd64:41800] Signal code: (-6)\n+## [i-capture-the-hostname:63776] *** Process received signal ***\n+## [i-capture-the-hostname:63776] Signal: Aborted (6)\n+## [i-capture-the-hostname:63776] Signal code: (-6)\n ## --------------------------------------------------------------------------\n ## The user-provided time limit for job execution has been reached:\n ## \n ## Timeout: 240 seconds\n ## \n ## The job will now be aborted. Please check your code and/or\n ## adjust/remove the job execution time limit (as specified by --timeout\n ## command line option or MPIEXEC_TIMEOUT environment variable).\n ## --------------------------------------------------------------------------\n-not ok 387 - ./rma/lockall_dt_flushlocal 2\n+not ok 388 - ./rma/lockall_dt_flushlocalall 2\n ---\n Directory: ./rma\n- File: lockall_dt_flushlocal\n+ File: lockall_dt_flushlocalall\n Num-procs: 2\n- Date: \"Sun Jan 5 00:52:53 2020\"\n+ Date: \"Mon Dec 3 16:21:04 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## malloc(): corrupted top size\n-## [profitbricks-build5-amd64:54159] *** Process received signal ***\n-## [profitbricks-build5-amd64:54159] Signal: Aborted (6)\n-## [profitbricks-build5-amd64:54159] Signal code: (-6)\n+## [i-capture-the-hostname:14144] *** Process received signal ***\n+## [i-capture-the-hostname:14144] Signal: Aborted (6)\n+## [i-capture-the-hostname:14144] Signal code: (-6)\n ## --------------------------------------------------------------------------\n ## The user-provided time limit for job execution has been reached:\n ## \n ## Timeout: 240 seconds\n ## \n ## The job will now be aborted. Please check your code and/or\n ## adjust/remove the job execution time limit (as specified by --timeout\n ## command line option or MPIEXEC_TIMEOUT environment variable).\n ## --------------------------------------------------------------------------\n-not ok 388 - ./rma/lockall_dt_flushlocalall 2\n+not ok 389 - ./rma/lock_contention_dt 2\n ---\n Directory: ./rma\n- File: lockall_dt_flushlocalall\n+ File: lock_contention_dt\n Num-procs: 2\n- Date: \"Sun Jan 5 00:52:55 2020\"\n+ Date: \"Mon Dec 3 16:21:06 2018\"\n ...\n ## Test output (expected 'No Errors'):\n-## [profitbricks-build5-amd64:09247] *** Process received signal ***\n-## [profitbricks-build5-amd64:09247] Signal: Segmentation fault (11)\n-## [profitbricks-build5-amd64:09247] Signal code: Address not mapped (1)\n-## [profitbricks-build5-amd64:09247] Failing at address: 0x564fe1409190\n-## [profitbricks-build5-amd64:09247] [ 0] /lib/x86_64-linux-gnu/libpthread.so.0(+0x126b0)[0x7f1500d4a6b0]\n-## [profitbricks-build5-amd64:09247] [ 1] /lib/x86_64-linux-gnu/libc.so.6(+0x13acc1)[0x7f1500caacc1]\n-## [profitbricks-build5-amd64:09247] [ 2] /usr/lib/x86_64-linux-gnu/libopen-pal.so.40(opal_generic_simple_unpack+0x4e6)[0x7f1500a44596]\n-## [profitbricks-build5-amd64:09247] [ 3] /usr/lib/x86_64-linux-gnu/openmpi/lib/openmpi3/mca_pml_ob1.so(mca_pml_ob1_recv_request_progress_frag+0x1c2)[0x7f14fe178b62]\n-## [profitbricks-build5-amd64:09247] [ 4] /usr/lib/x86_64-linux-gnu/openmpi/lib/openmpi3/mca_btl_vader.so(mca_btl_vader_poll_handle_frag+0x8f)[0x7f14fe1b46bf]\n-## [profitbricks-build5-amd64:09247] [ 5] /usr/lib/x86_64-linux-gnu/openmpi/lib/openmpi3/mca_btl_vader.so(+0x4b8a)[0x7f14fe1b4b8a]\n-## [profitbricks-build5-amd64:09247] [ 6] /usr/lib/x86_64-linux-gnu/libopen-pal.so.40(opal_progress+0x2c)[0x7f1500a28e0c]\n-## [profitbricks-build5-amd64:09247] [ 7] /usr/lib/x86_64-linux-gnu/libopen-pal.so.40(ompi_sync_wait_mt+0x115)[0x7f1500a2f835]\n-## [profitbricks-build5-amd64:09247] [ 8] /usr/lib/x86_64-linux-gnu/libmpi.so.40(ompi_request_default_wait+0x1d9)[0x7f1500daa9b9]\n-## [profitbricks-build5-amd64:09247] [ 9] /usr/lib/x86_64-linux-gnu/libmpi.so.40(ompi_coll_base_barrier_intra_two_procs+0xea)[0x7f1500e07c8a]\n-## [profitbricks-build5-amd64:09247] [10] /usr/lib/x86_64-linux-gnu/libmpi.so.40(MPI_Barrier+0xa8)[0x7f1500dc3978]\n-## [profitbricks-build5-amd64:09247] [11] ./lockall_dt_flushlocalall(+0x36ac)[0x564fdf2226ac]\n-## [profitbricks-build5-amd64:09247] [12] /lib/x86_64-linux-gnu/libc.so.6(__libc_start_main+0xeb)[0x7f1500b9409b]\n-## [profitbricks-build5-amd64:09247] [13] ./lockall_dt_flushlocalall(+0x39ca)[0x564fdf2229ca]\n-## [profitbricks-build5-amd64:09247] *** End of error message ***\n+## [i-capture-the-hostname:40682] *** Process received signal ***\n+## [i-capture-the-hostname:40682] Signal: Segmentation fault (11)\n+## [i-capture-the-hostname:40682] Signal code: Address not mapped (1)\n+## [i-capture-the-hostname:40682] Failing at address: 0x555b11725290\n+## [i-capture-the-hostname:40682] [ 0] /lib/x86_64-linux-gnu/libpthread.so.0(+0x126b0)[0x7fca9e9356b0]\n+## [i-capture-the-hostname:40682] [ 1] /lib/x86_64-linux-gnu/libc.so.6(+0x13ad67)[0x7fca9e895d67]\n+## [i-capture-the-hostname:40682] [ 2] /usr/lib/x86_64-linux-gnu/libopen-pal.so.40(opal_generic_simple_unpack+0x4e6)[0x7fca9e62f596]\n+## [i-capture-the-hostname:40682] [ 3] /usr/lib/x86_64-linux-gnu/openmpi/lib/openmpi3/mca_pml_ob1.so(mca_pml_ob1_recv_request_progress_frag+0x1c2)[0x7fca9c163b62]\n+## [i-capture-the-hostname:40682] [ 4] /usr/lib/x86_64-linux-gnu/openmpi/lib/openmpi3/mca_btl_vader.so(mca_btl_vader_poll_handle_frag+0x8f)[0x7fca9c19f6bf]\n+## [i-capture-the-hostname:40682] [ 5] /usr/lib/x86_64-linux-gnu/openmpi/lib/openmpi3/mca_btl_vader.so(+0x4b8a)[0x7fca9c19fb8a]\n+## [i-capture-the-hostname:40682] [ 6] /usr/lib/x86_64-linux-gnu/libopen-pal.so.40(opal_progress+0x2c)[0x7fca9e613e0c]\n+## [i-capture-the-hostname:40682] [ 7] /usr/lib/x86_64-linux-gnu/libopen-pal.so.40(ompi_sync_wait_mt+0x115)[0x7fca9e61a835]\n+## [i-capture-the-hostname:40682] [ 8] /usr/lib/x86_64-linux-gnu/libmpi.so.40(ompi_request_default_wait+0x1d9)[0x7fca9e9959b9]\n+## [i-capture-the-hostname:40682] [ 9] /usr/lib/x86_64-linux-gnu/libmpi.so.40(ompi_coll_base_barrier_intra_two_procs+0xea)[0x7fca9e9f2c8a]\n+## [i-capture-the-hostname:40682] [10] /usr/lib/x86_64-linux-gnu/libmpi.so.40(MPI_Barrier+0xa8)[0x7fca9e9ae978]\n+## [i-capture-the-hostname:40682] [11] ./lock_contention_dt(+0x2824)[0x555b1057c824]\n+## [i-capture-the-hostname:40682] [12] /lib/x86_64-linux-gnu/libc.so.6(__libc_start_main+0xeb)[0x7fca9e77f09b]\n+## [i-capture-the-hostname:40682] [13] ./lock_contention_dt(+0x28fa)[0x555b1057c8fa]\n+## [i-capture-the-hostname:40682] *** End of error message ***\n ## --------------------------------------------------------------------------\n ## Primary job terminated normally, but 1 process returned\n ## a non-zero exit code. Per user-direction, the job has been aborted.\n ## --------------------------------------------------------------------------\n ## --------------------------------------------------------------------------\n-## mpiexec.openmpi noticed that process rank 1 with PID 0 on node profitbricks-build5-amd64 exited on signal 11 (Segmentation fault).\n-## --------------------------------------------------------------------------\n-not ok 389 - ./rma/lock_contention_dt 2\n- ---\n- Directory: ./rma\n- File: lock_contention_dt\n- Num-procs: 2\n- Date: \"Sun Jan 5 00:56:56 2020\"\n- ...\n-## Test output (expected 'No Errors'):\n-## Data expected = ff but got p[1,0] = df\n-## Data expected = fe but got p[1,1] = de\n-## Data expected = fd but got p[1,2] = dd\n-## Data expected = fc but got p[1,3] = dc\n-## Data expected = fb but got p[1,4] = db\n-## Data expected = fa but got p[1,5] = da\n-## Data expected = f9 but got p[1,6] = d9\n-## Data expected = f8 but got p[1,7] = d8\n-## Data expected = f7 but got p[1,8] = d7\n-## free(): invalid next size (normal)\n-## [profitbricks-build5-amd64:10369] *** Process received signal ***\n-## [profitbricks-build5-amd64:10369] Signal: Aborted (6)\n-## [profitbricks-build5-amd64:10369] Signal code: (-6)\n-## --------------------------------------------------------------------------\n-## The user-provided time limit for job execution has been reached:\n-## \n-## Timeout: 240 seconds\n-## \n-## The job will now be aborted. Please check your code and/or\n-## adjust/remove the job execution time limit (as specified by --timeout\n-## command line option or MPIEXEC_TIMEOUT environment variable).\n+## mpiexec.openmpi noticed that process rank 1 with PID 0 on node i-capture-the-hostname exited on signal 11 (Segmentation fault).\n ## --------------------------------------------------------------------------\n ok 390 - ./rma/transpose4 2\n ok 391 - ./rma/fetchandadd 2\n ok 392 - ./rma/fetchandadd_tree 2\n ok 393 - ./rma/wintest 2\n ok 394 - ./rma/wintest_shm 2\n ok 395 - ./rma/contig_displ 1\n@@ -2183,15 +2204,15 @@\n ok 415 - ./rma/epochtest 2\n ok 416 - ./rma/locknull 2\n not ok 417 - ./rma/rmanull 2\n ---\n Directory: ./rma\n File: rmanull\n Num-procs: 2\n- Date: \"Sun Jan 5 00:57:58 2020\"\n+ Date: \"Mon Dec 3 16:22:08 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Lock beforePut: Error class 6 (MPI_ERR_RANK: invalid rank)\n ## Unlock after Put: Error class 6 (MPI_ERR_RANK: invalid rank)\n ## Lock beforeGet: Error class 6 (MPI_ERR_RANK: invalid rank)\n ## Unlock after Get: Error class 6 (MPI_ERR_RANK: invalid rank)\n ## Lock beforeAccumulate: Error class 6 (MPI_ERR_RANK: invalid rank)\n@@ -2260,21 +2281,21 @@\n ok 466 - ./rma/rput_local_comp 2\n ok 467 - ./rma/racc_local_comp 2\n not ok 468 - ./rma/win_info 2\n ---\n Directory: ./rma\n File: win_info\n Num-procs: 2\n- Date: \"Sun Jan 5 00:58:47 2020\"\n+ Date: \"Mon Dec 3 16:22:57 2018\"\n ...\n ## Test output (expected 'No Errors'):\n-## 0: no_locks is not defined\n-## 0: no_locks is not defined\n ## 1: no_locks is not defined\n ## 1: no_locks is not defined\n+## 0: no_locks is not defined\n+## 0: no_locks is not defined\n ok 469 - ./rma/linked_list_lockall 2\n ok 470 - ./rma/pscw_ordering 2\n ok 471 - ./rma/pscw_ordering_shm 2\n ok 472 - ./rma/linked_list_bench_lock_all 2\n ok 473 - ./rma/linked_list_bench_lock_excl 2\n ok 474 - ./rma/linked_list_bench_lock_shr 2\n ok 475 - ./rma/linked_list_bench_lock_shr_nocheck 2\n@@ -2283,15 +2304,15 @@\n ok 478 - ./rma/mutex_bench_shm 2\n ok 479 - ./rma/mutex_bench_shm_ordered 2\n not ok 480 - ./rma/rma-contig 2\n ---\n Directory: ./rma\n File: rma-contig\n Num-procs: 2\n- Date: \"Sun Jan 5 01:10:58 2020\"\n+ Date: \"Mon Dec 3 16:35:08 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## --------------------------------------------------------------------------\n ## The user-provided time limit for job execution has been reached:\n ## \n ## Timeout: 720 seconds\n ## \n@@ -2300,66 +2321,66 @@\n ## command line option or MPIEXEC_TIMEOUT environment variable).\n ## --------------------------------------------------------------------------\n not ok 481 - ./rma/badrma 2\n ---\n Directory: ./rma\n File: badrma\n Num-procs: 2\n- Date: \"Sun Jan 5 01:10:59 2020\"\n+ Date: \"Mon Dec 3 16:35:10 2018\"\n ...\n ## Test output (expected 'No Errors'):\n-## [profitbricks-build5-amd64:23555] *** An error occurred in MPI_Accumulate\n-## [profitbricks-build5-amd64:23555] *** reported by process [2679439361,0]\n-## [profitbricks-build5-amd64:23555] *** on win pt2pt window 3\n-## [profitbricks-build5-amd64:23555] *** MPI_ERR_ARG: invalid argument of some other kind\n-## [profitbricks-build5-amd64:23555] *** MPI_ERRORS_ARE_FATAL (processes in this win will now abort,\n-## [profitbricks-build5-amd64:23555] *** and potentially your MPI job)\n+## [i-capture-the-hostname:60307] *** An error occurred in MPI_Accumulate\n+## [i-capture-the-hostname:60307] *** reported by process [1713504257,0]\n+## [i-capture-the-hostname:60307] *** on win pt2pt window 3\n+## [i-capture-the-hostname:60307] *** MPI_ERR_ARG: invalid argument of some other kind\n+## [i-capture-the-hostname:60307] *** MPI_ERRORS_ARE_FATAL (processes in this win will now abort,\n+## [i-capture-the-hostname:60307] *** and potentially your MPI job)\n ok 482 - ./rma/acc-loc 2\n ok 483 - ./rma/fence_shm 2\n ok 484 - ./rma/win_shared_zerobyte 2\n not ok 485 - ./rma/win_shared_put_flush_get 2\n ---\n Directory: ./rma\n File: win_shared_put_flush_get\n Num-procs: 2\n- Date: \"Sun Jan 5 01:11:02 2020\"\n+ Date: \"Mon Dec 3 16:35:17 2018\"\n ...\n ## Test output (expected 'No Errors'):\n+## Error: must be run with four processes\n ## --------------------------------------------------------------------------\n-## MPI_ABORT was invoked on rank 1 in communicator MPI_COMM_WORLD\n+## MPI_ABORT was invoked on rank 0 in communicator MPI_COMM_WORLD\n ## with errorcode 1.\n ## \n ## NOTE: invoking MPI_ABORT causes Open MPI to kill all MPI processes.\n ## You may or may not see output from other processes, depending on\n ## exactly when Open MPI kills them.\n ## --------------------------------------------------------------------------\n-## Error: must be run with four processes\n-## [profitbricks-build5-amd64:24810] 1 more process has sent help message help-mpi-api.txt / mpi-abort\n-## [profitbricks-build5-amd64:24810] Set MCA parameter \"orte_base_help_aggregate\" to 0 to see all help / error messages\n+## [i-capture-the-hostname:60531] 1 more process has sent help message help-mpi-api.txt / mpi-abort\n+## [i-capture-the-hostname:60531] Set MCA parameter \"orte_base_help_aggregate\" to 0 to see all help / error messages\n ok 486 - ./rma/get-struct 2\n ok 487 - ./rma/at_complete 2\n ok 488 - ./rma/atomic_rmw_fop 2\n not ok 489 - ./rma/atomic_rmw_cas 2\n ---\n Directory: ./rma\n File: atomic_rmw_cas\n Num-procs: 2\n- Date: \"Sun Jan 5 01:11:06 2020\"\n+ Date: \"Mon Dec 3 16:35:23 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ok 490 - ./rma/atomic_rmw_gacc 2\n ok 491 - ./rma/atomic_get 2\n ok 493 - ./rma/aint 2 # SKIP requires MPI version 3.1\n ok 492 - ./rma/acc-pairtype 2\n not ok 493 - ./rma/manyget 2\n ---\n Directory: ./rma\n File: manyget\n Num-procs: 2\n- Date: \"Sun Jan 5 01:14:09 2020\"\n+ Date: \"Mon Dec 3 16:38:28 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## --------------------------------------------------------------------------\n ## The user-provided time limit for job execution has been reached:\n ## \n ## Timeout: 180 seconds\n ## \n@@ -2368,15 +2389,15 @@\n ## command line option or MPIEXEC_TIMEOUT environment variable).\n ## --------------------------------------------------------------------------\n not ok 494 - ./rma/derived-acc-flush_local 2\n ---\n Directory: ./rma\n File: derived-acc-flush_local\n Num-procs: 2\n- Date: \"Sun Jan 5 01:14:25 2020\"\n+ Date: \"Mon Dec 3 16:38:36 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## tar_buf[0] = 0, expected 500\n ## tar_buf[1] = 0, expected 500\n ## tar_buf[2] = 0, expected 500\n ## tar_buf[3] = 0, expected 500\n ## tar_buf[4] = 0, expected 500\n@@ -997378,855 +997399,807 @@\n ok 495 - ./rma/large-acc-flush_local 2\n ok 496 - ./rma/large-small-acc 2\n not ok 497 - ./rma/win_shared_put_flush_load 2\n ---\n Directory: ./rma\n File: win_shared_put_flush_load\n Num-procs: 2\n- Date: \"Sun Jan 5 01:14:57 2020\"\n+ Date: \"Mon Dec 3 16:39:16 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Error: must be run with three processes\n ## --------------------------------------------------------------------------\n ## MPI_ABORT was invoked on rank 1 in communicator MPI_COMM_WORLD\n ## with errorcode 1.\n ## \n ## NOTE: invoking MPI_ABORT causes Open MPI to kill all MPI processes.\n ## You may or may not see output from other processes, depending on\n ## exactly when Open MPI kills them.\n ## --------------------------------------------------------------------------\n-## [profitbricks-build5-amd64:29563] 1 more process has sent help message help-mpi-api.txt / mpi-abort\n-## [profitbricks-build5-amd64:29563] Set MCA parameter \"orte_base_help_aggregate\" to 0 to see all help / error messages\n+## [i-capture-the-hostname:45288] 1 more process has sent help message help-mpi-api.txt / mpi-abort\n+## [i-capture-the-hostname:45288] Set MCA parameter \"orte_base_help_aggregate\" to 0 to see all help / error messages\n not ok 498 - ./rma/win_shared_acc_flush_load 2\n ---\n Directory: ./rma\n File: win_shared_acc_flush_load\n Num-procs: 2\n- Date: \"Sun Jan 5 01:14:58 2020\"\n+ Date: \"Mon Dec 3 16:39:18 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Error: must be run with three processes\n ## --------------------------------------------------------------------------\n-## MPI_ABORT was invoked on rank 0 in communicator MPI_COMM_WORLD\n+## MPI_ABORT was invoked on rank 1 in communicator MPI_COMM_WORLD\n ## with errorcode 1.\n ## \n ## NOTE: invoking MPI_ABORT causes Open MPI to kill all MPI processes.\n ## You may or may not see output from other processes, depending on\n ## exactly when Open MPI kills them.\n ## --------------------------------------------------------------------------\n-## [profitbricks-build5-amd64:29815] 1 more process has sent help message help-mpi-api.txt / mpi-abort\n-## [profitbricks-build5-amd64:29815] Set MCA parameter \"orte_base_help_aggregate\" to 0 to see all help / error messages\n+## [i-capture-the-hostname:45419] 1 more process has sent help message help-mpi-api.txt / mpi-abort\n+## [i-capture-the-hostname:45419] Set MCA parameter \"orte_base_help_aggregate\" to 0 to see all help / error messages\n not ok 499 - ./rma/win_shared_gacc_flush_load 2\n ---\n Directory: ./rma\n File: win_shared_gacc_flush_load\n Num-procs: 2\n- Date: \"Sun Jan 5 01:14:59 2020\"\n+ Date: \"Mon Dec 3 16:39:20 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Error: must be run with three processes\n ## --------------------------------------------------------------------------\n ## MPI_ABORT was invoked on rank 1 in communicator MPI_COMM_WORLD\n ## with errorcode 1.\n ## \n ## NOTE: invoking MPI_ABORT causes Open MPI to kill all MPI processes.\n ## You may or may not see output from other processes, depending on\n ## exactly when Open MPI kills them.\n ## --------------------------------------------------------------------------\n-## [profitbricks-build5-amd64:29966] 1 more process has sent help message help-mpi-api.txt / mpi-abort\n-## [profitbricks-build5-amd64:29966] Set MCA parameter \"orte_base_help_aggregate\" to 0 to see all help / error messages\n+## [i-capture-the-hostname:45518] 1 more process has sent help message help-mpi-api.txt / mpi-abort\n+## [i-capture-the-hostname:45518] Set MCA parameter \"orte_base_help_aggregate\" to 0 to see all help / error messages\n not ok 500 - ./rma/win_shared_fop_flush_load 2\n ---\n Directory: ./rma\n File: win_shared_fop_flush_load\n Num-procs: 2\n- Date: \"Sun Jan 5 01:15:00 2020\"\n+ Date: \"Mon Dec 3 16:39:22 2018\"\n ...\n ## Test output (expected 'No Errors'):\n+## Error: must be run with three processes\n ## --------------------------------------------------------------------------\n ## MPI_ABORT was invoked on rank 1 in communicator MPI_COMM_WORLD\n ## with errorcode 1.\n ## \n ## NOTE: invoking MPI_ABORT causes Open MPI to kill all MPI processes.\n ## You may or may not see output from other processes, depending on\n ## exactly when Open MPI kills them.\n ## --------------------------------------------------------------------------\n-## Error: must be run with three processes\n-## [profitbricks-build5-amd64:30327] 1 more process has sent help message help-mpi-api.txt / mpi-abort\n-## [profitbricks-build5-amd64:30327] Set MCA parameter \"orte_base_help_aggregate\" to 0 to see all help / error messages\n+## [i-capture-the-hostname:45615] 1 more process has sent help message help-mpi-api.txt / mpi-abort\n+## [i-capture-the-hostname:45615] Set MCA parameter \"orte_base_help_aggregate\" to 0 to see all help / error messages\n not ok 501 - ./rma/win_shared_cas_flush_load 2\n ---\n Directory: ./rma\n File: win_shared_cas_flush_load\n Num-procs: 2\n- Date: \"Sun Jan 5 01:15:01 2020\"\n+ Date: \"Mon Dec 3 16:39:24 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Error: must be run with three processes\n ## --------------------------------------------------------------------------\n ## MPI_ABORT was invoked on rank 1 in communicator MPI_COMM_WORLD\n ## with errorcode 1.\n ## \n ## NOTE: invoking MPI_ABORT causes Open MPI to kill all MPI processes.\n ## You may or may not see output from other processes, depending on\n ## exactly when Open MPI kills them.\n ## --------------------------------------------------------------------------\n-## [profitbricks-build5-amd64:30683] 1 more process has sent help message help-mpi-api.txt / mpi-abort\n-## [profitbricks-build5-amd64:30683] Set MCA parameter \"orte_base_help_aggregate\" to 0 to see all help / error messages\n+## [i-capture-the-hostname:45755] 1 more process has sent help message help-mpi-api.txt / mpi-abort\n+## [i-capture-the-hostname:45755] Set MCA parameter \"orte_base_help_aggregate\" to 0 to see all help / error messages\n not ok 502 - ./rma/put_flush_get 2\n ---\n Directory: ./rma\n File: put_flush_get\n Num-procs: 2\n- Date: \"Sun Jan 5 01:15:03 2020\"\n+ Date: \"Mon Dec 3 16:39:27 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Error: must be run with three processes\n ## --------------------------------------------------------------------------\n-## MPI_ABORT was invoked on rank 0 in communicator MPI_COMM_WORLD\n+## MPI_ABORT was invoked on rank 1 in communicator MPI_COMM_WORLD\n ## with errorcode 1.\n ## \n ## NOTE: invoking MPI_ABORT causes Open MPI to kill all MPI processes.\n ## You may or may not see output from other processes, depending on\n ## exactly when Open MPI kills them.\n ## --------------------------------------------------------------------------\n-## [profitbricks-build5-amd64:30917] 1 more process has sent help message help-mpi-api.txt / mpi-abort\n-## [profitbricks-build5-amd64:30917] Set MCA parameter \"orte_base_help_aggregate\" to 0 to see all help / error messages\n+## [i-capture-the-hostname:45895] 1 more process has sent help message help-mpi-api.txt / mpi-abort\n+## [i-capture-the-hostname:45895] Set MCA parameter \"orte_base_help_aggregate\" to 0 to see all help / error messages\n not ok 503 - ./rma/acc_flush_get 2\n ---\n Directory: ./rma\n File: acc_flush_get\n Num-procs: 2\n- Date: \"Sun Jan 5 01:15:04 2020\"\n+ Date: \"Mon Dec 3 16:39:28 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Error: must be run with three processes\n ## --------------------------------------------------------------------------\n ## MPI_ABORT was invoked on rank 1 in communicator MPI_COMM_WORLD\n ## with errorcode 1.\n ## \n ## NOTE: invoking MPI_ABORT causes Open MPI to kill all MPI processes.\n ## You may or may not see output from other processes, depending on\n ## exactly when Open MPI kills them.\n ## --------------------------------------------------------------------------\n-## [profitbricks-build5-amd64:31120] 1 more process has sent help message help-mpi-api.txt / mpi-abort\n-## [profitbricks-build5-amd64:31120] Set MCA parameter \"orte_base_help_aggregate\" to 0 to see all help / error messages\n+## [i-capture-the-hostname:45926] 1 more process has sent help message help-mpi-api.txt / mpi-abort\n+## [i-capture-the-hostname:45926] Set MCA parameter \"orte_base_help_aggregate\" to 0 to see all help / error messages\n not ok 504 - ./rma/gacc_flush_get 2\n ---\n Directory: ./rma\n File: gacc_flush_get\n Num-procs: 2\n- Date: \"Sun Jan 5 01:15:05 2020\"\n+ Date: \"Mon Dec 3 16:39:30 2018\"\n ...\n ## Test output (expected 'No Errors'):\n+## Error: must be run with three processes\n ## --------------------------------------------------------------------------\n-## MPI_ABORT was invoked on rank 0 in communicator MPI_COMM_WORLD\n+## MPI_ABORT was invoked on rank 1 in communicator MPI_COMM_WORLD\n ## with errorcode 1.\n ## \n ## NOTE: invoking MPI_ABORT causes Open MPI to kill all MPI processes.\n ## You may or may not see output from other processes, depending on\n ## exactly when Open MPI kills them.\n ## --------------------------------------------------------------------------\n-## Error: must be run with three processes\n-## [profitbricks-build5-amd64:31342] 1 more process has sent help message help-mpi-api.txt / mpi-abort\n-## [profitbricks-build5-amd64:31342] Set MCA parameter \"orte_base_help_aggregate\" to 0 to see all help / error messages\n+## [i-capture-the-hostname:45957] 1 more process has sent help message help-mpi-api.txt / mpi-abort\n+## [i-capture-the-hostname:45957] Set MCA parameter \"orte_base_help_aggregate\" to 0 to see all help / error messages\n not ok 505 - ./rma/fop_flush_get 2\n ---\n Directory: ./rma\n File: fop_flush_get\n Num-procs: 2\n- Date: \"Sun Jan 5 01:15:06 2020\"\n+ Date: \"Mon Dec 3 16:39:32 2018\"\n ...\n ## Test output (expected 'No Errors'):\n+## Error: must be run with three processes\n ## --------------------------------------------------------------------------\n ## MPI_ABORT was invoked on rank 1 in communicator MPI_COMM_WORLD\n ## with errorcode 1.\n ## \n ## NOTE: invoking MPI_ABORT causes Open MPI to kill all MPI processes.\n ## You may or may not see output from other processes, depending on\n ## exactly when Open MPI kills them.\n ## --------------------------------------------------------------------------\n-## Error: must be run with three processes\n-## [profitbricks-build5-amd64:31706] 1 more process has sent help message help-mpi-api.txt / mpi-abort\n-## [profitbricks-build5-amd64:31706] Set MCA parameter \"orte_base_help_aggregate\" to 0 to see all help / error messages\n+## [i-capture-the-hostname:45988] 1 more process has sent help message help-mpi-api.txt / mpi-abort\n+## [i-capture-the-hostname:45988] Set MCA parameter \"orte_base_help_aggregate\" to 0 to see all help / error messages\n not ok 506 - ./rma/cas_flush_get 2\n ---\n Directory: ./rma\n File: cas_flush_get\n Num-procs: 2\n- Date: \"Sun Jan 5 01:15:08 2020\"\n+ Date: \"Mon Dec 3 16:39:35 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Error: must be run with three processes\n ## --------------------------------------------------------------------------\n-## MPI_ABORT was invoked on rank 0 in communicator MPI_COMM_WORLD\n+## MPI_ABORT was invoked on rank 1 in communicator MPI_COMM_WORLD\n ## with errorcode 1.\n ## \n ## NOTE: invoking MPI_ABORT causes Open MPI to kill all MPI processes.\n ## You may or may not see output from other processes, depending on\n ## exactly when Open MPI kills them.\n ## --------------------------------------------------------------------------\n-## [profitbricks-build5-amd64:31908] 1 more process has sent help message help-mpi-api.txt / mpi-abort\n-## [profitbricks-build5-amd64:31908] Set MCA parameter \"orte_base_help_aggregate\" to 0 to see all help / error messages\n+## [i-capture-the-hostname:46013] 1 more process has sent help message help-mpi-api.txt / mpi-abort\n+## [i-capture-the-hostname:46013] Set MCA parameter \"orte_base_help_aggregate\" to 0 to see all help / error messages\n ok 507 - ./spawn/namepub 2\n not ok 508 - ./spawn/spawn1 1\n ---\n Directory: ./spawn\n File: spawn1\n Num-procs: 1\n- Date: \"Sun Jan 5 01:15:11 2020\"\n+ Date: \"Mon Dec 3 16:39:40 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## --------------------------------------------------------------------------\n ## At least one pair of MPI processes are unable to reach each other for\n ## MPI communications. This means that no Open MPI device has indicated\n ## that it can be used to communicate between these processes. This is\n ## an error; Open MPI requires that all MPI processes be able to reach\n ## each other. This error can sometimes be the result of forgetting to\n ## specify the \"self\" BTL.\n ## \n-## Process 1 ([[47742,2],1]) is on host: profitbricks-build5-amd64\n-## Process 2 ([[47742,1],0]) is on host: profitbricks-build5-amd64\n+## Process 1 ([[14664,2],0]) is on host: i-capture-the-hostname\n+## Process 2 ([[14664,1],0]) is on host: i-capture-the-hostname\n ## BTLs attempted: vader self\n ## \n ## Your MPI job is now going to abort; sorry.\n ## --------------------------------------------------------------------------\n+## [i-capture-the-hostname:46135] [[14664,2],0] ORTE_ERROR_LOG: Unreachable in file dpm/dpm.c at line 433\n+## [i-capture-the-hostname:46136] [[14664,2],1] ORTE_ERROR_LOG: Unreachable in file dpm/dpm.c at line 433\n+## [i-capture-the-hostname:46122] [[14664,1],0] ORTE_ERROR_LOG: Unreachable in file dpm/dpm.c at line 433\n+## [i-capture-the-hostname:46122] *** An error occurred in MPI_Comm_spawn\n+## [i-capture-the-hostname:46122] *** reported by process [961019905,0]\n+## [i-capture-the-hostname:46122] *** on communicator MPI_COMM_WORLD\n+## [i-capture-the-hostname:46122] *** MPI_ERR_INTERN: internal error\n+## [i-capture-the-hostname:46122] *** MPI_ERRORS_ARE_FATAL (processes in this communicator will now abort,\n+## [i-capture-the-hostname:46122] *** and potentially your MPI job)\n ## --------------------------------------------------------------------------\n ## It looks like MPI_INIT failed for some reason; your parallel process is\n ## likely to abort. There are many reasons that a parallel process can\n ## fail during MPI_INIT; some of which are due to configuration or environment\n ## problems. This failure appears to be an internal failure; here's some\n ## additional information (which may only be relevant to an Open MPI\n ## developer):\n ## \n ## ompi_dpm_dyn_init() failed\n ## --> Returned \"Unreachable\" (-12) instead of \"Success\" (0)\n ## --------------------------------------------------------------------------\n-## [profitbricks-build5-amd64:32597] [[47742,2],1] ORTE_ERROR_LOG: Unreachable in file dpm/dpm.c at line 433\n-## [profitbricks-build5-amd64:32597] *** An error occurred in MPI_Init_thread\n-## [profitbricks-build5-amd64:32597] *** reported by process [3128819714,1]\n-## [profitbricks-build5-amd64:32597] *** on a NULL communicator\n-## [profitbricks-build5-amd64:32597] *** Unknown error\n-## [profitbricks-build5-amd64:32597] *** MPI_ERRORS_ARE_FATAL (processes in this communicator will now abort,\n-## [profitbricks-build5-amd64:32597] *** and potentially your MPI job)\n-## [profitbricks-build5-amd64:32594] [[47742,2],0] ORTE_ERROR_LOG: Unreachable in file dpm/dpm.c at line 433\n-## [profitbricks-build5-amd64:32477] [[47742,1],0] ORTE_ERROR_LOG: Unreachable in file dpm/dpm.c at line 433\n-## [profitbricks-build5-amd64:32477] *** An error occurred in MPI_Comm_spawn\n-## [profitbricks-build5-amd64:32477] *** reported by process [3128819713,0]\n-## [profitbricks-build5-amd64:32477] *** on communicator MPI_COMM_WORLD\n-## [profitbricks-build5-amd64:32477] *** MPI_ERR_INTERN: internal error\n-## [profitbricks-build5-amd64:32477] *** MPI_ERRORS_ARE_FATAL (processes in this communicator will now abort,\n-## [profitbricks-build5-amd64:32477] *** and potentially your MPI job)\n-## [profitbricks-build5-amd64:32428] 2 more processes have sent help message help-mca-bml-r2.txt / unreachable proc\n-## [profitbricks-build5-amd64:32428] Set MCA parameter \"orte_base_help_aggregate\" to 0 to see all help / error messages\n-## [profitbricks-build5-amd64:32428] 1 more process has sent help message help-mpi-runtime.txt / mpi_init:startup:internal-failure\n-## [profitbricks-build5-amd64:32428] 1 more process has sent help message help-mpi-errors.txt / mpi_errors_are_fatal unknown handle\n+## [i-capture-the-hostname:46135] *** An error occurred in MPI_Init_thread\n+## [i-capture-the-hostname:46135] *** reported by process [961019906,0]\n+## [i-capture-the-hostname:46135] *** on a NULL communicator\n+## [i-capture-the-hostname:46135] *** Unknown error\n+## [i-capture-the-hostname:46135] *** MPI_ERRORS_ARE_FATAL (processes in this communicator will now abort,\n+## [i-capture-the-hostname:46135] *** and potentially your MPI job)\n+## [i-capture-the-hostname:46105] 2 more processes have sent help message help-mca-bml-r2.txt / unreachable proc\n+## [i-capture-the-hostname:46105] Set MCA parameter \"orte_base_help_aggregate\" to 0 to see all help / error messages\n+## [i-capture-the-hostname:46105] 1 more process has sent help message help-mpi-runtime.txt / mpi_init:startup:internal-failure\n+## [i-capture-the-hostname:46105] 1 more process has sent help message help-mpi-errors.txt / mpi_errors_are_fatal unknown handle\n not ok 509 - ./spawn/spawn2 1\n ---\n Directory: ./spawn\n File: spawn2\n Num-procs: 1\n- Date: \"Sun Jan 5 01:15:13 2020\"\n+ Date: \"Mon Dec 3 16:39:42 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## --------------------------------------------------------------------------\n ## At least one pair of MPI processes are unable to reach each other for\n ## MPI communications. This means that no Open MPI device has indicated\n ## that it can be used to communicate between these processes. This is\n ## an error; Open MPI requires that all MPI processes be able to reach\n ## each other. This error can sometimes be the result of forgetting to\n ## specify the \"self\" BTL.\n ## \n-## Process 1 ([[17614,1],0]) is on host: profitbricks-build5-amd64\n-## Process 2 ([[17614,2],0]) is on host: profitbricks-build5-amd64\n-## BTLs attempted: self\n+## Process 1 ([[14622,2],0]) is on host: i-capture-the-hostname\n+## Process 2 ([[14622,1],0]) is on host: i-capture-the-hostname\n+## BTLs attempted: vader self\n ## \n ## Your MPI job is now going to abort; sorry.\n ## --------------------------------------------------------------------------\n-## [profitbricks-build5-amd64:32912] [[17614,1],0] ORTE_ERROR_LOG: Unreachable in file dpm/dpm.c at line 433\n-## [profitbricks-build5-amd64:32912] *** An error occurred in MPI_Comm_spawn\n-## [profitbricks-build5-amd64:32912] *** reported by process [1154351105,0]\n-## [profitbricks-build5-amd64:32912] *** on communicator MPI_COMM_WORLD\n-## [profitbricks-build5-amd64:32912] *** MPI_ERR_INTERN: internal error\n-## [profitbricks-build5-amd64:32912] *** MPI_ERRORS_ARE_FATAL (processes in this communicator will now abort,\n-## [profitbricks-build5-amd64:32912] *** and potentially your MPI job)\n+## [i-capture-the-hostname:46183] [[14622,2],0] ORTE_ERROR_LOG: Unreachable in file dpm/dpm.c at line 433\n ## --------------------------------------------------------------------------\n ## It looks like MPI_INIT failed for some reason; your parallel process is\n ## likely to abort. There are many reasons that a parallel process can\n ## fail during MPI_INIT; some of which are due to configuration or environment\n ## problems. This failure appears to be an internal failure; here's some\n ## additional information (which may only be relevant to an Open MPI\n ## developer):\n ## \n ## ompi_dpm_dyn_init() failed\n ## --> Returned \"Unreachable\" (-12) instead of \"Success\" (0)\n ## --------------------------------------------------------------------------\n-## [profitbricks-build5-amd64:33075] [[17614,2],1] ORTE_ERROR_LOG: Unreachable in file dpm/dpm.c at line 433\n-## [profitbricks-build5-amd64:33069] [[17614,2],0] ORTE_ERROR_LOG: Unreachable in file dpm/dpm.c at line 433\n-## [profitbricks-build5-amd64:32796] 2 more processes have sent help message help-mca-bml-r2.txt / unreachable proc\n-## [profitbricks-build5-amd64:32796] Set MCA parameter \"orte_base_help_aggregate\" to 0 to see all help / error messages\n-## [profitbricks-build5-amd64:32796] 1 more process has sent help message help-mpi-runtime.txt / mpi_init:startup:internal-failure\n+## [i-capture-the-hostname:46183] *** An error occurred in MPI_Init_thread\n+## [i-capture-the-hostname:46183] *** reported by process [958267394,0]\n+## [i-capture-the-hostname:46183] *** on a NULL communicator\n+## [i-capture-the-hostname:46183] *** Unknown error\n+## [i-capture-the-hostname:46183] *** MPI_ERRORS_ARE_FATAL (processes in this communicator will now abort,\n+## [i-capture-the-hostname:46183] *** and potentially your MPI job)\n+## [i-capture-the-hostname:46164] [[14622,1],0] ORTE_ERROR_LOG: Unreachable in file dpm/dpm.c at line 433\n+## [i-capture-the-hostname:46159] 2 more processes have sent help message help-mca-bml-r2.txt / unreachable proc\n+## [i-capture-the-hostname:46159] Set MCA parameter \"orte_base_help_aggregate\" to 0 to see all help / error messages\n not ok 510 - ./spawn/spawninfo1 1\n ---\n Directory: ./spawn\n File: spawninfo1\n Num-procs: 1\n- Date: \"Sun Jan 5 01:15:15 2020\"\n+ Date: \"Mon Dec 3 16:39:44 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## --------------------------------------------------------------------------\n ## mpiexec.openmpi was unable to find the specified executable file, and therefore\n ## did not launch the job. This error was first reported for process\n ## rank 0; it may have occurred for other processes as well.\n ## \n ## NOTE: A common cause for this error is misspelling a mpiexec.openmpi command\n ## line parameter option (remember that mpiexec.openmpi interprets the first\n ## unrecognized command line token as the executable).\n ## \n-## Node: profitbricks-build5-amd64\n+## Node: i-capture-the-hostname\n ## Executable: spawninfo1\n ## --------------------------------------------------------------------------\n ## 2 total processes failed to start\n not ok 511 - ./spawn/spawnminfo1 1\n ---\n Directory: ./spawn\n File: spawnminfo1\n Num-procs: 1\n- Date: \"Sun Jan 5 01:15:17 2020\"\n+ Date: \"Mon Dec 3 16:39:46 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## --------------------------------------------------------------------------\n ## mpiexec.openmpi was unable to find the specified executable file, and therefore\n ## did not launch the job. This error was first reported for process\n ## rank 0; it may have occurred for other processes as well.\n ## \n ## NOTE: A common cause for this error is misspelling a mpiexec.openmpi command\n ## line parameter option (remember that mpiexec.openmpi interprets the first\n ## unrecognized command line token as the executable).\n ## \n-## Node: profitbricks-build5-amd64\n+## Node: i-capture-the-hostname\n ## Executable: spawnminfo1\n ## --------------------------------------------------------------------------\n not ok 512 - ./spawn/spawnintra 1\n ---\n Directory: ./spawn\n File: spawnintra\n Num-procs: 1\n- Date: \"Sun Jan 5 01:15:19 2020\"\n+ Date: \"Mon Dec 3 16:39:47 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## --------------------------------------------------------------------------\n ## At least one pair of MPI processes are unable to reach each other for\n ## MPI communications. This means that no Open MPI device has indicated\n ## that it can be used to communicate between these processes. This is\n ## an error; Open MPI requires that all MPI processes be able to reach\n ## each other. This error can sometimes be the result of forgetting to\n ## specify the \"self\" BTL.\n ## \n-## Process 1 ([[17123,2],1]) is on host: profitbricks-build5-amd64\n-## Process 2 ([[17123,1],0]) is on host: profitbricks-build5-amd64\n-## BTLs attempted: vader self\n+## Process 1 ([[14779,1],0]) is on host: i-capture-the-hostname\n+## Process 2 ([[14779,2],0]) is on host: i-capture-the-hostname\n+## BTLs attempted: self\n ## \n ## Your MPI job is now going to abort; sorry.\n ## --------------------------------------------------------------------------\n-## [profitbricks-build5-amd64:34436] *** An error occurred in MPI_Comm_spawn\n-## [profitbricks-build5-amd64:34436] *** reported by process [1122172929,0]\n-## [profitbricks-build5-amd64:34436] *** on communicator MPI_COMM_WORLD\n-## [profitbricks-build5-amd64:34436] *** MPI_ERR_INTERN: internal error\n-## [profitbricks-build5-amd64:34436] *** MPI_ERRORS_ARE_FATAL (processes in this communicator will now abort,\n-## [profitbricks-build5-amd64:34436] *** and potentially your MPI job)\n+## [i-capture-the-hostname:46331] [[14779,1],0] ORTE_ERROR_LOG: Unreachable in file dpm/dpm.c at line 433\n+## [i-capture-the-hostname:46331] *** An error occurred in MPI_Comm_spawn\n+## [i-capture-the-hostname:46331] *** reported by process [968556545,0]\n+## [i-capture-the-hostname:46331] *** on communicator MPI_COMM_WORLD\n+## [i-capture-the-hostname:46331] *** MPI_ERR_INTERN: internal error\n+## [i-capture-the-hostname:46331] *** MPI_ERRORS_ARE_FATAL (processes in this communicator will now abort,\n+## [i-capture-the-hostname:46331] *** and potentially your MPI job)\n+## [i-capture-the-hostname:46344] [[14779,2],1] ORTE_ERROR_LOG: Unreachable in file dpm/dpm.c at line 433\n+## [i-capture-the-hostname:46342] [[14779,2],0] ORTE_ERROR_LOG: Unreachable in file dpm/dpm.c at line 433\n ## --------------------------------------------------------------------------\n ## It looks like MPI_INIT failed for some reason; your parallel process is\n ## likely to abort. There are many reasons that a parallel process can\n ## fail during MPI_INIT; some of which are due to configuration or environment\n ## problems. This failure appears to be an internal failure; here's some\n ## additional information (which may only be relevant to an Open MPI\n ## developer):\n ## \n ## ompi_dpm_dyn_init() failed\n ## --> Returned \"Unreachable\" (-12) instead of \"Success\" (0)\n ## --------------------------------------------------------------------------\n-## [profitbricks-build5-amd64:34436] [[17123,1],0] ORTE_ERROR_LOG: Unreachable in file dpm/dpm.c at line 433\n-## [profitbricks-build5-amd64:34586] [[17123,2],0] ORTE_ERROR_LOG: Unreachable in file dpm/dpm.c at line 433\n-## [profitbricks-build5-amd64:34588] [[17123,2],1] ORTE_ERROR_LOG: Unreachable in file dpm/dpm.c at line 433\n-## [profitbricks-build5-amd64:34586] *** An error occurred in MPI_Init_thread\n-## [profitbricks-build5-amd64:34586] *** reported by process [1122172930,0]\n-## [profitbricks-build5-amd64:34586] *** on a NULL communicator\n-## [profitbricks-build5-amd64:34586] *** Unknown error\n-## [profitbricks-build5-amd64:34586] *** MPI_ERRORS_ARE_FATAL (processes in this communicator will now abort,\n-## [profitbricks-build5-amd64:34586] *** and potentially your MPI job)\n-## [profitbricks-build5-amd64:34353] 2 more processes have sent help message help-mca-bml-r2.txt / unreachable proc\n-## [profitbricks-build5-amd64:34353] Set MCA parameter \"orte_base_help_aggregate\" to 0 to see all help / error messages\n-## [profitbricks-build5-amd64:34353] 1 more process has sent help message help-mpi-runtime.txt / mpi_init:startup:internal-failure\n-## [profitbricks-build5-amd64:34353] 1 more process has sent help message help-mpi-errors.txt / mpi_errors_are_fatal unknown handle\n+## [i-capture-the-hostname:46344] *** An error occurred in MPI_Init_thread\n+## [i-capture-the-hostname:46344] *** reported by process [968556546,1]\n+## [i-capture-the-hostname:46344] *** on a NULL communicator\n+## [i-capture-the-hostname:46344] *** Unknown error\n+## [i-capture-the-hostname:46344] *** MPI_ERRORS_ARE_FATAL (processes in this communicator will now abort,\n+## [i-capture-the-hostname:46344] *** and potentially your MPI job)\n+## [i-capture-the-hostname:46314] 2 more processes have sent help message help-mca-bml-r2.txt / unreachable proc\n+## [i-capture-the-hostname:46314] Set MCA parameter \"orte_base_help_aggregate\" to 0 to see all help / error messages\n+## [i-capture-the-hostname:46314] 1 more process has sent help message help-mpi-runtime.txt / mpi_init:startup:internal-failure\n+## [i-capture-the-hostname:46314] 1 more process has sent help message help-mpi-errors.txt / mpi_errors_are_fatal unknown handle\n not ok 513 - ./spawn/spawnintra 2\n ---\n Directory: ./spawn\n File: spawnintra\n Num-procs: 2\n- Date: \"Sun Jan 5 01:15:21 2020\"\n+ Date: \"Mon Dec 3 16:39:49 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## --------------------------------------------------------------------------\n ## At least one pair of MPI processes are unable to reach each other for\n ## MPI communications. This means that no Open MPI device has indicated\n ## that it can be used to communicate between these processes. This is\n ## an error; Open MPI requires that all MPI processes be able to reach\n ## each other. This error can sometimes be the result of forgetting to\n ## specify the \"self\" BTL.\n ## \n-## Process 1 ([[17183,1],1]) is on host: profitbricks-build5-amd64\n-## Process 2 ([[17183,2],0]) is on host: profitbricks-build5-amd64\n+## Process 1 ([[14412,1],1]) is on host: i-capture-the-hostname\n+## Process 2 ([[14412,2],0]) is on host: i-capture-the-hostname\n ## BTLs attempted: vader self\n ## \n ## Your MPI job is now going to abort; sorry.\n ## --------------------------------------------------------------------------\n-## [profitbricks-build5-amd64:34850] *** An error occurred in MPI_Comm_spawn\n-## [profitbricks-build5-amd64:34850] *** reported by process [1126105089,1]\n-## [profitbricks-build5-amd64:34850] *** on communicator MPI_COMM_WORLD\n-## [profitbricks-build5-amd64:34850] *** MPI_ERR_INTERN: internal error\n-## [profitbricks-build5-amd64:34850] *** MPI_ERRORS_ARE_FATAL (processes in this communicator will now abort,\n-## [profitbricks-build5-amd64:34850] *** and potentially your MPI job)\n+## [i-capture-the-hostname:46382] [[14412,1],1] ORTE_ERROR_LOG: Unreachable in file dpm/dpm.c at line 433\n+## [i-capture-the-hostname:46399] [[14412,2],1] ORTE_ERROR_LOG: Unreachable in file dpm/dpm.c at line 433\n+## [i-capture-the-hostname:46381] [[14412,1],0] ORTE_ERROR_LOG: Unreachable in file dpm/dpm.c at line 433\n+## [i-capture-the-hostname:46381] *** An error occurred in MPI_Comm_spawn\n+## [i-capture-the-hostname:46381] *** reported by process [944504833,0]\n+## [i-capture-the-hostname:46381] *** on communicator MPI_COMM_WORLD\n+## [i-capture-the-hostname:46381] *** MPI_ERR_INTERN: internal error\n+## [i-capture-the-hostname:46381] *** MPI_ERRORS_ARE_FATAL (processes in this communicator will now abort,\n+## [i-capture-the-hostname:46381] *** and potentially your MPI job)\n ## --------------------------------------------------------------------------\n ## It looks like MPI_INIT failed for some reason; your parallel process is\n ## likely to abort. There are many reasons that a parallel process can\n ## fail during MPI_INIT; some of which are due to configuration or environment\n ## problems. This failure appears to be an internal failure; here's some\n ## additional information (which may only be relevant to an Open MPI\n ## developer):\n ## \n ## ompi_dpm_dyn_init() failed\n ## --> Returned \"Unreachable\" (-12) instead of \"Success\" (0)\n ## --------------------------------------------------------------------------\n-## [profitbricks-build5-amd64:34974] [[17183,2],1] ORTE_ERROR_LOG: Unreachable in file dpm/dpm.c at line 433\n-## [profitbricks-build5-amd64:34850] [[17183,1],1] ORTE_ERROR_LOG: Unreachable in file dpm/dpm.c at line 433\n-## [profitbricks-build5-amd64:34849] [[17183,1],0] ORTE_ERROR_LOG: Unreachable in file dpm/dpm.c at line 433\n-## [profitbricks-build5-amd64:34966] [[17183,2],0] ORTE_ERROR_LOG: Unreachable in file dpm/dpm.c at line 433\n-## [profitbricks-build5-amd64:34974] *** An error occurred in MPI_Init_thread\n-## [profitbricks-build5-amd64:34974] *** reported by process [1126105090,1]\n-## [profitbricks-build5-amd64:34974] *** on a NULL communicator\n-## [profitbricks-build5-amd64:34974] *** Unknown error\n-## [profitbricks-build5-amd64:34974] *** MPI_ERRORS_ARE_FATAL (processes in this communicator will now abort,\n-## [profitbricks-build5-amd64:34974] *** and potentially your MPI job)\n-## [profitbricks-build5-amd64:34765] 3 more processes have sent help message help-mca-bml-r2.txt / unreachable proc\n-## [profitbricks-build5-amd64:34765] Set MCA parameter \"orte_base_help_aggregate\" to 0 to see all help / error messages\n-## [profitbricks-build5-amd64:34765] 1 more process has sent help message help-mpi-errors.txt / mpi_errors_are_fatal\n-## [profitbricks-build5-amd64:34765] 1 more process has sent help message help-mpi-runtime.txt / mpi_init:startup:internal-failure\n-## [profitbricks-build5-amd64:34765] 1 more process has sent help message help-mpi-errors.txt / mpi_errors_are_fatal unknown handle\n+## [i-capture-the-hostname:46398] [[14412,2],0] ORTE_ERROR_LOG: Unreachable in file dpm/dpm.c at line 433\n+## [i-capture-the-hostname:46399] *** An error occurred in MPI_Init_thread\n+## [i-capture-the-hostname:46399] *** reported by process [944504834,1]\n+## [i-capture-the-hostname:46399] *** on a NULL communicator\n+## [i-capture-the-hostname:46399] *** Unknown error\n+## [i-capture-the-hostname:46399] *** MPI_ERRORS_ARE_FATAL (processes in this communicator will now abort,\n+## [i-capture-the-hostname:46399] *** and potentially your MPI job)\n+## [i-capture-the-hostname:46365] 3 more processes have sent help message help-mca-bml-r2.txt / unreachable proc\n+## [i-capture-the-hostname:46365] Set MCA parameter \"orte_base_help_aggregate\" to 0 to see all help / error messages\n+## [i-capture-the-hostname:46365] 1 more process has sent help message help-mpi-errors.txt / mpi_errors_are_fatal\n+## [i-capture-the-hostname:46365] 1 more process has sent help message help-mpi-runtime.txt / mpi_init:startup:internal-failure\n+## [i-capture-the-hostname:46365] 1 more process has sent help message help-mpi-errors.txt / mpi_errors_are_fatal unknown handle\n not ok 514 - ./spawn/spawnargv 1\n ---\n Directory: ./spawn\n File: spawnargv\n Num-procs: 1\n- Date: \"Sun Jan 5 01:15:23 2020\"\n+ Date: \"Mon Dec 3 16:39:50 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## --------------------------------------------------------------------------\n ## At least one pair of MPI processes are unable to reach each other for\n ## MPI communications. This means that no Open MPI device has indicated\n ## that it can be used to communicate between these processes. This is\n ## an error; Open MPI requires that all MPI processes be able to reach\n ## each other. This error can sometimes be the result of forgetting to\n ## specify the \"self\" BTL.\n ## \n-## Process 1 ([[19493,2],0]) is on host: profitbricks-build5-amd64\n-## Process 2 ([[19493,1],0]) is on host: profitbricks-build5-amd64\n-## BTLs attempted: vader self\n+## Process 1 ([[14338,1],0]) is on host: i-capture-the-hostname\n+## Process 2 ([[14338,2],0]) is on host: i-capture-the-hostname\n+## BTLs attempted: self\n ## \n ## Your MPI job is now going to abort; sorry.\n ## --------------------------------------------------------------------------\n+## [i-capture-the-hostname:46424] [[14338,1],0] ORTE_ERROR_LOG: Unreachable in file dpm/dpm.c at line 433\n+## [i-capture-the-hostname:46424] *** An error occurred in MPI_Comm_spawn\n+## [i-capture-the-hostname:46424] *** reported by process [939655169,0]\n+## [i-capture-the-hostname:46424] *** on communicator MPI_COMM_WORLD\n+## [i-capture-the-hostname:46424] *** MPI_ERR_INTERN: internal error\n+## [i-capture-the-hostname:46424] *** MPI_ERRORS_ARE_FATAL (processes in this communicator will now abort,\n+## [i-capture-the-hostname:46424] *** and potentially your MPI job)\n+## [i-capture-the-hostname:46444] [[14338,2],0] ORTE_ERROR_LOG: Unreachable in file dpm/dpm.c at line 433\n+## [i-capture-the-hostname:46445] [[14338,2],1] ORTE_ERROR_LOG: Unreachable in file dpm/dpm.c at line 433\n ## --------------------------------------------------------------------------\n ## It looks like MPI_INIT failed for some reason; your parallel process is\n ## likely to abort. There are many reasons that a parallel process can\n ## fail during MPI_INIT; some of which are due to configuration or environment\n ## problems. This failure appears to be an internal failure; here's some\n ## additional information (which may only be relevant to an Open MPI\n ## developer):\n ## \n ## ompi_dpm_dyn_init() failed\n ## --> Returned \"Unreachable\" (-12) instead of \"Success\" (0)\n ## --------------------------------------------------------------------------\n-## [profitbricks-build5-amd64:35244] [[19493,2],0] ORTE_ERROR_LOG: Unreachable in file dpm/dpm.c at line 433\n-## [profitbricks-build5-amd64:35244] *** An error occurred in MPI_Init_thread\n-## [profitbricks-build5-amd64:35244] *** reported by process [1277493250,0]\n-## [profitbricks-build5-amd64:35244] *** on a NULL communicator\n-## [profitbricks-build5-amd64:35244] *** Unknown error\n-## [profitbricks-build5-amd64:35244] *** MPI_ERRORS_ARE_FATAL (processes in this communicator will now abort,\n-## [profitbricks-build5-amd64:35244] *** and potentially your MPI job)\n-## [profitbricks-build5-amd64:35095] [[19493,1],0] ORTE_ERROR_LOG: Unreachable in file dpm/dpm.c at line 433\n-## [profitbricks-build5-amd64:35095] *** An error occurred in MPI_Comm_spawn\n-## [profitbricks-build5-amd64:35095] *** reported by process [1277493249,0]\n-## [profitbricks-build5-amd64:35095] *** on communicator MPI_COMM_WORLD\n-## [profitbricks-build5-amd64:35095] *** MPI_ERR_INTERN: internal error\n-## [profitbricks-build5-amd64:35095] *** MPI_ERRORS_ARE_FATAL (processes in this communicator will now abort,\n-## [profitbricks-build5-amd64:35095] *** and potentially your MPI job)\n-## [profitbricks-build5-amd64:35246] [[19493,2],1] ORTE_ERROR_LOG: Unreachable in file dpm/dpm.c at line 433\n-## [profitbricks-build5-amd64:35063] 2 more processes have sent help message help-mca-bml-r2.txt / unreachable proc\n-## [profitbricks-build5-amd64:35063] Set MCA parameter \"orte_base_help_aggregate\" to 0 to see all help / error messages\n-## [profitbricks-build5-amd64:35063] 1 more process has sent help message help-mpi-runtime.txt / mpi_init:startup:internal-failure\n+## [i-capture-the-hostname:46444] *** An error occurred in MPI_Init_thread\n+## [i-capture-the-hostname:46444] *** reported by process [939655170,0]\n+## [i-capture-the-hostname:46444] *** on a NULL communicator\n+## [i-capture-the-hostname:46444] *** Unknown error\n+## [i-capture-the-hostname:46444] *** MPI_ERRORS_ARE_FATAL (processes in this communicator will now abort,\n+## [i-capture-the-hostname:46444] *** and potentially your MPI job)\n+## [i-capture-the-hostname:46419] 2 more processes have sent help message help-mca-bml-r2.txt / unreachable proc\n+## [i-capture-the-hostname:46419] Set MCA parameter \"orte_base_help_aggregate\" to 0 to see all help / error messages\n+## [i-capture-the-hostname:46419] 1 more process has sent help message help-mpi-runtime.txt / mpi_init:startup:internal-failure\n+## [i-capture-the-hostname:46419] 1 more process has sent help message help-mpi-errors.txt / mpi_errors_are_fatal unknown handle\n not ok 515 - ./spawn/spawnmanyarg 1\n ---\n Directory: ./spawn\n File: spawnmanyarg\n Num-procs: 1\n- Date: \"Sun Jan 5 01:15:24 2020\"\n+ Date: \"Mon Dec 3 16:39:52 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## --------------------------------------------------------------------------\n ## At least one pair of MPI processes are unable to reach each other for\n ## MPI communications. This means that no Open MPI device has indicated\n ## that it can be used to communicate between these processes. This is\n ## an error; Open MPI requires that all MPI processes be able to reach\n ## each other. This error can sometimes be the result of forgetting to\n ## specify the \"self\" BTL.\n ## \n-## Process 1 ([[20005,2],1]) is on host: profitbricks-build5-amd64\n-## Process 2 ([[20005,1],0]) is on host: profitbricks-build5-amd64\n+## Process 1 ([[14530,2],0]) is on host: i-capture-the-hostname\n+## Process 2 ([[14530,1],0]) is on host: i-capture-the-hostname\n ## BTLs attempted: vader self\n ## \n ## Your MPI job is now going to abort; sorry.\n ## --------------------------------------------------------------------------\n+## [i-capture-the-hostname:46571] [[14530,2],0] ORTE_ERROR_LOG: Unreachable in file dpm/dpm.c at line 433\n+## [i-capture-the-hostname:46575] [[14530,2],1] ORTE_ERROR_LOG: Unreachable in file dpm/dpm.c at line 433\n ## --------------------------------------------------------------------------\n ## It looks like MPI_INIT failed for some reason; your parallel process is\n ## likely to abort. There are many reasons that a parallel process can\n ## fail during MPI_INIT; some of which are due to configuration or environment\n ## problems. This failure appears to be an internal failure; here's some\n ## additional information (which may only be relevant to an Open MPI\n ## developer):\n ## \n ## ompi_dpm_dyn_init() failed\n ## --> Returned \"Unreachable\" (-12) instead of \"Success\" (0)\n ## --------------------------------------------------------------------------\n-## [profitbricks-build5-amd64:35871] [[20005,2],1] ORTE_ERROR_LOG: Unreachable in file dpm/dpm.c at line 433\n-## [profitbricks-build5-amd64:35866] [[20005,2],0] ORTE_ERROR_LOG: Unreachable in file dpm/dpm.c at line 433\n-## [profitbricks-build5-amd64:35871] *** An error occurred in MPI_Init_thread\n-## [profitbricks-build5-amd64:35871] *** reported by process [1311047682,1]\n-## [profitbricks-build5-amd64:35871] *** on a NULL communicator\n-## [profitbricks-build5-amd64:35871] *** Unknown error\n-## [profitbricks-build5-amd64:35871] *** MPI_ERRORS_ARE_FATAL (processes in this communicator will now abort,\n-## [profitbricks-build5-amd64:35871] *** and potentially your MPI job)\n-## [profitbricks-build5-amd64:35672] [[20005,1],0] ORTE_ERROR_LOG: Unreachable in file dpm/dpm.c at line 433\n-## [profitbricks-build5-amd64:35672] *** An error occurred in MPI_Comm_spawn\n-## [profitbricks-build5-amd64:35672] *** reported by process [1311047681,0]\n-## [profitbricks-build5-amd64:35672] *** on communicator MPI_COMM_WORLD\n-## [profitbricks-build5-amd64:35672] *** MPI_ERR_INTERN: internal error\n-## [profitbricks-build5-amd64:35672] *** MPI_ERRORS_ARE_FATAL (processes in this communicator will now abort,\n-## [profitbricks-build5-amd64:35672] *** and potentially your MPI job)\n-## [profitbricks-build5-amd64:35575] 2 more processes have sent help message help-mca-bml-r2.txt / unreachable proc\n-## [profitbricks-build5-amd64:35575] Set MCA parameter \"orte_base_help_aggregate\" to 0 to see all help / error messages\n-## [profitbricks-build5-amd64:35575] 1 more process has sent help message help-mpi-runtime.txt / mpi_init:startup:internal-failure\n-## [profitbricks-build5-amd64:35575] 1 more process has sent help message help-mpi-errors.txt / mpi_errors_are_fatal unknown handle\n+## [i-capture-the-hostname:46575] *** An error occurred in MPI_Init_thread\n+## [i-capture-the-hostname:46575] *** reported by process [952238082,1]\n+## [i-capture-the-hostname:46575] *** on a NULL communicator\n+## [i-capture-the-hostname:46575] *** Unknown error\n+## [i-capture-the-hostname:46575] *** MPI_ERRORS_ARE_FATAL (processes in this communicator will now abort,\n+## [i-capture-the-hostname:46575] *** and potentially your MPI job)\n+## [i-capture-the-hostname:46483] 2 more processes have sent help message help-mca-bml-r2.txt / unreachable proc\n+## [i-capture-the-hostname:46483] Set MCA parameter \"orte_base_help_aggregate\" to 0 to see all help / error messages\n+## [i-capture-the-hostname:46483] 1 more process has sent help message help-mpi-runtime.txt / mpi_init:startup:internal-failure\n+## [i-capture-the-hostname:46483] 1 more process has sent help message help-mpi-errors.txt / mpi_errors_are_fatal unknown handle\n not ok 516 - ./spawn/spawnmult2 2\n ---\n Directory: ./spawn\n File: spawnmult2\n Num-procs: 2\n- Date: \"Sun Jan 5 01:15:26 2020\"\n+ Date: \"Mon Dec 3 16:39:53 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## --------------------------------------------------------------------------\n ## At least one pair of MPI processes are unable to reach each other for\n ## MPI communications. This means that no Open MPI device has indicated\n ## that it can be used to communicate between these processes. This is\n ## an error; Open MPI requires that all MPI processes be able to reach\n ## each other. This error can sometimes be the result of forgetting to\n ## specify the \"self\" BTL.\n ## \n-## Process 1 ([[18925,1],0]) is on host: profitbricks-build5-amd64\n-## Process 2 ([[18925,2],0]) is on host: profitbricks-build5-amd64\n+## Process 1 ([[15263,1],0]) is on host: i-capture-the-hostname\n+## Process 2 ([[15263,2],0]) is on host: i-capture-the-hostname\n ## BTLs attempted: vader self\n ## \n ## Your MPI job is now going to abort; sorry.\n ## --------------------------------------------------------------------------\n-## [profitbricks-build5-amd64:36366] [[18925,2],1] ORTE_ERROR_LOG: Unreachable in file dpm/dpm.c at line 433\n-## [profitbricks-build5-amd64:36278] [[18925,1],0] ORTE_ERROR_LOG: Unreachable in file dpm/dpm.c at line 433\n-## --------------------------------------------------------------------------\n-## It looks like MPI_INIT failed for some reason; your parallel process is\n-## likely to abort. There are many reasons that a parallel process can\n-## fail during MPI_INIT; some of which are due to configuration or environment\n-## problems. This failure appears to be an internal failure; here's some\n-## additional information (which may only be relevant to an Open MPI\n-## developer):\n-## \n-## ompi_dpm_dyn_init() failed\n-## --> Returned \"Unreachable\" (-12) instead of \"Success\" (0)\n-## --------------------------------------------------------------------------\n-## [profitbricks-build5-amd64:36278] *** An error occurred in MPI_Comm_spawn_multiple\n-## [profitbricks-build5-amd64:36278] *** reported by process [1240268801,0]\n-## [profitbricks-build5-amd64:36278] *** on communicator MPI_COMM_WORLD\n-## [profitbricks-build5-amd64:36278] *** MPI_ERR_INTERN: internal error\n-## [profitbricks-build5-amd64:36278] *** MPI_ERRORS_ARE_FATAL (processes in this communicator will now abort,\n-## [profitbricks-build5-amd64:36278] *** and potentially your MPI job)\n-## [profitbricks-build5-amd64:36283] [[18925,1],1] ORTE_ERROR_LOG: Unreachable in file dpm/dpm.c at line 433\n-## [profitbricks-build5-amd64:36365] [[18925,2],0] ORTE_ERROR_LOG: Unreachable in file dpm/dpm.c at line 433\n-## [profitbricks-build5-amd64:36366] *** An error occurred in MPI_Init_thread\n-## [profitbricks-build5-amd64:36366] *** reported by process [1240268802,1]\n-## [profitbricks-build5-amd64:36366] *** on a NULL communicator\n-## [profitbricks-build5-amd64:36366] *** Unknown error\n-## [profitbricks-build5-amd64:36366] *** MPI_ERRORS_ARE_FATAL (processes in this communicator will now abort,\n-## [profitbricks-build5-amd64:36366] *** and potentially your MPI job)\n-## [profitbricks-build5-amd64:36159] 3 more processes have sent help message help-mca-bml-r2.txt / unreachable proc\n-## [profitbricks-build5-amd64:36159] Set MCA parameter \"orte_base_help_aggregate\" to 0 to see all help / error messages\n-## [profitbricks-build5-amd64:36159] 1 more process has sent help message help-mpi-runtime.txt / mpi_init:startup:internal-failure\n-## [profitbricks-build5-amd64:36159] 1 more process has sent help message help-mpi-errors.txt / mpi_errors_are_fatal\n-## [profitbricks-build5-amd64:36159] 1 more process has sent help message help-mpi-errors.txt / mpi_errors_are_fatal unknown handle\n+## [i-capture-the-hostname:46869] [[15263,1],0] ORTE_ERROR_LOG: Unreachable in file dpm/dpm.c at line 433\n+## [i-capture-the-hostname:46869] *** An error occurred in MPI_Comm_spawn_multiple\n+## [i-capture-the-hostname:46869] *** reported by process [1000275969,0]\n+## [i-capture-the-hostname:46869] *** on communicator MPI_COMM_WORLD\n+## [i-capture-the-hostname:46869] *** MPI_ERR_INTERN: internal error\n+## [i-capture-the-hostname:46869] *** MPI_ERRORS_ARE_FATAL (processes in this communicator will now abort,\n+## [i-capture-the-hostname:46869] *** and potentially your MPI job)\n+## [i-capture-the-hostname:46798] 3 more processes have sent help message help-mca-bml-r2.txt / unreachable proc\n+## [i-capture-the-hostname:46798] Set MCA parameter \"orte_base_help_aggregate\" to 0 to see all help / error messages\n not ok 517 - ./spawn/spaconacc 1\n ---\n Directory: ./spawn\n File: spaconacc\n Num-procs: 1\n- Date: \"Sun Jan 5 01:15:28 2020\"\n+ Date: \"Mon Dec 3 16:39:55 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## --------------------------------------------------------------------------\n ## At least one pair of MPI processes are unable to reach each other for\n ## MPI communications. This means that no Open MPI device has indicated\n ## that it can be used to communicate between these processes. This is\n ## an error; Open MPI requires that all MPI processes be able to reach\n ## each other. This error can sometimes be the result of forgetting to\n ## specify the \"self\" BTL.\n ## \n-## Process 1 ([[18951,1],0]) is on host: profitbricks-build5-amd64\n-## Process 2 ([[18951,2],0]) is on host: profitbricks-build5-amd64\n+## Process 1 ([[15061,1],0]) is on host: i-capture-the-hostname\n+## Process 2 ([[15061,2],0]) is on host: i-capture-the-hostname\n ## BTLs attempted: self\n ## \n ## Your MPI job is now going to abort; sorry.\n ## --------------------------------------------------------------------------\n+## [i-capture-the-hostname:47010] [[15061,1],0] ORTE_ERROR_LOG: Unreachable in file dpm/dpm.c at line 433\n ## MPI_Comm_spawn failed: MPI_ERR_INTERN: internal error\n-## [profitbricks-build5-amd64:36667] [[18951,1],0] ORTE_ERROR_LOG: Unreachable in file dpm/dpm.c at line 433\n+## [i-capture-the-hostname:47022] [[15061,2],0] ORTE_ERROR_LOG: Unreachable in file dpm/dpm.c at line 433\n ## --------------------------------------------------------------------------\n ## MPI_ABORT was invoked on rank 0 in communicator MPI_COMM_WORLD\n ## with errorcode 17.\n ## \n ## NOTE: invoking MPI_ABORT causes Open MPI to kill all MPI processes.\n ## You may or may not see output from other processes, depending on\n ## exactly when Open MPI kills them.\n ## --------------------------------------------------------------------------\n-## [profitbricks-build5-amd64:36718] [[18951,2],0] ORTE_ERROR_LOG: Unreachable in file dpm/dpm.c at line 433\n ## --------------------------------------------------------------------------\n ## It looks like MPI_INIT failed for some reason; your parallel process is\n ## likely to abort. There are many reasons that a parallel process can\n ## fail during MPI_INIT; some of which are due to configuration or environment\n ## problems. This failure appears to be an internal failure; here's some\n ## additional information (which may only be relevant to an Open MPI\n ## developer):\n ## \n ## ompi_dpm_dyn_init() failed\n ## --> Returned \"Unreachable\" (-12) instead of \"Success\" (0)\n ## --------------------------------------------------------------------------\n-## [profitbricks-build5-amd64:36718] *** An error occurred in MPI_Init\n-## [profitbricks-build5-amd64:36718] *** reported by process [1241972738,0]\n-## [profitbricks-build5-amd64:36718] *** on a NULL communicator\n-## [profitbricks-build5-amd64:36718] *** Unknown error\n-## [profitbricks-build5-amd64:36718] *** MPI_ERRORS_ARE_FATAL (processes in this communicator will now abort,\n-## [profitbricks-build5-amd64:36718] *** and potentially your MPI job)\n-## [profitbricks-build5-amd64:36565] 1 more process has sent help message help-mca-bml-r2.txt / unreachable proc\n-## [profitbricks-build5-amd64:36565] Set MCA parameter \"orte_base_help_aggregate\" to 0 to see all help / error messages\n+## [i-capture-the-hostname:47022] *** An error occurred in MPI_Init\n+## [i-capture-the-hostname:47022] *** reported by process [987037698,0]\n+## [i-capture-the-hostname:47022] *** on a NULL communicator\n+## [i-capture-the-hostname:47022] *** Unknown error\n+## [i-capture-the-hostname:47022] *** MPI_ERRORS_ARE_FATAL (processes in this communicator will now abort,\n+## [i-capture-the-hostname:47022] *** and potentially your MPI job)\n+## [i-capture-the-hostname:46980] 1 more process has sent help message help-mca-bml-r2.txt / unreachable proc\n+## [i-capture-the-hostname:46980] Set MCA parameter \"orte_base_help_aggregate\" to 0 to see all help / error messages\n not ok 518 - ./spawn/spaconacc2 1\n ---\n Directory: ./spawn\n File: spaconacc2\n Num-procs: 1\n- Date: \"Sun Jan 5 01:15:30 2020\"\n+ Date: \"Mon Dec 3 16:39:56 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## --------------------------------------------------------------------------\n ## At least one pair of MPI processes are unable to reach each other for\n ## MPI communications. This means that no Open MPI device has indicated\n ## that it can be used to communicate between these processes. This is\n ## an error; Open MPI requires that all MPI processes be able to reach\n ## each other. This error can sometimes be the result of forgetting to\n ## specify the \"self\" BTL.\n ## \n-## Process 1 ([[19251,1],0]) is on host: profitbricks-build5-amd64\n-## Process 2 ([[19251,2],0]) is on host: profitbricks-build5-amd64\n+## Process 1 ([[14994,1],0]) is on host: i-capture-the-hostname\n+## Process 2 ([[14994,2],0]) is on host: i-capture-the-hostname\n ## BTLs attempted: self\n ## \n ## Your MPI job is now going to abort; sorry.\n ## --------------------------------------------------------------------------\n+## [i-capture-the-hostname:47049] [[14994,1],0] ORTE_ERROR_LOG: Unreachable in file dpm/dpm.c at line 433\n+## [i-capture-the-hostname:47049] *** An error occurred in MPI_Comm_spawn\n+## [i-capture-the-hostname:47049] *** reported by process [982646785,0]\n+## [i-capture-the-hostname:47049] *** on communicator MPI_COMM_SELF\n+## [i-capture-the-hostname:47049] *** MPI_ERR_INTERN: internal error\n+## [i-capture-the-hostname:47049] *** MPI_ERRORS_ARE_FATAL (processes in this communicator will now abort,\n+## [i-capture-the-hostname:47049] *** and potentially your MPI job)\n ## --------------------------------------------------------------------------\n ## It looks like MPI_INIT failed for some reason; your parallel process is\n ## likely to abort. There are many reasons that a parallel process can\n ## fail during MPI_INIT; some of which are due to configuration or environment\n ## problems. This failure appears to be an internal failure; here's some\n ## additional information (which may only be relevant to an Open MPI\n ## developer):\n ## \n ## ompi_dpm_dyn_init() failed\n ## --> Returned \"Unreachable\" (-12) instead of \"Success\" (0)\n ## --------------------------------------------------------------------------\n-## [profitbricks-build5-amd64:36984] [[19251,2],0] ORTE_ERROR_LOG: Unreachable in file dpm/dpm.c at line 433\n-## [profitbricks-build5-amd64:36984] *** An error occurred in MPI_Init\n-## [profitbricks-build5-amd64:36984] *** reported by process [1261633538,0]\n-## [profitbricks-build5-amd64:36984] *** on a NULL communicator\n-## [profitbricks-build5-amd64:36984] *** Unknown error\n-## [profitbricks-build5-amd64:36984] *** MPI_ERRORS_ARE_FATAL (processes in this communicator will now abort,\n-## [profitbricks-build5-amd64:36984] *** and potentially your MPI job)\n-## [profitbricks-build5-amd64:36874] [[19251,1],0] ORTE_ERROR_LOG: Unreachable in file dpm/dpm.c at line 433\n-## [profitbricks-build5-amd64:36874] *** An error occurred in MPI_Comm_spawn\n-## [profitbricks-build5-amd64:36874] *** reported by process [1261633537,0]\n-## [profitbricks-build5-amd64:36874] *** on communicator MPI_COMM_SELF\n-## [profitbricks-build5-amd64:36874] *** MPI_ERR_INTERN: internal error\n-## [profitbricks-build5-amd64:36874] *** MPI_ERRORS_ARE_FATAL (processes in this communicator will now abort,\n-## [profitbricks-build5-amd64:36874] *** and potentially your MPI job)\n-## [profitbricks-build5-amd64:36833] 1 more process has sent help message help-mca-bml-r2.txt / unreachable proc\n-## [profitbricks-build5-amd64:36833] Set MCA parameter \"orte_base_help_aggregate\" to 0 to see all help / error messages\n+## [i-capture-the-hostname:47063] [[14994,2],0] ORTE_ERROR_LOG: Unreachable in file dpm/dpm.c at line 433\n+## [i-capture-the-hostname:47043] 1 more process has sent help message help-mca-bml-r2.txt / unreachable proc\n+## [i-capture-the-hostname:47043] Set MCA parameter \"orte_base_help_aggregate\" to 0 to see all help / error messages\n ok 519 - ./spawn/selfconacc 2\n not ok 520 - ./spawn/spaiccreate 2\n ---\n Directory: ./spawn\n File: spaiccreate\n Num-procs: 2\n- Date: \"Sun Jan 5 01:15:33 2020\"\n+ Date: \"Mon Dec 3 16:39:59 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## --------------------------------------------------------------------------\n ## At least one pair of MPI processes are unable to reach each other for\n ## MPI communications. This means that no Open MPI device has indicated\n ## that it can be used to communicate between these processes. This is\n ## an error; Open MPI requires that all MPI processes be able to reach\n ## each other. This error can sometimes be the result of forgetting to\n ## specify the \"self\" BTL.\n ## \n-## Process 1 ([[22167,2],1]) is on host: profitbricks-build5-amd64\n-## Process 2 ([[22167,1],0]) is on host: profitbricks-build5-amd64\n+## Process 1 ([[13590,1],0]) is on host: i-capture-the-hostname\n+## Process 2 ([[13590,2],0]) is on host: i-capture-the-hostname\n ## BTLs attempted: vader self\n ## \n ## Your MPI job is now going to abort; sorry.\n ## --------------------------------------------------------------------------\n-## [profitbricks-build5-amd64:37597] [[22167,2],1] ORTE_ERROR_LOG: Unreachable in file dpm/dpm.c at line 433\n+## [i-capture-the-hostname:47194] [[13590,1],0] ORTE_ERROR_LOG: Unreachable in file dpm/dpm.c at line 433\n+## [i-capture-the-hostname:47303] [[13590,2],0] ORTE_ERROR_LOG: Unreachable in file dpm/dpm.c at line 433\n+## [i-capture-the-hostname:47306] [[13590,2],1] ORTE_ERROR_LOG: Unreachable in file dpm/dpm.c at line 433\n ## --------------------------------------------------------------------------\n ## It looks like MPI_INIT failed for some reason; your parallel process is\n ## likely to abort. There are many reasons that a parallel process can\n ## fail during MPI_INIT; some of which are due to configuration or environment\n ## problems. This failure appears to be an internal failure; here's some\n ## additional information (which may only be relevant to an Open MPI\n ## developer):\n ## \n ## ompi_dpm_dyn_init() failed\n ## --> Returned \"Unreachable\" (-12) instead of \"Success\" (0)\n ## --------------------------------------------------------------------------\n-## [profitbricks-build5-amd64:37597] *** An error occurred in MPI_Init_thread\n-## [profitbricks-build5-amd64:37597] *** reported by process [1452736514,1]\n-## [profitbricks-build5-amd64:37597] *** on a NULL communicator\n-## [profitbricks-build5-amd64:37597] *** Unknown error\n-## [profitbricks-build5-amd64:37597] *** MPI_ERRORS_ARE_FATAL (processes in this communicator will now abort,\n-## [profitbricks-build5-amd64:37597] *** and potentially your MPI job)\n-## [profitbricks-build5-amd64:37596] [[22167,2],0] ORTE_ERROR_LOG: Unreachable in file dpm/dpm.c at line 433\n-## [profitbricks-build5-amd64:37501] [[22167,1],0] ORTE_ERROR_LOG: Unreachable in file dpm/dpm.c at line 433\n-## [profitbricks-build5-amd64:37501] *** An error occurred in MPI_Comm_spawn\n-## [profitbricks-build5-amd64:37501] *** reported by process [1452736513,0]\n-## [profitbricks-build5-amd64:37501] *** on communicator MPI_COMM_SELF\n-## [profitbricks-build5-amd64:37501] *** MPI_ERR_INTERN: internal error\n-## [profitbricks-build5-amd64:37501] *** MPI_ERRORS_ARE_FATAL (processes in this communicator will now abort,\n-## [profitbricks-build5-amd64:37501] *** and potentially your MPI job)\n-## [profitbricks-build5-amd64:37445] 2 more processes have sent help message help-mca-bml-r2.txt / unreachable proc\n-## [profitbricks-build5-amd64:37445] Set MCA parameter \"orte_base_help_aggregate\" to 0 to see all help / error messages\n-## [profitbricks-build5-amd64:37445] 1 more process has sent help message help-mpi-runtime.txt / mpi_init:startup:internal-failure\n-## [profitbricks-build5-amd64:37445] 1 more process has sent help message help-mpi-errors.txt / mpi_errors_are_fatal unknown handle\n+## [i-capture-the-hostname:47194] *** An error occurred in MPI_Comm_spawn\n+## [i-capture-the-hostname:47194] *** reported by process [890634241,0]\n+## [i-capture-the-hostname:47194] *** on communicator MPI_COMM_SELF\n+## [i-capture-the-hostname:47194] *** MPI_ERR_INTERN: internal error\n+## [i-capture-the-hostname:47194] *** MPI_ERRORS_ARE_FATAL (processes in this communicator will now abort,\n+## [i-capture-the-hostname:47194] *** and potentially your MPI job)\n+## [i-capture-the-hostname:47303] *** An error occurred in MPI_Init_thread\n+## [i-capture-the-hostname:47303] *** reported by process [890634242,0]\n+## [i-capture-the-hostname:47303] *** on a NULL communicator\n+## [i-capture-the-hostname:47303] *** Unknown error\n+## [i-capture-the-hostname:47303] *** MPI_ERRORS_ARE_FATAL (processes in this communicator will now abort,\n+## [i-capture-the-hostname:47303] *** and potentially your MPI job)\n+## [i-capture-the-hostname:47175] 2 more processes have sent help message help-mca-bml-r2.txt / unreachable proc\n+## [i-capture-the-hostname:47175] Set MCA parameter \"orte_base_help_aggregate\" to 0 to see all help / error messages\n+## [i-capture-the-hostname:47175] 1 more process has sent help message help-mpi-runtime.txt / mpi_init:startup:internal-failure\n+## [i-capture-the-hostname:47175] 1 more process has sent help message help-mpi-errors.txt / mpi_errors_are_fatal unknown handle\n not ok 521 - ./spawn/taskmaster 1\n ---\n Directory: ./spawn\n File: taskmaster\n Num-procs: 1\n- Date: \"Sun Jan 5 01:15:35 2020\"\n+ Date: \"Mon Dec 3 16:40:00 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## --------------------------------------------------------------------------\n ## At least one pair of MPI processes are unable to reach each other for\n ## MPI communications. This means that no Open MPI device has indicated\n ## that it can be used to communicate between these processes. This is\n ## an error; Open MPI requires that all MPI processes be able to reach\n ## each other. This error can sometimes be the result of forgetting to\n ## specify the \"self\" BTL.\n ## \n-## Process 1 ([[22462,2],0]) is on host: profitbricks-build5-amd64\n-## Process 2 ([[22462,1],0]) is on host: profitbricks-build5-amd64\n+## Process 1 ([[13351,1],0]) is on host: i-capture-the-hostname\n+## Process 2 ([[13351,2],0]) is on host: i-capture-the-hostname\n ## BTLs attempted: self\n ## \n ## Your MPI job is now going to abort; sorry.\n ## --------------------------------------------------------------------------\n-## [profitbricks-build5-amd64:37968] [[22462,2],0] ORTE_ERROR_LOG: Unreachable in file dpm/dpm.c at line 433\n+## [i-capture-the-hostname:47649] [[13351,2],0] ORTE_ERROR_LOG: Unreachable in file dpm/dpm.c at line 433\n+## [i-capture-the-hostname:47531] [[13351,1],0] ORTE_ERROR_LOG: Unreachable in file dpm/dpm.c at line 433\n ## --------------------------------------------------------------------------\n ## It looks like MPI_INIT failed for some reason; your parallel process is\n ## likely to abort. There are many reasons that a parallel process can\n ## fail during MPI_INIT; some of which are due to configuration or environment\n ## problems. This failure appears to be an internal failure; here's some\n ## additional information (which may only be relevant to an Open MPI\n ## developer):\n ## \n ## ompi_dpm_dyn_init() failed\n ## --> Returned \"Unreachable\" (-12) instead of \"Success\" (0)\n ## --------------------------------------------------------------------------\n-## [profitbricks-build5-amd64:37968] *** An error occurred in MPI_Init\n-## [profitbricks-build5-amd64:37968] *** reported by process [1472069634,0]\n-## [profitbricks-build5-amd64:37968] *** on a NULL communicator\n-## [profitbricks-build5-amd64:37968] *** Unknown error\n-## [profitbricks-build5-amd64:37968] *** MPI_ERRORS_ARE_FATAL (processes in this communicator will now abort,\n-## [profitbricks-build5-amd64:37968] *** and potentially your MPI job)\n-## [profitbricks-build5-amd64:37840] *** An error occurred in MPI_Comm_spawn\n-## [profitbricks-build5-amd64:37840] *** reported by process [1472069633,0]\n-## [profitbricks-build5-amd64:37840] *** on communicator MPI_COMM_WORLD\n-## [profitbricks-build5-amd64:37840] *** MPI_ERR_INTERN: internal error\n-## [profitbricks-build5-amd64:37840] *** MPI_ERRORS_ARE_FATAL (processes in this communicator will now abort,\n-## [profitbricks-build5-amd64:37840] *** and potentially your MPI job)\n-## [profitbricks-build5-amd64:37840] [[22462,1],0] ORTE_ERROR_LOG: Unreachable in file dpm/dpm.c at line 433\n-## [profitbricks-build5-amd64:37740] 1 more process has sent help message help-mca-bml-r2.txt / unreachable proc\n-## [profitbricks-build5-amd64:37740] Set MCA parameter \"orte_base_help_aggregate\" to 0 to see all help / error messages\n+## [i-capture-the-hostname:47531] *** An error occurred in MPI_Comm_spawn\n+## [i-capture-the-hostname:47531] *** reported by process [874971137,0]\n+## [i-capture-the-hostname:47531] *** on communicator MPI_COMM_WORLD\n+## [i-capture-the-hostname:47531] *** MPI_ERR_INTERN: internal error\n+## [i-capture-the-hostname:47531] *** MPI_ERRORS_ARE_FATAL (processes in this communicator will now abort,\n+## [i-capture-the-hostname:47531] *** and potentially your MPI job)\n+## [i-capture-the-hostname:47649] *** An error occurred in MPI_Init\n+## [i-capture-the-hostname:47649] *** reported by process [874971138,0]\n+## [i-capture-the-hostname:47649] *** on a NULL communicator\n+## [i-capture-the-hostname:47649] *** Unknown error\n+## [i-capture-the-hostname:47649] *** MPI_ERRORS_ARE_FATAL (processes in this communicator will now abort,\n+## [i-capture-the-hostname:47649] *** and potentially your MPI job)\n+## [i-capture-the-hostname:47478] 1 more process has sent help message help-mca-bml-r2.txt / unreachable proc\n+## [i-capture-the-hostname:47478] Set MCA parameter \"orte_base_help_aggregate\" to 0 to see all help / error messages\n not ok 522 - ./spawn/taskmaster 2\n ---\n Directory: ./spawn\n File: taskmaster\n Num-procs: 2\n- Date: \"Sun Jan 5 01:15:36 2020\"\n+ Date: \"Mon Dec 3 16:40:02 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## --------------------------------------------------------------------------\n ## At least one pair of MPI processes are unable to reach each other for\n ## MPI communications. This means that no Open MPI device has indicated\n ## that it can be used to communicate between these processes. This is\n ## an error; Open MPI requires that all MPI processes be able to reach\n ## each other. This error can sometimes be the result of forgetting to\n ## specify the \"self\" BTL.\n ## \n-## Process 1 ([[20979,1],1]) is on host: profitbricks-build5-amd64\n-## Process 2 ([[20979,2],0]) is on host: profitbricks-build5-amd64\n+## Process 1 ([[12569,1],1]) is on host: i-capture-the-hostname\n+## Process 2 ([[12569,2],0]) is on host: i-capture-the-hostname\n ## BTLs attempted: vader self\n ## \n ## Your MPI job is now going to abort; sorry.\n ## --------------------------------------------------------------------------\n-## [profitbricks-build5-amd64:38233] [[20979,1],1] ORTE_ERROR_LOG: Unreachable in file dpm/dpm.c at line 433\n-## [profitbricks-build5-amd64:38233] *** An error occurred in MPI_Comm_spawn\n-## [profitbricks-build5-amd64:38233] *** reported by process [1374879745,1]\n-## [profitbricks-build5-amd64:38233] *** on communicator MPI_COMM_WORLD\n-## [profitbricks-build5-amd64:38233] *** MPI_ERR_INTERN: internal error\n-## [profitbricks-build5-amd64:38233] *** MPI_ERRORS_ARE_FATAL (processes in this communicator will now abort,\n-## [profitbricks-build5-amd64:38233] *** and potentially your MPI job)\n-## [profitbricks-build5-amd64:38232] [[20979,1],0] ORTE_ERROR_LOG: Unreachable in file dpm/dpm.c at line 433\n-## --------------------------------------------------------------------------\n-## It looks like MPI_INIT failed for some reason; your parallel process is\n-## likely to abort. There are many reasons that a parallel process can\n-## fail during MPI_INIT; some of which are due to configuration or environment\n-## problems. This failure appears to be an internal failure; here's some\n-## additional information (which may only be relevant to an Open MPI\n-## developer):\n-## \n-## ompi_dpm_dyn_init() failed\n-## --> Returned \"Unreachable\" (-12) instead of \"Success\" (0)\n-## --------------------------------------------------------------------------\n-## [profitbricks-build5-amd64:38240] [[20979,2],0] ORTE_ERROR_LOG: Unreachable in file dpm/dpm.c at line 433\n-## [profitbricks-build5-amd64:38177] 2 more processes have sent help message help-mca-bml-r2.txt / unreachable proc\n-## [profitbricks-build5-amd64:38177] Set MCA parameter \"orte_base_help_aggregate\" to 0 to see all help / error messages\n-## [profitbricks-build5-amd64:38177] 1 more process has sent help message help-mpi-errors.txt / mpi_errors_are_fatal\n+## [i-capture-the-hostname:48848] [[12569,1],0] ORTE_ERROR_LOG: Unreachable in file dpm/dpm.c at line 433\n+## [i-capture-the-hostname:48848] *** An error occurred in MPI_Comm_spawn\n+## [i-capture-the-hostname:48848] *** reported by process [823721985,0]\n+## [i-capture-the-hostname:48848] *** on communicator MPI_COMM_WORLD\n+## [i-capture-the-hostname:48848] *** MPI_ERR_INTERN: internal error\n+## [i-capture-the-hostname:48848] *** MPI_ERRORS_ARE_FATAL (processes in this communicator will now abort,\n+## [i-capture-the-hostname:48848] *** and potentially your MPI job)\n+## [i-capture-the-hostname:48854] [[12569,1],1] ORTE_ERROR_LOG: Unreachable in file dpm/dpm.c at line 433\n+## [i-capture-the-hostname:48200] 2 more processes have sent help message help-mca-bml-r2.txt / unreachable proc\n+## [i-capture-the-hostname:48200] Set MCA parameter \"orte_base_help_aggregate\" to 0 to see all help / error messages\n+## [i-capture-the-hostname:48200] 1 more process has sent help message help-mpi-errors.txt / mpi_errors_are_fatal\n+## [warn] Epoll MOD(1) on fd 29 failed. Old events were 6; read change was 0 (none); write change was 2 (del); close change was 0 (none): Bad file descriptor\n not ok 523 - ./spawn/join 2\n ---\n Directory: ./spawn\n File: join\n Num-procs: 2\n- Date: \"Sun Jan 5 01:15:37 2020\"\n+ Date: \"Mon Dec 3 16:40:03 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## gethostbyname failed\n ## --------------------------------------------------------------------------\n ## MPI_ABORT was invoked on rank 0 in communicator MPI_COMM_WORLD\n ## with errorcode 1.\n ## \n@@ -998235,516 +998208,545 @@\n ## exactly when Open MPI kills them.\n ## --------------------------------------------------------------------------\n not ok 524 - ./spawn/disconnect_reconnect 2\n ---\n Directory: ./spawn\n File: disconnect_reconnect\n Num-procs: 2\n- Date: \"Sun Jan 5 01:15:39 2020\"\n+ Date: \"Mon Dec 3 16:40:05 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## --------------------------------------------------------------------------\n ## At least one pair of MPI processes are unable to reach each other for\n ## MPI communications. This means that no Open MPI device has indicated\n ## that it can be used to communicate between these processes. This is\n ## an error; Open MPI requires that all MPI processes be able to reach\n ## each other. This error can sometimes be the result of forgetting to\n ## specify the \"self\" BTL.\n ## \n-## Process 1 ([[21189,1],1]) is on host: profitbricks-build5-amd64\n-## Process 2 ([[21189,2],0]) is on host: profitbricks-build5-amd64\n+## Process 1 ([[19253,2],2]) is on host: i-capture-the-hostname\n+## Process 2 ([[19253,1],0]) is on host: i-capture-the-hostname\n ## BTLs attempted: vader self\n ## \n ## Your MPI job is now going to abort; sorry.\n ## --------------------------------------------------------------------------\n-## [profitbricks-build5-amd64:38469] [[21189,1],1] ORTE_ERROR_LOG: Unreachable in file dpm/dpm.c at line 433\n-## [profitbricks-build5-amd64:38615] [[21189,2],2] ORTE_ERROR_LOG: Unreachable in file dpm/dpm.c at line 433\n-## [profitbricks-build5-amd64:38469] *** An error occurred in MPI_Comm_spawn\n-## [profitbricks-build5-amd64:38469] *** reported by process [1388642305,1]\n-## [profitbricks-build5-amd64:38469] *** on communicator MPI_COMM_WORLD\n-## [profitbricks-build5-amd64:38469] *** MPI_ERR_INTERN: internal error\n-## [profitbricks-build5-amd64:38469] *** MPI_ERRORS_ARE_FATAL (processes in this communicator will now abort,\n-## [profitbricks-build5-amd64:38469] *** and potentially your MPI job)\n+## [i-capture-the-hostname:51399] [[19253,2],0] ORTE_ERROR_LOG: Unreachable in file dpm/dpm.c at line 433\n ## --------------------------------------------------------------------------\n ## It looks like MPI_INIT failed for some reason; your parallel process is\n ## likely to abort. There are many reasons that a parallel process can\n ## fail during MPI_INIT; some of which are due to configuration or environment\n ## problems. This failure appears to be an internal failure; here's some\n ## additional information (which may only be relevant to an Open MPI\n ## developer):\n ## \n ## ompi_dpm_dyn_init() failed\n ## --> Returned \"Unreachable\" (-12) instead of \"Success\" (0)\n ## --------------------------------------------------------------------------\n-## [profitbricks-build5-amd64:38468] [[21189,1],0] ORTE_ERROR_LOG: Unreachable in file dpm/dpm.c at line 433\n-## [profitbricks-build5-amd64:38615] *** An error occurred in MPI_Init_thread\n-## [profitbricks-build5-amd64:38615] *** reported by process [1388642306,2]\n-## [profitbricks-build5-amd64:38615] *** on a NULL communicator\n-## [profitbricks-build5-amd64:38615] *** Unknown error\n-## [profitbricks-build5-amd64:38615] *** MPI_ERRORS_ARE_FATAL (processes in this communicator will now abort,\n-## [profitbricks-build5-amd64:38615] *** and potentially your MPI job)\n-## [profitbricks-build5-amd64:38423] 4 more processes have sent help message help-mca-bml-r2.txt / unreachable proc\n-## [profitbricks-build5-amd64:38423] Set MCA parameter \"orte_base_help_aggregate\" to 0 to see all help / error messages\n-## [profitbricks-build5-amd64:38423] 1 more process has sent help message help-mpi-errors.txt / mpi_errors_are_fatal\n+## [i-capture-the-hostname:51043] [[19253,1],1] ORTE_ERROR_LOG: Unreachable in file dpm/dpm.c at line 433\n+## [i-capture-the-hostname:51043] *** An error occurred in MPI_Comm_spawn\n+## [i-capture-the-hostname:51043] *** reported by process [1261764609,1]\n+## [i-capture-the-hostname:51043] *** on communicator MPI_COMM_WORLD\n+## [i-capture-the-hostname:51043] *** MPI_ERR_INTERN: internal error\n+## [i-capture-the-hostname:51043] *** MPI_ERRORS_ARE_FATAL (processes in this communicator will now abort,\n+## [i-capture-the-hostname:51043] *** and potentially your MPI job)\n+## [i-capture-the-hostname:51399] *** An error occurred in MPI_Init_thread\n+## [i-capture-the-hostname:51399] *** reported by process [1261764610,0]\n+## [i-capture-the-hostname:51399] *** on a NULL communicator\n+## [i-capture-the-hostname:51399] *** Unknown error\n+## [i-capture-the-hostname:51399] *** MPI_ERRORS_ARE_FATAL (processes in this communicator will now abort,\n+## [i-capture-the-hostname:51399] *** and potentially your MPI job)\n+## [i-capture-the-hostname:51036] [[19253,1],0] ORTE_ERROR_LOG: Unreachable in file dpm/dpm.c at line 433\n+## [i-capture-the-hostname:51417] [[19253,2],2] ORTE_ERROR_LOG: Unreachable in file dpm/dpm.c at line 433\n+## [i-capture-the-hostname:50788] 4 more processes have sent help message help-mca-bml-r2.txt / unreachable proc\n+## [i-capture-the-hostname:50788] Set MCA parameter \"orte_base_help_aggregate\" to 0 to see all help / error messages\n+## [i-capture-the-hostname:50788] 1 more process has sent help message help-mpi-runtime.txt / mpi_init:startup:internal-failure\n+## [i-capture-the-hostname:50788] 1 more process has sent help message help-mpi-errors.txt / mpi_errors_are_fatal\n not ok 525 - ./spawn/disconnect_reconnect2 2\n ---\n Directory: ./spawn\n File: disconnect_reconnect2\n Num-procs: 2\n- Date: \"Sun Jan 5 01:15:41 2020\"\n+ Date: \"Mon Dec 3 16:40:06 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## --------------------------------------------------------------------------\n ## At least one pair of MPI processes are unable to reach each other for\n ## MPI communications. This means that no Open MPI device has indicated\n ## that it can be used to communicate between these processes. This is\n ## an error; Open MPI requires that all MPI processes be able to reach\n ## each other. This error can sometimes be the result of forgetting to\n ## specify the \"self\" BTL.\n ## \n-## Process 1 ([[23641,1],1]) is on host: profitbricks-build5-amd64\n-## Process 2 ([[23641,2],0]) is on host: profitbricks-build5-amd64\n+## Process 1 ([[18221,2],1]) is on host: i-capture-the-hostname\n+## Process 2 ([[18221,1],0]) is on host: i-capture-the-hostname\n ## BTLs attempted: vader self\n ## \n ## Your MPI job is now going to abort; sorry.\n ## --------------------------------------------------------------------------\n-## [profitbricks-build5-amd64:39150] [[23641,1],1] ORTE_ERROR_LOG: Unreachable in file dpm/dpm.c at line 433\n-## [profitbricks-build5-amd64:39150] *** An error occurred in MPI_Comm_spawn\n-## [profitbricks-build5-amd64:39150] *** reported by process [1549336577,1]\n-## [profitbricks-build5-amd64:39150] *** on communicator MPI_COMM_WORLD\n-## [profitbricks-build5-amd64:39150] *** MPI_ERR_INTERN: internal error\n-## [profitbricks-build5-amd64:39150] *** MPI_ERRORS_ARE_FATAL (processes in this communicator will now abort,\n-## [profitbricks-build5-amd64:39150] *** and potentially your MPI job)\n+## [i-capture-the-hostname:52518] [[18221,2],1] ORTE_ERROR_LOG: Unreachable in file dpm/dpm.c at line 433\n+## [i-capture-the-hostname:52119] [[18221,1],1] ORTE_ERROR_LOG: Unreachable in file dpm/dpm.c at line 433\n ## --------------------------------------------------------------------------\n ## It looks like MPI_INIT failed for some reason; your parallel process is\n ## likely to abort. There are many reasons that a parallel process can\n ## fail during MPI_INIT; some of which are due to configuration or environment\n ## problems. This failure appears to be an internal failure; here's some\n ## additional information (which may only be relevant to an Open MPI\n ## developer):\n ## \n ## ompi_dpm_dyn_init() failed\n ## --> Returned \"Unreachable\" (-12) instead of \"Success\" (0)\n ## --------------------------------------------------------------------------\n-## [profitbricks-build5-amd64:39148] [[23641,1],0] ORTE_ERROR_LOG: Unreachable in file dpm/dpm.c at line 433\n-## [profitbricks-build5-amd64:39271] [[23641,2],2] ORTE_ERROR_LOG: Unreachable in file dpm/dpm.c at line 433\n-## [profitbricks-build5-amd64:39270] [[23641,2],1] ORTE_ERROR_LOG: Unreachable in file dpm/dpm.c at line 433\n-## [profitbricks-build5-amd64:39051] 4 more processes have sent help message help-mca-bml-r2.txt / unreachable proc\n-## [profitbricks-build5-amd64:39051] Set MCA parameter \"orte_base_help_aggregate\" to 0 to see all help / error messages\n-## [profitbricks-build5-amd64:39051] 1 more process has sent help message help-mpi-errors.txt / mpi_errors_are_fatal\n-## [profitbricks-build5-amd64:39051] 1 more process has sent help message help-mpi-runtime.txt / mpi_init:startup:internal-failure\n+## [i-capture-the-hostname:52119] *** An error occurred in MPI_Comm_spawn\n+## [i-capture-the-hostname:52119] *** reported by process [1194131457,1]\n+## [i-capture-the-hostname:52119] *** on communicator MPI_COMM_WORLD\n+## [i-capture-the-hostname:52119] *** MPI_ERR_INTERN: internal error\n+## [i-capture-the-hostname:52119] *** MPI_ERRORS_ARE_FATAL (processes in this communicator will now abort,\n+## [i-capture-the-hostname:52119] *** and potentially your MPI job)\n+## [i-capture-the-hostname:52518] *** An error occurred in MPI_Init\n+## [i-capture-the-hostname:52518] *** reported by process [1194131458,1]\n+## [i-capture-the-hostname:52518] *** on a NULL communicator\n+## [i-capture-the-hostname:52518] *** Unknown error\n+## [i-capture-the-hostname:52518] *** MPI_ERRORS_ARE_FATAL (processes in this communicator will now abort,\n+## [i-capture-the-hostname:52518] *** and potentially your MPI job)\n+## [i-capture-the-hostname:51836] 4 more processes have sent help message help-mca-bml-r2.txt / unreachable proc\n+## [i-capture-the-hostname:51836] Set MCA parameter \"orte_base_help_aggregate\" to 0 to see all help / error messages\n not ok 526 - ./spawn/disconnect_reconnect3 2\n ---\n Directory: ./spawn\n File: disconnect_reconnect3\n Num-procs: 2\n- Date: \"Sun Jan 5 01:15:43 2020\"\n+ Date: \"Mon Dec 3 16:40:08 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## --------------------------------------------------------------------------\n ## At least one pair of MPI processes are unable to reach each other for\n ## MPI communications. This means that no Open MPI device has indicated\n ## that it can be used to communicate between these processes. This is\n ## an error; Open MPI requires that all MPI processes be able to reach\n ## each other. This error can sometimes be the result of forgetting to\n ## specify the \"self\" BTL.\n ## \n-## Process 1 ([[24307,2],0]) is on host: profitbricks-build5-amd64\n-## Process 2 ([[24307,1],0]) is on host: profitbricks-build5-amd64\n+## Process 1 ([[16910,2],0]) is on host: i-capture-the-hostname\n+## Process 2 ([[16910,1],0]) is on host: i-capture-the-hostname\n ## BTLs attempted: vader self\n ## \n ## Your MPI job is now going to abort; sorry.\n ## --------------------------------------------------------------------------\n-## [profitbricks-build5-amd64:39665] [[24307,2],0] ORTE_ERROR_LOG: Unreachable in file dpm/dpm.c at line 433\n-## [profitbricks-build5-amd64:39668] [[24307,2],1] ORTE_ERROR_LOG: Unreachable in file dpm/dpm.c at line 433\n-## [profitbricks-build5-amd64:39672] [[24307,2],2] ORTE_ERROR_LOG: Unreachable in file dpm/dpm.c at line 433\n-## [profitbricks-build5-amd64:39677] [[24307,2],3] ORTE_ERROR_LOG: Unreachable in file dpm/dpm.c at line 433\n-## [profitbricks-build5-amd64:39562] [[24307,1],0] ORTE_ERROR_LOG: Unreachable in file dpm/dpm.c at line 433\n-## [profitbricks-build5-amd64:39563] [[24307,1],1] ORTE_ERROR_LOG: Unreachable in file dpm/dpm.c at line 433\n-## [profitbricks-build5-amd64:39562] *** An error occurred in MPI_Comm_spawn\n-## [profitbricks-build5-amd64:39562] *** reported by process [1592983553,0]\n-## [profitbricks-build5-amd64:39562] *** on communicator MPI_COMM_WORLD\n-## [profitbricks-build5-amd64:39562] *** MPI_ERR_INTERN: internal error\n-## [profitbricks-build5-amd64:39562] *** MPI_ERRORS_ARE_FATAL (processes in this communicator will now abort,\n-## [profitbricks-build5-amd64:39562] *** and potentially your MPI job)\n+## [i-capture-the-hostname:53900] [[16910,2],0] ORTE_ERROR_LOG: Unreachable in file dpm/dpm.c at line 433\n+## [i-capture-the-hostname:53921] [[16910,2],2] ORTE_ERROR_LOG: Unreachable in file dpm/dpm.c at line 433\n ## --------------------------------------------------------------------------\n ## It looks like MPI_INIT failed for some reason; your parallel process is\n ## likely to abort. There are many reasons that a parallel process can\n ## fail during MPI_INIT; some of which are due to configuration or environment\n ## problems. This failure appears to be an internal failure; here's some\n ## additional information (which may only be relevant to an Open MPI\n ## developer):\n ## \n ## ompi_dpm_dyn_init() failed\n ## --> Returned \"Unreachable\" (-12) instead of \"Success\" (0)\n ## --------------------------------------------------------------------------\n-## [profitbricks-build5-amd64:39672] *** An error occurred in MPI_Init_thread\n-## [profitbricks-build5-amd64:39672] *** reported by process [1592983554,2]\n-## [profitbricks-build5-amd64:39672] *** on a NULL communicator\n-## [profitbricks-build5-amd64:39672] *** Unknown error\n-## [profitbricks-build5-amd64:39672] *** MPI_ERRORS_ARE_FATAL (processes in this communicator will now abort,\n-## [profitbricks-build5-amd64:39672] *** and potentially your MPI job)\n-## [profitbricks-build5-amd64:39457] 5 more processes have sent help message help-mca-bml-r2.txt / unreachable proc\n-## [profitbricks-build5-amd64:39457] Set MCA parameter \"orte_base_help_aggregate\" to 0 to see all help / error messages\n-## [profitbricks-build5-amd64:39457] 1 more process has sent help message help-mpi-errors.txt / mpi_errors_are_fatal\n-## [profitbricks-build5-amd64:39457] 3 more processes have sent help message help-mpi-runtime.txt / mpi_init:startup:internal-failure\n-## [profitbricks-build5-amd64:39457] 3 more processes have sent help message help-mpi-errors.txt / mpi_errors_are_fatal unknown handle\n+## [i-capture-the-hostname:53436] *** An error occurred in MPI_Comm_spawn\n+## [i-capture-the-hostname:53436] *** reported by process [1108213761,1]\n+## [i-capture-the-hostname:53436] *** on communicator MPI_COMM_WORLD\n+## [i-capture-the-hostname:53436] *** MPI_ERR_INTERN: internal error\n+## [i-capture-the-hostname:53436] *** MPI_ERRORS_ARE_FATAL (processes in this communicator will now abort,\n+## [i-capture-the-hostname:53436] *** and potentially your MPI job)\n+## [i-capture-the-hostname:53921] *** An error occurred in MPI_Init_thread\n+## [i-capture-the-hostname:53921] *** reported by process [1108213762,2]\n+## [i-capture-the-hostname:53921] *** on a NULL communicator\n+## [i-capture-the-hostname:53921] *** Unknown error\n+## [i-capture-the-hostname:53921] *** MPI_ERRORS_ARE_FATAL (processes in this communicator will now abort,\n+## [i-capture-the-hostname:53921] *** and potentially your MPI job)\n+## [i-capture-the-hostname:53428] [[16910,1],0] ORTE_ERROR_LOG: Unreachable in file dpm/dpm.c at line 433\n+## [i-capture-the-hostname:53436] [[16910,1],1] ORTE_ERROR_LOG: Unreachable in file dpm/dpm.c at line 433\n+## [i-capture-the-hostname:53930] [[16910,2],3] ORTE_ERROR_LOG: Unreachable in file dpm/dpm.c at line 433\n+## [i-capture-the-hostname:53911] [[16910,2],1] ORTE_ERROR_LOG: Unreachable in file dpm/dpm.c at line 433\n+## [i-capture-the-hostname:53087] 5 more processes have sent help message help-mca-bml-r2.txt / unreachable proc\n+## [i-capture-the-hostname:53087] Set MCA parameter \"orte_base_help_aggregate\" to 0 to see all help / error messages\n+## [i-capture-the-hostname:53087] 3 more processes have sent help message help-mpi-runtime.txt / mpi_init:startup:internal-failure\n+## [i-capture-the-hostname:53087] 1 more process has sent help message help-mpi-errors.txt / mpi_errors_are_fatal\n+## [i-capture-the-hostname:53087] 3 more processes have sent help message help-mpi-errors.txt / mpi_errors_are_fatal unknown handle\n not ok 527 - ./spawn/multiple_ports 2\n ---\n Directory: ./spawn\n File: multiple_ports\n Num-procs: 2\n- Date: \"Sun Jan 5 01:15:44 2020\"\n+ Date: \"Mon Dec 3 16:40:09 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Three processes needed to run this test.\n ## Three processes needed to run this test.\n not ok 528 - ./spawn/multiple_ports2 2\n ---\n Directory: ./spawn\n File: multiple_ports2\n Num-procs: 2\n- Date: \"Sun Jan 5 01:15:45 2020\"\n+ Date: \"Mon Dec 3 16:40:11 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Four processes needed to run this test.\n ## Four processes needed to run this test.\n not ok 529 - ./spawn/disconnect 2\n ---\n Directory: ./spawn\n File: disconnect\n Num-procs: 2\n- Date: \"Sun Jan 5 01:15:47 2020\"\n+ Date: \"Mon Dec 3 16:40:14 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## --------------------------------------------------------------------------\n ## At least one pair of MPI processes are unable to reach each other for\n ## MPI communications. This means that no Open MPI device has indicated\n ## that it can be used to communicate between these processes. This is\n ## an error; Open MPI requires that all MPI processes be able to reach\n ## each other. This error can sometimes be the result of forgetting to\n ## specify the \"self\" BTL.\n ## \n-## Process 1 ([[23265,2],2]) is on host: profitbricks-build5-amd64\n-## Process 2 ([[23265,1],0]) is on host: profitbricks-build5-amd64\n+## Process 1 ([[21815,1],0]) is on host: i-capture-the-hostname\n+## Process 2 ([[21815,2],0]) is on host: i-capture-the-hostname\n ## BTLs attempted: vader self\n ## \n ## Your MPI job is now going to abort; sorry.\n ## --------------------------------------------------------------------------\n-## [profitbricks-build5-amd64:40873] [[23265,2],2] ORTE_ERROR_LOG: Unreachable in file dpm/dpm.c at line 433\n+## [i-capture-the-hostname:55553] [[21815,1],0] ORTE_ERROR_LOG: Unreachable in file dpm/dpm.c at line 433\n+## [i-capture-the-hostname:55553] *** An error occurred in MPI_Comm_spawn\n+## [i-capture-the-hostname:55553] *** reported by process [1429667841,0]\n+## [i-capture-the-hostname:55553] *** on communicator MPI_COMM_WORLD\n+## [i-capture-the-hostname:55553] *** MPI_ERR_INTERN: internal error\n+## [i-capture-the-hostname:55553] *** MPI_ERRORS_ARE_FATAL (processes in this communicator will now abort,\n+## [i-capture-the-hostname:55553] *** and potentially your MPI job)\n+## [i-capture-the-hostname:55756] [[21815,2],0] ORTE_ERROR_LOG: Unreachable in file dpm/dpm.c at line 433\n+## [i-capture-the-hostname:55762] [[21815,2],1] ORTE_ERROR_LOG: Unreachable in file dpm/dpm.c at line 433\n+## [i-capture-the-hostname:55766] [[21815,2],2] ORTE_ERROR_LOG: Unreachable in file dpm/dpm.c at line 433\n+## [i-capture-the-hostname:55560] [[21815,1],1] ORTE_ERROR_LOG: Unreachable in file dpm/dpm.c at line 433\n ## --------------------------------------------------------------------------\n ## It looks like MPI_INIT failed for some reason; your parallel process is\n ## likely to abort. There are many reasons that a parallel process can\n ## fail during MPI_INIT; some of which are due to configuration or environment\n ## problems. This failure appears to be an internal failure; here's some\n ## additional information (which may only be relevant to an Open MPI\n ## developer):\n ## \n ## ompi_dpm_dyn_init() failed\n ## --> Returned \"Unreachable\" (-12) instead of \"Success\" (0)\n ## --------------------------------------------------------------------------\n-## [profitbricks-build5-amd64:40873] *** An error occurred in MPI_Init_thread\n-## [profitbricks-build5-amd64:40873] *** reported by process [1524695042,2]\n-## [profitbricks-build5-amd64:40873] *** on a NULL communicator\n-## [profitbricks-build5-amd64:40873] *** Unknown error\n-## [profitbricks-build5-amd64:40873] *** MPI_ERRORS_ARE_FATAL (processes in this communicator will now abort,\n-## [profitbricks-build5-amd64:40873] *** and potentially your MPI job)\n-## [profitbricks-build5-amd64:40499] 4 more processes have sent help message help-mca-bml-r2.txt / unreachable proc\n-## [profitbricks-build5-amd64:40499] Set MCA parameter \"orte_base_help_aggregate\" to 0 to see all help / error messages\n+## [i-capture-the-hostname:55766] *** An error occurred in MPI_Init_thread\n+## [i-capture-the-hostname:55766] *** reported by process [1429667842,2]\n+## [i-capture-the-hostname:55766] *** on a NULL communicator\n+## [i-capture-the-hostname:55766] *** Unknown error\n+## [i-capture-the-hostname:55766] *** MPI_ERRORS_ARE_FATAL (processes in this communicator will now abort,\n+## [i-capture-the-hostname:55766] *** and potentially your MPI job)\n+## [i-capture-the-hostname:55398] 4 more processes have sent help message help-mca-bml-r2.txt / unreachable proc\n+## [i-capture-the-hostname:55398] Set MCA parameter \"orte_base_help_aggregate\" to 0 to see all help / error messages\n+## [i-capture-the-hostname:55398] 1 more process has sent help message help-mpi-errors.txt / mpi_errors_are_fatal\n+## [i-capture-the-hostname:55398] 2 more processes have sent help message help-mpi-runtime.txt / mpi_init:startup:internal-failure\n+## [i-capture-the-hostname:55398] 2 more processes have sent help message help-mpi-errors.txt / mpi_errors_are_fatal unknown handle\n not ok 530 - ./spawn/disconnect2 2\n ---\n Directory: ./spawn\n File: disconnect2\n Num-procs: 2\n- Date: \"Sun Jan 5 01:15:49 2020\"\n+ Date: \"Mon Dec 3 16:40:16 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## --------------------------------------------------------------------------\n ## At least one pair of MPI processes are unable to reach each other for\n ## MPI communications. This means that no Open MPI device has indicated\n ## that it can be used to communicate between these processes. This is\n ## an error; Open MPI requires that all MPI processes be able to reach\n ## each other. This error can sometimes be the result of forgetting to\n ## specify the \"self\" BTL.\n ## \n-## Process 1 ([[25777,1],1]) is on host: profitbricks-build5-amd64\n-## Process 2 ([[25777,2],0]) is on host: profitbricks-build5-amd64\n+## Process 1 ([[22521,2],0]) is on host: i-capture-the-hostname\n+## Process 2 ([[22521,1],0]) is on host: i-capture-the-hostname\n ## BTLs attempted: vader self\n ## \n ## Your MPI job is now going to abort; sorry.\n ## --------------------------------------------------------------------------\n-## [profitbricks-build5-amd64:41184] [[25777,1],1] ORTE_ERROR_LOG: Unreachable in file dpm/dpm.c at line 433\n-## [profitbricks-build5-amd64:41184] *** An error occurred in MPI_Comm_spawn\n-## [profitbricks-build5-amd64:41184] *** reported by process [1689321473,1]\n-## [profitbricks-build5-amd64:41184] *** on communicator MPI_COMM_WORLD\n-## [profitbricks-build5-amd64:41184] *** MPI_ERR_INTERN: internal error\n-## [profitbricks-build5-amd64:41184] *** MPI_ERRORS_ARE_FATAL (processes in this communicator will now abort,\n-## [profitbricks-build5-amd64:41184] *** and potentially your MPI job)\n-## [profitbricks-build5-amd64:41311] [[25777,2],2] ORTE_ERROR_LOG: Unreachable in file dpm/dpm.c at line 433\n ## --------------------------------------------------------------------------\n ## It looks like MPI_INIT failed for some reason; your parallel process is\n ## likely to abort. There are many reasons that a parallel process can\n ## fail during MPI_INIT; some of which are due to configuration or environment\n ## problems. This failure appears to be an internal failure; here's some\n ## additional information (which may only be relevant to an Open MPI\n ## developer):\n ## \n ## ompi_dpm_dyn_init() failed\n ## --> Returned \"Unreachable\" (-12) instead of \"Success\" (0)\n ## --------------------------------------------------------------------------\n-## [profitbricks-build5-amd64:41301] [[25777,2],0] ORTE_ERROR_LOG: Unreachable in file dpm/dpm.c at line 433\n-## [profitbricks-build5-amd64:41301] *** An error occurred in MPI_Init_thread\n-## [profitbricks-build5-amd64:41301] *** reported by process [1689321474,0]\n-## [profitbricks-build5-amd64:41301] *** on a NULL communicator\n-## [profitbricks-build5-amd64:41301] *** Unknown error\n-## [profitbricks-build5-amd64:41301] *** MPI_ERRORS_ARE_FATAL (processes in this communicator will now abort,\n-## [profitbricks-build5-amd64:41301] *** and potentially your MPI job)\n-## [profitbricks-build5-amd64:41182] [[25777,1],0] ORTE_ERROR_LOG: Unreachable in file dpm/dpm.c at line 433\n-## [profitbricks-build5-amd64:41307] [[25777,2],1] ORTE_ERROR_LOG: Unreachable in file dpm/dpm.c at line 433\n-## [profitbricks-build5-amd64:41059] 4 more processes have sent help message help-mca-bml-r2.txt / unreachable proc\n-## [profitbricks-build5-amd64:41059] Set MCA parameter \"orte_base_help_aggregate\" to 0 to see all help / error messages\n-## [profitbricks-build5-amd64:41059] 1 more process has sent help message help-mpi-errors.txt / mpi_errors_are_fatal\n-## [profitbricks-build5-amd64:41059] 2 more processes have sent help message help-mpi-runtime.txt / mpi_init:startup:internal-failure\n-## [profitbricks-build5-amd64:41059] 1 more process has sent help message help-mpi-errors.txt / mpi_errors_are_fatal unknown handle\n+## [i-capture-the-hostname:56076] [[22521,1],0] ORTE_ERROR_LOG: Unreachable in file dpm/dpm.c at line 433\n+## [i-capture-the-hostname:56085] [[22521,1],1] ORTE_ERROR_LOG: Unreachable in file dpm/dpm.c at line 433\n+## [i-capture-the-hostname:56224] [[22521,2],0] ORTE_ERROR_LOG: Unreachable in file dpm/dpm.c at line 433\n+## [i-capture-the-hostname:56233] [[22521,2],2] ORTE_ERROR_LOG: Unreachable in file dpm/dpm.c at line 433\n+## [i-capture-the-hostname:56229] [[22521,2],1] ORTE_ERROR_LOG: Unreachable in file dpm/dpm.c at line 433\n+## [i-capture-the-hostname:56085] *** An error occurred in MPI_Comm_spawn\n+## [i-capture-the-hostname:56085] *** reported by process [1475936257,1]\n+## [i-capture-the-hostname:56085] *** on communicator MPI_COMM_WORLD\n+## [i-capture-the-hostname:56085] *** MPI_ERR_INTERN: internal error\n+## [i-capture-the-hostname:56085] *** MPI_ERRORS_ARE_FATAL (processes in this communicator will now abort,\n+## [i-capture-the-hostname:56085] *** and potentially your MPI job)\n+## [i-capture-the-hostname:56224] *** An error occurred in MPI_Init_thread\n+## [i-capture-the-hostname:56224] *** reported by process [1475936258,0]\n+## [i-capture-the-hostname:56224] *** on a NULL communicator\n+## [i-capture-the-hostname:56224] *** Unknown error\n+## [i-capture-the-hostname:56224] *** MPI_ERRORS_ARE_FATAL (processes in this communicator will now abort,\n+## [i-capture-the-hostname:56224] *** and potentially your MPI job)\n+## [i-capture-the-hostname:55976] 4 more processes have sent help message help-mca-bml-r2.txt / unreachable proc\n+## [i-capture-the-hostname:55976] Set MCA parameter \"orte_base_help_aggregate\" to 0 to see all help / error messages\n+## [i-capture-the-hostname:55976] 2 more processes have sent help message help-mpi-runtime.txt / mpi_init:startup:internal-failure\n+## [i-capture-the-hostname:55976] 1 more process has sent help message help-mpi-errors.txt / mpi_errors_are_fatal\n+## [i-capture-the-hostname:55976] 2 more processes have sent help message help-mpi-errors.txt / mpi_errors_are_fatal unknown handle\n not ok 531 - ./spawn/disconnect3 2\n ---\n Directory: ./spawn\n File: disconnect3\n Num-procs: 2\n- Date: \"Sun Jan 5 01:15:50 2020\"\n+ Date: \"Mon Dec 3 16:40:18 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## --------------------------------------------------------------------------\n ## At least one pair of MPI processes are unable to reach each other for\n ## MPI communications. This means that no Open MPI device has indicated\n ## that it can be used to communicate between these processes. This is\n ## an error; Open MPI requires that all MPI processes be able to reach\n ## each other. This error can sometimes be the result of forgetting to\n ## specify the \"self\" BTL.\n ## \n-## Process 1 ([[26263,2],0]) is on host: profitbricks-build5-amd64\n-## Process 2 ([[26263,1],0]) is on host: profitbricks-build5-amd64\n+## Process 1 ([[20933,1],0]) is on host: i-capture-the-hostname\n+## Process 2 ([[20933,2],0]) is on host: i-capture-the-hostname\n ## BTLs attempted: vader self\n ## \n ## Your MPI job is now going to abort; sorry.\n ## --------------------------------------------------------------------------\n-## [profitbricks-build5-amd64:41806] [[26263,2],0] ORTE_ERROR_LOG: Unreachable in file dpm/dpm.c at line 433\n-## [profitbricks-build5-amd64:41692] [[26263,1],1] ORTE_ERROR_LOG: Unreachable in file dpm/dpm.c at line 433\n-## [profitbricks-build5-amd64:41810] [[26263,2],1] ORTE_ERROR_LOG: Unreachable in file dpm/dpm.c at line 433\n+## [i-capture-the-hostname:56575] [[20933,1],0] ORTE_ERROR_LOG: Unreachable in file dpm/dpm.c at line 433\n+## [i-capture-the-hostname:56575] *** An error occurred in MPI_Comm_spawn\n+## [i-capture-the-hostname:56575] *** reported by process [1371865089,0]\n+## [i-capture-the-hostname:56575] *** on communicator MPI_COMM_WORLD\n+## [i-capture-the-hostname:56575] *** MPI_ERR_INTERN: internal error\n+## [i-capture-the-hostname:56575] *** MPI_ERRORS_ARE_FATAL (processes in this communicator will now abort,\n+## [i-capture-the-hostname:56575] *** and potentially your MPI job)\n+## [i-capture-the-hostname:56578] [[20933,1],1] ORTE_ERROR_LOG: Unreachable in file dpm/dpm.c at line 433\n ## --------------------------------------------------------------------------\n ## It looks like MPI_INIT failed for some reason; your parallel process is\n ## likely to abort. There are many reasons that a parallel process can\n ## fail during MPI_INIT; some of which are due to configuration or environment\n ## problems. This failure appears to be an internal failure; here's some\n ## additional information (which may only be relevant to an Open MPI\n ## developer):\n ## \n ## ompi_dpm_dyn_init() failed\n ## --> Returned \"Unreachable\" (-12) instead of \"Success\" (0)\n ## --------------------------------------------------------------------------\n-## [profitbricks-build5-amd64:41692] *** An error occurred in MPI_Comm_spawn\n-## [profitbricks-build5-amd64:41692] *** reported by process [1721171969,1]\n-## [profitbricks-build5-amd64:41692] *** on communicator MPI_COMM_WORLD\n-## [profitbricks-build5-amd64:41692] *** MPI_ERR_INTERN: internal error\n-## [profitbricks-build5-amd64:41692] *** MPI_ERRORS_ARE_FATAL (processes in this communicator will now abort,\n-## [profitbricks-build5-amd64:41692] *** and potentially your MPI job)\n-## [profitbricks-build5-amd64:41811] [[26263,2],2] ORTE_ERROR_LOG: Unreachable in file dpm/dpm.c at line 433\n-## [profitbricks-build5-amd64:41810] *** An error occurred in MPI_Init_thread\n-## [profitbricks-build5-amd64:41810] *** reported by process [1721171970,1]\n-## [profitbricks-build5-amd64:41810] *** on a NULL communicator\n-## [profitbricks-build5-amd64:41810] *** Unknown error\n-## [profitbricks-build5-amd64:41810] *** MPI_ERRORS_ARE_FATAL (processes in this communicator will now abort,\n-## [profitbricks-build5-amd64:41810] *** and potentially your MPI job)\n-## [profitbricks-build5-amd64:41541] 4 more processes have sent help message help-mca-bml-r2.txt / unreachable proc\n-## [profitbricks-build5-amd64:41541] Set MCA parameter \"orte_base_help_aggregate\" to 0 to see all help / error messages\n-## [profitbricks-build5-amd64:41541] 2 more processes have sent help message help-mpi-runtime.txt / mpi_init:startup:internal-failure\n-## [profitbricks-build5-amd64:41541] 2 more processes have sent help message help-mpi-errors.txt / mpi_errors_are_fatal unknown handle\n-## [warn] Epoll MOD(1) on fd 31 failed. Old events were 6; read change was 0 (none); write change was 2 (del); close change was 0 (none): Bad file descriptor\n+## [i-capture-the-hostname:56698] [[20933,2],2] ORTE_ERROR_LOG: Unreachable in file dpm/dpm.c at line 433\n+## [i-capture-the-hostname:56695] [[20933,2],0] ORTE_ERROR_LOG: Unreachable in file dpm/dpm.c at line 433\n+## [i-capture-the-hostname:56697] [[20933,2],1] ORTE_ERROR_LOG: Unreachable in file dpm/dpm.c at line 433\n+## [i-capture-the-hostname:56468] 4 more processes have sent help message help-mca-bml-r2.txt / unreachable proc\n+## [i-capture-the-hostname:56468] Set MCA parameter \"orte_base_help_aggregate\" to 0 to see all help / error messages\n+## [i-capture-the-hostname:56468] 1 more process has sent help message help-mpi-errors.txt / mpi_errors_are_fatal\n+## [i-capture-the-hostname:56468] 2 more processes have sent help message help-mpi-runtime.txt / mpi_init:startup:internal-failure\n not ok 532 - ./spawn/concurrent_spawns 1\n ---\n Directory: ./spawn\n File: concurrent_spawns\n Num-procs: 1\n- Date: \"Sun Jan 5 01:15:52 2020\"\n+ Date: \"Mon Dec 3 16:40:22 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## --------------------------------------------------------------------------\n ## At least one pair of MPI processes are unable to reach each other for\n ## MPI communications. This means that no Open MPI device has indicated\n ## that it can be used to communicate between these processes. This is\n ## an error; Open MPI requires that all MPI processes be able to reach\n ## each other. This error can sometimes be the result of forgetting to\n ## specify the \"self\" BTL.\n ## \n-## Process 1 ([[24801,1],0]) is on host: profitbricks-build5-amd64\n-## Process 2 ([[24801,2],0]) is on host: profitbricks-build5-amd64\n+## Process 1 ([[21366,2],0]) is on host: i-capture-the-hostname\n+## Process 2 ([[21366,1],0]) is on host: i-capture-the-hostname\n ## BTLs attempted: self\n ## \n ## Your MPI job is now going to abort; sorry.\n ## --------------------------------------------------------------------------\n-## [profitbricks-build5-amd64:42156] [[24801,1],0] ORTE_ERROR_LOG: Unreachable in file dpm/dpm.c at line 433\n-## [profitbricks-build5-amd64:42156] *** An error occurred in MPI_Comm_spawn\n-## [profitbricks-build5-amd64:42156] *** reported by process [1625358337,0]\n-## [profitbricks-build5-amd64:42156] *** on communicator MPI_COMM_WORLD\n-## [profitbricks-build5-amd64:42156] *** MPI_ERR_INTERN: internal error\n-## [profitbricks-build5-amd64:42156] *** MPI_ERRORS_ARE_FATAL (processes in this communicator will now abort,\n-## [profitbricks-build5-amd64:42156] *** and potentially your MPI job)\n-## [profitbricks-build5-amd64:42035] 1 more process has sent help message help-mca-bml-r2.txt / unreachable proc\n-## [profitbricks-build5-amd64:42035] Set MCA parameter \"orte_base_help_aggregate\" to 0 to see all help / error messages\n+## [i-capture-the-hostname:57123] [[21366,2],0] ORTE_ERROR_LOG: Unreachable in file dpm/dpm.c at line 433\n+## --------------------------------------------------------------------------\n+## It looks like MPI_INIT failed for some reason; your parallel process is\n+## likely to abort. There are many reasons that a parallel process can\n+## fail during MPI_INIT; some of which are due to configuration or environment\n+## problems. This failure appears to be an internal failure; here's some\n+## additional information (which may only be relevant to an Open MPI\n+## developer):\n+## \n+## ompi_dpm_dyn_init() failed\n+## --> Returned \"Unreachable\" (-12) instead of \"Success\" (0)\n+## --------------------------------------------------------------------------\n+## [i-capture-the-hostname:56968] [[21366,1],0] ORTE_ERROR_LOG: Unreachable in file dpm/dpm.c at line 433\n+## [i-capture-the-hostname:56968] *** An error occurred in MPI_Comm_spawn\n+## [i-capture-the-hostname:56968] *** reported by process [1400242177,0]\n+## [i-capture-the-hostname:56968] *** on communicator MPI_COMM_WORLD\n+## [i-capture-the-hostname:56968] *** MPI_ERR_INTERN: internal error\n+## [i-capture-the-hostname:56968] *** MPI_ERRORS_ARE_FATAL (processes in this communicator will now abort,\n+## [i-capture-the-hostname:56968] *** and potentially your MPI job)\n+## [i-capture-the-hostname:57123] *** An error occurred in MPI_Init_thread\n+## [i-capture-the-hostname:57123] *** reported by process [1400242178,0]\n+## [i-capture-the-hostname:57123] *** on a NULL communicator\n+## [i-capture-the-hostname:57123] *** Unknown error\n+## [i-capture-the-hostname:57123] *** MPI_ERRORS_ARE_FATAL (processes in this communicator will now abort,\n+## [i-capture-the-hostname:57123] *** and potentially your MPI job)\n+## [i-capture-the-hostname:56871] 1 more process has sent help message help-mca-bml-r2.txt / unreachable proc\n+## [i-capture-the-hostname:56871] Set MCA parameter \"orte_base_help_aggregate\" to 0 to see all help / error messages\n ok 533 - ./spawn/pgroup_connect_test 2\n ok 534 - ./spawn/pgroup_intercomm_test 2\n not ok 535 - ./spawn/spawn-rootargs 2\n ---\n Directory: ./spawn\n File: spawn-rootargs\n Num-procs: 2\n- Date: \"Sun Jan 5 01:15:56 2020\"\n+ Date: \"Mon Dec 3 16:40:28 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## --------------------------------------------------------------------------\n ## mpiexec.openmpi was unable to launch the specified application as it could not access\n ## or execute an executable:\n ## \n ## Executable: ./spawn-nullargs\n-## Node: profitbricks-build5-amd64\n+## Node: i-capture-the-hostname\n ## \n ## while attempting to start process rank 0.\n ## --------------------------------------------------------------------------\n ## 5 total processes failed to start\n not ok 536 - ./topo/cartmap1 2\n ---\n Directory: ./topo\n File: cartmap1\n Num-procs: 2\n- Date: \"Sun Jan 5 01:15:58 2020\"\n+ Date: \"Mon Dec 3 16:40:30 2018\"\n ...\n ## Test output (expected 'No Errors'):\n-## Found 2 errors\n ## rank outside of input communicator not UNDEFINED\n ## rank outside of input communicator not UNDEFINED\n+## Found 2 errors\n ok 537 - ./topo/cartzero 2\n ok 538 - ./topo/cartshift1 2\n not ok 539 - ./topo/cartsuball 2\n ---\n Directory: ./topo\n File: cartsuball\n Num-procs: 2\n- Date: \"Sun Jan 5 01:16:02 2020\"\n+ Date: \"Mon Dec 3 16:40:35 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Found 1 errors\n ## cart sub to size 0 did not give null\n ok 540 - ./topo/cartcreates 2\n ok 541 - ./topo/dims1 2\n ok 542 - ./topo/dims2 1\n ok 543 - ./topo/dims3 1\n ok 544 - ./topo/dims4 1\n not ok 545 - ./topo/graphmap1 2\n ---\n Directory: ./topo\n File: graphmap1\n Num-procs: 2\n- Date: \"Sun Jan 5 01:16:08 2020\"\n+ Date: \"Mon Dec 3 16:40:45 2018\"\n ...\n ## Test output (expected 'No Errors'):\n-## Found 1 errors\n ## Graph map with no local nodes did not return MPI_UNDEFINED\n+## Found 1 errors\n ok 546 - ./topo/topotest 2\n ok 547 - ./topo/topodup 2\n ok 548 - ./topo/graphcr 2\n ok 549 - ./topo/graphcr2 2\n not ok 550 - ./topo/distgraph1 2\n ---\n Directory: ./topo\n File: distgraph1\n Num-procs: 2\n- Date: \"Sun Jan 5 01:16:14 2020\"\n+ Date: \"Mon Dec 3 16:40:55 2018\"\n ...\n ## Test output (expected 'No Errors'):\n-## Error: Cannot partition 16 elements in 17 parts\n-## [profitbricks-build5-amd64:52290] *** Process received signal ***\n-## [profitbricks-build5-amd64:52290] Signal: Segmentation fault (11)\n-## [profitbricks-build5-amd64:52290] Signal code: Address not mapped (1)\n-## [profitbricks-build5-amd64:52290] Failing at address: (nil)\n-## [profitbricks-build5-amd64:52290] [ 0] /lib/x86_64-linux-gnu/libpthread.so.0(+0x126b0)[0x7fa3a945a6b0]\n-## [profitbricks-build5-amd64:52290] [ 1] /usr/lib/x86_64-linux-gnu/openmpi/lib/openmpi3/mca_topo_treematch.so(split_com_mat+0xab)[0x7fa3a5d6de8b]\n-## [profitbricks-build5-amd64:52290] [ 2] /usr/lib/x86_64-linux-gnu/openmpi/lib/openmpi3/mca_topo_treematch.so(kpartition_build_level_topology+0xa3)[0x7fa3a5d6e423]\n-## [profitbricks-build5-amd64:52290] [ 3] /usr/lib/x86_64-linux-gnu/openmpi/lib/openmpi3/mca_topo_treematch.so(kpartition_build_tree_from_topology+0x172)[0x7fa3a5d6e7c2]\n-## [profitbricks-build5-amd64:52290] [ 4] /usr/lib/x86_64-linux-gnu/openmpi/lib/openmpi3/mca_topo_treematch.so(tm_build_tree_from_topology+0xbd)[0x7fa3a5d6b28d]\n-## [profitbricks-build5-amd64:52290] [ 5] /usr/lib/x86_64-linux-gnu/openmpi/lib/openmpi3/mca_topo_treematch.so(mca_topo_treematch_dist_graph_create+0x2959)[0x7fa3a5d619e9]\n-## [profitbricks-build5-amd64:52290] [ 6] /usr/lib/x86_64-linux-gnu/libmpi.so.40(MPI_Dist_graph_create+0x21d)[0x7fa3a94d93cd]\n-## [profitbricks-build5-amd64:52290] [ 7] ./distgraph1(+0x39e7)[0x556bf805a9e7]\n-## [profitbricks-build5-amd64:52290] [ 8] /lib/x86_64-linux-gnu/libc.so.6(__libc_start_main+0xeb)[0x7fa3a92a409b]\n-## [profitbricks-build5-amd64:52290] [ 9] ./distgraph1(+0x497a)[0x556bf805b97a]\n-## [profitbricks-build5-amd64:52290] *** End of error message ***\n+## Error: Cannot partition 15 elements in 2 parts\n+## [i-capture-the-hostname:60427] *** Process received signal ***\n+## [i-capture-the-hostname:60427] Signal: Segmentation fault (11)\n+## [i-capture-the-hostname:60427] Signal code: Address not mapped (1)\n+## [i-capture-the-hostname:60427] Failing at address: (nil)\n+## [i-capture-the-hostname:60427] [ 0] /lib/x86_64-linux-gnu/libpthread.so.0(+0x126b0)[0x7fd3796d56b0]\n+## [i-capture-the-hostname:60427] [ 1] /usr/lib/x86_64-linux-gnu/openmpi/lib/openmpi3/mca_topo_treematch.so(split_com_mat+0xab)[0x7fd375fe8e8b]\n+## [i-capture-the-hostname:60427] [ 2] /usr/lib/x86_64-linux-gnu/openmpi/lib/openmpi3/mca_topo_treematch.so(kpartition_build_level_topology+0xa3)[0x7fd375fe9423]\n+## [i-capture-the-hostname:60427] [ 3] /usr/lib/x86_64-linux-gnu/openmpi/lib/openmpi3/mca_topo_treematch.so(kpartition_build_tree_from_topology+0x172)[0x7fd375fe97c2]\n+## [i-capture-the-hostname:60427] [ 4] /usr/lib/x86_64-linux-gnu/openmpi/lib/openmpi3/mca_topo_treematch.so(tm_build_tree_from_topology+0xbd)[0x7fd375fe628d]\n+## [i-capture-the-hostname:60427] [ 5] /usr/lib/x86_64-linux-gnu/openmpi/lib/openmpi3/mca_topo_treematch.so(mca_topo_treematch_dist_graph_create+0x2959)[0x7fd375fdc9e9]\n+## [i-capture-the-hostname:60427] [ 6] /usr/lib/x86_64-linux-gnu/libmpi.so.40(MPI_Dist_graph_create+0x21d)[0x7fd3797543cd]\n+## [i-capture-the-hostname:60427] [ 7] ./distgraph1(+0x39e7)[0x55bccf6859e7]\n+## [i-capture-the-hostname:60427] [ 8] /lib/x86_64-linux-gnu/libc.so.6(__libc_start_main+0xeb)[0x7fd37951f09b]\n+## [i-capture-the-hostname:60427] [ 9] ./distgraph1(+0x497a)[0x55bccf68697a]\n+## [i-capture-the-hostname:60427] *** End of error message ***\n ## --------------------------------------------------------------------------\n ## Primary job terminated normally, but 1 process returned\n ## a non-zero exit code. Per user-direction, the job has been aborted.\n ## --------------------------------------------------------------------------\n ## --------------------------------------------------------------------------\n-## mpiexec.openmpi noticed that process rank 0 with PID 0 on node profitbricks-build5-amd64 exited on signal 11 (Segmentation fault).\n+## mpiexec.openmpi noticed that process rank 0 with PID 0 on node i-capture-the-hostname exited on signal 11 (Segmentation fault).\n ## --------------------------------------------------------------------------\n not ok 551 - ./topo/dgraph_unwgt 2\n ---\n Directory: ./topo\n File: dgraph_unwgt\n Num-procs: 2\n- Date: \"Sun Jan 5 01:16:17 2020\"\n+ Date: \"Mon Dec 3 16:40:58 2018\"\n ...\n ## Test output (expected 'No Errors'):\n-## Error: Cannot partition 16 elements in 17 parts\n-## [profitbricks-build5-amd64:52743] *** Process received signal ***\n-## [profitbricks-build5-amd64:52743] Signal: Segmentation fault (11)\n-## [profitbricks-build5-amd64:52743] Signal code: Address not mapped (1)\n-## [profitbricks-build5-amd64:52743] Failing at address: (nil)\n-## [profitbricks-build5-amd64:52743] [ 0] /lib/x86_64-linux-gnu/libpthread.so.0(+0x126b0)[0x7f30493126b0]\n-## [profitbricks-build5-amd64:52743] [ 1] /usr/lib/x86_64-linux-gnu/openmpi/lib/openmpi3/mca_topo_treematch.so(split_com_mat+0xab)[0x7f3045c25e8b]\n-## [profitbricks-build5-amd64:52743] [ 2] /usr/lib/x86_64-linux-gnu/openmpi/lib/openmpi3/mca_topo_treematch.so(kpartition_build_level_topology+0xa3)[0x7f3045c26423]\n-## [profitbricks-build5-amd64:52743] [ 3] /usr/lib/x86_64-linux-gnu/openmpi/lib/openmpi3/mca_topo_treematch.so(kpartition_build_tree_from_topology+0x172)[0x7f3045c267c2]\n-## [profitbricks-build5-amd64:52743] [ 4] /usr/lib/x86_64-linux-gnu/openmpi/lib/openmpi3/mca_topo_treematch.so(tm_build_tree_from_topology+0xbd)[0x7f3045c2328d]\n-## [profitbricks-build5-amd64:52743] [ 5] /usr/lib/x86_64-linux-gnu/openmpi/lib/openmpi3/mca_topo_treematch.so(mca_topo_treematch_dist_graph_create+0x2959)[0x7f3045c199e9]\n-## [profitbricks-build5-amd64:52743] [ 6] /usr/lib/x86_64-linux-gnu/libmpi.so.40(MPI_Dist_graph_create+0x21d)[0x7f30493913cd]\n-## [profitbricks-build5-amd64:52743] [ 7] ./dgraph_unwgt(+0x25e7)[0x556c34e935e7]\n-## [profitbricks-build5-amd64:52743] [ 8] /lib/x86_64-linux-gnu/libc.so.6(__libc_start_main+0xeb)[0x7f304915c09b]\n-## [profitbricks-build5-amd64:52743] [ 9] ./dgraph_unwgt(+0x278a)[0x556c34e9378a]\n-## [profitbricks-build5-amd64:52743] *** End of error message ***\n+## Error: Cannot partition 15 elements in 2 parts\n+## [i-capture-the-hostname:60703] *** Process received signal ***\n+## [i-capture-the-hostname:60703] Signal: Segmentation fault (11)\n+## [i-capture-the-hostname:60703] Signal code: Address not mapped (1)\n+## [i-capture-the-hostname:60703] Failing at address: (nil)\n+## [i-capture-the-hostname:60703] [ 0] /lib/x86_64-linux-gnu/libpthread.so.0(+0x126b0)[0x7f1d29a356b0]\n+## [i-capture-the-hostname:60703] [ 1] /usr/lib/x86_64-linux-gnu/openmpi/lib/openmpi3/mca_topo_treematch.so(split_com_mat+0xab)[0x7f1d26348e8b]\n+## [i-capture-the-hostname:60703] [ 2] /usr/lib/x86_64-linux-gnu/openmpi/lib/openmpi3/mca_topo_treematch.so(kpartition_build_level_topology+0xa3)[0x7f1d26349423]\n+## [i-capture-the-hostname:60703] [ 3] /usr/lib/x86_64-linux-gnu/openmpi/lib/openmpi3/mca_topo_treematch.so(kpartition_build_tree_from_topology+0x172)[0x7f1d263497c2]\n+## [i-capture-the-hostname:60703] [ 4] /usr/lib/x86_64-linux-gnu/openmpi/lib/openmpi3/mca_topo_treematch.so(tm_build_tree_from_topology+0xbd)[0x7f1d2634628d]\n+## [i-capture-the-hostname:60703] [ 5] /usr/lib/x86_64-linux-gnu/openmpi/lib/openmpi3/mca_topo_treematch.so(mca_topo_treematch_dist_graph_create+0x2959)[0x7f1d2633c9e9]\n+## [i-capture-the-hostname:60703] [ 6] /usr/lib/x86_64-linux-gnu/libmpi.so.40(MPI_Dist_graph_create+0x21d)[0x7f1d29ab43cd]\n+## [i-capture-the-hostname:60703] [ 7] ./dgraph_unwgt(+0x25e7)[0x55b68d46f5e7]\n+## [i-capture-the-hostname:60703] [ 8] /lib/x86_64-linux-gnu/libc.so.6(__libc_start_main+0xeb)[0x7f1d2987f09b]\n+## [i-capture-the-hostname:60703] [ 9] ./dgraph_unwgt(+0x278a)[0x55b68d46f78a]\n+## [i-capture-the-hostname:60703] *** End of error message ***\n ## --------------------------------------------------------------------------\n ## Primary job terminated normally, but 1 process returned\n ## a non-zero exit code. Per user-direction, the job has been aborted.\n ## --------------------------------------------------------------------------\n ## --------------------------------------------------------------------------\n-## mpiexec.openmpi noticed that process rank 0 with PID 0 on node profitbricks-build5-amd64 exited on signal 11 (Segmentation fault).\n+## mpiexec.openmpi noticed that process rank 0 with PID 0 on node i-capture-the-hostname exited on signal 11 (Segmentation fault).\n ## --------------------------------------------------------------------------\n ok 552 - ./topo/neighb_coll 2\n ok 553 - ./io/rdwrord 2\n ok 554 - ./io/rdwrzero 2\n ok 555 - ./io/getextent 2\n ok 556 - ./io/setinfo 2\n not ok 557 - ./io/setviewcur 2\n ---\n Directory: ./io\n File: setviewcur\n Num-procs: 2\n- Date: \"Sun Jan 5 01:16:24 2020\"\n+ Date: \"Mon Dec 3 16:41:08 2018\"\n ...\n ## Test output (expected 'No Errors'):\n+## Found 2 errors\n ## mca_fbtl_posix_preadv: error in (p)readv:Invalid argument\n ## 0: buf[0] = -1\n ## mca_fbtl_posix_preadv: error in (p)readv:Invalid argument\n ## 1: buf[0] = -1\n-## Found 2 errors\n ok 558 - ./io/i_noncontig 2\n ok 559 - ./io/async 2\n ok 560 - ./io/async_any 2\n not ok 561 - ./io/userioerr 1\n ---\n Directory: ./io\n File: userioerr\n Num-procs: 1\n- Date: \"Sun Jan 5 01:16:30 2020\"\n+ Date: \"Mon Dec 3 16:41:16 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Improper use of FILE Mode, Using WRONLY for Read!\n ## No Errors\n ok 562 - ./io/resized 1\n ok 563 - ./io/resized2 1\n ok 564 - ./io/hindexed_io 1\n@@ -998786,63 +998788,63 @@\n ok 588 - ./f77/datatype/typenamef 1\n ok 589 - ./f77/datatype/typename3f 1\n not ok 590 - ./f77/datatype/typesnamef 1\n ---\n Directory: ./f77/datatype\n File: typesnamef\n Num-procs: 1\n- Date: \"Sun Jan 5 01:17:04 2020\"\n+ Date: \"Mon Dec 3 16:42:03 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## (type2) Expected length 0, got 17\n ## (type2) Datatype name is not all blank\n ## Found 2 errors\n not ok 591 - ./f77/datatype/typecntsf 1\n ---\n Directory: ./f77/datatype\n File: typecntsf\n Num-procs: 1\n- Date: \"Sun Jan 5 01:17:07 2020\"\n+ Date: \"Mon Dec 3 16:42:07 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## \n ## Program received signal SIGSEGV: Segmentation fault - invalid memory reference.\n ## \n ## Backtrace for this error:\n-## #0 0x7fb55383d89d in ???\n-## #1 0x7fb55383cae3 in ???\n-## #2 0x7fb5534778df in ???\n-## #3 0x7fb55357c63e in ???\n-## #4 0x7fb552cf0415 in ???\n-## #5 0x7fb553aeb05e in ???\n-## #6 0x7fb553b1cb26 in ???\n-## #7 0x7fb553bed6d3 in ???\n-## #8 0x562440b7f53b in explore_\n+## #0 0x7f5f53d4089d in ???\n+## #1 0x7f5f53d3fae3 in ???\n+## #2 0x7f5f5397a8df in ???\n+## #3 0x7f5f53a7ce90 in ???\n+## #4 0x7f5f531f3415 in ???\n+## #5 0x7f5f53fee05e in ???\n+## #6 0x7f5f5401fb26 in ???\n+## #7 0x7f5f540f06d3 in ???\n+## #8 0x55610e8de53b in explore_\n ## \tat ../../../f77/datatype/typecntsf.f:52\n-## #9 0x562440b7f27c in MAIN__\n+## #9 0x55610e8de27c in MAIN__\n ## \tat ../../../f77/datatype/typecntsf.f:24\n-## #10 0x562440b7f27c in main\n+## #10 0x55610e8de27c in main\n ## \tat ../../../f77/datatype/typecntsf.f:33\n ## --------------------------------------------------------------------------\n ## Primary job terminated normally, but 1 process returned\n ## a non-zero exit code. Per user-direction, the job has been aborted.\n ## --------------------------------------------------------------------------\n ## --------------------------------------------------------------------------\n-## mpiexec.openmpi noticed that process rank 0 with PID 0 on node profitbricks-build5-amd64 exited on signal 11 (Segmentation fault).\n+## mpiexec.openmpi noticed that process rank 0 with PID 0 on node i-capture-the-hostname exited on signal 11 (Segmentation fault).\n ## --------------------------------------------------------------------------\n ok 592 - ./f77/datatype/typem2f 1\n ok 593 - ./f77/datatype/typesubf 1\n ok 594 - ./f77/datatype/packef 1\n ok 595 - ./f77/datatype/gaddressf 1\n not ok 596 - ./f77/datatype/allctypesf 1\n ---\n Directory: ./f77/datatype\n File: allctypesf\n Num-procs: 1\n- Date: \"Sun Jan 5 01:17:13 2020\"\n+ Date: \"Mon Dec 3 16:42:16 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## For datatype MPI_C_FLOAT_COMPLEX found name MPI_C_COMPLEX\n ## Found 1 errors\n ok 597 - ./f77/datatype/hindex1f 1\n ok 598 - ./f77/datatype/hindexed_blockf 1\n ok 599 - ./f77/datatype/bottom 2\n@@ -998850,25 +998852,26 @@\n ok 601 - ./f77/pt2pt/greqf 1\n ok 602 - ./f77/pt2pt/allpairf 2\n not ok 603 - ./f77/pt2pt/mprobef 2\n ---\n Directory: ./f77/pt2pt\n File: mprobef\n Num-procs: 2\n- Date: \"Sun Jan 5 01:17:22 2020\"\n+ Date: \"Mon Dec 3 16:42:28 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## s1(MPI_ERROR) != MPI_ERR_DIMS at T4 Mprobe().\n ## s2(MPI_ERROR) != MPI_ERR_OTHER at T4 Mrecv().\n ## s1(MPI_ERROR) != MPI_ERR_DIMS at T5 Mprobe().\n ## s2(MPI_ERROR) != MPI_ERR_OTHER at T5 Imrecv().\n ## s1(MPI_ERROR) != MPI_ERR_DIMS at T6 Improbe().\n ## s2(MPI_ERROR) != MPI_ERR_OTHER at T6 Mrecv().\n ## s1(MPI_ERROR) != MPI_ERR_DIMS at T7 Improbe().\n ## s2(MPI_ERROR) != MPI_ERR_OTHER at T7 Imrecv().\n+## Found 24 errors\n ## s1(MPI_ERROR) != MPI_ERR_DIMS at T0 Mprobe().\n ## s2(MPI_ERROR) != MPI_ERR_OTHER at T0 Mrecv().\n ## s1(MPI_ERROR) != MPI_ERR_DIMS at T1 Mprobe().\n ## s2(MPI_ERROR) != MPI_ERR_OTHER at T1 Imrecv().\n ## s1(MPI_ERROR) != MPI_ERR_DIMS at T2 Improbe().\n ## s2(MPI_ERROR) != MPI_ERR_OTHER at T2 Mrecv().\n ## s1(MPI_ERROR) != MPI_ERR_DIMS at T3 Improbe().\n@@ -998877,237 +998880,253 @@\n ## s2(MPI_ERROR) != MPI_ERR_OTHER at T4 Mrecv().\n ## s1(MPI_ERROR) != MPI_ERR_DIMS at T5 Mprobe().\n ## s2(MPI_ERROR) != MPI_ERR_OTHER at T5 Imrecv().\n ## s1(MPI_ERROR) != MPI_ERR_DIMS at T6 Improbe().\n ## s2(MPI_ERROR) != MPI_ERR_OTHER at T6 Mrecv().\n ## s1(MPI_ERROR) != MPI_ERR_DIMS at T7 Improbe().\n ## s2(MPI_ERROR) != MPI_ERR_OTHER at T7 Imrecv().\n-## Found 24 errors\n ok 604 - ./f77/info/infotestf 1\n ok 605 - ./f77/info/infotest2f 1\n ok 606 - ./f77/spawn/namepubf 2\n not ok 607 - ./f77/spawn/spawnf 1\n ---\n Directory: ./f77/spawn\n File: spawnf\n Num-procs: 1\n- Date: \"Sun Jan 5 01:17:27 2020\"\n+ Date: \"Mon Dec 3 16:42:36 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## --------------------------------------------------------------------------\n ## At least one pair of MPI processes are unable to reach each other for\n ## MPI communications. This means that no Open MPI device has indicated\n ## that it can be used to communicate between these processes. This is\n ## an error; Open MPI requires that all MPI processes be able to reach\n ## each other. This error can sometimes be the result of forgetting to\n ## specify the \"self\" BTL.\n ## \n-## Process 1 ([[55085,1],0]) is on host: profitbricks-build5-amd64\n-## Process 2 ([[55085,2],0]) is on host: profitbricks-build5-amd64\n+## Process 1 ([[39012,1],0]) is on host: i-capture-the-hostname\n+## Process 2 ([[39012,2],0]) is on host: i-capture-the-hostname\n ## BTLs attempted: self\n ## \n ## Your MPI job is now going to abort; sorry.\n ## --------------------------------------------------------------------------\n-## [profitbricks-build5-amd64:05143] [[55085,1],0] ORTE_ERROR_LOG: Unreachable in file dpm/dpm.c at line 433\n-## [profitbricks-build5-amd64:05162] [[55085,2],0] ORTE_ERROR_LOG: Unreachable in file dpm/dpm.c at line 433\n-## [profitbricks-build5-amd64:05163] [[55085,2],1] ORTE_ERROR_LOG: Unreachable in file dpm/dpm.c at line 433\n+## [i-capture-the-hostname:05460] [[39012,1],0] ORTE_ERROR_LOG: Unreachable in file dpm/dpm.c at line 433\n+## [i-capture-the-hostname:05501] [[39012,2],0] ORTE_ERROR_LOG: Unreachable in file dpm/dpm.c at line 433\n+## [i-capture-the-hostname:05503] [[39012,2],1] ORTE_ERROR_LOG: Unreachable in file dpm/dpm.c at line 433\n ## --------------------------------------------------------------------------\n ## It looks like MPI_INIT failed for some reason; your parallel process is\n ## likely to abort. There are many reasons that a parallel process can\n ## fail during MPI_INIT; some of which are due to configuration or environment\n ## problems. This failure appears to be an internal failure; here's some\n ## additional information (which may only be relevant to an Open MPI\n ## developer):\n ## \n ## ompi_dpm_dyn_init() failed\n ## --> Returned \"Unreachable\" (-12) instead of \"Success\" (0)\n ## --------------------------------------------------------------------------\n-## [profitbricks-build5-amd64:05143] *** An error occurred in MPI_Comm_spawn\n-## [profitbricks-build5-amd64:05143] *** reported by process [3610050561,0]\n-## [profitbricks-build5-amd64:05143] *** on communicator MPI_COMM_WORLD\n-## [profitbricks-build5-amd64:05143] *** MPI_ERR_INTERN: internal error\n-## [profitbricks-build5-amd64:05143] *** MPI_ERRORS_ARE_FATAL (processes in this communicator will now abort,\n-## [profitbricks-build5-amd64:05143] *** and potentially your MPI job)\n-## [profitbricks-build5-amd64:05162] *** An error occurred in MPI_Init\n-## [profitbricks-build5-amd64:05162] *** reported by process [3610050562,0]\n-## [profitbricks-build5-amd64:05162] *** on a NULL communicator\n-## [profitbricks-build5-amd64:05162] *** Unknown error\n-## [profitbricks-build5-amd64:05162] *** MPI_ERRORS_ARE_FATAL (processes in this communicator will now abort,\n-## [profitbricks-build5-amd64:05162] *** and potentially your MPI job)\n-## [profitbricks-build5-amd64:05119] 2 more processes have sent help message help-mca-bml-r2.txt / unreachable proc\n-## [profitbricks-build5-amd64:05119] Set MCA parameter \"orte_base_help_aggregate\" to 0 to see all help / error messages\n-## [profitbricks-build5-amd64:05119] 1 more process has sent help message help-mpi-runtime.txt / mpi_init:startup:internal-failure\n-## [profitbricks-build5-amd64:05119] 1 more process has sent help message help-mpi-errors.txt / mpi_errors_are_fatal unknown handle\n+## [i-capture-the-hostname:05460] *** An error occurred in MPI_Comm_spawn\n+## [i-capture-the-hostname:05460] *** reported by process [2556690433,0]\n+## [i-capture-the-hostname:05460] *** on communicator MPI_COMM_WORLD\n+## [i-capture-the-hostname:05460] *** MPI_ERR_INTERN: internal error\n+## [i-capture-the-hostname:05460] *** MPI_ERRORS_ARE_FATAL (processes in this communicator will now abort,\n+## [i-capture-the-hostname:05460] *** and potentially your MPI job)\n+## [i-capture-the-hostname:05501] *** An error occurred in MPI_Init\n+## [i-capture-the-hostname:05501] *** reported by process [2556690434,0]\n+## [i-capture-the-hostname:05501] *** on a NULL communicator\n+## [i-capture-the-hostname:05501] *** Unknown error\n+## [i-capture-the-hostname:05501] *** MPI_ERRORS_ARE_FATAL (processes in this communicator will now abort,\n+## [i-capture-the-hostname:05501] *** and potentially your MPI job)\n+## [i-capture-the-hostname:05429] 2 more processes have sent help message help-mca-bml-r2.txt / unreachable proc\n+## [i-capture-the-hostname:05429] Set MCA parameter \"orte_base_help_aggregate\" to 0 to see all help / error messages\n+## [i-capture-the-hostname:05429] 1 more process has sent help message help-mpi-runtime.txt / mpi_init:startup:internal-failure\n+## [i-capture-the-hostname:05429] 1 more process has sent help message help-mpi-errors.txt / mpi_errors_are_fatal unknown handle\n not ok 608 - ./f77/spawn/spawnargvf 1\n ---\n Directory: ./f77/spawn\n File: spawnargvf\n Num-procs: 1\n- Date: \"Sun Jan 5 01:17:29 2020\"\n+ Date: \"Mon Dec 3 16:42:38 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## --------------------------------------------------------------------------\n ## At least one pair of MPI processes are unable to reach each other for\n ## MPI communications. This means that no Open MPI device has indicated\n ## that it can be used to communicate between these processes. This is\n ## an error; Open MPI requires that all MPI processes be able to reach\n ## each other. This error can sometimes be the result of forgetting to\n ## specify the \"self\" BTL.\n ## \n-## Process 1 ([[53436,2],1]) is on host: profitbricks-build5-amd64\n-## Process 2 ([[53436,1],0]) is on host: profitbricks-build5-amd64\n-## BTLs attempted: vader self\n+## Process 1 ([[39149,1],0]) is on host: i-capture-the-hostname\n+## Process 2 ([[39149,2],0]) is on host: i-capture-the-hostname\n+## BTLs attempted: self\n ## \n ## Your MPI job is now going to abort; sorry.\n ## --------------------------------------------------------------------------\n+## [i-capture-the-hostname:05592] [[39149,1],0] ORTE_ERROR_LOG: Unreachable in file dpm/dpm.c at line 433\n+## [i-capture-the-hostname:05592] *** An error occurred in MPI_Comm_spawn\n+## [i-capture-the-hostname:05592] *** reported by process [2565668865,0]\n+## [i-capture-the-hostname:05592] *** on communicator MPI_COMM_WORLD\n+## [i-capture-the-hostname:05592] *** MPI_ERR_INTERN: internal error\n+## [i-capture-the-hostname:05592] *** MPI_ERRORS_ARE_FATAL (processes in this communicator will now abort,\n+## [i-capture-the-hostname:05592] *** and potentially your MPI job)\n+## [i-capture-the-hostname:05635] [[39149,2],1] ORTE_ERROR_LOG: Unreachable in file dpm/dpm.c at line 433\n+## [i-capture-the-hostname:05632] [[39149,2],0] ORTE_ERROR_LOG: Unreachable in file dpm/dpm.c at line 433\n ## --------------------------------------------------------------------------\n ## It looks like MPI_INIT failed for some reason; your parallel process is\n ## likely to abort. There are many reasons that a parallel process can\n ## fail during MPI_INIT; some of which are due to configuration or environment\n ## problems. This failure appears to be an internal failure; here's some\n ## additional information (which may only be relevant to an Open MPI\n ## developer):\n ## \n ## ompi_dpm_dyn_init() failed\n ## --> Returned \"Unreachable\" (-12) instead of \"Success\" (0)\n ## --------------------------------------------------------------------------\n-## [profitbricks-build5-amd64:05418] [[53436,2],1] ORTE_ERROR_LOG: Unreachable in file dpm/dpm.c at line 433\n-## [profitbricks-build5-amd64:05418] *** An error occurred in MPI_Init\n-## [profitbricks-build5-amd64:05418] *** reported by process [3501981698,1]\n-## [profitbricks-build5-amd64:05418] *** on a NULL communicator\n-## [profitbricks-build5-amd64:05418] *** Unknown error\n-## [profitbricks-build5-amd64:05418] *** MPI_ERRORS_ARE_FATAL (processes in this communicator will now abort,\n-## [profitbricks-build5-amd64:05418] *** and potentially your MPI job)\n-## [profitbricks-build5-amd64:05258] *** An error occurred in MPI_Comm_spawn\n-## [profitbricks-build5-amd64:05258] *** reported by process [3501981697,0]\n-## [profitbricks-build5-amd64:05258] *** on communicator MPI_COMM_WORLD\n-## [profitbricks-build5-amd64:05258] *** MPI_ERR_INTERN: internal error\n-## [profitbricks-build5-amd64:05258] *** MPI_ERRORS_ARE_FATAL (processes in this communicator will now abort,\n-## [profitbricks-build5-amd64:05258] *** and potentially your MPI job)\n-## [profitbricks-build5-amd64:05258] [[53436,1],0] ORTE_ERROR_LOG: Unreachable in file dpm/dpm.c at line 433\n-## [profitbricks-build5-amd64:05414] [[53436,2],0] ORTE_ERROR_LOG: Unreachable in file dpm/dpm.c at line 433\n-## [profitbricks-build5-amd64:05230] 2 more processes have sent help message help-mca-bml-r2.txt / unreachable proc\n-## [profitbricks-build5-amd64:05230] Set MCA parameter \"orte_base_help_aggregate\" to 0 to see all help / error messages\n-## [profitbricks-build5-amd64:05230] 1 more process has sent help message help-mpi-runtime.txt / mpi_init:startup:internal-failure\n+## [i-capture-the-hostname:05632] *** An error occurred in MPI_Init\n+## [i-capture-the-hostname:05632] *** reported by process [2565668866,0]\n+## [i-capture-the-hostname:05632] *** on a NULL communicator\n+## [i-capture-the-hostname:05632] *** Unknown error\n+## [i-capture-the-hostname:05632] *** MPI_ERRORS_ARE_FATAL (processes in this communicator will now abort,\n+## [i-capture-the-hostname:05632] *** and potentially your MPI job)\n+## [i-capture-the-hostname:05564] 2 more processes have sent help message help-mca-bml-r2.txt / unreachable proc\n+## [i-capture-the-hostname:05564] Set MCA parameter \"orte_base_help_aggregate\" to 0 to see all help / error messages\n+## [i-capture-the-hostname:05564] 1 more process has sent help message help-mpi-runtime.txt / mpi_init:startup:internal-failure\n+## [i-capture-the-hostname:05564] 1 more process has sent help message help-mpi-errors.txt / mpi_errors_are_fatal unknown handle\n ok 609 - ./f77/spawn/connaccf 2\n not ok 610 - ./f77/spawn/spawnmultf 1\n ---\n Directory: ./f77/spawn\n File: spawnmultf\n Num-procs: 1\n- Date: \"Sun Jan 5 01:17:32 2020\"\n+ Date: \"Mon Dec 3 16:42:42 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## --------------------------------------------------------------------------\n ## At least one pair of MPI processes are unable to reach each other for\n ## MPI communications. This means that no Open MPI device has indicated\n ## that it can be used to communicate between these processes. This is\n ## an error; Open MPI requires that all MPI processes be able to reach\n ## each other. This error can sometimes be the result of forgetting to\n ## specify the \"self\" BTL.\n ## \n-## Process 1 ([[54051,2],0]) is on host: profitbricks-build5-amd64\n-## Process 2 ([[54051,1],0]) is on host: profitbricks-build5-amd64\n+## Process 1 ([[39509,2],1]) is on host: i-capture-the-hostname\n+## Process 2 ([[39509,1],0]) is on host: i-capture-the-hostname\n ## BTLs attempted: vader self\n ## \n ## Your MPI job is now going to abort; sorry.\n ## --------------------------------------------------------------------------\n+## [i-capture-the-hostname:05989] [[39509,2],1] ORTE_ERROR_LOG: Unreachable in file dpm/dpm.c at line 433\n+## [i-capture-the-hostname:05924] [[39509,1],0] ORTE_ERROR_LOG: Unreachable in file dpm/dpm.c at line 433\n+## [i-capture-the-hostname:05984] [[39509,2],0] ORTE_ERROR_LOG: Unreachable in file dpm/dpm.c at line 433\n ## --------------------------------------------------------------------------\n ## It looks like MPI_INIT failed for some reason; your parallel process is\n ## likely to abort. There are many reasons that a parallel process can\n ## fail during MPI_INIT; some of which are due to configuration or environment\n ## problems. This failure appears to be an internal failure; here's some\n ## additional information (which may only be relevant to an Open MPI\n ## developer):\n ## \n ## ompi_dpm_dyn_init() failed\n ## --> Returned \"Unreachable\" (-12) instead of \"Success\" (0)\n ## --------------------------------------------------------------------------\n-## [profitbricks-build5-amd64:06382] [[54051,2],0] ORTE_ERROR_LOG: Unreachable in file dpm/dpm.c at line 433\n-## [profitbricks-build5-amd64:06386] [[54051,2],1] ORTE_ERROR_LOG: Unreachable in file dpm/dpm.c at line 433\n-## [profitbricks-build5-amd64:06382] *** An error occurred in MPI_Init\n-## [profitbricks-build5-amd64:06382] *** reported by process [3542286338,0]\n-## [profitbricks-build5-amd64:06382] *** on a NULL communicator\n-## [profitbricks-build5-amd64:06382] *** Unknown error\n-## [profitbricks-build5-amd64:06382] *** MPI_ERRORS_ARE_FATAL (processes in this communicator will now abort,\n-## [profitbricks-build5-amd64:06382] *** and potentially your MPI job)\n-## [profitbricks-build5-amd64:06192] [[54051,1],0] ORTE_ERROR_LOG: Unreachable in file dpm/dpm.c at line 433\n-## [profitbricks-build5-amd64:06192] *** An error occurred in MPI_Comm_spawn_multiple\n-## [profitbricks-build5-amd64:06192] *** reported by process [3542286337,0]\n-## [profitbricks-build5-amd64:06192] *** on communicator MPI_COMM_WORLD\n-## [profitbricks-build5-amd64:06192] *** MPI_ERR_INTERN: internal error\n-## [profitbricks-build5-amd64:06192] *** MPI_ERRORS_ARE_FATAL (processes in this communicator will now abort,\n-## [profitbricks-build5-amd64:06192] *** and potentially your MPI job)\n-## [profitbricks-build5-amd64:06129] 2 more processes have sent help message help-mca-bml-r2.txt / unreachable proc\n-## [profitbricks-build5-amd64:06129] Set MCA parameter \"orte_base_help_aggregate\" to 0 to see all help / error messages\n-## [profitbricks-build5-amd64:06129] 1 more process has sent help message help-mpi-runtime.txt / mpi_init:startup:internal-failure\n-## [profitbricks-build5-amd64:06129] 1 more process has sent help message help-mpi-errors.txt / mpi_errors_are_fatal unknown handle\n+## [i-capture-the-hostname:05924] *** An error occurred in MPI_Comm_spawn_multiple\n+## [i-capture-the-hostname:05924] *** reported by process [2589261825,0]\n+## [i-capture-the-hostname:05924] *** on communicator MPI_COMM_WORLD\n+## [i-capture-the-hostname:05924] *** MPI_ERR_INTERN: internal error\n+## [i-capture-the-hostname:05924] *** MPI_ERRORS_ARE_FATAL (processes in this communicator will now abort,\n+## [i-capture-the-hostname:05924] *** and potentially your MPI job)\n+## [i-capture-the-hostname:05984] *** An error occurred in MPI_Init\n+## [i-capture-the-hostname:05984] *** reported by process [2589261826,0]\n+## [i-capture-the-hostname:05984] *** on a NULL communicator\n+## [i-capture-the-hostname:05984] *** Unknown error\n+## [i-capture-the-hostname:05984] *** MPI_ERRORS_ARE_FATAL (processes in this communicator will now abort,\n+## [i-capture-the-hostname:05984] *** and potentially your MPI job)\n+## [i-capture-the-hostname:05892] 2 more processes have sent help message help-mca-bml-r2.txt / unreachable proc\n+## [i-capture-the-hostname:05892] Set MCA parameter \"orte_base_help_aggregate\" to 0 to see all help / error messages\n+## [i-capture-the-hostname:05892] 1 more process has sent help message help-mpi-runtime.txt / mpi_init:startup:internal-failure\n+## [i-capture-the-hostname:05892] 1 more process has sent help message help-mpi-errors.txt / mpi_errors_are_fatal unknown handle\n ## [warn] Epoll MOD(1) on fd 29 failed. Old events were 6; read change was 0 (none); write change was 2 (del); close change was 0 (none): Bad file descriptor\n not ok 611 - ./f77/spawn/spawnmult2f 2\n ---\n Directory: ./f77/spawn\n File: spawnmult2f\n Num-procs: 2\n- Date: \"Sun Jan 5 01:17:34 2020\"\n+ Date: \"Mon Dec 3 16:42:44 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## --------------------------------------------------------------------------\n ## At least one pair of MPI processes are unable to reach each other for\n ## MPI communications. This means that no Open MPI device has indicated\n ## that it can be used to communicate between these processes. This is\n ## an error; Open MPI requires that all MPI processes be able to reach\n ## each other. This error can sometimes be the result of forgetting to\n ## specify the \"self\" BTL.\n ## \n-## Process 1 ([[56658,1],1]) is on host: profitbricks-build5-amd64\n-## Process 2 ([[56658,2],0]) is on host: profitbricks-build5-amd64\n+## Process 1 ([[39594,1],1]) is on host: i-capture-the-hostname\n+## Process 2 ([[39594,2],0]) is on host: i-capture-the-hostname\n ## BTLs attempted: vader self\n ## \n ## Your MPI job is now going to abort; sorry.\n ## --------------------------------------------------------------------------\n-## [profitbricks-build5-amd64:06724] [[56658,2],1] ORTE_ERROR_LOG: Unreachable in file dpm/dpm.c at line 433\n-## [profitbricks-build5-amd64:06720] [[56658,2],0] ORTE_ERROR_LOG: Unreachable in file dpm/dpm.c at line 433\n ## --------------------------------------------------------------------------\n ## It looks like MPI_INIT failed for some reason; your parallel process is\n ## likely to abort. There are many reasons that a parallel process can\n ## fail during MPI_INIT; some of which are due to configuration or environment\n ## problems. This failure appears to be an internal failure; here's some\n ## additional information (which may only be relevant to an Open MPI\n ## developer):\n ## \n ## ompi_dpm_dyn_init() failed\n ## --> Returned \"Unreachable\" (-12) instead of \"Success\" (0)\n ## --------------------------------------------------------------------------\n-## [profitbricks-build5-amd64:06724] *** An error occurred in MPI_Init\n-## [profitbricks-build5-amd64:06724] *** reported by process [3713138690,1]\n-## [profitbricks-build5-amd64:06724] *** on a NULL communicator\n-## [profitbricks-build5-amd64:06724] *** Unknown error\n-## [profitbricks-build5-amd64:06724] *** MPI_ERRORS_ARE_FATAL (processes in this communicator will now abort,\n-## [profitbricks-build5-amd64:06724] *** and potentially your MPI job)\n-## [profitbricks-build5-amd64:06613] *** An error occurred in MPI_Comm_spawn_multiple\n-## [profitbricks-build5-amd64:06613] *** reported by process [3713138689,0]\n-## [profitbricks-build5-amd64:06613] *** on communicator MPI_COMM_WORLD\n-## [profitbricks-build5-amd64:06613] *** MPI_ERR_INTERN: internal error\n-## [profitbricks-build5-amd64:06613] *** MPI_ERRORS_ARE_FATAL (processes in this communicator will now abort,\n-## [profitbricks-build5-amd64:06613] *** and potentially your MPI job)\n-## [profitbricks-build5-amd64:06615] [[56658,1],1] ORTE_ERROR_LOG: Unreachable in file dpm/dpm.c at line 433\n-## [profitbricks-build5-amd64:06613] [[56658,1],0] ORTE_ERROR_LOG: Unreachable in file dpm/dpm.c at line 433\n-## [profitbricks-build5-amd64:06528] 3 more processes have sent help message help-mca-bml-r2.txt / unreachable proc\n-## [profitbricks-build5-amd64:06528] Set MCA parameter \"orte_base_help_aggregate\" to 0 to see all help / error messages\n-## [profitbricks-build5-amd64:06528] 1 more process has sent help message help-mpi-runtime.txt / mpi_init:startup:internal-failure\n-## [profitbricks-build5-amd64:06528] 1 more process has sent help message help-mpi-errors.txt / mpi_errors_are_fatal unknown handle\n-## [profitbricks-build5-amd64:06528] 1 more process has sent help message help-mpi-errors.txt / mpi_errors_are_fatal\n+## [i-capture-the-hostname:06230] [[39594,1],1] ORTE_ERROR_LOG: Unreachable in file dpm/dpm.c at line 433\n+## [i-capture-the-hostname:06352] [[39594,2],0] ORTE_ERROR_LOG: Unreachable in file dpm/dpm.c at line 433\n+## [i-capture-the-hostname:06224] [[39594,1],0] ORTE_ERROR_LOG: Unreachable in file dpm/dpm.c at line 433\n+## [i-capture-the-hostname:06230] *** An error occurred in MPI_Comm_spawn_multiple\n+## [i-capture-the-hostname:06230] *** reported by process [2594832385,1]\n+## [i-capture-the-hostname:06230] *** on communicator MPI_COMM_WORLD\n+## [i-capture-the-hostname:06230] *** MPI_ERR_INTERN: internal error\n+## [i-capture-the-hostname:06230] *** MPI_ERRORS_ARE_FATAL (processes in this communicator will now abort,\n+## [i-capture-the-hostname:06230] *** and potentially your MPI job)\n+## [i-capture-the-hostname:06358] [[39594,2],1] ORTE_ERROR_LOG: Unreachable in file dpm/dpm.c at line 433\n+## [i-capture-the-hostname:06352] *** An error occurred in MPI_Init\n+## [i-capture-the-hostname:06352] *** reported by process [2594832386,0]\n+## [i-capture-the-hostname:06352] *** on a NULL communicator\n+## [i-capture-the-hostname:06352] *** Unknown error\n+## [i-capture-the-hostname:06352] *** MPI_ERRORS_ARE_FATAL (processes in this communicator will now abort,\n+## [i-capture-the-hostname:06352] *** and potentially your MPI job)\n+## [i-capture-the-hostname:06139] 3 more processes have sent help message help-mca-bml-r2.txt / unreachable proc\n+## [i-capture-the-hostname:06139] Set MCA parameter \"orte_base_help_aggregate\" to 0 to see all help / error messages\n+## [i-capture-the-hostname:06139] 1 more process has sent help message help-mpi-runtime.txt / mpi_init:startup:internal-failure\n+## [i-capture-the-hostname:06139] 1 more process has sent help message help-mpi-errors.txt / mpi_errors_are_fatal\n+## [i-capture-the-hostname:06139] 1 more process has sent help message help-mpi-errors.txt / mpi_errors_are_fatal unknown handle\n ok 612 - ./f77/io/iwriteatf 2\n-ok 613 - ./f77/io/iwritef 2\n+not ok 613 - ./f77/io/iwritef 2\n+ ---\n+ Directory: ./f77/io\n+ File: iwritef\n+ Num-procs: 2\n+ Date: \"Mon Dec 3 16:47:11 2018\"\n+ ...\n+## Test output (expected 'No Errors'):\n+## --------------------------------------------------------------------------\n+## The user-provided time limit for job execution has been reached:\n+## \n+## Timeout: 180 seconds\n+## \n+## The job will now be aborted. Please check your code and/or\n+## adjust/remove the job execution time limit (as specified by --timeout\n+## command line option or MPIEXEC_TIMEOUT environment variable).\n+## --------------------------------------------------------------------------\n not ok 614 - ./f77/io/iwriteshf 2\n ---\n Directory: ./f77/io\n File: iwriteshf\n Num-procs: 2\n- Date: \"Sun Jan 5 01:23:52 2020\"\n+ Date: \"Mon Dec 3 16:50:12 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## --------------------------------------------------------------------------\n ## The user-provided time limit for job execution has been reached:\n ## \n ## Timeout: 180 seconds\n ## \n@@ -999117,62 +999136,126 @@\n ## --------------------------------------------------------------------------\n ok 615 - ./f77/io/writef 2\n ok 616 - ./f77/io/writeatf 2\n ok 617 - ./f77/io/writeallf 2\n ok 618 - ./f77/io/writeshf 2\n ok 619 - ./f77/io/writeordf 2\n ok 620 - ./f77/io/writeatallf 2\n-ok 621 - ./f77/io/writeatallbef 2\n-ok 622 - ./f77/io/writeallbef 2\n-ok 623 - ./f77/io/writeordbef 2\n+not ok 621 - ./f77/io/writeatallbef 2\n+ ---\n+ Directory: ./f77/io\n+ File: writeatallbef\n+ Num-procs: 2\n+ Date: \"Mon Dec 3 16:53:37 2018\"\n+ ...\n+## Test output (expected 'No Errors'):\n+## --------------------------------------------------------------------------\n+## The user-provided time limit for job execution has been reached:\n+## \n+## Timeout: 180 seconds\n+## \n+## The job will now be aborted. Please check your code and/or\n+## adjust/remove the job execution time limit (as specified by --timeout\n+## command line option or MPIEXEC_TIMEOUT environment variable).\n+## --------------------------------------------------------------------------\n+not ok 622 - ./f77/io/writeallbef 2\n+ ---\n+ Directory: ./f77/io\n+ File: writeallbef\n+ Num-procs: 2\n+ Date: \"Mon Dec 3 16:56:38 2018\"\n+ ...\n+## Test output (expected 'No Errors'):\n+## --------------------------------------------------------------------------\n+## The user-provided time limit for job execution has been reached:\n+## \n+## Timeout: 180 seconds\n+## \n+## The job will now be aborted. Please check your code and/or\n+## adjust/remove the job execution time limit (as specified by --timeout\n+## command line option or MPIEXEC_TIMEOUT environment variable).\n+## --------------------------------------------------------------------------\n+not ok 623 - ./f77/io/writeordbef 2\n+ ---\n+ Directory: ./f77/io\n+ File: writeordbef\n+ Num-procs: 2\n+ Date: \"Mon Dec 3 16:59:40 2018\"\n+ ...\n+## Test output (expected 'No Errors'):\n+## --------------------------------------------------------------------------\n+## The user-provided time limit for job execution has been reached:\n+## \n+## Timeout: 180 seconds\n+## \n+## The job will now be aborted. Please check your code and/or\n+## adjust/remove the job execution time limit (as specified by --timeout\n+## command line option or MPIEXEC_TIMEOUT environment variable).\n+## --------------------------------------------------------------------------\n ok 624 - ./f77/io/fileerrf 1\n ok 625 - ./f77/io/fileinfof 2\n ok 626 - ./f77/io/shpositionf 2\n-ok 627 - ./f77/io/atomicityf 2\n+not ok 627 - ./f77/io/atomicityf 2\n+ ---\n+ Directory: ./f77/io\n+ File: atomicityf\n+ Num-procs: 2\n+ Date: \"Mon Dec 3 16:59:49 2018\"\n+ ...\n+## Test output (expected 'No Errors'):\n+## --------------------------------------------------------------------------\n+## MPI_ABORT was invoked on rank 1 in communicator MPI_COMM_WORLD\n+## with errorcode 1.\n+## \n+## NOTE: invoking MPI_ABORT causes Open MPI to kill all MPI processes.\n+## You may or may not see output from other processes, depending on\n+## exactly when Open MPI kills them.\n+## --------------------------------------------------------------------------\n+## (noncontig)Process 1 : readbuf( 2277 ) is 0 , should be 10\n not ok 628 - ./f77/io/miscfilef 2\n ---\n Directory: ./f77/io\n File: miscfilef\n Num-procs: 2\n- Date: \"Sun Jan 5 01:32:14 2020\"\n+ Date: \"Mon Dec 3 16:59:52 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## File pointer position = 275 , should be 20\n-## File pointer position in bytes = 2680 , should be 1000\n ## File pointer position = 275 , should be 20\n ## File pointer position in bytes = 2680 , should be 1000\n+## File pointer position in bytes = 2680 , should be 1000\n ## Found 4 errors\n not ok 629 - ./f77/io/setviewcurf 2\n ---\n Directory: ./f77/io\n File: setviewcurf\n Num-procs: 2\n- Date: \"Sun Jan 5 01:32:15 2020\"\n+ Date: \"Mon Dec 3 16:59:54 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## 0 : buf(1) = -1\n-## Found 2 errors\n ## 1 : buf(1) = -1\n ## mca_fbtl_posix_preadv: error in (p)readv:Invalid argument\n ## mca_fbtl_posix_preadv: error in (p)readv:Invalid argument\n+## Found 2 errors\n ok 630 - ./f77/io/c2f2ciof 1\n ok 631 - ./f77/io/c2fmultio 1\n ok 646 - ./f77/io/i_setviewcurf 2 # SKIP requires MPI version 3.1\n ok 647 - ./f77/io/iwriteatallf 2 # SKIP requires MPI version 3.1\n ok 632 - ./f77/rma/winscale1f 2\n ok 633 - ./f77/rma/winfencef 2\n ok 634 - ./f77/rma/wingetf 2\n ok 635 - ./f77/rma/winscale2f 2\n ok 636 - ./f77/rma/winerrf 1\n not ok 637 - ./f77/rma/winnamef 1\n ---\n Directory: ./f77/rma\n File: winnamef\n Num-procs: 1\n- Date: \"Sun Jan 5 01:32:24 2020\"\n+ Date: \"Mon Dec 3 17:00:12 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Did not get empty name from new window\n ## Found 1 errors\n ok 638 - ./f77/rma/wingroupf 2\n ok 639 - ./f77/rma/winaccf 2\n ok 640 - ./f77/rma/c2f2cwinf 1\n@@ -999189,91 +999272,91 @@\n ok 650 - ./f77/ext/allocmemf 1\n ok 651 - ./f77/topo/cartcrf 2\n not ok 652 - ./f77/topo/dgraph_wgtf 2\n ---\n Directory: ./f77/topo\n File: dgraph_wgtf\n Num-procs: 2\n- Date: \"Sun Jan 5 01:32:41 2020\"\n+ Date: \"Mon Dec 3 17:00:51 2018\"\n ...\n ## Test output (expected 'No Errors'):\n-## Error: Cannot partition 16 elements in 17 parts\n+## Error: Cannot partition 15 elements in 2 parts\n ## \n ## Program received signal SIGSEGV: Segmentation fault - invalid memory reference.\n ## \n ## Backtrace for this error:\n-## #0 0x7f40c401d89d in ???\n-## #1 0x7f40c401cae3 in ???\n-## #2 0x7f40c3c578df in ???\n-## #3 0x7f40c0895e8b in ???\n-## #4 0x7f40c0896422 in ???\n-## #5 0x7f40c08967c1 in ???\n-## #6 0x7f40c089328c in ???\n-## #7 0x7f40c08899e8 in ???\n-## #8 0x7f40c42d93cc in ???\n-## #9 0x7f40c43c4bca in ???\n-## #10 0x55ac5ff3aa1b in dgraph_unwgt\n+## #0 0x7f30d2fd089d in ???\n+## #1 0x7f30d2fcfae3 in ???\n+## #2 0x7f30d2c0a8df in ???\n+## #3 0x7f30d0148e8b in ???\n+## #4 0x7f30d0149422 in ???\n+## #5 0x7f30d01497c1 in ???\n+## #6 0x7f30d014628c in ???\n+## #7 0x7f30d013c9e8 in ???\n+## #8 0x7f30d328c3cc in ???\n+## #9 0x7f30d3377bca in ???\n+## #10 0x561e752aea1b in dgraph_unwgt\n ## \tat ../../../f77/topo/dgraph_wgtf.f:157\n-## #11 0x55ac5ff3a1fe in main\n+## #11 0x561e752ae1fe in main\n ## \tat ../../../f77/topo/dgraph_wgtf.f:201\n ## --------------------------------------------------------------------------\n ## Primary job terminated normally, but 1 process returned\n ## a non-zero exit code. Per user-direction, the job has been aborted.\n ## --------------------------------------------------------------------------\n ## --------------------------------------------------------------------------\n-## mpiexec.openmpi noticed that process rank 0 with PID 0 on node profitbricks-build5-amd64 exited on signal 11 (Segmentation fault).\n+## mpiexec.openmpi noticed that process rank 0 with PID 0 on node i-capture-the-hostname exited on signal 11 (Segmentation fault).\n ## --------------------------------------------------------------------------\n not ok 653 - ./f77/topo/dgraph_unwgtf 2\n ---\n Directory: ./f77/topo\n File: dgraph_unwgtf\n Num-procs: 2\n- Date: \"Sun Jan 5 01:32:43 2020\"\n+ Date: \"Mon Dec 3 17:00:55 2018\"\n ...\n ## Test output (expected 'No Errors'):\n-## Error: Cannot partition 16 elements in 17 parts\n+## Error: Cannot partition 15 elements in 2 parts\n ## \n ## Program received signal SIGSEGV: Segmentation fault - invalid memory reference.\n ## \n ## Backtrace for this error:\n-## #0 0x7ffabbf0589d in ???\n-## #1 0x7ffabbf04ae3 in ???\n-## #2 0x7ffabbb3f8df in ???\n-## #3 0x7ffab877de8b in ???\n-## #4 0x7ffab877e422 in ???\n-## #5 0x7ffab877e7c1 in ???\n-## #6 0x7ffab877b28c in ???\n-## #7 0x7ffab87719e8 in ???\n-## #8 0x7ffabc1c13cc in ???\n-## #9 0x7ffabc2acbca in ???\n-## #10 0x55f07e31590d in dgraph_unwgt\n+## #0 0x7fd401c2889d in ???\n+## #1 0x7fd401c27ae3 in ???\n+## #2 0x7fd4018628df in ???\n+## #3 0x7fd3fa498e8b in ???\n+## #4 0x7fd3fa499422 in ???\n+## #5 0x7fd3fa4997c1 in ???\n+## #6 0x7fd3fa49628c in ???\n+## #7 0x7fd3fa48c9e8 in ???\n+## #8 0x7fd401ee43cc in ???\n+## #9 0x7fd401fcfbca in ???\n+## #10 0x55bebef6a90d in dgraph_unwgt\n ## \tat ../../../f77/topo/dgraph_unwgtf.f:141\n-## #11 0x55f07e3151fe in main\n+## #11 0x55bebef6a1fe in main\n ## \tat ../../../f77/topo/dgraph_unwgtf.f:216\n ## --------------------------------------------------------------------------\n ## Primary job terminated normally, but 1 process returned\n ## a non-zero exit code. Per user-direction, the job has been aborted.\n ## --------------------------------------------------------------------------\n ## --------------------------------------------------------------------------\n-## mpiexec.openmpi noticed that process rank 0 with PID 0 on node profitbricks-build5-amd64 exited on signal 11 (Segmentation fault).\n+## mpiexec.openmpi noticed that process rank 0 with PID 0 on node i-capture-the-hostname exited on signal 11 (Segmentation fault).\n ## --------------------------------------------------------------------------\n ok 654 - ./f77/profile/profile1f 2\n not ok 655 - ./cxx/attr/attrtx \n ---\n Directory: ./cxx/attr\n File: attrtx\n Num-procs: \n- Date: \"Sun Jan 5 01:32:45 2020\"\n+ Date: \"Mon Dec 3 17:01:03 2018\"\n ...\n ## Test output (expected 'No Errors'):\n-## Failed to build attrtx; make[2]: Entering directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/attr'\n+## Failed to build attrtx; make[2]: Entering directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/attr'\n ## (cd ../../cxx/util && make mtest.o)\n-## make[3]: Entering directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/util'\n+## make[3]: Entering directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/util'\n ## depbase=`echo mtest.o | sed 's|[^/]*$|.deps/&|;s|\\.o$||'`;\\\n-## /usr/bin/mpicxx.openmpi -DHAVE_CONFIG_H -I. -I../../../cxx/util -I../../include -I../../include -I../../../include -Wdate-time -D_FORTIFY_SOURCE=2 -g -O2 -ffile-prefix-map=/build/1st/mpi-testsuite-3.2+dfsg=. -fstack-protector-strong -Wformat -Werror=format-security -MT mtest.o -MD -MP -MF $depbase.Tpo -c -o mtest.o ../../../cxx/util/mtest.cxx &&\\\n+## /usr/bin/mpicxx.openmpi -DHAVE_CONFIG_H -I. -I../../../cxx/util -I../../include -I../../include -I../../../include -Wdate-time -D_FORTIFY_SOURCE=2 -g -O2 -ffile-prefix-map=/build/mpi-testsuite-3.2+dfsg/2nd=. -fstack-protector-strong -Wformat -Werror=format-security -MT mtest.o -MD -MP -MF $depbase.Tpo -c -o mtest.o ../../../cxx/util/mtest.cxx &&\\\n ## mv -f $depbase.Tpo $depbase.Po\n ## ../../../cxx/util/mtest.cxx: In function 'void* MTestTypeContigInitRecv(MTestDatatype*)':\n ## ../../../cxx/util/mtest.cxx:189:42: warning: 'int MPI_Type_extent(MPI_Datatype, MPI_Aint*)' is deprecated: MPI_Type_extent is superseded by MPI_Type_get_extent in MPI-2.0 [-Wdeprecated-declarations]\n ## MPI_Type_extent( mtype->datatype, &size );\n ## ^\n ## In file included from ../../../cxx/util/mtest.cxx:7:\n ## /usr/lib/x86_64-linux-gnu/openmpi/include/mpi.h:1758:20: note: declared here\n@@ -999305,30 +999388,30 @@\n ## ../../../cxx/util/mtest.cxx:848:50: error: 'NULL_COPY_FN' is not a member of 'MPI::Win'\n ## mem_keyval = MPI::Win::Create_keyval( MPI::Win::NULL_COPY_FN,\n ## ^~~~~~~~~~~~\n ## ../../../cxx/util/mtest.cxx:849:22: error: 'NULL_DELETE_FN' is not a member of 'MPI::Win'\n ## MPI::Win::NULL_DELETE_FN, 0 );\n ## ^~~~~~~~~~~~~~\n ## make[3]: *** [Makefile:416: mtest.o] Error 1\n-## make[3]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/util'\n+## make[3]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/util'\n ## make[2]: *** [Makefile:695: ../../cxx/util/mtest.o] Error 2\n-## make[2]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/attr'\n+## make[2]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/attr'\n not ok 656 - ./cxx/attr/attricx \n ---\n Directory: ./cxx/attr\n File: attricx\n Num-procs: \n- Date: \"Sun Jan 5 01:32:46 2020\"\n+ Date: \"Mon Dec 3 17:01:08 2018\"\n ...\n ## Test output (expected 'No Errors'):\n-## Failed to build attricx; make[2]: Entering directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/attr'\n+## Failed to build attricx; make[2]: Entering directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/attr'\n ## (cd ../../cxx/util && make mtest.o)\n-## make[3]: Entering directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/util'\n+## make[3]: Entering directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/util'\n ## depbase=`echo mtest.o | sed 's|[^/]*$|.deps/&|;s|\\.o$||'`;\\\n-## /usr/bin/mpicxx.openmpi -DHAVE_CONFIG_H -I. -I../../../cxx/util -I../../include -I../../include -I../../../include -Wdate-time -D_FORTIFY_SOURCE=2 -g -O2 -ffile-prefix-map=/build/1st/mpi-testsuite-3.2+dfsg=. -fstack-protector-strong -Wformat -Werror=format-security -MT mtest.o -MD -MP -MF $depbase.Tpo -c -o mtest.o ../../../cxx/util/mtest.cxx &&\\\n+## /usr/bin/mpicxx.openmpi -DHAVE_CONFIG_H -I. -I../../../cxx/util -I../../include -I../../include -I../../../include -Wdate-time -D_FORTIFY_SOURCE=2 -g -O2 -ffile-prefix-map=/build/mpi-testsuite-3.2+dfsg/2nd=. -fstack-protector-strong -Wformat -Werror=format-security -MT mtest.o -MD -MP -MF $depbase.Tpo -c -o mtest.o ../../../cxx/util/mtest.cxx &&\\\n ## mv -f $depbase.Tpo $depbase.Po\n ## ../../../cxx/util/mtest.cxx: In function 'void* MTestTypeContigInitRecv(MTestDatatype*)':\n ## ../../../cxx/util/mtest.cxx:189:42: warning: 'int MPI_Type_extent(MPI_Datatype, MPI_Aint*)' is deprecated: MPI_Type_extent is superseded by MPI_Type_get_extent in MPI-2.0 [-Wdeprecated-declarations]\n ## MPI_Type_extent( mtype->datatype, &size );\n ## ^\n ## In file included from ../../../cxx/util/mtest.cxx:7:\n ## /usr/lib/x86_64-linux-gnu/openmpi/include/mpi.h:1758:20: note: declared here\n@@ -999360,30 +999443,30 @@\n ## ../../../cxx/util/mtest.cxx:848:50: error: 'NULL_COPY_FN' is not a member of 'MPI::Win'\n ## mem_keyval = MPI::Win::Create_keyval( MPI::Win::NULL_COPY_FN,\n ## ^~~~~~~~~~~~\n ## ../../../cxx/util/mtest.cxx:849:22: error: 'NULL_DELETE_FN' is not a member of 'MPI::Win'\n ## MPI::Win::NULL_DELETE_FN, 0 );\n ## ^~~~~~~~~~~~~~\n ## make[3]: *** [Makefile:416: mtest.o] Error 1\n-## make[3]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/util'\n+## make[3]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/util'\n ## make[2]: *** [Makefile:695: ../../cxx/util/mtest.o] Error 2\n-## make[2]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/attr'\n+## make[2]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/attr'\n not ok 657 - ./cxx/attr/baseattrcommx \n ---\n Directory: ./cxx/attr\n File: baseattrcommx\n Num-procs: \n- Date: \"Sun Jan 5 01:32:48 2020\"\n+ Date: \"Mon Dec 3 17:01:15 2018\"\n ...\n ## Test output (expected 'No Errors'):\n-## Failed to build baseattrcommx; make[2]: Entering directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/attr'\n+## Failed to build baseattrcommx; make[2]: Entering directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/attr'\n ## (cd ../../cxx/util && make mtest.o)\n-## make[3]: Entering directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/util'\n+## make[3]: Entering directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/util'\n ## depbase=`echo mtest.o | sed 's|[^/]*$|.deps/&|;s|\\.o$||'`;\\\n-## /usr/bin/mpicxx.openmpi -DHAVE_CONFIG_H -I. -I../../../cxx/util -I../../include -I../../include -I../../../include -Wdate-time -D_FORTIFY_SOURCE=2 -g -O2 -ffile-prefix-map=/build/1st/mpi-testsuite-3.2+dfsg=. -fstack-protector-strong -Wformat -Werror=format-security -MT mtest.o -MD -MP -MF $depbase.Tpo -c -o mtest.o ../../../cxx/util/mtest.cxx &&\\\n+## /usr/bin/mpicxx.openmpi -DHAVE_CONFIG_H -I. -I../../../cxx/util -I../../include -I../../include -I../../../include -Wdate-time -D_FORTIFY_SOURCE=2 -g -O2 -ffile-prefix-map=/build/mpi-testsuite-3.2+dfsg/2nd=. -fstack-protector-strong -Wformat -Werror=format-security -MT mtest.o -MD -MP -MF $depbase.Tpo -c -o mtest.o ../../../cxx/util/mtest.cxx &&\\\n ## mv -f $depbase.Tpo $depbase.Po\n ## ../../../cxx/util/mtest.cxx: In function 'void* MTestTypeContigInitRecv(MTestDatatype*)':\n ## ../../../cxx/util/mtest.cxx:189:42: warning: 'int MPI_Type_extent(MPI_Datatype, MPI_Aint*)' is deprecated: MPI_Type_extent is superseded by MPI_Type_get_extent in MPI-2.0 [-Wdeprecated-declarations]\n ## MPI_Type_extent( mtype->datatype, &size );\n ## ^\n ## In file included from ../../../cxx/util/mtest.cxx:7:\n ## /usr/lib/x86_64-linux-gnu/openmpi/include/mpi.h:1758:20: note: declared here\n@@ -999415,30 +999498,30 @@\n ## ../../../cxx/util/mtest.cxx:848:50: error: 'NULL_COPY_FN' is not a member of 'MPI::Win'\n ## mem_keyval = MPI::Win::Create_keyval( MPI::Win::NULL_COPY_FN,\n ## ^~~~~~~~~~~~\n ## ../../../cxx/util/mtest.cxx:849:22: error: 'NULL_DELETE_FN' is not a member of 'MPI::Win'\n ## MPI::Win::NULL_DELETE_FN, 0 );\n ## ^~~~~~~~~~~~~~\n ## make[3]: *** [Makefile:416: mtest.o] Error 1\n-## make[3]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/util'\n+## make[3]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/util'\n ## make[2]: *** [Makefile:695: ../../cxx/util/mtest.o] Error 2\n-## make[2]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/attr'\n+## make[2]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/attr'\n not ok 658 - ./cxx/attr/fkeyvalcommx \n ---\n Directory: ./cxx/attr\n File: fkeyvalcommx\n Num-procs: \n- Date: \"Sun Jan 5 01:32:49 2020\"\n+ Date: \"Mon Dec 3 17:01:20 2018\"\n ...\n ## Test output (expected 'No Errors'):\n-## Failed to build fkeyvalcommx; make[2]: Entering directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/attr'\n+## Failed to build fkeyvalcommx; make[2]: Entering directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/attr'\n ## (cd ../../cxx/util && make mtest.o)\n-## make[3]: Entering directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/util'\n+## make[3]: Entering directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/util'\n ## depbase=`echo mtest.o | sed 's|[^/]*$|.deps/&|;s|\\.o$||'`;\\\n-## /usr/bin/mpicxx.openmpi -DHAVE_CONFIG_H -I. -I../../../cxx/util -I../../include -I../../include -I../../../include -Wdate-time -D_FORTIFY_SOURCE=2 -g -O2 -ffile-prefix-map=/build/1st/mpi-testsuite-3.2+dfsg=. -fstack-protector-strong -Wformat -Werror=format-security -MT mtest.o -MD -MP -MF $depbase.Tpo -c -o mtest.o ../../../cxx/util/mtest.cxx &&\\\n+## /usr/bin/mpicxx.openmpi -DHAVE_CONFIG_H -I. -I../../../cxx/util -I../../include -I../../include -I../../../include -Wdate-time -D_FORTIFY_SOURCE=2 -g -O2 -ffile-prefix-map=/build/mpi-testsuite-3.2+dfsg/2nd=. -fstack-protector-strong -Wformat -Werror=format-security -MT mtest.o -MD -MP -MF $depbase.Tpo -c -o mtest.o ../../../cxx/util/mtest.cxx &&\\\n ## mv -f $depbase.Tpo $depbase.Po\n ## ../../../cxx/util/mtest.cxx: In function 'void* MTestTypeContigInitRecv(MTestDatatype*)':\n ## ../../../cxx/util/mtest.cxx:189:42: warning: 'int MPI_Type_extent(MPI_Datatype, MPI_Aint*)' is deprecated: MPI_Type_extent is superseded by MPI_Type_get_extent in MPI-2.0 [-Wdeprecated-declarations]\n ## MPI_Type_extent( mtype->datatype, &size );\n ## ^\n ## In file included from ../../../cxx/util/mtest.cxx:7:\n ## /usr/lib/x86_64-linux-gnu/openmpi/include/mpi.h:1758:20: note: declared here\n@@ -999470,28 +999553,28 @@\n ## ../../../cxx/util/mtest.cxx:848:50: error: 'NULL_COPY_FN' is not a member of 'MPI::Win'\n ## mem_keyval = MPI::Win::Create_keyval( MPI::Win::NULL_COPY_FN,\n ## ^~~~~~~~~~~~\n ## ../../../cxx/util/mtest.cxx:849:22: error: 'NULL_DELETE_FN' is not a member of 'MPI::Win'\n ## MPI::Win::NULL_DELETE_FN, 0 );\n ## ^~~~~~~~~~~~~~\n ## make[3]: *** [Makefile:416: mtest.o] Error 1\n-## make[3]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/util'\n+## make[3]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/util'\n ## make[2]: *** [Makefile:695: ../../cxx/util/mtest.o] Error 2\n-## make[2]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/attr'\n+## make[2]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/attr'\n not ok 659 - ./cxx/attr/fkeyvaltypex \n ---\n Directory: ./cxx/attr\n File: fkeyvaltypex\n Num-procs: \n- Date: \"Sun Jan 5 01:32:50 2020\"\n+ Date: \"Mon Dec 3 17:01:25 2018\"\n ...\n ## Test output (expected 'No Errors'):\n-## Failed to build fkeyvaltypex; make[2]: Entering directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/attr'\n+## Failed to build fkeyvaltypex; make[2]: Entering directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/attr'\n ## depbase=`echo fkeyvaltypex.o | sed 's|[^/]*$|.deps/&|;s|\\.o$||'`;\\\n-## /usr/bin/mpicxx.openmpi -DHAVE_CONFIG_H -I. -I../../../cxx/attr -I../../include -I../../include -I../../../include -Wdate-time -D_FORTIFY_SOURCE=2 -g -O2 -ffile-prefix-map=/build/1st/mpi-testsuite-3.2+dfsg=. -fstack-protector-strong -Wformat -Werror=format-security -MT fkeyvaltypex.o -MD -MP -MF $depbase.Tpo -c -o fkeyvaltypex.o ../../../cxx/attr/fkeyvaltypex.cxx &&\\\n+## /usr/bin/mpicxx.openmpi -DHAVE_CONFIG_H -I. -I../../../cxx/attr -I../../include -I../../include -I../../../include -Wdate-time -D_FORTIFY_SOURCE=2 -g -O2 -ffile-prefix-map=/build/mpi-testsuite-3.2+dfsg/2nd=. -fstack-protector-strong -Wformat -Werror=format-security -MT fkeyvaltypex.o -MD -MP -MF $depbase.Tpo -c -o fkeyvaltypex.o ../../../cxx/attr/fkeyvaltypex.cxx &&\\\n ## mv -f $depbase.Tpo $depbase.Po\n ## ../../../cxx/attr/fkeyvaltypex.cxx: In function 'int main(int, char**)':\n ## ../../../cxx/attr/fkeyvaltypex.cxx:67:71: error: no matching function for call to 'Create_keyval(int (&)(const MPI::Datatype&, int, void*, void*, void*, bool&), int (&)(MPI::Datatype&, int, void*, void*), void*)'\n ## keyval = MPI::Datatype::Create_keyval( copy_fn, delete_fn, (void *)0 );\n ## ^\n ## In file included from /usr/lib/x86_64-linux-gnu/openmpi/include/openmpi/ompi/mpi/cxx/mpicxx.h:269,\n ## from /usr/lib/x86_64-linux-gnu/openmpi/include/mpi.h:2684,\n@@ -999536,28 +999619,28 @@\n ## ../../../cxx/attr/fkeyvaltypex.cxx:79:18: error: 'NULL_COPY_FN' is not a member of 'MPI::Datatype'\n ## MPI::Datatype::NULL_COPY_FN, MPI::Datatype::NULL_DELETE_FN,\n ## ^~~~~~~~~~~~\n ## ../../../cxx/attr/fkeyvaltypex.cxx:79:47: error: 'NULL_DELETE_FN' is not a member of 'MPI::Datatype'\n ## MPI::Datatype::NULL_COPY_FN, MPI::Datatype::NULL_DELETE_FN,\n ## ^~~~~~~~~~~~~~\n ## make[2]: *** [Makefile:475: fkeyvaltypex.o] Error 1\n-## make[2]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/attr'\n+## make[2]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/attr'\n not ok 660 - ./cxx/pt2pt/bsend1cxx \n ---\n Directory: ./cxx/pt2pt\n File: bsend1cxx\n Num-procs: \n- Date: \"Sun Jan 5 01:32:51 2020\"\n+ Date: \"Mon Dec 3 17:01:31 2018\"\n ...\n ## Test output (expected 'No Errors'):\n-## Failed to build bsend1cxx; make[2]: Entering directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/pt2pt'\n+## Failed to build bsend1cxx; make[2]: Entering directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/pt2pt'\n ## (cd ../../cxx/util && make mtest.o)\n-## make[3]: Entering directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/util'\n+## make[3]: Entering directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/util'\n ## depbase=`echo mtest.o | sed 's|[^/]*$|.deps/&|;s|\\.o$||'`;\\\n-## /usr/bin/mpicxx.openmpi -DHAVE_CONFIG_H -I. -I../../../cxx/util -I../../include -I../../include -I../../../include -Wdate-time -D_FORTIFY_SOURCE=2 -g -O2 -ffile-prefix-map=/build/1st/mpi-testsuite-3.2+dfsg=. -fstack-protector-strong -Wformat -Werror=format-security -MT mtest.o -MD -MP -MF $depbase.Tpo -c -o mtest.o ../../../cxx/util/mtest.cxx &&\\\n+## /usr/bin/mpicxx.openmpi -DHAVE_CONFIG_H -I. -I../../../cxx/util -I../../include -I../../include -I../../../include -Wdate-time -D_FORTIFY_SOURCE=2 -g -O2 -ffile-prefix-map=/build/mpi-testsuite-3.2+dfsg/2nd=. -fstack-protector-strong -Wformat -Werror=format-security -MT mtest.o -MD -MP -MF $depbase.Tpo -c -o mtest.o ../../../cxx/util/mtest.cxx &&\\\n ## mv -f $depbase.Tpo $depbase.Po\n ## ../../../cxx/util/mtest.cxx: In function 'void* MTestTypeContigInitRecv(MTestDatatype*)':\n ## ../../../cxx/util/mtest.cxx:189:42: warning: 'int MPI_Type_extent(MPI_Datatype, MPI_Aint*)' is deprecated: MPI_Type_extent is superseded by MPI_Type_get_extent in MPI-2.0 [-Wdeprecated-declarations]\n ## MPI_Type_extent( mtype->datatype, &size );\n ## ^\n ## In file included from ../../../cxx/util/mtest.cxx:7:\n ## /usr/lib/x86_64-linux-gnu/openmpi/include/mpi.h:1758:20: note: declared here\n@@ -999589,30 +999672,30 @@\n ## ../../../cxx/util/mtest.cxx:848:50: error: 'NULL_COPY_FN' is not a member of 'MPI::Win'\n ## mem_keyval = MPI::Win::Create_keyval( MPI::Win::NULL_COPY_FN,\n ## ^~~~~~~~~~~~\n ## ../../../cxx/util/mtest.cxx:849:22: error: 'NULL_DELETE_FN' is not a member of 'MPI::Win'\n ## MPI::Win::NULL_DELETE_FN, 0 );\n ## ^~~~~~~~~~~~~~\n ## make[3]: *** [Makefile:416: mtest.o] Error 1\n-## make[3]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/util'\n+## make[3]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/util'\n ## make[2]: *** [Makefile:650: ../../cxx/util/mtest.o] Error 2\n-## make[2]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/pt2pt'\n+## make[2]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/pt2pt'\n not ok 661 - ./cxx/pt2pt/sendrecvx \n ---\n Directory: ./cxx/pt2pt\n File: sendrecvx\n Num-procs: \n- Date: \"Sun Jan 5 01:32:52 2020\"\n+ Date: \"Mon Dec 3 17:01:36 2018\"\n ...\n ## Test output (expected 'No Errors'):\n-## Failed to build sendrecvx; make[2]: Entering directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/pt2pt'\n+## Failed to build sendrecvx; make[2]: Entering directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/pt2pt'\n ## (cd ../../cxx/util && make mtest.o)\n-## make[3]: Entering directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/util'\n+## make[3]: Entering directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/util'\n ## depbase=`echo mtest.o | sed 's|[^/]*$|.deps/&|;s|\\.o$||'`;\\\n-## /usr/bin/mpicxx.openmpi -DHAVE_CONFIG_H -I. -I../../../cxx/util -I../../include -I../../include -I../../../include -Wdate-time -D_FORTIFY_SOURCE=2 -g -O2 -ffile-prefix-map=/build/1st/mpi-testsuite-3.2+dfsg=. -fstack-protector-strong -Wformat -Werror=format-security -MT mtest.o -MD -MP -MF $depbase.Tpo -c -o mtest.o ../../../cxx/util/mtest.cxx &&\\\n+## /usr/bin/mpicxx.openmpi -DHAVE_CONFIG_H -I. -I../../../cxx/util -I../../include -I../../include -I../../../include -Wdate-time -D_FORTIFY_SOURCE=2 -g -O2 -ffile-prefix-map=/build/mpi-testsuite-3.2+dfsg/2nd=. -fstack-protector-strong -Wformat -Werror=format-security -MT mtest.o -MD -MP -MF $depbase.Tpo -c -o mtest.o ../../../cxx/util/mtest.cxx &&\\\n ## mv -f $depbase.Tpo $depbase.Po\n ## ../../../cxx/util/mtest.cxx: In function 'void* MTestTypeContigInitRecv(MTestDatatype*)':\n ## ../../../cxx/util/mtest.cxx:189:42: warning: 'int MPI_Type_extent(MPI_Datatype, MPI_Aint*)' is deprecated: MPI_Type_extent is superseded by MPI_Type_get_extent in MPI-2.0 [-Wdeprecated-declarations]\n ## MPI_Type_extent( mtype->datatype, &size );\n ## ^\n ## In file included from ../../../cxx/util/mtest.cxx:7:\n ## /usr/lib/x86_64-linux-gnu/openmpi/include/mpi.h:1758:20: note: declared here\n@@ -999644,30 +999727,30 @@\n ## ../../../cxx/util/mtest.cxx:848:50: error: 'NULL_COPY_FN' is not a member of 'MPI::Win'\n ## mem_keyval = MPI::Win::Create_keyval( MPI::Win::NULL_COPY_FN,\n ## ^~~~~~~~~~~~\n ## ../../../cxx/util/mtest.cxx:849:22: error: 'NULL_DELETE_FN' is not a member of 'MPI::Win'\n ## MPI::Win::NULL_DELETE_FN, 0 );\n ## ^~~~~~~~~~~~~~\n ## make[3]: *** [Makefile:416: mtest.o] Error 1\n-## make[3]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/util'\n+## make[3]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/util'\n ## make[2]: *** [Makefile:650: ../../cxx/util/mtest.o] Error 2\n-## make[2]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/pt2pt'\n+## make[2]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/pt2pt'\n not ok 662 - ./cxx/comm/commname2 \n ---\n Directory: ./cxx/comm\n File: commname2\n Num-procs: \n- Date: \"Sun Jan 5 01:32:54 2020\"\n+ Date: \"Mon Dec 3 17:01:42 2018\"\n ...\n ## Test output (expected 'No Errors'):\n-## Failed to build commname2; make[2]: Entering directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/comm'\n+## Failed to build commname2; make[2]: Entering directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/comm'\n ## (cd ../../cxx/util && make mtest.o)\n-## make[3]: Entering directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/util'\n+## make[3]: Entering directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/util'\n ## depbase=`echo mtest.o | sed 's|[^/]*$|.deps/&|;s|\\.o$||'`;\\\n-## /usr/bin/mpicxx.openmpi -DHAVE_CONFIG_H -I. -I../../../cxx/util -I../../include -I../../include -I../../../include -Wdate-time -D_FORTIFY_SOURCE=2 -g -O2 -ffile-prefix-map=/build/1st/mpi-testsuite-3.2+dfsg=. -fstack-protector-strong -Wformat -Werror=format-security -MT mtest.o -MD -MP -MF $depbase.Tpo -c -o mtest.o ../../../cxx/util/mtest.cxx &&\\\n+## /usr/bin/mpicxx.openmpi -DHAVE_CONFIG_H -I. -I../../../cxx/util -I../../include -I../../include -I../../../include -Wdate-time -D_FORTIFY_SOURCE=2 -g -O2 -ffile-prefix-map=/build/mpi-testsuite-3.2+dfsg/2nd=. -fstack-protector-strong -Wformat -Werror=format-security -MT mtest.o -MD -MP -MF $depbase.Tpo -c -o mtest.o ../../../cxx/util/mtest.cxx &&\\\n ## mv -f $depbase.Tpo $depbase.Po\n ## ../../../cxx/util/mtest.cxx: In function 'void* MTestTypeContigInitRecv(MTestDatatype*)':\n ## ../../../cxx/util/mtest.cxx:189:42: warning: 'int MPI_Type_extent(MPI_Datatype, MPI_Aint*)' is deprecated: MPI_Type_extent is superseded by MPI_Type_get_extent in MPI-2.0 [-Wdeprecated-declarations]\n ## MPI_Type_extent( mtype->datatype, &size );\n ## ^\n ## In file included from ../../../cxx/util/mtest.cxx:7:\n ## /usr/lib/x86_64-linux-gnu/openmpi/include/mpi.h:1758:20: note: declared here\n@@ -999699,30 +999782,30 @@\n ## ../../../cxx/util/mtest.cxx:848:50: error: 'NULL_COPY_FN' is not a member of 'MPI::Win'\n ## mem_keyval = MPI::Win::Create_keyval( MPI::Win::NULL_COPY_FN,\n ## ^~~~~~~~~~~~\n ## ../../../cxx/util/mtest.cxx:849:22: error: 'NULL_DELETE_FN' is not a member of 'MPI::Win'\n ## MPI::Win::NULL_DELETE_FN, 0 );\n ## ^~~~~~~~~~~~~~\n ## make[3]: *** [Makefile:416: mtest.o] Error 1\n-## make[3]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/util'\n+## make[3]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/util'\n ## make[2]: *** [Makefile:640: ../../cxx/util/mtest.o] Error 2\n-## make[2]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/comm'\n+## make[2]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/comm'\n not ok 663 - ./cxx/coll/arcomplex \n ---\n Directory: ./cxx/coll\n File: arcomplex\n Num-procs: \n- Date: \"Sun Jan 5 01:32:55 2020\"\n+ Date: \"Mon Dec 3 17:01:48 2018\"\n ...\n ## Test output (expected 'No Errors'):\n-## Failed to build arcomplex; make[2]: Entering directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/coll'\n+## Failed to build arcomplex; make[2]: Entering directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/coll'\n ## (cd ../../cxx/util && make mtest.o)\n-## make[3]: Entering directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/util'\n+## make[3]: Entering directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/util'\n ## depbase=`echo mtest.o | sed 's|[^/]*$|.deps/&|;s|\\.o$||'`;\\\n-## /usr/bin/mpicxx.openmpi -DHAVE_CONFIG_H -I. -I../../../cxx/util -I../../include -I../../include -I../../../include -Wdate-time -D_FORTIFY_SOURCE=2 -g -O2 -ffile-prefix-map=/build/1st/mpi-testsuite-3.2+dfsg=. -fstack-protector-strong -Wformat -Werror=format-security -MT mtest.o -MD -MP -MF $depbase.Tpo -c -o mtest.o ../../../cxx/util/mtest.cxx &&\\\n+## /usr/bin/mpicxx.openmpi -DHAVE_CONFIG_H -I. -I../../../cxx/util -I../../include -I../../include -I../../../include -Wdate-time -D_FORTIFY_SOURCE=2 -g -O2 -ffile-prefix-map=/build/mpi-testsuite-3.2+dfsg/2nd=. -fstack-protector-strong -Wformat -Werror=format-security -MT mtest.o -MD -MP -MF $depbase.Tpo -c -o mtest.o ../../../cxx/util/mtest.cxx &&\\\n ## mv -f $depbase.Tpo $depbase.Po\n ## ../../../cxx/util/mtest.cxx: In function 'void* MTestTypeContigInitRecv(MTestDatatype*)':\n ## ../../../cxx/util/mtest.cxx:189:42: warning: 'int MPI_Type_extent(MPI_Datatype, MPI_Aint*)' is deprecated: MPI_Type_extent is superseded by MPI_Type_get_extent in MPI-2.0 [-Wdeprecated-declarations]\n ## MPI_Type_extent( mtype->datatype, &size );\n ## ^\n ## In file included from ../../../cxx/util/mtest.cxx:7:\n ## /usr/lib/x86_64-linux-gnu/openmpi/include/mpi.h:1758:20: note: declared here\n@@ -999754,30 +999837,30 @@\n ## ../../../cxx/util/mtest.cxx:848:50: error: 'NULL_COPY_FN' is not a member of 'MPI::Win'\n ## mem_keyval = MPI::Win::Create_keyval( MPI::Win::NULL_COPY_FN,\n ## ^~~~~~~~~~~~\n ## ../../../cxx/util/mtest.cxx:849:22: error: 'NULL_DELETE_FN' is not a member of 'MPI::Win'\n ## MPI::Win::NULL_DELETE_FN, 0 );\n ## ^~~~~~~~~~~~~~\n ## make[3]: *** [Makefile:416: mtest.o] Error 1\n-## make[3]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/util'\n+## make[3]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/util'\n ## make[2]: *** [Makefile:865: ../../cxx/util/mtest.o] Error 2\n-## make[2]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/coll'\n+## make[2]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/coll'\n not ok 664 - ./cxx/coll/uallredx \n ---\n Directory: ./cxx/coll\n File: uallredx\n Num-procs: \n- Date: \"Sun Jan 5 01:32:56 2020\"\n+ Date: \"Mon Dec 3 17:01:55 2018\"\n ...\n ## Test output (expected 'No Errors'):\n-## Failed to build uallredx; make[2]: Entering directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/coll'\n+## Failed to build uallredx; make[2]: Entering directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/coll'\n ## (cd ../../cxx/util && make mtest.o)\n-## make[3]: Entering directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/util'\n+## make[3]: Entering directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/util'\n ## depbase=`echo mtest.o | sed 's|[^/]*$|.deps/&|;s|\\.o$||'`;\\\n-## /usr/bin/mpicxx.openmpi -DHAVE_CONFIG_H -I. -I../../../cxx/util -I../../include -I../../include -I../../../include -Wdate-time -D_FORTIFY_SOURCE=2 -g -O2 -ffile-prefix-map=/build/1st/mpi-testsuite-3.2+dfsg=. -fstack-protector-strong -Wformat -Werror=format-security -MT mtest.o -MD -MP -MF $depbase.Tpo -c -o mtest.o ../../../cxx/util/mtest.cxx &&\\\n+## /usr/bin/mpicxx.openmpi -DHAVE_CONFIG_H -I. -I../../../cxx/util -I../../include -I../../include -I../../../include -Wdate-time -D_FORTIFY_SOURCE=2 -g -O2 -ffile-prefix-map=/build/mpi-testsuite-3.2+dfsg/2nd=. -fstack-protector-strong -Wformat -Werror=format-security -MT mtest.o -MD -MP -MF $depbase.Tpo -c -o mtest.o ../../../cxx/util/mtest.cxx &&\\\n ## mv -f $depbase.Tpo $depbase.Po\n ## ../../../cxx/util/mtest.cxx: In function 'void* MTestTypeContigInitRecv(MTestDatatype*)':\n ## ../../../cxx/util/mtest.cxx:189:42: warning: 'int MPI_Type_extent(MPI_Datatype, MPI_Aint*)' is deprecated: MPI_Type_extent is superseded by MPI_Type_get_extent in MPI-2.0 [-Wdeprecated-declarations]\n ## MPI_Type_extent( mtype->datatype, &size );\n ## ^\n ## In file included from ../../../cxx/util/mtest.cxx:7:\n ## /usr/lib/x86_64-linux-gnu/openmpi/include/mpi.h:1758:20: note: declared here\n@@ -999809,30 +999892,30 @@\n ## ../../../cxx/util/mtest.cxx:848:50: error: 'NULL_COPY_FN' is not a member of 'MPI::Win'\n ## mem_keyval = MPI::Win::Create_keyval( MPI::Win::NULL_COPY_FN,\n ## ^~~~~~~~~~~~\n ## ../../../cxx/util/mtest.cxx:849:22: error: 'NULL_DELETE_FN' is not a member of 'MPI::Win'\n ## MPI::Win::NULL_DELETE_FN, 0 );\n ## ^~~~~~~~~~~~~~\n ## make[3]: *** [Makefile:416: mtest.o] Error 1\n-## make[3]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/util'\n+## make[3]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/util'\n ## make[2]: *** [Makefile:865: ../../cxx/util/mtest.o] Error 2\n-## make[2]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/coll'\n+## make[2]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/coll'\n not ok 665 - ./cxx/coll/uallreduce \n ---\n Directory: ./cxx/coll\n File: uallreduce\n Num-procs: \n- Date: \"Sun Jan 5 01:32:57 2020\"\n+ Date: \"Mon Dec 3 17:02:00 2018\"\n ...\n ## Test output (expected 'No Errors'):\n-## Failed to build uallreduce; make[2]: Entering directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/coll'\n+## Failed to build uallreduce; make[2]: Entering directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/coll'\n ## (cd ../../cxx/util && make mtest.o)\n-## make[3]: Entering directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/util'\n+## make[3]: Entering directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/util'\n ## depbase=`echo mtest.o | sed 's|[^/]*$|.deps/&|;s|\\.o$||'`;\\\n-## /usr/bin/mpicxx.openmpi -DHAVE_CONFIG_H -I. -I../../../cxx/util -I../../include -I../../include -I../../../include -Wdate-time -D_FORTIFY_SOURCE=2 -g -O2 -ffile-prefix-map=/build/1st/mpi-testsuite-3.2+dfsg=. -fstack-protector-strong -Wformat -Werror=format-security -MT mtest.o -MD -MP -MF $depbase.Tpo -c -o mtest.o ../../../cxx/util/mtest.cxx &&\\\n+## /usr/bin/mpicxx.openmpi -DHAVE_CONFIG_H -I. -I../../../cxx/util -I../../include -I../../include -I../../../include -Wdate-time -D_FORTIFY_SOURCE=2 -g -O2 -ffile-prefix-map=/build/mpi-testsuite-3.2+dfsg/2nd=. -fstack-protector-strong -Wformat -Werror=format-security -MT mtest.o -MD -MP -MF $depbase.Tpo -c -o mtest.o ../../../cxx/util/mtest.cxx &&\\\n ## mv -f $depbase.Tpo $depbase.Po\n ## ../../../cxx/util/mtest.cxx: In function 'void* MTestTypeContigInitRecv(MTestDatatype*)':\n ## ../../../cxx/util/mtest.cxx:189:42: warning: 'int MPI_Type_extent(MPI_Datatype, MPI_Aint*)' is deprecated: MPI_Type_extent is superseded by MPI_Type_get_extent in MPI-2.0 [-Wdeprecated-declarations]\n ## MPI_Type_extent( mtype->datatype, &size );\n ## ^\n ## In file included from ../../../cxx/util/mtest.cxx:7:\n ## /usr/lib/x86_64-linux-gnu/openmpi/include/mpi.h:1758:20: note: declared here\n@@ -999864,30 +999947,30 @@\n ## ../../../cxx/util/mtest.cxx:848:50: error: 'NULL_COPY_FN' is not a member of 'MPI::Win'\n ## mem_keyval = MPI::Win::Create_keyval( MPI::Win::NULL_COPY_FN,\n ## ^~~~~~~~~~~~\n ## ../../../cxx/util/mtest.cxx:849:22: error: 'NULL_DELETE_FN' is not a member of 'MPI::Win'\n ## MPI::Win::NULL_DELETE_FN, 0 );\n ## ^~~~~~~~~~~~~~\n ## make[3]: *** [Makefile:416: mtest.o] Error 1\n-## make[3]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/util'\n+## make[3]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/util'\n ## make[2]: *** [Makefile:865: ../../cxx/util/mtest.o] Error 2\n-## make[2]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/coll'\n+## make[2]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/coll'\n not ok 666 - ./cxx/coll/ureduce \n ---\n Directory: ./cxx/coll\n File: ureduce\n Num-procs: \n- Date: \"Sun Jan 5 01:32:58 2020\"\n+ Date: \"Mon Dec 3 17:02:05 2018\"\n ...\n ## Test output (expected 'No Errors'):\n-## Failed to build ureduce; make[2]: Entering directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/coll'\n+## Failed to build ureduce; make[2]: Entering directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/coll'\n ## (cd ../../cxx/util && make mtest.o)\n-## make[3]: Entering directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/util'\n+## make[3]: Entering directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/util'\n ## depbase=`echo mtest.o | sed 's|[^/]*$|.deps/&|;s|\\.o$||'`;\\\n-## /usr/bin/mpicxx.openmpi -DHAVE_CONFIG_H -I. -I../../../cxx/util -I../../include -I../../include -I../../../include -Wdate-time -D_FORTIFY_SOURCE=2 -g -O2 -ffile-prefix-map=/build/1st/mpi-testsuite-3.2+dfsg=. -fstack-protector-strong -Wformat -Werror=format-security -MT mtest.o -MD -MP -MF $depbase.Tpo -c -o mtest.o ../../../cxx/util/mtest.cxx &&\\\n+## /usr/bin/mpicxx.openmpi -DHAVE_CONFIG_H -I. -I../../../cxx/util -I../../include -I../../include -I../../../include -Wdate-time -D_FORTIFY_SOURCE=2 -g -O2 -ffile-prefix-map=/build/mpi-testsuite-3.2+dfsg/2nd=. -fstack-protector-strong -Wformat -Werror=format-security -MT mtest.o -MD -MP -MF $depbase.Tpo -c -o mtest.o ../../../cxx/util/mtest.cxx &&\\\n ## mv -f $depbase.Tpo $depbase.Po\n ## ../../../cxx/util/mtest.cxx: In function 'void* MTestTypeContigInitRecv(MTestDatatype*)':\n ## ../../../cxx/util/mtest.cxx:189:42: warning: 'int MPI_Type_extent(MPI_Datatype, MPI_Aint*)' is deprecated: MPI_Type_extent is superseded by MPI_Type_get_extent in MPI-2.0 [-Wdeprecated-declarations]\n ## MPI_Type_extent( mtype->datatype, &size );\n ## ^\n ## In file included from ../../../cxx/util/mtest.cxx:7:\n ## /usr/lib/x86_64-linux-gnu/openmpi/include/mpi.h:1758:20: note: declared here\n@@ -999919,30 +1000002,30 @@\n ## ../../../cxx/util/mtest.cxx:848:50: error: 'NULL_COPY_FN' is not a member of 'MPI::Win'\n ## mem_keyval = MPI::Win::Create_keyval( MPI::Win::NULL_COPY_FN,\n ## ^~~~~~~~~~~~\n ## ../../../cxx/util/mtest.cxx:849:22: error: 'NULL_DELETE_FN' is not a member of 'MPI::Win'\n ## MPI::Win::NULL_DELETE_FN, 0 );\n ## ^~~~~~~~~~~~~~\n ## make[3]: *** [Makefile:416: mtest.o] Error 1\n-## make[3]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/util'\n+## make[3]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/util'\n ## make[2]: *** [Makefile:865: ../../cxx/util/mtest.o] Error 2\n-## make[2]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/coll'\n+## make[2]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/coll'\n not ok 667 - ./cxx/coll/ureducelocal \n ---\n Directory: ./cxx/coll\n File: ureducelocal\n Num-procs: \n- Date: \"Sun Jan 5 01:32:59 2020\"\n+ Date: \"Mon Dec 3 17:02:10 2018\"\n ...\n ## Test output (expected 'No Errors'):\n-## Failed to build ureducelocal; make[2]: Entering directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/coll'\n+## Failed to build ureducelocal; make[2]: Entering directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/coll'\n ## (cd ../../cxx/util && make mtest.o)\n-## make[3]: Entering directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/util'\n+## make[3]: Entering directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/util'\n ## depbase=`echo mtest.o | sed 's|[^/]*$|.deps/&|;s|\\.o$||'`;\\\n-## /usr/bin/mpicxx.openmpi -DHAVE_CONFIG_H -I. -I../../../cxx/util -I../../include -I../../include -I../../../include -Wdate-time -D_FORTIFY_SOURCE=2 -g -O2 -ffile-prefix-map=/build/1st/mpi-testsuite-3.2+dfsg=. -fstack-protector-strong -Wformat -Werror=format-security -MT mtest.o -MD -MP -MF $depbase.Tpo -c -o mtest.o ../../../cxx/util/mtest.cxx &&\\\n+## /usr/bin/mpicxx.openmpi -DHAVE_CONFIG_H -I. -I../../../cxx/util -I../../include -I../../include -I../../../include -Wdate-time -D_FORTIFY_SOURCE=2 -g -O2 -ffile-prefix-map=/build/mpi-testsuite-3.2+dfsg/2nd=. -fstack-protector-strong -Wformat -Werror=format-security -MT mtest.o -MD -MP -MF $depbase.Tpo -c -o mtest.o ../../../cxx/util/mtest.cxx &&\\\n ## mv -f $depbase.Tpo $depbase.Po\n ## ../../../cxx/util/mtest.cxx: In function 'void* MTestTypeContigInitRecv(MTestDatatype*)':\n ## ../../../cxx/util/mtest.cxx:189:42: warning: 'int MPI_Type_extent(MPI_Datatype, MPI_Aint*)' is deprecated: MPI_Type_extent is superseded by MPI_Type_get_extent in MPI-2.0 [-Wdeprecated-declarations]\n ## MPI_Type_extent( mtype->datatype, &size );\n ## ^\n ## In file included from ../../../cxx/util/mtest.cxx:7:\n ## /usr/lib/x86_64-linux-gnu/openmpi/include/mpi.h:1758:20: note: declared here\n@@ -999974,30 +1000057,30 @@\n ## ../../../cxx/util/mtest.cxx:848:50: error: 'NULL_COPY_FN' is not a member of 'MPI::Win'\n ## mem_keyval = MPI::Win::Create_keyval( MPI::Win::NULL_COPY_FN,\n ## ^~~~~~~~~~~~\n ## ../../../cxx/util/mtest.cxx:849:22: error: 'NULL_DELETE_FN' is not a member of 'MPI::Win'\n ## MPI::Win::NULL_DELETE_FN, 0 );\n ## ^~~~~~~~~~~~~~\n ## make[3]: *** [Makefile:416: mtest.o] Error 1\n-## make[3]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/util'\n+## make[3]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/util'\n ## make[2]: *** [Makefile:865: ../../cxx/util/mtest.o] Error 2\n-## make[2]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/coll'\n+## make[2]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/coll'\n not ok 668 - ./cxx/coll/uscan \n ---\n Directory: ./cxx/coll\n File: uscan\n Num-procs: \n- Date: \"Sun Jan 5 01:33:00 2020\"\n+ Date: \"Mon Dec 3 17:02:17 2018\"\n ...\n ## Test output (expected 'No Errors'):\n-## Failed to build uscan; make[2]: Entering directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/coll'\n+## Failed to build uscan; make[2]: Entering directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/coll'\n ## (cd ../../cxx/util && make mtest.o)\n-## make[3]: Entering directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/util'\n+## make[3]: Entering directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/util'\n ## depbase=`echo mtest.o | sed 's|[^/]*$|.deps/&|;s|\\.o$||'`;\\\n-## /usr/bin/mpicxx.openmpi -DHAVE_CONFIG_H -I. -I../../../cxx/util -I../../include -I../../include -I../../../include -Wdate-time -D_FORTIFY_SOURCE=2 -g -O2 -ffile-prefix-map=/build/1st/mpi-testsuite-3.2+dfsg=. -fstack-protector-strong -Wformat -Werror=format-security -MT mtest.o -MD -MP -MF $depbase.Tpo -c -o mtest.o ../../../cxx/util/mtest.cxx &&\\\n+## /usr/bin/mpicxx.openmpi -DHAVE_CONFIG_H -I. -I../../../cxx/util -I../../include -I../../include -I../../../include -Wdate-time -D_FORTIFY_SOURCE=2 -g -O2 -ffile-prefix-map=/build/mpi-testsuite-3.2+dfsg/2nd=. -fstack-protector-strong -Wformat -Werror=format-security -MT mtest.o -MD -MP -MF $depbase.Tpo -c -o mtest.o ../../../cxx/util/mtest.cxx &&\\\n ## mv -f $depbase.Tpo $depbase.Po\n ## ../../../cxx/util/mtest.cxx: In function 'void* MTestTypeContigInitRecv(MTestDatatype*)':\n ## ../../../cxx/util/mtest.cxx:189:42: warning: 'int MPI_Type_extent(MPI_Datatype, MPI_Aint*)' is deprecated: MPI_Type_extent is superseded by MPI_Type_get_extent in MPI-2.0 [-Wdeprecated-declarations]\n ## MPI_Type_extent( mtype->datatype, &size );\n ## ^\n ## In file included from ../../../cxx/util/mtest.cxx:7:\n ## /usr/lib/x86_64-linux-gnu/openmpi/include/mpi.h:1758:20: note: declared here\n@@ -1000029,30 +1000112,30 @@\n ## ../../../cxx/util/mtest.cxx:848:50: error: 'NULL_COPY_FN' is not a member of 'MPI::Win'\n ## mem_keyval = MPI::Win::Create_keyval( MPI::Win::NULL_COPY_FN,\n ## ^~~~~~~~~~~~\n ## ../../../cxx/util/mtest.cxx:849:22: error: 'NULL_DELETE_FN' is not a member of 'MPI::Win'\n ## MPI::Win::NULL_DELETE_FN, 0 );\n ## ^~~~~~~~~~~~~~\n ## make[3]: *** [Makefile:416: mtest.o] Error 1\n-## make[3]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/util'\n+## make[3]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/util'\n ## make[2]: *** [Makefile:865: ../../cxx/util/mtest.o] Error 2\n-## make[2]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/coll'\n+## make[2]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/coll'\n not ok 669 - ./cxx/coll/uexscan \n ---\n Directory: ./cxx/coll\n File: uexscan\n Num-procs: \n- Date: \"Sun Jan 5 01:33:01 2020\"\n+ Date: \"Mon Dec 3 17:02:23 2018\"\n ...\n ## Test output (expected 'No Errors'):\n-## Failed to build uexscan; make[2]: Entering directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/coll'\n+## Failed to build uexscan; make[2]: Entering directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/coll'\n ## (cd ../../cxx/util && make mtest.o)\n-## make[3]: Entering directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/util'\n+## make[3]: Entering directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/util'\n ## depbase=`echo mtest.o | sed 's|[^/]*$|.deps/&|;s|\\.o$||'`;\\\n-## /usr/bin/mpicxx.openmpi -DHAVE_CONFIG_H -I. -I../../../cxx/util -I../../include -I../../include -I../../../include -Wdate-time -D_FORTIFY_SOURCE=2 -g -O2 -ffile-prefix-map=/build/1st/mpi-testsuite-3.2+dfsg=. -fstack-protector-strong -Wformat -Werror=format-security -MT mtest.o -MD -MP -MF $depbase.Tpo -c -o mtest.o ../../../cxx/util/mtest.cxx &&\\\n+## /usr/bin/mpicxx.openmpi -DHAVE_CONFIG_H -I. -I../../../cxx/util -I../../include -I../../include -I../../../include -Wdate-time -D_FORTIFY_SOURCE=2 -g -O2 -ffile-prefix-map=/build/mpi-testsuite-3.2+dfsg/2nd=. -fstack-protector-strong -Wformat -Werror=format-security -MT mtest.o -MD -MP -MF $depbase.Tpo -c -o mtest.o ../../../cxx/util/mtest.cxx &&\\\n ## mv -f $depbase.Tpo $depbase.Po\n ## ../../../cxx/util/mtest.cxx: In function 'void* MTestTypeContigInitRecv(MTestDatatype*)':\n ## ../../../cxx/util/mtest.cxx:189:42: warning: 'int MPI_Type_extent(MPI_Datatype, MPI_Aint*)' is deprecated: MPI_Type_extent is superseded by MPI_Type_get_extent in MPI-2.0 [-Wdeprecated-declarations]\n ## MPI_Type_extent( mtype->datatype, &size );\n ## ^\n ## In file included from ../../../cxx/util/mtest.cxx:7:\n ## /usr/lib/x86_64-linux-gnu/openmpi/include/mpi.h:1758:20: note: declared here\n@@ -1000084,30 +1000167,30 @@\n ## ../../../cxx/util/mtest.cxx:848:50: error: 'NULL_COPY_FN' is not a member of 'MPI::Win'\n ## mem_keyval = MPI::Win::Create_keyval( MPI::Win::NULL_COPY_FN,\n ## ^~~~~~~~~~~~\n ## ../../../cxx/util/mtest.cxx:849:22: error: 'NULL_DELETE_FN' is not a member of 'MPI::Win'\n ## MPI::Win::NULL_DELETE_FN, 0 );\n ## ^~~~~~~~~~~~~~\n ## make[3]: *** [Makefile:416: mtest.o] Error 1\n-## make[3]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/util'\n+## make[3]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/util'\n ## make[2]: *** [Makefile:865: ../../cxx/util/mtest.o] Error 2\n-## make[2]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/coll'\n+## make[2]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/coll'\n not ok 670 - ./cxx/coll/alltoallw2x \n ---\n Directory: ./cxx/coll\n File: alltoallw2x\n Num-procs: \n- Date: \"Sun Jan 5 01:33:03 2020\"\n+ Date: \"Mon Dec 3 17:02:28 2018\"\n ...\n ## Test output (expected 'No Errors'):\n-## Failed to build alltoallw2x; make[2]: Entering directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/coll'\n+## Failed to build alltoallw2x; make[2]: Entering directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/coll'\n ## (cd ../../cxx/util && make mtest.o)\n-## make[3]: Entering directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/util'\n+## make[3]: Entering directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/util'\n ## depbase=`echo mtest.o | sed 's|[^/]*$|.deps/&|;s|\\.o$||'`;\\\n-## /usr/bin/mpicxx.openmpi -DHAVE_CONFIG_H -I. -I../../../cxx/util -I../../include -I../../include -I../../../include -Wdate-time -D_FORTIFY_SOURCE=2 -g -O2 -ffile-prefix-map=/build/1st/mpi-testsuite-3.2+dfsg=. -fstack-protector-strong -Wformat -Werror=format-security -MT mtest.o -MD -MP -MF $depbase.Tpo -c -o mtest.o ../../../cxx/util/mtest.cxx &&\\\n+## /usr/bin/mpicxx.openmpi -DHAVE_CONFIG_H -I. -I../../../cxx/util -I../../include -I../../include -I../../../include -Wdate-time -D_FORTIFY_SOURCE=2 -g -O2 -ffile-prefix-map=/build/mpi-testsuite-3.2+dfsg/2nd=. -fstack-protector-strong -Wformat -Werror=format-security -MT mtest.o -MD -MP -MF $depbase.Tpo -c -o mtest.o ../../../cxx/util/mtest.cxx &&\\\n ## mv -f $depbase.Tpo $depbase.Po\n ## ../../../cxx/util/mtest.cxx: In function 'void* MTestTypeContigInitRecv(MTestDatatype*)':\n ## ../../../cxx/util/mtest.cxx:189:42: warning: 'int MPI_Type_extent(MPI_Datatype, MPI_Aint*)' is deprecated: MPI_Type_extent is superseded by MPI_Type_get_extent in MPI-2.0 [-Wdeprecated-declarations]\n ## MPI_Type_extent( mtype->datatype, &size );\n ## ^\n ## In file included from ../../../cxx/util/mtest.cxx:7:\n ## /usr/lib/x86_64-linux-gnu/openmpi/include/mpi.h:1758:20: note: declared here\n@@ -1000139,30 +1000222,30 @@\n ## ../../../cxx/util/mtest.cxx:848:50: error: 'NULL_COPY_FN' is not a member of 'MPI::Win'\n ## mem_keyval = MPI::Win::Create_keyval( MPI::Win::NULL_COPY_FN,\n ## ^~~~~~~~~~~~\n ## ../../../cxx/util/mtest.cxx:849:22: error: 'NULL_DELETE_FN' is not a member of 'MPI::Win'\n ## MPI::Win::NULL_DELETE_FN, 0 );\n ## ^~~~~~~~~~~~~~\n ## make[3]: *** [Makefile:416: mtest.o] Error 1\n-## make[3]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/util'\n+## make[3]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/util'\n ## make[2]: *** [Makefile:865: ../../cxx/util/mtest.o] Error 2\n-## make[2]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/coll'\n+## make[2]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/coll'\n not ok 671 - ./cxx/coll/icbcastx \n ---\n Directory: ./cxx/coll\n File: icbcastx\n Num-procs: \n- Date: \"Sun Jan 5 01:33:04 2020\"\n+ Date: \"Mon Dec 3 17:02:33 2018\"\n ...\n ## Test output (expected 'No Errors'):\n-## Failed to build icbcastx; make[2]: Entering directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/coll'\n+## Failed to build icbcastx; make[2]: Entering directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/coll'\n ## (cd ../../cxx/util && make mtest.o)\n-## make[3]: Entering directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/util'\n+## make[3]: Entering directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/util'\n ## depbase=`echo mtest.o | sed 's|[^/]*$|.deps/&|;s|\\.o$||'`;\\\n-## /usr/bin/mpicxx.openmpi -DHAVE_CONFIG_H -I. -I../../../cxx/util -I../../include -I../../include -I../../../include -Wdate-time -D_FORTIFY_SOURCE=2 -g -O2 -ffile-prefix-map=/build/1st/mpi-testsuite-3.2+dfsg=. -fstack-protector-strong -Wformat -Werror=format-security -MT mtest.o -MD -MP -MF $depbase.Tpo -c -o mtest.o ../../../cxx/util/mtest.cxx &&\\\n+## /usr/bin/mpicxx.openmpi -DHAVE_CONFIG_H -I. -I../../../cxx/util -I../../include -I../../include -I../../../include -Wdate-time -D_FORTIFY_SOURCE=2 -g -O2 -ffile-prefix-map=/build/mpi-testsuite-3.2+dfsg/2nd=. -fstack-protector-strong -Wformat -Werror=format-security -MT mtest.o -MD -MP -MF $depbase.Tpo -c -o mtest.o ../../../cxx/util/mtest.cxx &&\\\n ## mv -f $depbase.Tpo $depbase.Po\n ## ../../../cxx/util/mtest.cxx: In function 'void* MTestTypeContigInitRecv(MTestDatatype*)':\n ## ../../../cxx/util/mtest.cxx:189:42: warning: 'int MPI_Type_extent(MPI_Datatype, MPI_Aint*)' is deprecated: MPI_Type_extent is superseded by MPI_Type_get_extent in MPI-2.0 [-Wdeprecated-declarations]\n ## MPI_Type_extent( mtype->datatype, &size );\n ## ^\n ## In file included from ../../../cxx/util/mtest.cxx:7:\n ## /usr/lib/x86_64-linux-gnu/openmpi/include/mpi.h:1758:20: note: declared here\n@@ -1000194,33 +1000277,33 @@\n ## ../../../cxx/util/mtest.cxx:848:50: error: 'NULL_COPY_FN' is not a member of 'MPI::Win'\n ## mem_keyval = MPI::Win::Create_keyval( MPI::Win::NULL_COPY_FN,\n ## ^~~~~~~~~~~~\n ## ../../../cxx/util/mtest.cxx:849:22: error: 'NULL_DELETE_FN' is not a member of 'MPI::Win'\n ## MPI::Win::NULL_DELETE_FN, 0 );\n ## ^~~~~~~~~~~~~~\n ## make[3]: *** [Makefile:416: mtest.o] Error 1\n-## make[3]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/util'\n+## make[3]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/util'\n ## make[2]: *** [Makefile:865: ../../cxx/util/mtest.o] Error 2\n-## make[2]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/coll'\n+## make[2]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/coll'\n not ok 672 - ./cxx/coll/icbcastx \n ---\n Directory: ./cxx/coll\n File: icbcastx\n Num-procs: \n- Date: \"Sun Jan 5 01:33:07 2020\"\n+ Date: \"Mon Dec 3 17:02:48 2018\"\n ...\n ## Test output (expected 'No Errors'):\n-## Failed to build icbcastx; make[2]: Entering directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/coll'\n+## Failed to build icbcastx; make[2]: Entering directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/coll'\n ## depbase=`echo icbcastx.o | sed 's|[^/]*$|.deps/&|;s|\\.o$||'`;\\\n-## /usr/bin/mpicxx.openmpi -DHAVE_CONFIG_H -I. -I../../../cxx/coll -I../../include -I../../include -I../../../include -Wdate-time -D_FORTIFY_SOURCE=2 -g -O2 -ffile-prefix-map=/build/1st/mpi-testsuite-3.2+dfsg=. -fstack-protector-strong -Wformat -Werror=format-security -MT icbcastx.o -MD -MP -MF $depbase.Tpo -c -o icbcastx.o ../../../cxx/coll/icbcastx.cxx &&\\\n+## /usr/bin/mpicxx.openmpi -DHAVE_CONFIG_H -I. -I../../../cxx/coll -I../../include -I../../include -I../../../include -Wdate-time -D_FORTIFY_SOURCE=2 -g -O2 -ffile-prefix-map=/build/mpi-testsuite-3.2+dfsg/2nd=. -fstack-protector-strong -Wformat -Werror=format-security -MT icbcastx.o -MD -MP -MF $depbase.Tpo -c -o icbcastx.o ../../../cxx/coll/icbcastx.cxx &&\\\n ## mv -f $depbase.Tpo $depbase.Po\n ## (cd ../../cxx/util && make mtest.o)\n-## make[3]: Entering directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/util'\n+## make[3]: Entering directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/util'\n ## depbase=`echo mtest.o | sed 's|[^/]*$|.deps/&|;s|\\.o$||'`;\\\n-## /usr/bin/mpicxx.openmpi -DHAVE_CONFIG_H -I. -I../../../cxx/util -I../../include -I../../include -I../../../include -Wdate-time -D_FORTIFY_SOURCE=2 -g -O2 -ffile-prefix-map=/build/1st/mpi-testsuite-3.2+dfsg=. -fstack-protector-strong -Wformat -Werror=format-security -MT mtest.o -MD -MP -MF $depbase.Tpo -c -o mtest.o ../../../cxx/util/mtest.cxx &&\\\n+## /usr/bin/mpicxx.openmpi -DHAVE_CONFIG_H -I. -I../../../cxx/util -I../../include -I../../include -I../../../include -Wdate-time -D_FORTIFY_SOURCE=2 -g -O2 -ffile-prefix-map=/build/mpi-testsuite-3.2+dfsg/2nd=. -fstack-protector-strong -Wformat -Werror=format-security -MT mtest.o -MD -MP -MF $depbase.Tpo -c -o mtest.o ../../../cxx/util/mtest.cxx &&\\\n ## mv -f $depbase.Tpo $depbase.Po\n ## ../../../cxx/util/mtest.cxx: In function 'void* MTestTypeContigInitRecv(MTestDatatype*)':\n ## ../../../cxx/util/mtest.cxx:189:42: warning: 'int MPI_Type_extent(MPI_Datatype, MPI_Aint*)' is deprecated: MPI_Type_extent is superseded by MPI_Type_get_extent in MPI-2.0 [-Wdeprecated-declarations]\n ## MPI_Type_extent( mtype->datatype, &size );\n ## ^\n ## In file included from ../../../cxx/util/mtest.cxx:7:\n ## /usr/lib/x86_64-linux-gnu/openmpi/include/mpi.h:1758:20: note: declared here\n@@ -1000252,30 +1000335,30 @@\n ## ../../../cxx/util/mtest.cxx:848:50: error: 'NULL_COPY_FN' is not a member of 'MPI::Win'\n ## mem_keyval = MPI::Win::Create_keyval( MPI::Win::NULL_COPY_FN,\n ## ^~~~~~~~~~~~\n ## ../../../cxx/util/mtest.cxx:849:22: error: 'NULL_DELETE_FN' is not a member of 'MPI::Win'\n ## MPI::Win::NULL_DELETE_FN, 0 );\n ## ^~~~~~~~~~~~~~\n ## make[3]: *** [Makefile:416: mtest.o] Error 1\n-## make[3]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/util'\n+## make[3]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/util'\n ## make[2]: *** [Makefile:865: ../../cxx/util/mtest.o] Error 2\n-## make[2]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/coll'\n+## make[2]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/coll'\n not ok 673 - ./cxx/coll/icallreducex \n ---\n Directory: ./cxx/coll\n File: icallreducex\n Num-procs: \n- Date: \"Sun Jan 5 01:33:08 2020\"\n+ Date: \"Mon Dec 3 17:02:54 2018\"\n ...\n ## Test output (expected 'No Errors'):\n-## Failed to build icallreducex; make[2]: Entering directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/coll'\n+## Failed to build icallreducex; make[2]: Entering directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/coll'\n ## (cd ../../cxx/util && make mtest.o)\n-## make[3]: Entering directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/util'\n+## make[3]: Entering directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/util'\n ## depbase=`echo mtest.o | sed 's|[^/]*$|.deps/&|;s|\\.o$||'`;\\\n-## /usr/bin/mpicxx.openmpi -DHAVE_CONFIG_H -I. -I../../../cxx/util -I../../include -I../../include -I../../../include -Wdate-time -D_FORTIFY_SOURCE=2 -g -O2 -ffile-prefix-map=/build/1st/mpi-testsuite-3.2+dfsg=. -fstack-protector-strong -Wformat -Werror=format-security -MT mtest.o -MD -MP -MF $depbase.Tpo -c -o mtest.o ../../../cxx/util/mtest.cxx &&\\\n+## /usr/bin/mpicxx.openmpi -DHAVE_CONFIG_H -I. -I../../../cxx/util -I../../include -I../../include -I../../../include -Wdate-time -D_FORTIFY_SOURCE=2 -g -O2 -ffile-prefix-map=/build/mpi-testsuite-3.2+dfsg/2nd=. -fstack-protector-strong -Wformat -Werror=format-security -MT mtest.o -MD -MP -MF $depbase.Tpo -c -o mtest.o ../../../cxx/util/mtest.cxx &&\\\n ## mv -f $depbase.Tpo $depbase.Po\n ## ../../../cxx/util/mtest.cxx: In function 'void* MTestTypeContigInitRecv(MTestDatatype*)':\n ## ../../../cxx/util/mtest.cxx:189:42: warning: 'int MPI_Type_extent(MPI_Datatype, MPI_Aint*)' is deprecated: MPI_Type_extent is superseded by MPI_Type_get_extent in MPI-2.0 [-Wdeprecated-declarations]\n ## MPI_Type_extent( mtype->datatype, &size );\n ## ^\n ## In file included from ../../../cxx/util/mtest.cxx:7:\n ## /usr/lib/x86_64-linux-gnu/openmpi/include/mpi.h:1758:20: note: declared here\n@@ -1000307,30 +1000390,30 @@\n ## ../../../cxx/util/mtest.cxx:848:50: error: 'NULL_COPY_FN' is not a member of 'MPI::Win'\n ## mem_keyval = MPI::Win::Create_keyval( MPI::Win::NULL_COPY_FN,\n ## ^~~~~~~~~~~~\n ## ../../../cxx/util/mtest.cxx:849:22: error: 'NULL_DELETE_FN' is not a member of 'MPI::Win'\n ## MPI::Win::NULL_DELETE_FN, 0 );\n ## ^~~~~~~~~~~~~~\n ## make[3]: *** [Makefile:416: mtest.o] Error 1\n-## make[3]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/util'\n+## make[3]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/util'\n ## make[2]: *** [Makefile:865: ../../cxx/util/mtest.o] Error 2\n-## make[2]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/coll'\n+## make[2]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/coll'\n not ok 674 - ./cxx/coll/icreducex \n ---\n Directory: ./cxx/coll\n File: icreducex\n Num-procs: \n- Date: \"Sun Jan 5 01:33:09 2020\"\n+ Date: \"Mon Dec 3 17:03:00 2018\"\n ...\n ## Test output (expected 'No Errors'):\n-## Failed to build icreducex; make[2]: Entering directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/coll'\n+## Failed to build icreducex; make[2]: Entering directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/coll'\n ## (cd ../../cxx/util && make mtest.o)\n-## make[3]: Entering directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/util'\n+## make[3]: Entering directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/util'\n ## depbase=`echo mtest.o | sed 's|[^/]*$|.deps/&|;s|\\.o$||'`;\\\n-## /usr/bin/mpicxx.openmpi -DHAVE_CONFIG_H -I. -I../../../cxx/util -I../../include -I../../include -I../../../include -Wdate-time -D_FORTIFY_SOURCE=2 -g -O2 -ffile-prefix-map=/build/1st/mpi-testsuite-3.2+dfsg=. -fstack-protector-strong -Wformat -Werror=format-security -MT mtest.o -MD -MP -MF $depbase.Tpo -c -o mtest.o ../../../cxx/util/mtest.cxx &&\\\n+## /usr/bin/mpicxx.openmpi -DHAVE_CONFIG_H -I. -I../../../cxx/util -I../../include -I../../include -I../../../include -Wdate-time -D_FORTIFY_SOURCE=2 -g -O2 -ffile-prefix-map=/build/mpi-testsuite-3.2+dfsg/2nd=. -fstack-protector-strong -Wformat -Werror=format-security -MT mtest.o -MD -MP -MF $depbase.Tpo -c -o mtest.o ../../../cxx/util/mtest.cxx &&\\\n ## mv -f $depbase.Tpo $depbase.Po\n ## ../../../cxx/util/mtest.cxx: In function 'void* MTestTypeContigInitRecv(MTestDatatype*)':\n ## ../../../cxx/util/mtest.cxx:189:42: warning: 'int MPI_Type_extent(MPI_Datatype, MPI_Aint*)' is deprecated: MPI_Type_extent is superseded by MPI_Type_get_extent in MPI-2.0 [-Wdeprecated-declarations]\n ## MPI_Type_extent( mtype->datatype, &size );\n ## ^\n ## In file included from ../../../cxx/util/mtest.cxx:7:\n ## /usr/lib/x86_64-linux-gnu/openmpi/include/mpi.h:1758:20: note: declared here\n@@ -1000362,30 +1000445,30 @@\n ## ../../../cxx/util/mtest.cxx:848:50: error: 'NULL_COPY_FN' is not a member of 'MPI::Win'\n ## mem_keyval = MPI::Win::Create_keyval( MPI::Win::NULL_COPY_FN,\n ## ^~~~~~~~~~~~\n ## ../../../cxx/util/mtest.cxx:849:22: error: 'NULL_DELETE_FN' is not a member of 'MPI::Win'\n ## MPI::Win::NULL_DELETE_FN, 0 );\n ## ^~~~~~~~~~~~~~\n ## make[3]: *** [Makefile:416: mtest.o] Error 1\n-## make[3]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/util'\n+## make[3]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/util'\n ## make[2]: *** [Makefile:865: ../../cxx/util/mtest.o] Error 2\n-## make[2]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/coll'\n+## make[2]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/coll'\n not ok 675 - ./cxx/coll/icscatterx \n ---\n Directory: ./cxx/coll\n File: icscatterx\n Num-procs: \n- Date: \"Sun Jan 5 01:33:10 2020\"\n+ Date: \"Mon Dec 3 17:03:06 2018\"\n ...\n ## Test output (expected 'No Errors'):\n-## Failed to build icscatterx; make[2]: Entering directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/coll'\n+## Failed to build icscatterx; make[2]: Entering directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/coll'\n ## (cd ../../cxx/util && make mtest.o)\n-## make[3]: Entering directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/util'\n+## make[3]: Entering directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/util'\n ## depbase=`echo mtest.o | sed 's|[^/]*$|.deps/&|;s|\\.o$||'`;\\\n-## /usr/bin/mpicxx.openmpi -DHAVE_CONFIG_H -I. -I../../../cxx/util -I../../include -I../../include -I../../../include -Wdate-time -D_FORTIFY_SOURCE=2 -g -O2 -ffile-prefix-map=/build/1st/mpi-testsuite-3.2+dfsg=. -fstack-protector-strong -Wformat -Werror=format-security -MT mtest.o -MD -MP -MF $depbase.Tpo -c -o mtest.o ../../../cxx/util/mtest.cxx &&\\\n+## /usr/bin/mpicxx.openmpi -DHAVE_CONFIG_H -I. -I../../../cxx/util -I../../include -I../../include -I../../../include -Wdate-time -D_FORTIFY_SOURCE=2 -g -O2 -ffile-prefix-map=/build/mpi-testsuite-3.2+dfsg/2nd=. -fstack-protector-strong -Wformat -Werror=format-security -MT mtest.o -MD -MP -MF $depbase.Tpo -c -o mtest.o ../../../cxx/util/mtest.cxx &&\\\n ## mv -f $depbase.Tpo $depbase.Po\n ## ../../../cxx/util/mtest.cxx: In function 'void* MTestTypeContigInitRecv(MTestDatatype*)':\n ## ../../../cxx/util/mtest.cxx:189:42: warning: 'int MPI_Type_extent(MPI_Datatype, MPI_Aint*)' is deprecated: MPI_Type_extent is superseded by MPI_Type_get_extent in MPI-2.0 [-Wdeprecated-declarations]\n ## MPI_Type_extent( mtype->datatype, &size );\n ## ^\n ## In file included from ../../../cxx/util/mtest.cxx:7:\n ## /usr/lib/x86_64-linux-gnu/openmpi/include/mpi.h:1758:20: note: declared here\n@@ -1000417,30 +1000500,30 @@\n ## ../../../cxx/util/mtest.cxx:848:50: error: 'NULL_COPY_FN' is not a member of 'MPI::Win'\n ## mem_keyval = MPI::Win::Create_keyval( MPI::Win::NULL_COPY_FN,\n ## ^~~~~~~~~~~~\n ## ../../../cxx/util/mtest.cxx:849:22: error: 'NULL_DELETE_FN' is not a member of 'MPI::Win'\n ## MPI::Win::NULL_DELETE_FN, 0 );\n ## ^~~~~~~~~~~~~~\n ## make[3]: *** [Makefile:416: mtest.o] Error 1\n-## make[3]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/util'\n+## make[3]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/util'\n ## make[2]: *** [Makefile:865: ../../cxx/util/mtest.o] Error 2\n-## make[2]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/coll'\n+## make[2]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/coll'\n not ok 676 - ./cxx/coll/icgatherx \n ---\n Directory: ./cxx/coll\n File: icgatherx\n Num-procs: \n- Date: \"Sun Jan 5 01:33:11 2020\"\n+ Date: \"Mon Dec 3 17:03:12 2018\"\n ...\n ## Test output (expected 'No Errors'):\n-## Failed to build icgatherx; make[2]: Entering directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/coll'\n+## Failed to build icgatherx; make[2]: Entering directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/coll'\n ## (cd ../../cxx/util && make mtest.o)\n-## make[3]: Entering directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/util'\n+## make[3]: Entering directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/util'\n ## depbase=`echo mtest.o | sed 's|[^/]*$|.deps/&|;s|\\.o$||'`;\\\n-## /usr/bin/mpicxx.openmpi -DHAVE_CONFIG_H -I. -I../../../cxx/util -I../../include -I../../include -I../../../include -Wdate-time -D_FORTIFY_SOURCE=2 -g -O2 -ffile-prefix-map=/build/1st/mpi-testsuite-3.2+dfsg=. -fstack-protector-strong -Wformat -Werror=format-security -MT mtest.o -MD -MP -MF $depbase.Tpo -c -o mtest.o ../../../cxx/util/mtest.cxx &&\\\n+## /usr/bin/mpicxx.openmpi -DHAVE_CONFIG_H -I. -I../../../cxx/util -I../../include -I../../include -I../../../include -Wdate-time -D_FORTIFY_SOURCE=2 -g -O2 -ffile-prefix-map=/build/mpi-testsuite-3.2+dfsg/2nd=. -fstack-protector-strong -Wformat -Werror=format-security -MT mtest.o -MD -MP -MF $depbase.Tpo -c -o mtest.o ../../../cxx/util/mtest.cxx &&\\\n ## mv -f $depbase.Tpo $depbase.Po\n ## ../../../cxx/util/mtest.cxx: In function 'void* MTestTypeContigInitRecv(MTestDatatype*)':\n ## ../../../cxx/util/mtest.cxx:189:42: warning: 'int MPI_Type_extent(MPI_Datatype, MPI_Aint*)' is deprecated: MPI_Type_extent is superseded by MPI_Type_get_extent in MPI-2.0 [-Wdeprecated-declarations]\n ## MPI_Type_extent( mtype->datatype, &size );\n ## ^\n ## In file included from ../../../cxx/util/mtest.cxx:7:\n ## /usr/lib/x86_64-linux-gnu/openmpi/include/mpi.h:1758:20: note: declared here\n@@ -1000472,30 +1000555,30 @@\n ## ../../../cxx/util/mtest.cxx:848:50: error: 'NULL_COPY_FN' is not a member of 'MPI::Win'\n ## mem_keyval = MPI::Win::Create_keyval( MPI::Win::NULL_COPY_FN,\n ## ^~~~~~~~~~~~\n ## ../../../cxx/util/mtest.cxx:849:22: error: 'NULL_DELETE_FN' is not a member of 'MPI::Win'\n ## MPI::Win::NULL_DELETE_FN, 0 );\n ## ^~~~~~~~~~~~~~\n ## make[3]: *** [Makefile:416: mtest.o] Error 1\n-## make[3]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/util'\n+## make[3]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/util'\n ## make[2]: *** [Makefile:865: ../../cxx/util/mtest.o] Error 2\n-## make[2]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/coll'\n+## make[2]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/coll'\n not ok 677 - ./cxx/coll/icallgatherx \n ---\n Directory: ./cxx/coll\n File: icallgatherx\n Num-procs: \n- Date: \"Sun Jan 5 01:33:12 2020\"\n+ Date: \"Mon Dec 3 17:03:18 2018\"\n ...\n ## Test output (expected 'No Errors'):\n-## Failed to build icallgatherx; make[2]: Entering directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/coll'\n+## Failed to build icallgatherx; make[2]: Entering directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/coll'\n ## (cd ../../cxx/util && make mtest.o)\n-## make[3]: Entering directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/util'\n+## make[3]: Entering directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/util'\n ## depbase=`echo mtest.o | sed 's|[^/]*$|.deps/&|;s|\\.o$||'`;\\\n-## /usr/bin/mpicxx.openmpi -DHAVE_CONFIG_H -I. -I../../../cxx/util -I../../include -I../../include -I../../../include -Wdate-time -D_FORTIFY_SOURCE=2 -g -O2 -ffile-prefix-map=/build/1st/mpi-testsuite-3.2+dfsg=. -fstack-protector-strong -Wformat -Werror=format-security -MT mtest.o -MD -MP -MF $depbase.Tpo -c -o mtest.o ../../../cxx/util/mtest.cxx &&\\\n+## /usr/bin/mpicxx.openmpi -DHAVE_CONFIG_H -I. -I../../../cxx/util -I../../include -I../../include -I../../../include -Wdate-time -D_FORTIFY_SOURCE=2 -g -O2 -ffile-prefix-map=/build/mpi-testsuite-3.2+dfsg/2nd=. -fstack-protector-strong -Wformat -Werror=format-security -MT mtest.o -MD -MP -MF $depbase.Tpo -c -o mtest.o ../../../cxx/util/mtest.cxx &&\\\n ## mv -f $depbase.Tpo $depbase.Po\n ## ../../../cxx/util/mtest.cxx: In function 'void* MTestTypeContigInitRecv(MTestDatatype*)':\n ## ../../../cxx/util/mtest.cxx:189:42: warning: 'int MPI_Type_extent(MPI_Datatype, MPI_Aint*)' is deprecated: MPI_Type_extent is superseded by MPI_Type_get_extent in MPI-2.0 [-Wdeprecated-declarations]\n ## MPI_Type_extent( mtype->datatype, &size );\n ## ^\n ## In file included from ../../../cxx/util/mtest.cxx:7:\n ## /usr/lib/x86_64-linux-gnu/openmpi/include/mpi.h:1758:20: note: declared here\n@@ -1000527,30 +1000610,30 @@\n ## ../../../cxx/util/mtest.cxx:848:50: error: 'NULL_COPY_FN' is not a member of 'MPI::Win'\n ## mem_keyval = MPI::Win::Create_keyval( MPI::Win::NULL_COPY_FN,\n ## ^~~~~~~~~~~~\n ## ../../../cxx/util/mtest.cxx:849:22: error: 'NULL_DELETE_FN' is not a member of 'MPI::Win'\n ## MPI::Win::NULL_DELETE_FN, 0 );\n ## ^~~~~~~~~~~~~~\n ## make[3]: *** [Makefile:416: mtest.o] Error 1\n-## make[3]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/util'\n+## make[3]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/util'\n ## make[2]: *** [Makefile:865: ../../cxx/util/mtest.o] Error 2\n-## make[2]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/coll'\n+## make[2]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/coll'\n not ok 678 - ./cxx/coll/icbarrierx \n ---\n Directory: ./cxx/coll\n File: icbarrierx\n Num-procs: \n- Date: \"Sun Jan 5 01:33:13 2020\"\n+ Date: \"Mon Dec 3 17:03:22 2018\"\n ...\n ## Test output (expected 'No Errors'):\n-## Failed to build icbarrierx; make[2]: Entering directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/coll'\n+## Failed to build icbarrierx; make[2]: Entering directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/coll'\n ## (cd ../../cxx/util && make mtest.o)\n-## make[3]: Entering directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/util'\n+## make[3]: Entering directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/util'\n ## depbase=`echo mtest.o | sed 's|[^/]*$|.deps/&|;s|\\.o$||'`;\\\n-## /usr/bin/mpicxx.openmpi -DHAVE_CONFIG_H -I. -I../../../cxx/util -I../../include -I../../include -I../../../include -Wdate-time -D_FORTIFY_SOURCE=2 -g -O2 -ffile-prefix-map=/build/1st/mpi-testsuite-3.2+dfsg=. -fstack-protector-strong -Wformat -Werror=format-security -MT mtest.o -MD -MP -MF $depbase.Tpo -c -o mtest.o ../../../cxx/util/mtest.cxx &&\\\n+## /usr/bin/mpicxx.openmpi -DHAVE_CONFIG_H -I. -I../../../cxx/util -I../../include -I../../include -I../../../include -Wdate-time -D_FORTIFY_SOURCE=2 -g -O2 -ffile-prefix-map=/build/mpi-testsuite-3.2+dfsg/2nd=. -fstack-protector-strong -Wformat -Werror=format-security -MT mtest.o -MD -MP -MF $depbase.Tpo -c -o mtest.o ../../../cxx/util/mtest.cxx &&\\\n ## mv -f $depbase.Tpo $depbase.Po\n ## ../../../cxx/util/mtest.cxx: In function 'void* MTestTypeContigInitRecv(MTestDatatype*)':\n ## ../../../cxx/util/mtest.cxx:189:42: warning: 'int MPI_Type_extent(MPI_Datatype, MPI_Aint*)' is deprecated: MPI_Type_extent is superseded by MPI_Type_get_extent in MPI-2.0 [-Wdeprecated-declarations]\n ## MPI_Type_extent( mtype->datatype, &size );\n ## ^\n ## In file included from ../../../cxx/util/mtest.cxx:7:\n ## /usr/lib/x86_64-linux-gnu/openmpi/include/mpi.h:1758:20: note: declared here\n@@ -1000582,30 +1000665,30 @@\n ## ../../../cxx/util/mtest.cxx:848:50: error: 'NULL_COPY_FN' is not a member of 'MPI::Win'\n ## mem_keyval = MPI::Win::Create_keyval( MPI::Win::NULL_COPY_FN,\n ## ^~~~~~~~~~~~\n ## ../../../cxx/util/mtest.cxx:849:22: error: 'NULL_DELETE_FN' is not a member of 'MPI::Win'\n ## MPI::Win::NULL_DELETE_FN, 0 );\n ## ^~~~~~~~~~~~~~\n ## make[3]: *** [Makefile:416: mtest.o] Error 1\n-## make[3]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/util'\n+## make[3]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/util'\n ## make[2]: *** [Makefile:865: ../../cxx/util/mtest.o] Error 2\n-## make[2]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/coll'\n+## make[2]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/coll'\n not ok 679 - ./cxx/coll/icallgathervx \n ---\n Directory: ./cxx/coll\n File: icallgathervx\n Num-procs: \n- Date: \"Sun Jan 5 01:33:15 2020\"\n+ Date: \"Mon Dec 3 17:03:27 2018\"\n ...\n ## Test output (expected 'No Errors'):\n-## Failed to build icallgathervx; make[2]: Entering directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/coll'\n+## Failed to build icallgathervx; make[2]: Entering directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/coll'\n ## (cd ../../cxx/util && make mtest.o)\n-## make[3]: Entering directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/util'\n+## make[3]: Entering directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/util'\n ## depbase=`echo mtest.o | sed 's|[^/]*$|.deps/&|;s|\\.o$||'`;\\\n-## /usr/bin/mpicxx.openmpi -DHAVE_CONFIG_H -I. -I../../../cxx/util -I../../include -I../../include -I../../../include -Wdate-time -D_FORTIFY_SOURCE=2 -g -O2 -ffile-prefix-map=/build/1st/mpi-testsuite-3.2+dfsg=. -fstack-protector-strong -Wformat -Werror=format-security -MT mtest.o -MD -MP -MF $depbase.Tpo -c -o mtest.o ../../../cxx/util/mtest.cxx &&\\\n+## /usr/bin/mpicxx.openmpi -DHAVE_CONFIG_H -I. -I../../../cxx/util -I../../include -I../../include -I../../../include -Wdate-time -D_FORTIFY_SOURCE=2 -g -O2 -ffile-prefix-map=/build/mpi-testsuite-3.2+dfsg/2nd=. -fstack-protector-strong -Wformat -Werror=format-security -MT mtest.o -MD -MP -MF $depbase.Tpo -c -o mtest.o ../../../cxx/util/mtest.cxx &&\\\n ## mv -f $depbase.Tpo $depbase.Po\n ## ../../../cxx/util/mtest.cxx: In function 'void* MTestTypeContigInitRecv(MTestDatatype*)':\n ## ../../../cxx/util/mtest.cxx:189:42: warning: 'int MPI_Type_extent(MPI_Datatype, MPI_Aint*)' is deprecated: MPI_Type_extent is superseded by MPI_Type_get_extent in MPI-2.0 [-Wdeprecated-declarations]\n ## MPI_Type_extent( mtype->datatype, &size );\n ## ^\n ## In file included from ../../../cxx/util/mtest.cxx:7:\n ## /usr/lib/x86_64-linux-gnu/openmpi/include/mpi.h:1758:20: note: declared here\n@@ -1000637,30 +1000720,30 @@\n ## ../../../cxx/util/mtest.cxx:848:50: error: 'NULL_COPY_FN' is not a member of 'MPI::Win'\n ## mem_keyval = MPI::Win::Create_keyval( MPI::Win::NULL_COPY_FN,\n ## ^~~~~~~~~~~~\n ## ../../../cxx/util/mtest.cxx:849:22: error: 'NULL_DELETE_FN' is not a member of 'MPI::Win'\n ## MPI::Win::NULL_DELETE_FN, 0 );\n ## ^~~~~~~~~~~~~~\n ## make[3]: *** [Makefile:416: mtest.o] Error 1\n-## make[3]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/util'\n+## make[3]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/util'\n ## make[2]: *** [Makefile:865: ../../cxx/util/mtest.o] Error 2\n-## make[2]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/coll'\n+## make[2]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/coll'\n not ok 680 - ./cxx/coll/icgathervx \n ---\n Directory: ./cxx/coll\n File: icgathervx\n Num-procs: \n- Date: \"Sun Jan 5 01:33:16 2020\"\n+ Date: \"Mon Dec 3 17:03:31 2018\"\n ...\n ## Test output (expected 'No Errors'):\n-## Failed to build icgathervx; make[2]: Entering directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/coll'\n+## Failed to build icgathervx; make[2]: Entering directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/coll'\n ## (cd ../../cxx/util && make mtest.o)\n-## make[3]: Entering directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/util'\n+## make[3]: Entering directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/util'\n ## depbase=`echo mtest.o | sed 's|[^/]*$|.deps/&|;s|\\.o$||'`;\\\n-## /usr/bin/mpicxx.openmpi -DHAVE_CONFIG_H -I. -I../../../cxx/util -I../../include -I../../include -I../../../include -Wdate-time -D_FORTIFY_SOURCE=2 -g -O2 -ffile-prefix-map=/build/1st/mpi-testsuite-3.2+dfsg=. -fstack-protector-strong -Wformat -Werror=format-security -MT mtest.o -MD -MP -MF $depbase.Tpo -c -o mtest.o ../../../cxx/util/mtest.cxx &&\\\n+## /usr/bin/mpicxx.openmpi -DHAVE_CONFIG_H -I. -I../../../cxx/util -I../../include -I../../include -I../../../include -Wdate-time -D_FORTIFY_SOURCE=2 -g -O2 -ffile-prefix-map=/build/mpi-testsuite-3.2+dfsg/2nd=. -fstack-protector-strong -Wformat -Werror=format-security -MT mtest.o -MD -MP -MF $depbase.Tpo -c -o mtest.o ../../../cxx/util/mtest.cxx &&\\\n ## mv -f $depbase.Tpo $depbase.Po\n ## ../../../cxx/util/mtest.cxx: In function 'void* MTestTypeContigInitRecv(MTestDatatype*)':\n ## ../../../cxx/util/mtest.cxx:189:42: warning: 'int MPI_Type_extent(MPI_Datatype, MPI_Aint*)' is deprecated: MPI_Type_extent is superseded by MPI_Type_get_extent in MPI-2.0 [-Wdeprecated-declarations]\n ## MPI_Type_extent( mtype->datatype, &size );\n ## ^\n ## In file included from ../../../cxx/util/mtest.cxx:7:\n ## /usr/lib/x86_64-linux-gnu/openmpi/include/mpi.h:1758:20: note: declared here\n@@ -1000692,30 +1000775,30 @@\n ## ../../../cxx/util/mtest.cxx:848:50: error: 'NULL_COPY_FN' is not a member of 'MPI::Win'\n ## mem_keyval = MPI::Win::Create_keyval( MPI::Win::NULL_COPY_FN,\n ## ^~~~~~~~~~~~\n ## ../../../cxx/util/mtest.cxx:849:22: error: 'NULL_DELETE_FN' is not a member of 'MPI::Win'\n ## MPI::Win::NULL_DELETE_FN, 0 );\n ## ^~~~~~~~~~~~~~\n ## make[3]: *** [Makefile:416: mtest.o] Error 1\n-## make[3]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/util'\n+## make[3]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/util'\n ## make[2]: *** [Makefile:865: ../../cxx/util/mtest.o] Error 2\n-## make[2]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/coll'\n+## make[2]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/coll'\n not ok 681 - ./cxx/coll/icscattervx \n ---\n Directory: ./cxx/coll\n File: icscattervx\n Num-procs: \n- Date: \"Sun Jan 5 01:33:17 2020\"\n+ Date: \"Mon Dec 3 17:03:36 2018\"\n ...\n ## Test output (expected 'No Errors'):\n-## Failed to build icscattervx; make[2]: Entering directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/coll'\n+## Failed to build icscattervx; make[2]: Entering directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/coll'\n ## (cd ../../cxx/util && make mtest.o)\n-## make[3]: Entering directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/util'\n+## make[3]: Entering directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/util'\n ## depbase=`echo mtest.o | sed 's|[^/]*$|.deps/&|;s|\\.o$||'`;\\\n-## /usr/bin/mpicxx.openmpi -DHAVE_CONFIG_H -I. -I../../../cxx/util -I../../include -I../../include -I../../../include -Wdate-time -D_FORTIFY_SOURCE=2 -g -O2 -ffile-prefix-map=/build/1st/mpi-testsuite-3.2+dfsg=. -fstack-protector-strong -Wformat -Werror=format-security -MT mtest.o -MD -MP -MF $depbase.Tpo -c -o mtest.o ../../../cxx/util/mtest.cxx &&\\\n+## /usr/bin/mpicxx.openmpi -DHAVE_CONFIG_H -I. -I../../../cxx/util -I../../include -I../../include -I../../../include -Wdate-time -D_FORTIFY_SOURCE=2 -g -O2 -ffile-prefix-map=/build/mpi-testsuite-3.2+dfsg/2nd=. -fstack-protector-strong -Wformat -Werror=format-security -MT mtest.o -MD -MP -MF $depbase.Tpo -c -o mtest.o ../../../cxx/util/mtest.cxx &&\\\n ## mv -f $depbase.Tpo $depbase.Po\n ## ../../../cxx/util/mtest.cxx: In function 'void* MTestTypeContigInitRecv(MTestDatatype*)':\n ## ../../../cxx/util/mtest.cxx:189:42: warning: 'int MPI_Type_extent(MPI_Datatype, MPI_Aint*)' is deprecated: MPI_Type_extent is superseded by MPI_Type_get_extent in MPI-2.0 [-Wdeprecated-declarations]\n ## MPI_Type_extent( mtype->datatype, &size );\n ## ^\n ## In file included from ../../../cxx/util/mtest.cxx:7:\n ## /usr/lib/x86_64-linux-gnu/openmpi/include/mpi.h:1758:20: note: declared here\n@@ -1000747,30 +1000830,30 @@\n ## ../../../cxx/util/mtest.cxx:848:50: error: 'NULL_COPY_FN' is not a member of 'MPI::Win'\n ## mem_keyval = MPI::Win::Create_keyval( MPI::Win::NULL_COPY_FN,\n ## ^~~~~~~~~~~~\n ## ../../../cxx/util/mtest.cxx:849:22: error: 'NULL_DELETE_FN' is not a member of 'MPI::Win'\n ## MPI::Win::NULL_DELETE_FN, 0 );\n ## ^~~~~~~~~~~~~~\n ## make[3]: *** [Makefile:416: mtest.o] Error 1\n-## make[3]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/util'\n+## make[3]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/util'\n ## make[2]: *** [Makefile:865: ../../cxx/util/mtest.o] Error 2\n-## make[2]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/coll'\n+## make[2]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/coll'\n not ok 682 - ./cxx/coll/icalltoallx \n ---\n Directory: ./cxx/coll\n File: icalltoallx\n Num-procs: \n- Date: \"Sun Jan 5 01:33:18 2020\"\n+ Date: \"Mon Dec 3 17:03:40 2018\"\n ...\n ## Test output (expected 'No Errors'):\n-## Failed to build icalltoallx; make[2]: Entering directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/coll'\n+## Failed to build icalltoallx; make[2]: Entering directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/coll'\n ## (cd ../../cxx/util && make mtest.o)\n-## make[3]: Entering directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/util'\n+## make[3]: Entering directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/util'\n ## depbase=`echo mtest.o | sed 's|[^/]*$|.deps/&|;s|\\.o$||'`;\\\n-## /usr/bin/mpicxx.openmpi -DHAVE_CONFIG_H -I. -I../../../cxx/util -I../../include -I../../include -I../../../include -Wdate-time -D_FORTIFY_SOURCE=2 -g -O2 -ffile-prefix-map=/build/1st/mpi-testsuite-3.2+dfsg=. -fstack-protector-strong -Wformat -Werror=format-security -MT mtest.o -MD -MP -MF $depbase.Tpo -c -o mtest.o ../../../cxx/util/mtest.cxx &&\\\n+## /usr/bin/mpicxx.openmpi -DHAVE_CONFIG_H -I. -I../../../cxx/util -I../../include -I../../include -I../../../include -Wdate-time -D_FORTIFY_SOURCE=2 -g -O2 -ffile-prefix-map=/build/mpi-testsuite-3.2+dfsg/2nd=. -fstack-protector-strong -Wformat -Werror=format-security -MT mtest.o -MD -MP -MF $depbase.Tpo -c -o mtest.o ../../../cxx/util/mtest.cxx &&\\\n ## mv -f $depbase.Tpo $depbase.Po\n ## ../../../cxx/util/mtest.cxx: In function 'void* MTestTypeContigInitRecv(MTestDatatype*)':\n ## ../../../cxx/util/mtest.cxx:189:42: warning: 'int MPI_Type_extent(MPI_Datatype, MPI_Aint*)' is deprecated: MPI_Type_extent is superseded by MPI_Type_get_extent in MPI-2.0 [-Wdeprecated-declarations]\n ## MPI_Type_extent( mtype->datatype, &size );\n ## ^\n ## In file included from ../../../cxx/util/mtest.cxx:7:\n ## /usr/lib/x86_64-linux-gnu/openmpi/include/mpi.h:1758:20: note: declared here\n@@ -1000802,30 +1000885,30 @@\n ## ../../../cxx/util/mtest.cxx:848:50: error: 'NULL_COPY_FN' is not a member of 'MPI::Win'\n ## mem_keyval = MPI::Win::Create_keyval( MPI::Win::NULL_COPY_FN,\n ## ^~~~~~~~~~~~\n ## ../../../cxx/util/mtest.cxx:849:22: error: 'NULL_DELETE_FN' is not a member of 'MPI::Win'\n ## MPI::Win::NULL_DELETE_FN, 0 );\n ## ^~~~~~~~~~~~~~\n ## make[3]: *** [Makefile:416: mtest.o] Error 1\n-## make[3]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/util'\n+## make[3]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/util'\n ## make[2]: *** [Makefile:865: ../../cxx/util/mtest.o] Error 2\n-## make[2]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/coll'\n+## make[2]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/coll'\n not ok 683 - ./cxx/coll/reduceboolx \n ---\n Directory: ./cxx/coll\n File: reduceboolx\n Num-procs: \n- Date: \"Sun Jan 5 01:33:19 2020\"\n+ Date: \"Mon Dec 3 17:03:45 2018\"\n ...\n ## Test output (expected 'No Errors'):\n-## Failed to build reduceboolx; make[2]: Entering directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/coll'\n+## Failed to build reduceboolx; make[2]: Entering directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/coll'\n ## (cd ../../cxx/util && make mtest.o)\n-## make[3]: Entering directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/util'\n+## make[3]: Entering directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/util'\n ## depbase=`echo mtest.o | sed 's|[^/]*$|.deps/&|;s|\\.o$||'`;\\\n-## /usr/bin/mpicxx.openmpi -DHAVE_CONFIG_H -I. -I../../../cxx/util -I../../include -I../../include -I../../../include -Wdate-time -D_FORTIFY_SOURCE=2 -g -O2 -ffile-prefix-map=/build/1st/mpi-testsuite-3.2+dfsg=. -fstack-protector-strong -Wformat -Werror=format-security -MT mtest.o -MD -MP -MF $depbase.Tpo -c -o mtest.o ../../../cxx/util/mtest.cxx &&\\\n+## /usr/bin/mpicxx.openmpi -DHAVE_CONFIG_H -I. -I../../../cxx/util -I../../include -I../../include -I../../../include -Wdate-time -D_FORTIFY_SOURCE=2 -g -O2 -ffile-prefix-map=/build/mpi-testsuite-3.2+dfsg/2nd=. -fstack-protector-strong -Wformat -Werror=format-security -MT mtest.o -MD -MP -MF $depbase.Tpo -c -o mtest.o ../../../cxx/util/mtest.cxx &&\\\n ## mv -f $depbase.Tpo $depbase.Po\n ## ../../../cxx/util/mtest.cxx: In function 'void* MTestTypeContigInitRecv(MTestDatatype*)':\n ## ../../../cxx/util/mtest.cxx:189:42: warning: 'int MPI_Type_extent(MPI_Datatype, MPI_Aint*)' is deprecated: MPI_Type_extent is superseded by MPI_Type_get_extent in MPI-2.0 [-Wdeprecated-declarations]\n ## MPI_Type_extent( mtype->datatype, &size );\n ## ^\n ## In file included from ../../../cxx/util/mtest.cxx:7:\n ## /usr/lib/x86_64-linux-gnu/openmpi/include/mpi.h:1758:20: note: declared here\n@@ -1000857,53 +1000940,53 @@\n ## ../../../cxx/util/mtest.cxx:848:50: error: 'NULL_COPY_FN' is not a member of 'MPI::Win'\n ## mem_keyval = MPI::Win::Create_keyval( MPI::Win::NULL_COPY_FN,\n ## ^~~~~~~~~~~~\n ## ../../../cxx/util/mtest.cxx:849:22: error: 'NULL_DELETE_FN' is not a member of 'MPI::Win'\n ## MPI::Win::NULL_DELETE_FN, 0 );\n ## ^~~~~~~~~~~~~~\n ## make[3]: *** [Makefile:416: mtest.o] Error 1\n-## make[3]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/util'\n+## make[3]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/util'\n ## make[2]: *** [Makefile:865: ../../cxx/util/mtest.o] Error 2\n-## make[2]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/coll'\n+## make[2]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/coll'\n not ok 684 - ./cxx/coll/redscatblk \n ---\n Directory: ./cxx/coll\n File: redscatblk\n Num-procs: \n- Date: \"Sun Jan 5 01:33:20 2020\"\n+ Date: \"Mon Dec 3 17:03:49 2018\"\n ...\n ## Test output (expected 'No Errors'):\n-## Failed to build redscatblk; make[2]: Entering directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/coll'\n+## Failed to build redscatblk; make[2]: Entering directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/coll'\n ## depbase=`echo redscatblk.o | sed 's|[^/]*$|.deps/&|;s|\\.o$||'`;\\\n-## /usr/bin/mpicxx.openmpi -DHAVE_CONFIG_H -I. -I../../../cxx/coll -I../../include -I../../include -I../../../include -Wdate-time -D_FORTIFY_SOURCE=2 -g -O2 -ffile-prefix-map=/build/1st/mpi-testsuite-3.2+dfsg=. -fstack-protector-strong -Wformat -Werror=format-security -MT redscatblk.o -MD -MP -MF $depbase.Tpo -c -o redscatblk.o ../../../cxx/coll/redscatblk.cxx &&\\\n+## /usr/bin/mpicxx.openmpi -DHAVE_CONFIG_H -I. -I../../../cxx/coll -I../../include -I../../include -I../../../include -Wdate-time -D_FORTIFY_SOURCE=2 -g -O2 -ffile-prefix-map=/build/mpi-testsuite-3.2+dfsg/2nd=. -fstack-protector-strong -Wformat -Werror=format-security -MT redscatblk.o -MD -MP -MF $depbase.Tpo -c -o redscatblk.o ../../../cxx/coll/redscatblk.cxx &&\\\n ## mv -f $depbase.Tpo $depbase.Po\n ## ../../../cxx/coll/redscatblk.cxx: In function 'int main(int, char**)':\n ## ../../../cxx/coll/redscatblk.cxx:42:21: error: 'class MPI::Intracomm' has no member named 'Reduce_scatter_block'; did you mean 'Reduce_scatter'?\n ## MPI::COMM_WORLD.Reduce_scatter_block(sendbuf, recvbuf, mycount, MPI::INT, MPI::SUM);\n ## ^~~~~~~~~~~~~~~~~~~~\n ## Reduce_scatter\n ## ../../../cxx/coll/redscatblk.cxx:54:21: error: 'class MPI::Intracomm' has no member named 'Reduce_scatter_block'; did you mean 'Reduce_scatter'?\n ## MPI::COMM_WORLD.Reduce_scatter_block(MPI_IN_PLACE, sendbuf, mycount, MPI::INT, MPI::SUM);\n ## ^~~~~~~~~~~~~~~~~~~~\n ## Reduce_scatter\n ## make[2]: *** [Makefile:645: redscatblk.o] Error 1\n-## make[2]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/coll'\n+## make[2]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/coll'\n not ok 685 - ./cxx/errhan/commcallx \n ---\n Directory: ./cxx/errhan\n File: commcallx\n Num-procs: \n- Date: \"Sun Jan 5 01:33:21 2020\"\n+ Date: \"Mon Dec 3 17:03:52 2018\"\n ...\n ## Test output (expected 'No Errors'):\n-## Failed to build commcallx; make[2]: Entering directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/errhan'\n+## Failed to build commcallx; make[2]: Entering directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/errhan'\n ## (cd ../../cxx/util && make mtest.o)\n-## make[3]: Entering directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/util'\n+## make[3]: Entering directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/util'\n ## depbase=`echo mtest.o | sed 's|[^/]*$|.deps/&|;s|\\.o$||'`;\\\n-## /usr/bin/mpicxx.openmpi -DHAVE_CONFIG_H -I. -I../../../cxx/util -I../../include -I../../include -I../../../include -Wdate-time -D_FORTIFY_SOURCE=2 -g -O2 -ffile-prefix-map=/build/1st/mpi-testsuite-3.2+dfsg=. -fstack-protector-strong -Wformat -Werror=format-security -MT mtest.o -MD -MP -MF $depbase.Tpo -c -o mtest.o ../../../cxx/util/mtest.cxx &&\\\n+## /usr/bin/mpicxx.openmpi -DHAVE_CONFIG_H -I. -I../../../cxx/util -I../../include -I../../include -I../../../include -Wdate-time -D_FORTIFY_SOURCE=2 -g -O2 -ffile-prefix-map=/build/mpi-testsuite-3.2+dfsg/2nd=. -fstack-protector-strong -Wformat -Werror=format-security -MT mtest.o -MD -MP -MF $depbase.Tpo -c -o mtest.o ../../../cxx/util/mtest.cxx &&\\\n ## mv -f $depbase.Tpo $depbase.Po\n ## ../../../cxx/util/mtest.cxx: In function 'void* MTestTypeContigInitRecv(MTestDatatype*)':\n ## ../../../cxx/util/mtest.cxx:189:42: warning: 'int MPI_Type_extent(MPI_Datatype, MPI_Aint*)' is deprecated: MPI_Type_extent is superseded by MPI_Type_get_extent in MPI-2.0 [-Wdeprecated-declarations]\n ## MPI_Type_extent( mtype->datatype, &size );\n ## ^\n ## In file included from ../../../cxx/util/mtest.cxx:7:\n ## /usr/lib/x86_64-linux-gnu/openmpi/include/mpi.h:1758:20: note: declared here\n@@ -1000935,30 +1001018,30 @@\n ## ../../../cxx/util/mtest.cxx:848:50: error: 'NULL_COPY_FN' is not a member of 'MPI::Win'\n ## mem_keyval = MPI::Win::Create_keyval( MPI::Win::NULL_COPY_FN,\n ## ^~~~~~~~~~~~\n ## ../../../cxx/util/mtest.cxx:849:22: error: 'NULL_DELETE_FN' is not a member of 'MPI::Win'\n ## MPI::Win::NULL_DELETE_FN, 0 );\n ## ^~~~~~~~~~~~~~\n ## make[3]: *** [Makefile:416: mtest.o] Error 1\n-## make[3]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/util'\n+## make[3]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/util'\n ## make[2]: *** [Makefile:640: ../../cxx/util/mtest.o] Error 2\n-## make[2]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/errhan'\n+## make[2]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/errhan'\n not ok 686 - ./cxx/init/baseenv \n ---\n Directory: ./cxx/init\n File: baseenv\n Num-procs: \n- Date: \"Sun Jan 5 01:33:22 2020\"\n+ Date: \"Mon Dec 3 17:03:57 2018\"\n ...\n ## Test output (expected 'No Errors'):\n-## Failed to build baseenv; make[2]: Entering directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/init'\n+## Failed to build baseenv; make[2]: Entering directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/init'\n ## (cd ../../cxx/util && make mtest.o)\n-## make[3]: Entering directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/util'\n+## make[3]: Entering directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/util'\n ## depbase=`echo mtest.o | sed 's|[^/]*$|.deps/&|;s|\\.o$||'`;\\\n-## /usr/bin/mpicxx.openmpi -DHAVE_CONFIG_H -I. -I../../../cxx/util -I../../include -I../../include -I../../../include -Wdate-time -D_FORTIFY_SOURCE=2 -g -O2 -ffile-prefix-map=/build/1st/mpi-testsuite-3.2+dfsg=. -fstack-protector-strong -Wformat -Werror=format-security -MT mtest.o -MD -MP -MF $depbase.Tpo -c -o mtest.o ../../../cxx/util/mtest.cxx &&\\\n+## /usr/bin/mpicxx.openmpi -DHAVE_CONFIG_H -I. -I../../../cxx/util -I../../include -I../../include -I../../../include -Wdate-time -D_FORTIFY_SOURCE=2 -g -O2 -ffile-prefix-map=/build/mpi-testsuite-3.2+dfsg/2nd=. -fstack-protector-strong -Wformat -Werror=format-security -MT mtest.o -MD -MP -MF $depbase.Tpo -c -o mtest.o ../../../cxx/util/mtest.cxx &&\\\n ## mv -f $depbase.Tpo $depbase.Po\n ## ../../../cxx/util/mtest.cxx: In function 'void* MTestTypeContigInitRecv(MTestDatatype*)':\n ## ../../../cxx/util/mtest.cxx:189:42: warning: 'int MPI_Type_extent(MPI_Datatype, MPI_Aint*)' is deprecated: MPI_Type_extent is superseded by MPI_Type_get_extent in MPI-2.0 [-Wdeprecated-declarations]\n ## MPI_Type_extent( mtype->datatype, &size );\n ## ^\n ## In file included from ../../../cxx/util/mtest.cxx:7:\n ## /usr/lib/x86_64-linux-gnu/openmpi/include/mpi.h:1758:20: note: declared here\n@@ -1000990,30 +1001073,30 @@\n ## ../../../cxx/util/mtest.cxx:848:50: error: 'NULL_COPY_FN' is not a member of 'MPI::Win'\n ## mem_keyval = MPI::Win::Create_keyval( MPI::Win::NULL_COPY_FN,\n ## ^~~~~~~~~~~~\n ## ../../../cxx/util/mtest.cxx:849:22: error: 'NULL_DELETE_FN' is not a member of 'MPI::Win'\n ## MPI::Win::NULL_DELETE_FN, 0 );\n ## ^~~~~~~~~~~~~~\n ## make[3]: *** [Makefile:416: mtest.o] Error 1\n-## make[3]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/util'\n+## make[3]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/util'\n ## make[2]: *** [Makefile:663: ../../cxx/util/mtest.o] Error 2\n-## make[2]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/init'\n+## make[2]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/init'\n not ok 687 - ./cxx/init/initstatx \n ---\n Directory: ./cxx/init\n File: initstatx\n Num-procs: \n- Date: \"Sun Jan 5 01:33:23 2020\"\n+ Date: \"Mon Dec 3 17:04:01 2018\"\n ...\n ## Test output (expected 'No Errors'):\n-## Failed to build initstatx; make[2]: Entering directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/init'\n+## Failed to build initstatx; make[2]: Entering directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/init'\n ## (cd ../../cxx/util && make mtest.o)\n-## make[3]: Entering directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/util'\n+## make[3]: Entering directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/util'\n ## depbase=`echo mtest.o | sed 's|[^/]*$|.deps/&|;s|\\.o$||'`;\\\n-## /usr/bin/mpicxx.openmpi -DHAVE_CONFIG_H -I. -I../../../cxx/util -I../../include -I../../include -I../../../include -Wdate-time -D_FORTIFY_SOURCE=2 -g -O2 -ffile-prefix-map=/build/1st/mpi-testsuite-3.2+dfsg=. -fstack-protector-strong -Wformat -Werror=format-security -MT mtest.o -MD -MP -MF $depbase.Tpo -c -o mtest.o ../../../cxx/util/mtest.cxx &&\\\n+## /usr/bin/mpicxx.openmpi -DHAVE_CONFIG_H -I. -I../../../cxx/util -I../../include -I../../include -I../../../include -Wdate-time -D_FORTIFY_SOURCE=2 -g -O2 -ffile-prefix-map=/build/mpi-testsuite-3.2+dfsg/2nd=. -fstack-protector-strong -Wformat -Werror=format-security -MT mtest.o -MD -MP -MF $depbase.Tpo -c -o mtest.o ../../../cxx/util/mtest.cxx &&\\\n ## mv -f $depbase.Tpo $depbase.Po\n ## ../../../cxx/util/mtest.cxx: In function 'void* MTestTypeContigInitRecv(MTestDatatype*)':\n ## ../../../cxx/util/mtest.cxx:189:42: warning: 'int MPI_Type_extent(MPI_Datatype, MPI_Aint*)' is deprecated: MPI_Type_extent is superseded by MPI_Type_get_extent in MPI-2.0 [-Wdeprecated-declarations]\n ## MPI_Type_extent( mtype->datatype, &size );\n ## ^\n ## In file included from ../../../cxx/util/mtest.cxx:7:\n ## /usr/lib/x86_64-linux-gnu/openmpi/include/mpi.h:1758:20: note: declared here\n@@ -1001045,30 +1001128,30 @@\n ## ../../../cxx/util/mtest.cxx:848:50: error: 'NULL_COPY_FN' is not a member of 'MPI::Win'\n ## mem_keyval = MPI::Win::Create_keyval( MPI::Win::NULL_COPY_FN,\n ## ^~~~~~~~~~~~\n ## ../../../cxx/util/mtest.cxx:849:22: error: 'NULL_DELETE_FN' is not a member of 'MPI::Win'\n ## MPI::Win::NULL_DELETE_FN, 0 );\n ## ^~~~~~~~~~~~~~\n ## make[3]: *** [Makefile:416: mtest.o] Error 1\n-## make[3]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/util'\n+## make[3]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/util'\n ## make[2]: *** [Makefile:663: ../../cxx/util/mtest.o] Error 2\n-## make[2]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/init'\n+## make[2]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/init'\n not ok 688 - ./cxx/init/initstat2x \n ---\n Directory: ./cxx/init\n File: initstat2x\n Num-procs: \n- Date: \"Sun Jan 5 01:33:24 2020\"\n+ Date: \"Mon Dec 3 17:04:05 2018\"\n ...\n ## Test output (expected 'No Errors'):\n-## Failed to build initstat2x; make[2]: Entering directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/init'\n+## Failed to build initstat2x; make[2]: Entering directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/init'\n ## (cd ../../cxx/util && make mtest.o)\n-## make[3]: Entering directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/util'\n+## make[3]: Entering directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/util'\n ## depbase=`echo mtest.o | sed 's|[^/]*$|.deps/&|;s|\\.o$||'`;\\\n-## /usr/bin/mpicxx.openmpi -DHAVE_CONFIG_H -I. -I../../../cxx/util -I../../include -I../../include -I../../../include -Wdate-time -D_FORTIFY_SOURCE=2 -g -O2 -ffile-prefix-map=/build/1st/mpi-testsuite-3.2+dfsg=. -fstack-protector-strong -Wformat -Werror=format-security -MT mtest.o -MD -MP -MF $depbase.Tpo -c -o mtest.o ../../../cxx/util/mtest.cxx &&\\\n+## /usr/bin/mpicxx.openmpi -DHAVE_CONFIG_H -I. -I../../../cxx/util -I../../include -I../../include -I../../../include -Wdate-time -D_FORTIFY_SOURCE=2 -g -O2 -ffile-prefix-map=/build/mpi-testsuite-3.2+dfsg/2nd=. -fstack-protector-strong -Wformat -Werror=format-security -MT mtest.o -MD -MP -MF $depbase.Tpo -c -o mtest.o ../../../cxx/util/mtest.cxx &&\\\n ## mv -f $depbase.Tpo $depbase.Po\n ## ../../../cxx/util/mtest.cxx: In function 'void* MTestTypeContigInitRecv(MTestDatatype*)':\n ## ../../../cxx/util/mtest.cxx:189:42: warning: 'int MPI_Type_extent(MPI_Datatype, MPI_Aint*)' is deprecated: MPI_Type_extent is superseded by MPI_Type_get_extent in MPI-2.0 [-Wdeprecated-declarations]\n ## MPI_Type_extent( mtype->datatype, &size );\n ## ^\n ## In file included from ../../../cxx/util/mtest.cxx:7:\n ## /usr/lib/x86_64-linux-gnu/openmpi/include/mpi.h:1758:20: note: declared here\n@@ -1001100,30 +1001183,30 @@\n ## ../../../cxx/util/mtest.cxx:848:50: error: 'NULL_COPY_FN' is not a member of 'MPI::Win'\n ## mem_keyval = MPI::Win::Create_keyval( MPI::Win::NULL_COPY_FN,\n ## ^~~~~~~~~~~~\n ## ../../../cxx/util/mtest.cxx:849:22: error: 'NULL_DELETE_FN' is not a member of 'MPI::Win'\n ## MPI::Win::NULL_DELETE_FN, 0 );\n ## ^~~~~~~~~~~~~~\n ## make[3]: *** [Makefile:416: mtest.o] Error 1\n-## make[3]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/util'\n+## make[3]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/util'\n ## make[2]: *** [Makefile:663: ../../cxx/util/mtest.o] Error 2\n-## make[2]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/init'\n+## make[2]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/init'\n not ok 689 - ./cxx/info/infodupx \n ---\n Directory: ./cxx/info\n File: infodupx\n Num-procs: \n- Date: \"Sun Jan 5 01:33:25 2020\"\n+ Date: \"Mon Dec 3 17:04:09 2018\"\n ...\n ## Test output (expected 'No Errors'):\n-## Failed to build infodupx; make[2]: Entering directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/info'\n+## Failed to build infodupx; make[2]: Entering directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/info'\n ## (cd ../../cxx/util && make mtest.o)\n-## make[3]: Entering directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/util'\n+## make[3]: Entering directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/util'\n ## depbase=`echo mtest.o | sed 's|[^/]*$|.deps/&|;s|\\.o$||'`;\\\n-## /usr/bin/mpicxx.openmpi -DHAVE_CONFIG_H -I. -I../../../cxx/util -I../../include -I../../include -I../../../include -Wdate-time -D_FORTIFY_SOURCE=2 -g -O2 -ffile-prefix-map=/build/1st/mpi-testsuite-3.2+dfsg=. -fstack-protector-strong -Wformat -Werror=format-security -MT mtest.o -MD -MP -MF $depbase.Tpo -c -o mtest.o ../../../cxx/util/mtest.cxx &&\\\n+## /usr/bin/mpicxx.openmpi -DHAVE_CONFIG_H -I. -I../../../cxx/util -I../../include -I../../include -I../../../include -Wdate-time -D_FORTIFY_SOURCE=2 -g -O2 -ffile-prefix-map=/build/mpi-testsuite-3.2+dfsg/2nd=. -fstack-protector-strong -Wformat -Werror=format-security -MT mtest.o -MD -MP -MF $depbase.Tpo -c -o mtest.o ../../../cxx/util/mtest.cxx &&\\\n ## mv -f $depbase.Tpo $depbase.Po\n ## ../../../cxx/util/mtest.cxx: In function 'void* MTestTypeContigInitRecv(MTestDatatype*)':\n ## ../../../cxx/util/mtest.cxx:189:42: warning: 'int MPI_Type_extent(MPI_Datatype, MPI_Aint*)' is deprecated: MPI_Type_extent is superseded by MPI_Type_get_extent in MPI-2.0 [-Wdeprecated-declarations]\n ## MPI_Type_extent( mtype->datatype, &size );\n ## ^\n ## In file included from ../../../cxx/util/mtest.cxx:7:\n ## /usr/lib/x86_64-linux-gnu/openmpi/include/mpi.h:1758:20: note: declared here\n@@ -1001155,30 +1001238,30 @@\n ## ../../../cxx/util/mtest.cxx:848:50: error: 'NULL_COPY_FN' is not a member of 'MPI::Win'\n ## mem_keyval = MPI::Win::Create_keyval( MPI::Win::NULL_COPY_FN,\n ## ^~~~~~~~~~~~\n ## ../../../cxx/util/mtest.cxx:849:22: error: 'NULL_DELETE_FN' is not a member of 'MPI::Win'\n ## MPI::Win::NULL_DELETE_FN, 0 );\n ## ^~~~~~~~~~~~~~\n ## make[3]: *** [Makefile:416: mtest.o] Error 1\n-## make[3]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/util'\n+## make[3]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/util'\n ## make[2]: *** [Makefile:673: ../../cxx/util/mtest.o] Error 2\n-## make[2]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/info'\n+## make[2]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/info'\n not ok 690 - ./cxx/info/infodelx \n ---\n Directory: ./cxx/info\n File: infodelx\n Num-procs: \n- Date: \"Sun Jan 5 01:33:26 2020\"\n+ Date: \"Mon Dec 3 17:04:14 2018\"\n ...\n ## Test output (expected 'No Errors'):\n-## Failed to build infodelx; make[2]: Entering directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/info'\n+## Failed to build infodelx; make[2]: Entering directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/info'\n ## (cd ../../cxx/util && make mtest.o)\n-## make[3]: Entering directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/util'\n+## make[3]: Entering directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/util'\n ## depbase=`echo mtest.o | sed 's|[^/]*$|.deps/&|;s|\\.o$||'`;\\\n-## /usr/bin/mpicxx.openmpi -DHAVE_CONFIG_H -I. -I../../../cxx/util -I../../include -I../../include -I../../../include -Wdate-time -D_FORTIFY_SOURCE=2 -g -O2 -ffile-prefix-map=/build/1st/mpi-testsuite-3.2+dfsg=. -fstack-protector-strong -Wformat -Werror=format-security -MT mtest.o -MD -MP -MF $depbase.Tpo -c -o mtest.o ../../../cxx/util/mtest.cxx &&\\\n+## /usr/bin/mpicxx.openmpi -DHAVE_CONFIG_H -I. -I../../../cxx/util -I../../include -I../../include -I../../../include -Wdate-time -D_FORTIFY_SOURCE=2 -g -O2 -ffile-prefix-map=/build/mpi-testsuite-3.2+dfsg/2nd=. -fstack-protector-strong -Wformat -Werror=format-security -MT mtest.o -MD -MP -MF $depbase.Tpo -c -o mtest.o ../../../cxx/util/mtest.cxx &&\\\n ## mv -f $depbase.Tpo $depbase.Po\n ## ../../../cxx/util/mtest.cxx: In function 'void* MTestTypeContigInitRecv(MTestDatatype*)':\n ## ../../../cxx/util/mtest.cxx:189:42: warning: 'int MPI_Type_extent(MPI_Datatype, MPI_Aint*)' is deprecated: MPI_Type_extent is superseded by MPI_Type_get_extent in MPI-2.0 [-Wdeprecated-declarations]\n ## MPI_Type_extent( mtype->datatype, &size );\n ## ^\n ## In file included from ../../../cxx/util/mtest.cxx:7:\n ## /usr/lib/x86_64-linux-gnu/openmpi/include/mpi.h:1758:20: note: declared here\n@@ -1001210,30 +1001293,30 @@\n ## ../../../cxx/util/mtest.cxx:848:50: error: 'NULL_COPY_FN' is not a member of 'MPI::Win'\n ## mem_keyval = MPI::Win::Create_keyval( MPI::Win::NULL_COPY_FN,\n ## ^~~~~~~~~~~~\n ## ../../../cxx/util/mtest.cxx:849:22: error: 'NULL_DELETE_FN' is not a member of 'MPI::Win'\n ## MPI::Win::NULL_DELETE_FN, 0 );\n ## ^~~~~~~~~~~~~~\n ## make[3]: *** [Makefile:416: mtest.o] Error 1\n-## make[3]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/util'\n+## make[3]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/util'\n ## make[2]: *** [Makefile:673: ../../cxx/util/mtest.o] Error 2\n-## make[2]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/info'\n+## make[2]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/info'\n not ok 691 - ./cxx/info/infovallenx \n ---\n Directory: ./cxx/info\n File: infovallenx\n Num-procs: \n- Date: \"Sun Jan 5 01:33:27 2020\"\n+ Date: \"Mon Dec 3 17:04:17 2018\"\n ...\n ## Test output (expected 'No Errors'):\n-## Failed to build infovallenx; make[2]: Entering directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/info'\n+## Failed to build infovallenx; make[2]: Entering directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/info'\n ## (cd ../../cxx/util && make mtest.o)\n-## make[3]: Entering directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/util'\n+## make[3]: Entering directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/util'\n ## depbase=`echo mtest.o | sed 's|[^/]*$|.deps/&|;s|\\.o$||'`;\\\n-## /usr/bin/mpicxx.openmpi -DHAVE_CONFIG_H -I. -I../../../cxx/util -I../../include -I../../include -I../../../include -Wdate-time -D_FORTIFY_SOURCE=2 -g -O2 -ffile-prefix-map=/build/1st/mpi-testsuite-3.2+dfsg=. -fstack-protector-strong -Wformat -Werror=format-security -MT mtest.o -MD -MP -MF $depbase.Tpo -c -o mtest.o ../../../cxx/util/mtest.cxx &&\\\n+## /usr/bin/mpicxx.openmpi -DHAVE_CONFIG_H -I. -I../../../cxx/util -I../../include -I../../include -I../../../include -Wdate-time -D_FORTIFY_SOURCE=2 -g -O2 -ffile-prefix-map=/build/mpi-testsuite-3.2+dfsg/2nd=. -fstack-protector-strong -Wformat -Werror=format-security -MT mtest.o -MD -MP -MF $depbase.Tpo -c -o mtest.o ../../../cxx/util/mtest.cxx &&\\\n ## mv -f $depbase.Tpo $depbase.Po\n ## ../../../cxx/util/mtest.cxx: In function 'void* MTestTypeContigInitRecv(MTestDatatype*)':\n ## ../../../cxx/util/mtest.cxx:189:42: warning: 'int MPI_Type_extent(MPI_Datatype, MPI_Aint*)' is deprecated: MPI_Type_extent is superseded by MPI_Type_get_extent in MPI-2.0 [-Wdeprecated-declarations]\n ## MPI_Type_extent( mtype->datatype, &size );\n ## ^\n ## In file included from ../../../cxx/util/mtest.cxx:7:\n ## /usr/lib/x86_64-linux-gnu/openmpi/include/mpi.h:1758:20: note: declared here\n@@ -1001265,30 +1001348,30 @@\n ## ../../../cxx/util/mtest.cxx:848:50: error: 'NULL_COPY_FN' is not a member of 'MPI::Win'\n ## mem_keyval = MPI::Win::Create_keyval( MPI::Win::NULL_COPY_FN,\n ## ^~~~~~~~~~~~\n ## ../../../cxx/util/mtest.cxx:849:22: error: 'NULL_DELETE_FN' is not a member of 'MPI::Win'\n ## MPI::Win::NULL_DELETE_FN, 0 );\n ## ^~~~~~~~~~~~~~\n ## make[3]: *** [Makefile:416: mtest.o] Error 1\n-## make[3]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/util'\n+## make[3]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/util'\n ## make[2]: *** [Makefile:673: ../../cxx/util/mtest.o] Error 2\n-## make[2]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/info'\n+## make[2]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/info'\n not ok 692 - ./cxx/info/infoorderx \n ---\n Directory: ./cxx/info\n File: infoorderx\n Num-procs: \n- Date: \"Sun Jan 5 01:33:28 2020\"\n+ Date: \"Mon Dec 3 17:04:22 2018\"\n ...\n ## Test output (expected 'No Errors'):\n-## Failed to build infoorderx; make[2]: Entering directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/info'\n+## Failed to build infoorderx; make[2]: Entering directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/info'\n ## (cd ../../cxx/util && make mtest.o)\n-## make[3]: Entering directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/util'\n+## make[3]: Entering directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/util'\n ## depbase=`echo mtest.o | sed 's|[^/]*$|.deps/&|;s|\\.o$||'`;\\\n-## /usr/bin/mpicxx.openmpi -DHAVE_CONFIG_H -I. -I../../../cxx/util -I../../include -I../../include -I../../../include -Wdate-time -D_FORTIFY_SOURCE=2 -g -O2 -ffile-prefix-map=/build/1st/mpi-testsuite-3.2+dfsg=. -fstack-protector-strong -Wformat -Werror=format-security -MT mtest.o -MD -MP -MF $depbase.Tpo -c -o mtest.o ../../../cxx/util/mtest.cxx &&\\\n+## /usr/bin/mpicxx.openmpi -DHAVE_CONFIG_H -I. -I../../../cxx/util -I../../include -I../../include -I../../../include -Wdate-time -D_FORTIFY_SOURCE=2 -g -O2 -ffile-prefix-map=/build/mpi-testsuite-3.2+dfsg/2nd=. -fstack-protector-strong -Wformat -Werror=format-security -MT mtest.o -MD -MP -MF $depbase.Tpo -c -o mtest.o ../../../cxx/util/mtest.cxx &&\\\n ## mv -f $depbase.Tpo $depbase.Po\n ## ../../../cxx/util/mtest.cxx: In function 'void* MTestTypeContigInitRecv(MTestDatatype*)':\n ## ../../../cxx/util/mtest.cxx:189:42: warning: 'int MPI_Type_extent(MPI_Datatype, MPI_Aint*)' is deprecated: MPI_Type_extent is superseded by MPI_Type_get_extent in MPI-2.0 [-Wdeprecated-declarations]\n ## MPI_Type_extent( mtype->datatype, &size );\n ## ^\n ## In file included from ../../../cxx/util/mtest.cxx:7:\n ## /usr/lib/x86_64-linux-gnu/openmpi/include/mpi.h:1758:20: note: declared here\n@@ -1001320,30 +1001403,30 @@\n ## ../../../cxx/util/mtest.cxx:848:50: error: 'NULL_COPY_FN' is not a member of 'MPI::Win'\n ## mem_keyval = MPI::Win::Create_keyval( MPI::Win::NULL_COPY_FN,\n ## ^~~~~~~~~~~~\n ## ../../../cxx/util/mtest.cxx:849:22: error: 'NULL_DELETE_FN' is not a member of 'MPI::Win'\n ## MPI::Win::NULL_DELETE_FN, 0 );\n ## ^~~~~~~~~~~~~~\n ## make[3]: *** [Makefile:416: mtest.o] Error 1\n-## make[3]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/util'\n+## make[3]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/util'\n ## make[2]: *** [Makefile:673: ../../cxx/util/mtest.o] Error 2\n-## make[2]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/info'\n+## make[2]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/info'\n not ok 693 - ./cxx/datatype/packsizex \n ---\n Directory: ./cxx/datatype\n File: packsizex\n Num-procs: \n- Date: \"Sun Jan 5 01:33:29 2020\"\n+ Date: \"Mon Dec 3 17:04:26 2018\"\n ...\n ## Test output (expected 'No Errors'):\n-## Failed to build packsizex; make[2]: Entering directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/datatype'\n+## Failed to build packsizex; make[2]: Entering directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/datatype'\n ## (cd ../../cxx/util && make mtest.o)\n-## make[3]: Entering directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/util'\n+## make[3]: Entering directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/util'\n ## depbase=`echo mtest.o | sed 's|[^/]*$|.deps/&|;s|\\.o$||'`;\\\n-## /usr/bin/mpicxx.openmpi -DHAVE_CONFIG_H -I. -I../../../cxx/util -I../../include -I../../include -I../../../include -Wdate-time -D_FORTIFY_SOURCE=2 -g -O2 -ffile-prefix-map=/build/1st/mpi-testsuite-3.2+dfsg=. -fstack-protector-strong -Wformat -Werror=format-security -MT mtest.o -MD -MP -MF $depbase.Tpo -c -o mtest.o ../../../cxx/util/mtest.cxx &&\\\n+## /usr/bin/mpicxx.openmpi -DHAVE_CONFIG_H -I. -I../../../cxx/util -I../../include -I../../include -I../../../include -Wdate-time -D_FORTIFY_SOURCE=2 -g -O2 -ffile-prefix-map=/build/mpi-testsuite-3.2+dfsg/2nd=. -fstack-protector-strong -Wformat -Werror=format-security -MT mtest.o -MD -MP -MF $depbase.Tpo -c -o mtest.o ../../../cxx/util/mtest.cxx &&\\\n ## mv -f $depbase.Tpo $depbase.Po\n ## ../../../cxx/util/mtest.cxx: In function 'void* MTestTypeContigInitRecv(MTestDatatype*)':\n ## ../../../cxx/util/mtest.cxx:189:42: warning: 'int MPI_Type_extent(MPI_Datatype, MPI_Aint*)' is deprecated: MPI_Type_extent is superseded by MPI_Type_get_extent in MPI-2.0 [-Wdeprecated-declarations]\n ## MPI_Type_extent( mtype->datatype, &size );\n ## ^\n ## In file included from ../../../cxx/util/mtest.cxx:7:\n ## /usr/lib/x86_64-linux-gnu/openmpi/include/mpi.h:1758:20: note: declared here\n@@ -1001375,30 +1001458,30 @@\n ## ../../../cxx/util/mtest.cxx:848:50: error: 'NULL_COPY_FN' is not a member of 'MPI::Win'\n ## mem_keyval = MPI::Win::Create_keyval( MPI::Win::NULL_COPY_FN,\n ## ^~~~~~~~~~~~\n ## ../../../cxx/util/mtest.cxx:849:22: error: 'NULL_DELETE_FN' is not a member of 'MPI::Win'\n ## MPI::Win::NULL_DELETE_FN, 0 );\n ## ^~~~~~~~~~~~~~\n ## make[3]: *** [Makefile:416: mtest.o] Error 1\n-## make[3]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/util'\n+## make[3]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/util'\n ## make[2]: *** [Makefile:673: ../../cxx/util/mtest.o] Error 2\n-## make[2]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/datatype'\n+## make[2]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/datatype'\n not ok 694 - ./cxx/datatype/typecntsx \n ---\n Directory: ./cxx/datatype\n File: typecntsx\n Num-procs: \n- Date: \"Sun Jan 5 01:33:30 2020\"\n+ Date: \"Mon Dec 3 17:04:31 2018\"\n ...\n ## Test output (expected 'No Errors'):\n-## Failed to build typecntsx; make[2]: Entering directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/datatype'\n+## Failed to build typecntsx; make[2]: Entering directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/datatype'\n ## (cd ../../cxx/util && make mtest.o)\n-## make[3]: Entering directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/util'\n+## make[3]: Entering directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/util'\n ## depbase=`echo mtest.o | sed 's|[^/]*$|.deps/&|;s|\\.o$||'`;\\\n-## /usr/bin/mpicxx.openmpi -DHAVE_CONFIG_H -I. -I../../../cxx/util -I../../include -I../../include -I../../../include -Wdate-time -D_FORTIFY_SOURCE=2 -g -O2 -ffile-prefix-map=/build/1st/mpi-testsuite-3.2+dfsg=. -fstack-protector-strong -Wformat -Werror=format-security -MT mtest.o -MD -MP -MF $depbase.Tpo -c -o mtest.o ../../../cxx/util/mtest.cxx &&\\\n+## /usr/bin/mpicxx.openmpi -DHAVE_CONFIG_H -I. -I../../../cxx/util -I../../include -I../../include -I../../../include -Wdate-time -D_FORTIFY_SOURCE=2 -g -O2 -ffile-prefix-map=/build/mpi-testsuite-3.2+dfsg/2nd=. -fstack-protector-strong -Wformat -Werror=format-security -MT mtest.o -MD -MP -MF $depbase.Tpo -c -o mtest.o ../../../cxx/util/mtest.cxx &&\\\n ## mv -f $depbase.Tpo $depbase.Po\n ## ../../../cxx/util/mtest.cxx: In function 'void* MTestTypeContigInitRecv(MTestDatatype*)':\n ## ../../../cxx/util/mtest.cxx:189:42: warning: 'int MPI_Type_extent(MPI_Datatype, MPI_Aint*)' is deprecated: MPI_Type_extent is superseded by MPI_Type_get_extent in MPI-2.0 [-Wdeprecated-declarations]\n ## MPI_Type_extent( mtype->datatype, &size );\n ## ^\n ## In file included from ../../../cxx/util/mtest.cxx:7:\n ## /usr/lib/x86_64-linux-gnu/openmpi/include/mpi.h:1758:20: note: declared here\n@@ -1001430,28 +1001513,28 @@\n ## ../../../cxx/util/mtest.cxx:848:50: error: 'NULL_COPY_FN' is not a member of 'MPI::Win'\n ## mem_keyval = MPI::Win::Create_keyval( MPI::Win::NULL_COPY_FN,\n ## ^~~~~~~~~~~~\n ## ../../../cxx/util/mtest.cxx:849:22: error: 'NULL_DELETE_FN' is not a member of 'MPI::Win'\n ## MPI::Win::NULL_DELETE_FN, 0 );\n ## ^~~~~~~~~~~~~~\n ## make[3]: *** [Makefile:416: mtest.o] Error 1\n-## make[3]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/util'\n+## make[3]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/util'\n ## make[2]: *** [Makefile:673: ../../cxx/util/mtest.o] Error 2\n-## make[2]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/datatype'\n+## make[2]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/datatype'\n not ok 695 - ./cxx/datatype/typenamex \n ---\n Directory: ./cxx/datatype\n File: typenamex\n Num-procs: \n- Date: \"Sun Jan 5 01:33:31 2020\"\n+ Date: \"Mon Dec 3 17:04:34 2018\"\n ...\n ## Test output (expected 'No Errors'):\n-## Failed to build typenamex; make[2]: Entering directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/datatype'\n+## Failed to build typenamex; make[2]: Entering directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/datatype'\n ## depbase=`echo typenamex.o | sed 's|[^/]*$|.deps/&|;s|\\.o$||'`;\\\n-## /usr/bin/mpicxx.openmpi -DHAVE_CONFIG_H -I. -I../../../cxx/datatype -I../../include -I../../include -I../../../include -Wdate-time -D_FORTIFY_SOURCE=2 -g -O2 -ffile-prefix-map=/build/1st/mpi-testsuite-3.2+dfsg=. -fstack-protector-strong -Wformat -Werror=format-security -MT typenamex.o -MD -MP -MF $depbase.Tpo -c -o typenamex.o ../../../cxx/datatype/typenamex.cxx &&\\\n+## /usr/bin/mpicxx.openmpi -DHAVE_CONFIG_H -I. -I../../../cxx/datatype -I../../include -I../../include -I../../../include -Wdate-time -D_FORTIFY_SOURCE=2 -g -O2 -ffile-prefix-map=/build/mpi-testsuite-3.2+dfsg/2nd=. -fstack-protector-strong -Wformat -Werror=format-security -MT typenamex.o -MD -MP -MF $depbase.Tpo -c -o typenamex.o ../../../cxx/datatype/typenamex.cxx &&\\\n ## mv -f $depbase.Tpo $depbase.Po\n ## ../../../cxx/datatype/typenamex.cxx: In function 'int main(int, char**)':\n ## ../../../cxx/datatype/typenamex.cxx:78:30: error: passing 'const MPI::Datatype' as 'this' argument discards qualifiers [-fpermissive]\n ## MPI::INT.Set_name( \"int\" );\n ## ^\n ## In file included from /usr/lib/x86_64-linux-gnu/openmpi/include/openmpi/ompi/mpi/cxx/mpicxx.h:269,\n ## from /usr/lib/x86_64-linux-gnu/openmpi/include/mpi.h:2684,\n@@ -1001499,26 +1001582,26 @@\n ## { MPI::INTEGER16, \"MPI_INTEGER16\" },\n ## ^~~~~~~~~\n ## ../../../cxx/datatype/typenamex.cxx:158:9: note: suggested alternative: 'INTEGER1'\n ## { MPI::INTEGER16, \"MPI_INTEGER16\" },\n ## ^~~~~~~~~\n ## INTEGER1\n ## make[2]: *** [Makefile:453: typenamex.o] Error 1\n-## make[2]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/datatype'\n+## make[2]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/datatype'\n not ok 696 - ./cxx/datatype/typemiscx \n ---\n Directory: ./cxx/datatype\n File: typemiscx\n Num-procs: \n- Date: \"Sun Jan 5 01:33:32 2020\"\n+ Date: \"Mon Dec 3 17:04:37 2018\"\n ...\n ## Test output (expected 'No Errors'):\n-## Failed to build typemiscx; make[2]: Entering directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/datatype'\n+## Failed to build typemiscx; make[2]: Entering directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/datatype'\n ## depbase=`echo typemiscx.o | sed 's|[^/]*$|.deps/&|;s|\\.o$||'`;\\\n-## /usr/bin/mpicxx.openmpi -DHAVE_CONFIG_H -I. -I../../../cxx/datatype -I../../include -I../../include -I../../../include -Wdate-time -D_FORTIFY_SOURCE=2 -g -O2 -ffile-prefix-map=/build/1st/mpi-testsuite-3.2+dfsg=. -fstack-protector-strong -Wformat -Werror=format-security -MT typemiscx.o -MD -MP -MF $depbase.Tpo -c -o typemiscx.o ../../../cxx/datatype/typemiscx.cxx &&\\\n+## /usr/bin/mpicxx.openmpi -DHAVE_CONFIG_H -I. -I../../../cxx/datatype -I../../include -I../../include -I../../../include -Wdate-time -D_FORTIFY_SOURCE=2 -g -O2 -ffile-prefix-map=/build/mpi-testsuite-3.2+dfsg/2nd=. -fstack-protector-strong -Wformat -Werror=format-security -MT typemiscx.o -MD -MP -MF $depbase.Tpo -c -o typemiscx.o ../../../cxx/datatype/typemiscx.cxx &&\\\n ## mv -f $depbase.Tpo $depbase.Po\n ## ../../../cxx/datatype/typemiscx.cxx: In function 'int main(int, char**)':\n ## ../../../cxx/datatype/typemiscx.cxx:49:14: error: 'REAL16' is not a member of 'MPI'\n ## if (MPI::REAL16 != MPI::DATATYPE_NULL) {\n ## ^~~~~~\n ## ../../../cxx/datatype/typemiscx.cxx:49:14: note: suggested alternative: 'REAL8'\n ## if (MPI::REAL16 != MPI::DATATYPE_NULL) {\n@@ -1001584,28 +1001667,28 @@\n ## MPI::INTEGER8.Get_envelope( num_ints, num_adds, num_types, combiner );\n ## ^~~~~~~~\n ## ../../../cxx/datatype/typemiscx.cxx:99:7: note: suggested alternative: 'INTEGER4'\n ## MPI::INTEGER8.Get_envelope( num_ints, num_adds, num_types, combiner );\n ## ^~~~~~~~\n ## INTEGER4\n ## make[2]: *** [Makefile:453: typemiscx.o] Error 1\n-## make[2]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/datatype'\n+## make[2]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/datatype'\n not ok 697 - ./cxx/io/iwriteatx \n ---\n Directory: ./cxx/io\n File: iwriteatx\n Num-procs: \n- Date: \"Sun Jan 5 01:33:33 2020\"\n+ Date: \"Mon Dec 3 17:04:42 2018\"\n ...\n ## Test output (expected 'No Errors'):\n-## Failed to build iwriteatx; make[2]: Entering directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/io'\n+## Failed to build iwriteatx; make[2]: Entering directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/io'\n ## (cd ../../cxx/util && make mtest.o)\n-## make[3]: Entering directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/util'\n+## make[3]: Entering directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/util'\n ## depbase=`echo mtest.o | sed 's|[^/]*$|.deps/&|;s|\\.o$||'`;\\\n-## /usr/bin/mpicxx.openmpi -DHAVE_CONFIG_H -I. -I../../../cxx/util -I../../include -I../../include -I../../../include -Wdate-time -D_FORTIFY_SOURCE=2 -g -O2 -ffile-prefix-map=/build/1st/mpi-testsuite-3.2+dfsg=. -fstack-protector-strong -Wformat -Werror=format-security -MT mtest.o -MD -MP -MF $depbase.Tpo -c -o mtest.o ../../../cxx/util/mtest.cxx &&\\\n+## /usr/bin/mpicxx.openmpi -DHAVE_CONFIG_H -I. -I../../../cxx/util -I../../include -I../../include -I../../../include -Wdate-time -D_FORTIFY_SOURCE=2 -g -O2 -ffile-prefix-map=/build/mpi-testsuite-3.2+dfsg/2nd=. -fstack-protector-strong -Wformat -Werror=format-security -MT mtest.o -MD -MP -MF $depbase.Tpo -c -o mtest.o ../../../cxx/util/mtest.cxx &&\\\n ## mv -f $depbase.Tpo $depbase.Po\n ## ../../../cxx/util/mtest.cxx: In function 'void* MTestTypeContigInitRecv(MTestDatatype*)':\n ## ../../../cxx/util/mtest.cxx:189:42: warning: 'int MPI_Type_extent(MPI_Datatype, MPI_Aint*)' is deprecated: MPI_Type_extent is superseded by MPI_Type_get_extent in MPI-2.0 [-Wdeprecated-declarations]\n ## MPI_Type_extent( mtype->datatype, &size );\n ## ^\n ## In file included from ../../../cxx/util/mtest.cxx:7:\n ## /usr/lib/x86_64-linux-gnu/openmpi/include/mpi.h:1758:20: note: declared here\n@@ -1001637,30 +1001720,30 @@\n ## ../../../cxx/util/mtest.cxx:848:50: error: 'NULL_COPY_FN' is not a member of 'MPI::Win'\n ## mem_keyval = MPI::Win::Create_keyval( MPI::Win::NULL_COPY_FN,\n ## ^~~~~~~~~~~~\n ## ../../../cxx/util/mtest.cxx:849:22: error: 'NULL_DELETE_FN' is not a member of 'MPI::Win'\n ## MPI::Win::NULL_DELETE_FN, 0 );\n ## ^~~~~~~~~~~~~~\n ## make[3]: *** [Makefile:416: mtest.o] Error 1\n-## make[3]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/util'\n+## make[3]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/util'\n ## make[2]: *** [Makefile:989: ../../cxx/util/mtest.o] Error 2\n-## make[2]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/io'\n+## make[2]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/io'\n not ok 698 - ./cxx/io/iwritex \n ---\n Directory: ./cxx/io\n File: iwritex\n Num-procs: \n- Date: \"Sun Jan 5 01:33:34 2020\"\n+ Date: \"Mon Dec 3 17:04:46 2018\"\n ...\n ## Test output (expected 'No Errors'):\n-## Failed to build iwritex; make[2]: Entering directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/io'\n+## Failed to build iwritex; make[2]: Entering directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/io'\n ## (cd ../../cxx/util && make mtest.o)\n-## make[3]: Entering directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/util'\n+## make[3]: Entering directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/util'\n ## depbase=`echo mtest.o | sed 's|[^/]*$|.deps/&|;s|\\.o$||'`;\\\n-## /usr/bin/mpicxx.openmpi -DHAVE_CONFIG_H -I. -I../../../cxx/util -I../../include -I../../include -I../../../include -Wdate-time -D_FORTIFY_SOURCE=2 -g -O2 -ffile-prefix-map=/build/1st/mpi-testsuite-3.2+dfsg=. -fstack-protector-strong -Wformat -Werror=format-security -MT mtest.o -MD -MP -MF $depbase.Tpo -c -o mtest.o ../../../cxx/util/mtest.cxx &&\\\n+## /usr/bin/mpicxx.openmpi -DHAVE_CONFIG_H -I. -I../../../cxx/util -I../../include -I../../include -I../../../include -Wdate-time -D_FORTIFY_SOURCE=2 -g -O2 -ffile-prefix-map=/build/mpi-testsuite-3.2+dfsg/2nd=. -fstack-protector-strong -Wformat -Werror=format-security -MT mtest.o -MD -MP -MF $depbase.Tpo -c -o mtest.o ../../../cxx/util/mtest.cxx &&\\\n ## mv -f $depbase.Tpo $depbase.Po\n ## ../../../cxx/util/mtest.cxx: In function 'void* MTestTypeContigInitRecv(MTestDatatype*)':\n ## ../../../cxx/util/mtest.cxx:189:42: warning: 'int MPI_Type_extent(MPI_Datatype, MPI_Aint*)' is deprecated: MPI_Type_extent is superseded by MPI_Type_get_extent in MPI-2.0 [-Wdeprecated-declarations]\n ## MPI_Type_extent( mtype->datatype, &size );\n ## ^\n ## In file included from ../../../cxx/util/mtest.cxx:7:\n ## /usr/lib/x86_64-linux-gnu/openmpi/include/mpi.h:1758:20: note: declared here\n@@ -1001692,30 +1001775,30 @@\n ## ../../../cxx/util/mtest.cxx:848:50: error: 'NULL_COPY_FN' is not a member of 'MPI::Win'\n ## mem_keyval = MPI::Win::Create_keyval( MPI::Win::NULL_COPY_FN,\n ## ^~~~~~~~~~~~\n ## ../../../cxx/util/mtest.cxx:849:22: error: 'NULL_DELETE_FN' is not a member of 'MPI::Win'\n ## MPI::Win::NULL_DELETE_FN, 0 );\n ## ^~~~~~~~~~~~~~\n ## make[3]: *** [Makefile:416: mtest.o] Error 1\n-## make[3]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/util'\n+## make[3]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/util'\n ## make[2]: *** [Makefile:989: ../../cxx/util/mtest.o] Error 2\n-## make[2]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/io'\n+## make[2]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/io'\n not ok 699 - ./cxx/io/iwriteshx \n ---\n Directory: ./cxx/io\n File: iwriteshx\n Num-procs: \n- Date: \"Sun Jan 5 01:33:35 2020\"\n+ Date: \"Mon Dec 3 17:04:49 2018\"\n ...\n ## Test output (expected 'No Errors'):\n-## Failed to build iwriteshx; make[2]: Entering directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/io'\n+## Failed to build iwriteshx; make[2]: Entering directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/io'\n ## (cd ../../cxx/util && make mtest.o)\n-## make[3]: Entering directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/util'\n+## make[3]: Entering directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/util'\n ## depbase=`echo mtest.o | sed 's|[^/]*$|.deps/&|;s|\\.o$||'`;\\\n-## /usr/bin/mpicxx.openmpi -DHAVE_CONFIG_H -I. -I../../../cxx/util -I../../include -I../../include -I../../../include -Wdate-time -D_FORTIFY_SOURCE=2 -g -O2 -ffile-prefix-map=/build/1st/mpi-testsuite-3.2+dfsg=. -fstack-protector-strong -Wformat -Werror=format-security -MT mtest.o -MD -MP -MF $depbase.Tpo -c -o mtest.o ../../../cxx/util/mtest.cxx &&\\\n+## /usr/bin/mpicxx.openmpi -DHAVE_CONFIG_H -I. -I../../../cxx/util -I../../include -I../../include -I../../../include -Wdate-time -D_FORTIFY_SOURCE=2 -g -O2 -ffile-prefix-map=/build/mpi-testsuite-3.2+dfsg/2nd=. -fstack-protector-strong -Wformat -Werror=format-security -MT mtest.o -MD -MP -MF $depbase.Tpo -c -o mtest.o ../../../cxx/util/mtest.cxx &&\\\n ## mv -f $depbase.Tpo $depbase.Po\n ## ../../../cxx/util/mtest.cxx: In function 'void* MTestTypeContigInitRecv(MTestDatatype*)':\n ## ../../../cxx/util/mtest.cxx:189:42: warning: 'int MPI_Type_extent(MPI_Datatype, MPI_Aint*)' is deprecated: MPI_Type_extent is superseded by MPI_Type_get_extent in MPI-2.0 [-Wdeprecated-declarations]\n ## MPI_Type_extent( mtype->datatype, &size );\n ## ^\n ## In file included from ../../../cxx/util/mtest.cxx:7:\n ## /usr/lib/x86_64-linux-gnu/openmpi/include/mpi.h:1758:20: note: declared here\n@@ -1001747,30 +1001830,30 @@\n ## ../../../cxx/util/mtest.cxx:848:50: error: 'NULL_COPY_FN' is not a member of 'MPI::Win'\n ## mem_keyval = MPI::Win::Create_keyval( MPI::Win::NULL_COPY_FN,\n ## ^~~~~~~~~~~~\n ## ../../../cxx/util/mtest.cxx:849:22: error: 'NULL_DELETE_FN' is not a member of 'MPI::Win'\n ## MPI::Win::NULL_DELETE_FN, 0 );\n ## ^~~~~~~~~~~~~~\n ## make[3]: *** [Makefile:416: mtest.o] Error 1\n-## make[3]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/util'\n+## make[3]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/util'\n ## make[2]: *** [Makefile:989: ../../cxx/util/mtest.o] Error 2\n-## make[2]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/io'\n+## make[2]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/io'\n not ok 700 - ./cxx/io/writex \n ---\n Directory: ./cxx/io\n File: writex\n Num-procs: \n- Date: \"Sun Jan 5 01:33:36 2020\"\n+ Date: \"Mon Dec 3 17:04:53 2018\"\n ...\n ## Test output (expected 'No Errors'):\n-## Failed to build writex; make[2]: Entering directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/io'\n+## Failed to build writex; make[2]: Entering directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/io'\n ## (cd ../../cxx/util && make mtest.o)\n-## make[3]: Entering directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/util'\n+## make[3]: Entering directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/util'\n ## depbase=`echo mtest.o | sed 's|[^/]*$|.deps/&|;s|\\.o$||'`;\\\n-## /usr/bin/mpicxx.openmpi -DHAVE_CONFIG_H -I. -I../../../cxx/util -I../../include -I../../include -I../../../include -Wdate-time -D_FORTIFY_SOURCE=2 -g -O2 -ffile-prefix-map=/build/1st/mpi-testsuite-3.2+dfsg=. -fstack-protector-strong -Wformat -Werror=format-security -MT mtest.o -MD -MP -MF $depbase.Tpo -c -o mtest.o ../../../cxx/util/mtest.cxx &&\\\n+## /usr/bin/mpicxx.openmpi -DHAVE_CONFIG_H -I. -I../../../cxx/util -I../../include -I../../include -I../../../include -Wdate-time -D_FORTIFY_SOURCE=2 -g -O2 -ffile-prefix-map=/build/mpi-testsuite-3.2+dfsg/2nd=. -fstack-protector-strong -Wformat -Werror=format-security -MT mtest.o -MD -MP -MF $depbase.Tpo -c -o mtest.o ../../../cxx/util/mtest.cxx &&\\\n ## mv -f $depbase.Tpo $depbase.Po\n ## ../../../cxx/util/mtest.cxx: In function 'void* MTestTypeContigInitRecv(MTestDatatype*)':\n ## ../../../cxx/util/mtest.cxx:189:42: warning: 'int MPI_Type_extent(MPI_Datatype, MPI_Aint*)' is deprecated: MPI_Type_extent is superseded by MPI_Type_get_extent in MPI-2.0 [-Wdeprecated-declarations]\n ## MPI_Type_extent( mtype->datatype, &size );\n ## ^\n ## In file included from ../../../cxx/util/mtest.cxx:7:\n ## /usr/lib/x86_64-linux-gnu/openmpi/include/mpi.h:1758:20: note: declared here\n@@ -1001802,30 +1001885,30 @@\n ## ../../../cxx/util/mtest.cxx:848:50: error: 'NULL_COPY_FN' is not a member of 'MPI::Win'\n ## mem_keyval = MPI::Win::Create_keyval( MPI::Win::NULL_COPY_FN,\n ## ^~~~~~~~~~~~\n ## ../../../cxx/util/mtest.cxx:849:22: error: 'NULL_DELETE_FN' is not a member of 'MPI::Win'\n ## MPI::Win::NULL_DELETE_FN, 0 );\n ## ^~~~~~~~~~~~~~\n ## make[3]: *** [Makefile:416: mtest.o] Error 1\n-## make[3]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/util'\n+## make[3]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/util'\n ## make[2]: *** [Makefile:989: ../../cxx/util/mtest.o] Error 2\n-## make[2]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/io'\n+## make[2]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/io'\n not ok 701 - ./cxx/io/writeatx \n ---\n Directory: ./cxx/io\n File: writeatx\n Num-procs: \n- Date: \"Sun Jan 5 01:33:37 2020\"\n+ Date: \"Mon Dec 3 17:04:57 2018\"\n ...\n ## Test output (expected 'No Errors'):\n-## Failed to build writeatx; make[2]: Entering directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/io'\n+## Failed to build writeatx; make[2]: Entering directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/io'\n ## (cd ../../cxx/util && make mtest.o)\n-## make[3]: Entering directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/util'\n+## make[3]: Entering directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/util'\n ## depbase=`echo mtest.o | sed 's|[^/]*$|.deps/&|;s|\\.o$||'`;\\\n-## /usr/bin/mpicxx.openmpi -DHAVE_CONFIG_H -I. -I../../../cxx/util -I../../include -I../../include -I../../../include -Wdate-time -D_FORTIFY_SOURCE=2 -g -O2 -ffile-prefix-map=/build/1st/mpi-testsuite-3.2+dfsg=. -fstack-protector-strong -Wformat -Werror=format-security -MT mtest.o -MD -MP -MF $depbase.Tpo -c -o mtest.o ../../../cxx/util/mtest.cxx &&\\\n+## /usr/bin/mpicxx.openmpi -DHAVE_CONFIG_H -I. -I../../../cxx/util -I../../include -I../../include -I../../../include -Wdate-time -D_FORTIFY_SOURCE=2 -g -O2 -ffile-prefix-map=/build/mpi-testsuite-3.2+dfsg/2nd=. -fstack-protector-strong -Wformat -Werror=format-security -MT mtest.o -MD -MP -MF $depbase.Tpo -c -o mtest.o ../../../cxx/util/mtest.cxx &&\\\n ## mv -f $depbase.Tpo $depbase.Po\n ## ../../../cxx/util/mtest.cxx: In function 'void* MTestTypeContigInitRecv(MTestDatatype*)':\n ## ../../../cxx/util/mtest.cxx:189:42: warning: 'int MPI_Type_extent(MPI_Datatype, MPI_Aint*)' is deprecated: MPI_Type_extent is superseded by MPI_Type_get_extent in MPI-2.0 [-Wdeprecated-declarations]\n ## MPI_Type_extent( mtype->datatype, &size );\n ## ^\n ## In file included from ../../../cxx/util/mtest.cxx:7:\n ## /usr/lib/x86_64-linux-gnu/openmpi/include/mpi.h:1758:20: note: declared here\n@@ -1001857,30 +1001940,30 @@\n ## ../../../cxx/util/mtest.cxx:848:50: error: 'NULL_COPY_FN' is not a member of 'MPI::Win'\n ## mem_keyval = MPI::Win::Create_keyval( MPI::Win::NULL_COPY_FN,\n ## ^~~~~~~~~~~~\n ## ../../../cxx/util/mtest.cxx:849:22: error: 'NULL_DELETE_FN' is not a member of 'MPI::Win'\n ## MPI::Win::NULL_DELETE_FN, 0 );\n ## ^~~~~~~~~~~~~~\n ## make[3]: *** [Makefile:416: mtest.o] Error 1\n-## make[3]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/util'\n+## make[3]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/util'\n ## make[2]: *** [Makefile:989: ../../cxx/util/mtest.o] Error 2\n-## make[2]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/io'\n+## make[2]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/io'\n not ok 702 - ./cxx/io/writeallx \n ---\n Directory: ./cxx/io\n File: writeallx\n Num-procs: \n- Date: \"Sun Jan 5 01:33:39 2020\"\n+ Date: \"Mon Dec 3 17:05:02 2018\"\n ...\n ## Test output (expected 'No Errors'):\n-## Failed to build writeallx; make[2]: Entering directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/io'\n+## Failed to build writeallx; make[2]: Entering directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/io'\n ## (cd ../../cxx/util && make mtest.o)\n-## make[3]: Entering directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/util'\n+## make[3]: Entering directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/util'\n ## depbase=`echo mtest.o | sed 's|[^/]*$|.deps/&|;s|\\.o$||'`;\\\n-## /usr/bin/mpicxx.openmpi -DHAVE_CONFIG_H -I. -I../../../cxx/util -I../../include -I../../include -I../../../include -Wdate-time -D_FORTIFY_SOURCE=2 -g -O2 -ffile-prefix-map=/build/1st/mpi-testsuite-3.2+dfsg=. -fstack-protector-strong -Wformat -Werror=format-security -MT mtest.o -MD -MP -MF $depbase.Tpo -c -o mtest.o ../../../cxx/util/mtest.cxx &&\\\n+## /usr/bin/mpicxx.openmpi -DHAVE_CONFIG_H -I. -I../../../cxx/util -I../../include -I../../include -I../../../include -Wdate-time -D_FORTIFY_SOURCE=2 -g -O2 -ffile-prefix-map=/build/mpi-testsuite-3.2+dfsg/2nd=. -fstack-protector-strong -Wformat -Werror=format-security -MT mtest.o -MD -MP -MF $depbase.Tpo -c -o mtest.o ../../../cxx/util/mtest.cxx &&\\\n ## mv -f $depbase.Tpo $depbase.Po\n ## ../../../cxx/util/mtest.cxx: In function 'void* MTestTypeContigInitRecv(MTestDatatype*)':\n ## ../../../cxx/util/mtest.cxx:189:42: warning: 'int MPI_Type_extent(MPI_Datatype, MPI_Aint*)' is deprecated: MPI_Type_extent is superseded by MPI_Type_get_extent in MPI-2.0 [-Wdeprecated-declarations]\n ## MPI_Type_extent( mtype->datatype, &size );\n ## ^\n ## In file included from ../../../cxx/util/mtest.cxx:7:\n ## /usr/lib/x86_64-linux-gnu/openmpi/include/mpi.h:1758:20: note: declared here\n@@ -1001912,30 +1001995,30 @@\n ## ../../../cxx/util/mtest.cxx:848:50: error: 'NULL_COPY_FN' is not a member of 'MPI::Win'\n ## mem_keyval = MPI::Win::Create_keyval( MPI::Win::NULL_COPY_FN,\n ## ^~~~~~~~~~~~\n ## ../../../cxx/util/mtest.cxx:849:22: error: 'NULL_DELETE_FN' is not a member of 'MPI::Win'\n ## MPI::Win::NULL_DELETE_FN, 0 );\n ## ^~~~~~~~~~~~~~\n ## make[3]: *** [Makefile:416: mtest.o] Error 1\n-## make[3]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/util'\n+## make[3]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/util'\n ## make[2]: *** [Makefile:989: ../../cxx/util/mtest.o] Error 2\n-## make[2]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/io'\n+## make[2]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/io'\n not ok 703 - ./cxx/io/writeshx \n ---\n Directory: ./cxx/io\n File: writeshx\n Num-procs: \n- Date: \"Sun Jan 5 01:33:40 2020\"\n+ Date: \"Mon Dec 3 17:05:06 2018\"\n ...\n ## Test output (expected 'No Errors'):\n-## Failed to build writeshx; make[2]: Entering directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/io'\n+## Failed to build writeshx; make[2]: Entering directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/io'\n ## (cd ../../cxx/util && make mtest.o)\n-## make[3]: Entering directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/util'\n+## make[3]: Entering directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/util'\n ## depbase=`echo mtest.o | sed 's|[^/]*$|.deps/&|;s|\\.o$||'`;\\\n-## /usr/bin/mpicxx.openmpi -DHAVE_CONFIG_H -I. -I../../../cxx/util -I../../include -I../../include -I../../../include -Wdate-time -D_FORTIFY_SOURCE=2 -g -O2 -ffile-prefix-map=/build/1st/mpi-testsuite-3.2+dfsg=. -fstack-protector-strong -Wformat -Werror=format-security -MT mtest.o -MD -MP -MF $depbase.Tpo -c -o mtest.o ../../../cxx/util/mtest.cxx &&\\\n+## /usr/bin/mpicxx.openmpi -DHAVE_CONFIG_H -I. -I../../../cxx/util -I../../include -I../../include -I../../../include -Wdate-time -D_FORTIFY_SOURCE=2 -g -O2 -ffile-prefix-map=/build/mpi-testsuite-3.2+dfsg/2nd=. -fstack-protector-strong -Wformat -Werror=format-security -MT mtest.o -MD -MP -MF $depbase.Tpo -c -o mtest.o ../../../cxx/util/mtest.cxx &&\\\n ## mv -f $depbase.Tpo $depbase.Po\n ## ../../../cxx/util/mtest.cxx: In function 'void* MTestTypeContigInitRecv(MTestDatatype*)':\n ## ../../../cxx/util/mtest.cxx:189:42: warning: 'int MPI_Type_extent(MPI_Datatype, MPI_Aint*)' is deprecated: MPI_Type_extent is superseded by MPI_Type_get_extent in MPI-2.0 [-Wdeprecated-declarations]\n ## MPI_Type_extent( mtype->datatype, &size );\n ## ^\n ## In file included from ../../../cxx/util/mtest.cxx:7:\n ## /usr/lib/x86_64-linux-gnu/openmpi/include/mpi.h:1758:20: note: declared here\n@@ -1001967,30 +1002050,30 @@\n ## ../../../cxx/util/mtest.cxx:848:50: error: 'NULL_COPY_FN' is not a member of 'MPI::Win'\n ## mem_keyval = MPI::Win::Create_keyval( MPI::Win::NULL_COPY_FN,\n ## ^~~~~~~~~~~~\n ## ../../../cxx/util/mtest.cxx:849:22: error: 'NULL_DELETE_FN' is not a member of 'MPI::Win'\n ## MPI::Win::NULL_DELETE_FN, 0 );\n ## ^~~~~~~~~~~~~~\n ## make[3]: *** [Makefile:416: mtest.o] Error 1\n-## make[3]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/util'\n+## make[3]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/util'\n ## make[2]: *** [Makefile:989: ../../cxx/util/mtest.o] Error 2\n-## make[2]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/io'\n+## make[2]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/io'\n not ok 704 - ./cxx/io/writeordx \n ---\n Directory: ./cxx/io\n File: writeordx\n Num-procs: \n- Date: \"Sun Jan 5 01:33:41 2020\"\n+ Date: \"Mon Dec 3 17:05:10 2018\"\n ...\n ## Test output (expected 'No Errors'):\n-## Failed to build writeordx; make[2]: Entering directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/io'\n+## Failed to build writeordx; make[2]: Entering directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/io'\n ## (cd ../../cxx/util && make mtest.o)\n-## make[3]: Entering directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/util'\n+## make[3]: Entering directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/util'\n ## depbase=`echo mtest.o | sed 's|[^/]*$|.deps/&|;s|\\.o$||'`;\\\n-## /usr/bin/mpicxx.openmpi -DHAVE_CONFIG_H -I. -I../../../cxx/util -I../../include -I../../include -I../../../include -Wdate-time -D_FORTIFY_SOURCE=2 -g -O2 -ffile-prefix-map=/build/1st/mpi-testsuite-3.2+dfsg=. -fstack-protector-strong -Wformat -Werror=format-security -MT mtest.o -MD -MP -MF $depbase.Tpo -c -o mtest.o ../../../cxx/util/mtest.cxx &&\\\n+## /usr/bin/mpicxx.openmpi -DHAVE_CONFIG_H -I. -I../../../cxx/util -I../../include -I../../include -I../../../include -Wdate-time -D_FORTIFY_SOURCE=2 -g -O2 -ffile-prefix-map=/build/mpi-testsuite-3.2+dfsg/2nd=. -fstack-protector-strong -Wformat -Werror=format-security -MT mtest.o -MD -MP -MF $depbase.Tpo -c -o mtest.o ../../../cxx/util/mtest.cxx &&\\\n ## mv -f $depbase.Tpo $depbase.Po\n ## ../../../cxx/util/mtest.cxx: In function 'void* MTestTypeContigInitRecv(MTestDatatype*)':\n ## ../../../cxx/util/mtest.cxx:189:42: warning: 'int MPI_Type_extent(MPI_Datatype, MPI_Aint*)' is deprecated: MPI_Type_extent is superseded by MPI_Type_get_extent in MPI-2.0 [-Wdeprecated-declarations]\n ## MPI_Type_extent( mtype->datatype, &size );\n ## ^\n ## In file included from ../../../cxx/util/mtest.cxx:7:\n ## /usr/lib/x86_64-linux-gnu/openmpi/include/mpi.h:1758:20: note: declared here\n@@ -1002022,30 +1002105,30 @@\n ## ../../../cxx/util/mtest.cxx:848:50: error: 'NULL_COPY_FN' is not a member of 'MPI::Win'\n ## mem_keyval = MPI::Win::Create_keyval( MPI::Win::NULL_COPY_FN,\n ## ^~~~~~~~~~~~\n ## ../../../cxx/util/mtest.cxx:849:22: error: 'NULL_DELETE_FN' is not a member of 'MPI::Win'\n ## MPI::Win::NULL_DELETE_FN, 0 );\n ## ^~~~~~~~~~~~~~\n ## make[3]: *** [Makefile:416: mtest.o] Error 1\n-## make[3]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/util'\n+## make[3]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/util'\n ## make[2]: *** [Makefile:989: ../../cxx/util/mtest.o] Error 2\n-## make[2]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/io'\n+## make[2]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/io'\n not ok 705 - ./cxx/io/writeatallx \n ---\n Directory: ./cxx/io\n File: writeatallx\n Num-procs: \n- Date: \"Sun Jan 5 01:33:42 2020\"\n+ Date: \"Mon Dec 3 17:05:14 2018\"\n ...\n ## Test output (expected 'No Errors'):\n-## Failed to build writeatallx; make[2]: Entering directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/io'\n+## Failed to build writeatallx; make[2]: Entering directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/io'\n ## (cd ../../cxx/util && make mtest.o)\n-## make[3]: Entering directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/util'\n+## make[3]: Entering directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/util'\n ## depbase=`echo mtest.o | sed 's|[^/]*$|.deps/&|;s|\\.o$||'`;\\\n-## /usr/bin/mpicxx.openmpi -DHAVE_CONFIG_H -I. -I../../../cxx/util -I../../include -I../../include -I../../../include -Wdate-time -D_FORTIFY_SOURCE=2 -g -O2 -ffile-prefix-map=/build/1st/mpi-testsuite-3.2+dfsg=. -fstack-protector-strong -Wformat -Werror=format-security -MT mtest.o -MD -MP -MF $depbase.Tpo -c -o mtest.o ../../../cxx/util/mtest.cxx &&\\\n+## /usr/bin/mpicxx.openmpi -DHAVE_CONFIG_H -I. -I../../../cxx/util -I../../include -I../../include -I../../../include -Wdate-time -D_FORTIFY_SOURCE=2 -g -O2 -ffile-prefix-map=/build/mpi-testsuite-3.2+dfsg/2nd=. -fstack-protector-strong -Wformat -Werror=format-security -MT mtest.o -MD -MP -MF $depbase.Tpo -c -o mtest.o ../../../cxx/util/mtest.cxx &&\\\n ## mv -f $depbase.Tpo $depbase.Po\n ## ../../../cxx/util/mtest.cxx: In function 'void* MTestTypeContigInitRecv(MTestDatatype*)':\n ## ../../../cxx/util/mtest.cxx:189:42: warning: 'int MPI_Type_extent(MPI_Datatype, MPI_Aint*)' is deprecated: MPI_Type_extent is superseded by MPI_Type_get_extent in MPI-2.0 [-Wdeprecated-declarations]\n ## MPI_Type_extent( mtype->datatype, &size );\n ## ^\n ## In file included from ../../../cxx/util/mtest.cxx:7:\n ## /usr/lib/x86_64-linux-gnu/openmpi/include/mpi.h:1758:20: note: declared here\n@@ -1002077,30 +1002160,30 @@\n ## ../../../cxx/util/mtest.cxx:848:50: error: 'NULL_COPY_FN' is not a member of 'MPI::Win'\n ## mem_keyval = MPI::Win::Create_keyval( MPI::Win::NULL_COPY_FN,\n ## ^~~~~~~~~~~~\n ## ../../../cxx/util/mtest.cxx:849:22: error: 'NULL_DELETE_FN' is not a member of 'MPI::Win'\n ## MPI::Win::NULL_DELETE_FN, 0 );\n ## ^~~~~~~~~~~~~~\n ## make[3]: *** [Makefile:416: mtest.o] Error 1\n-## make[3]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/util'\n+## make[3]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/util'\n ## make[2]: *** [Makefile:989: ../../cxx/util/mtest.o] Error 2\n-## make[2]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/io'\n+## make[2]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/io'\n not ok 706 - ./cxx/io/writeatallbex \n ---\n Directory: ./cxx/io\n File: writeatallbex\n Num-procs: \n- Date: \"Sun Jan 5 01:33:43 2020\"\n+ Date: \"Mon Dec 3 17:05:16 2018\"\n ...\n ## Test output (expected 'No Errors'):\n-## Failed to build writeatallbex; make[2]: Entering directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/io'\n+## Failed to build writeatallbex; make[2]: Entering directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/io'\n ## (cd ../../cxx/util && make mtest.o)\n-## make[3]: Entering directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/util'\n+## make[3]: Entering directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/util'\n ## depbase=`echo mtest.o | sed 's|[^/]*$|.deps/&|;s|\\.o$||'`;\\\n-## /usr/bin/mpicxx.openmpi -DHAVE_CONFIG_H -I. -I../../../cxx/util -I../../include -I../../include -I../../../include -Wdate-time -D_FORTIFY_SOURCE=2 -g -O2 -ffile-prefix-map=/build/1st/mpi-testsuite-3.2+dfsg=. -fstack-protector-strong -Wformat -Werror=format-security -MT mtest.o -MD -MP -MF $depbase.Tpo -c -o mtest.o ../../../cxx/util/mtest.cxx &&\\\n+## /usr/bin/mpicxx.openmpi -DHAVE_CONFIG_H -I. -I../../../cxx/util -I../../include -I../../include -I../../../include -Wdate-time -D_FORTIFY_SOURCE=2 -g -O2 -ffile-prefix-map=/build/mpi-testsuite-3.2+dfsg/2nd=. -fstack-protector-strong -Wformat -Werror=format-security -MT mtest.o -MD -MP -MF $depbase.Tpo -c -o mtest.o ../../../cxx/util/mtest.cxx &&\\\n ## mv -f $depbase.Tpo $depbase.Po\n ## ../../../cxx/util/mtest.cxx: In function 'void* MTestTypeContigInitRecv(MTestDatatype*)':\n ## ../../../cxx/util/mtest.cxx:189:42: warning: 'int MPI_Type_extent(MPI_Datatype, MPI_Aint*)' is deprecated: MPI_Type_extent is superseded by MPI_Type_get_extent in MPI-2.0 [-Wdeprecated-declarations]\n ## MPI_Type_extent( mtype->datatype, &size );\n ## ^\n ## In file included from ../../../cxx/util/mtest.cxx:7:\n ## /usr/lib/x86_64-linux-gnu/openmpi/include/mpi.h:1758:20: note: declared here\n@@ -1002132,30 +1002215,30 @@\n ## ../../../cxx/util/mtest.cxx:848:50: error: 'NULL_COPY_FN' is not a member of 'MPI::Win'\n ## mem_keyval = MPI::Win::Create_keyval( MPI::Win::NULL_COPY_FN,\n ## ^~~~~~~~~~~~\n ## ../../../cxx/util/mtest.cxx:849:22: error: 'NULL_DELETE_FN' is not a member of 'MPI::Win'\n ## MPI::Win::NULL_DELETE_FN, 0 );\n ## ^~~~~~~~~~~~~~\n ## make[3]: *** [Makefile:416: mtest.o] Error 1\n-## make[3]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/util'\n+## make[3]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/util'\n ## make[2]: *** [Makefile:989: ../../cxx/util/mtest.o] Error 2\n-## make[2]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/io'\n+## make[2]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/io'\n not ok 707 - ./cxx/io/writeallbex \n ---\n Directory: ./cxx/io\n File: writeallbex\n Num-procs: \n- Date: \"Sun Jan 5 01:33:44 2020\"\n+ Date: \"Mon Dec 3 17:05:20 2018\"\n ...\n ## Test output (expected 'No Errors'):\n-## Failed to build writeallbex; make[2]: Entering directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/io'\n+## Failed to build writeallbex; make[2]: Entering directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/io'\n ## (cd ../../cxx/util && make mtest.o)\n-## make[3]: Entering directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/util'\n+## make[3]: Entering directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/util'\n ## depbase=`echo mtest.o | sed 's|[^/]*$|.deps/&|;s|\\.o$||'`;\\\n-## /usr/bin/mpicxx.openmpi -DHAVE_CONFIG_H -I. -I../../../cxx/util -I../../include -I../../include -I../../../include -Wdate-time -D_FORTIFY_SOURCE=2 -g -O2 -ffile-prefix-map=/build/1st/mpi-testsuite-3.2+dfsg=. -fstack-protector-strong -Wformat -Werror=format-security -MT mtest.o -MD -MP -MF $depbase.Tpo -c -o mtest.o ../../../cxx/util/mtest.cxx &&\\\n+## /usr/bin/mpicxx.openmpi -DHAVE_CONFIG_H -I. -I../../../cxx/util -I../../include -I../../include -I../../../include -Wdate-time -D_FORTIFY_SOURCE=2 -g -O2 -ffile-prefix-map=/build/mpi-testsuite-3.2+dfsg/2nd=. -fstack-protector-strong -Wformat -Werror=format-security -MT mtest.o -MD -MP -MF $depbase.Tpo -c -o mtest.o ../../../cxx/util/mtest.cxx &&\\\n ## mv -f $depbase.Tpo $depbase.Po\n ## ../../../cxx/util/mtest.cxx: In function 'void* MTestTypeContigInitRecv(MTestDatatype*)':\n ## ../../../cxx/util/mtest.cxx:189:42: warning: 'int MPI_Type_extent(MPI_Datatype, MPI_Aint*)' is deprecated: MPI_Type_extent is superseded by MPI_Type_get_extent in MPI-2.0 [-Wdeprecated-declarations]\n ## MPI_Type_extent( mtype->datatype, &size );\n ## ^\n ## In file included from ../../../cxx/util/mtest.cxx:7:\n ## /usr/lib/x86_64-linux-gnu/openmpi/include/mpi.h:1758:20: note: declared here\n@@ -1002187,30 +1002270,30 @@\n ## ../../../cxx/util/mtest.cxx:848:50: error: 'NULL_COPY_FN' is not a member of 'MPI::Win'\n ## mem_keyval = MPI::Win::Create_keyval( MPI::Win::NULL_COPY_FN,\n ## ^~~~~~~~~~~~\n ## ../../../cxx/util/mtest.cxx:849:22: error: 'NULL_DELETE_FN' is not a member of 'MPI::Win'\n ## MPI::Win::NULL_DELETE_FN, 0 );\n ## ^~~~~~~~~~~~~~\n ## make[3]: *** [Makefile:416: mtest.o] Error 1\n-## make[3]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/util'\n+## make[3]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/util'\n ## make[2]: *** [Makefile:989: ../../cxx/util/mtest.o] Error 2\n-## make[2]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/io'\n+## make[2]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/io'\n not ok 708 - ./cxx/io/writeordbex \n ---\n Directory: ./cxx/io\n File: writeordbex\n Num-procs: \n- Date: \"Sun Jan 5 01:33:45 2020\"\n+ Date: \"Mon Dec 3 17:05:25 2018\"\n ...\n ## Test output (expected 'No Errors'):\n-## Failed to build writeordbex; make[2]: Entering directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/io'\n+## Failed to build writeordbex; make[2]: Entering directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/io'\n ## (cd ../../cxx/util && make mtest.o)\n-## make[3]: Entering directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/util'\n+## make[3]: Entering directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/util'\n ## depbase=`echo mtest.o | sed 's|[^/]*$|.deps/&|;s|\\.o$||'`;\\\n-## /usr/bin/mpicxx.openmpi -DHAVE_CONFIG_H -I. -I../../../cxx/util -I../../include -I../../include -I../../../include -Wdate-time -D_FORTIFY_SOURCE=2 -g -O2 -ffile-prefix-map=/build/1st/mpi-testsuite-3.2+dfsg=. -fstack-protector-strong -Wformat -Werror=format-security -MT mtest.o -MD -MP -MF $depbase.Tpo -c -o mtest.o ../../../cxx/util/mtest.cxx &&\\\n+## /usr/bin/mpicxx.openmpi -DHAVE_CONFIG_H -I. -I../../../cxx/util -I../../include -I../../include -I../../../include -Wdate-time -D_FORTIFY_SOURCE=2 -g -O2 -ffile-prefix-map=/build/mpi-testsuite-3.2+dfsg/2nd=. -fstack-protector-strong -Wformat -Werror=format-security -MT mtest.o -MD -MP -MF $depbase.Tpo -c -o mtest.o ../../../cxx/util/mtest.cxx &&\\\n ## mv -f $depbase.Tpo $depbase.Po\n ## ../../../cxx/util/mtest.cxx: In function 'void* MTestTypeContigInitRecv(MTestDatatype*)':\n ## ../../../cxx/util/mtest.cxx:189:42: warning: 'int MPI_Type_extent(MPI_Datatype, MPI_Aint*)' is deprecated: MPI_Type_extent is superseded by MPI_Type_get_extent in MPI-2.0 [-Wdeprecated-declarations]\n ## MPI_Type_extent( mtype->datatype, &size );\n ## ^\n ## In file included from ../../../cxx/util/mtest.cxx:7:\n ## /usr/lib/x86_64-linux-gnu/openmpi/include/mpi.h:1758:20: note: declared here\n@@ -1002242,30 +1002325,30 @@\n ## ../../../cxx/util/mtest.cxx:848:50: error: 'NULL_COPY_FN' is not a member of 'MPI::Win'\n ## mem_keyval = MPI::Win::Create_keyval( MPI::Win::NULL_COPY_FN,\n ## ^~~~~~~~~~~~\n ## ../../../cxx/util/mtest.cxx:849:22: error: 'NULL_DELETE_FN' is not a member of 'MPI::Win'\n ## MPI::Win::NULL_DELETE_FN, 0 );\n ## ^~~~~~~~~~~~~~\n ## make[3]: *** [Makefile:416: mtest.o] Error 1\n-## make[3]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/util'\n+## make[3]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/util'\n ## make[2]: *** [Makefile:989: ../../cxx/util/mtest.o] Error 2\n-## make[2]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/io'\n+## make[2]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/io'\n not ok 709 - ./cxx/io/iwriteatnosx \n ---\n Directory: ./cxx/io\n File: iwriteatnosx\n Num-procs: \n- Date: \"Sun Jan 5 01:33:46 2020\"\n+ Date: \"Mon Dec 3 17:05:29 2018\"\n ...\n ## Test output (expected 'No Errors'):\n-## Failed to build iwriteatnosx; make[2]: Entering directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/io'\n+## Failed to build iwriteatnosx; make[2]: Entering directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/io'\n ## (cd ../../cxx/util && make mtest.o)\n-## make[3]: Entering directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/util'\n+## make[3]: Entering directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/util'\n ## depbase=`echo mtest.o | sed 's|[^/]*$|.deps/&|;s|\\.o$||'`;\\\n-## /usr/bin/mpicxx.openmpi -DHAVE_CONFIG_H -I. -I../../../cxx/util -I../../include -I../../include -I../../../include -Wdate-time -D_FORTIFY_SOURCE=2 -g -O2 -ffile-prefix-map=/build/1st/mpi-testsuite-3.2+dfsg=. -fstack-protector-strong -Wformat -Werror=format-security -MT mtest.o -MD -MP -MF $depbase.Tpo -c -o mtest.o ../../../cxx/util/mtest.cxx &&\\\n+## /usr/bin/mpicxx.openmpi -DHAVE_CONFIG_H -I. -I../../../cxx/util -I../../include -I../../include -I../../../include -Wdate-time -D_FORTIFY_SOURCE=2 -g -O2 -ffile-prefix-map=/build/mpi-testsuite-3.2+dfsg/2nd=. -fstack-protector-strong -Wformat -Werror=format-security -MT mtest.o -MD -MP -MF $depbase.Tpo -c -o mtest.o ../../../cxx/util/mtest.cxx &&\\\n ## mv -f $depbase.Tpo $depbase.Po\n ## ../../../cxx/util/mtest.cxx: In function 'void* MTestTypeContigInitRecv(MTestDatatype*)':\n ## ../../../cxx/util/mtest.cxx:189:42: warning: 'int MPI_Type_extent(MPI_Datatype, MPI_Aint*)' is deprecated: MPI_Type_extent is superseded by MPI_Type_get_extent in MPI-2.0 [-Wdeprecated-declarations]\n ## MPI_Type_extent( mtype->datatype, &size );\n ## ^\n ## In file included from ../../../cxx/util/mtest.cxx:7:\n ## /usr/lib/x86_64-linux-gnu/openmpi/include/mpi.h:1758:20: note: declared here\n@@ -1002297,30 +1002380,30 @@\n ## ../../../cxx/util/mtest.cxx:848:50: error: 'NULL_COPY_FN' is not a member of 'MPI::Win'\n ## mem_keyval = MPI::Win::Create_keyval( MPI::Win::NULL_COPY_FN,\n ## ^~~~~~~~~~~~\n ## ../../../cxx/util/mtest.cxx:849:22: error: 'NULL_DELETE_FN' is not a member of 'MPI::Win'\n ## MPI::Win::NULL_DELETE_FN, 0 );\n ## ^~~~~~~~~~~~~~\n ## make[3]: *** [Makefile:416: mtest.o] Error 1\n-## make[3]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/util'\n+## make[3]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/util'\n ## make[2]: *** [Makefile:989: ../../cxx/util/mtest.o] Error 2\n-## make[2]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/io'\n+## make[2]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/io'\n not ok 710 - ./cxx/io/iwritenosx \n ---\n Directory: ./cxx/io\n File: iwritenosx\n Num-procs: \n- Date: \"Sun Jan 5 01:33:47 2020\"\n+ Date: \"Mon Dec 3 17:05:33 2018\"\n ...\n ## Test output (expected 'No Errors'):\n-## Failed to build iwritenosx; make[2]: Entering directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/io'\n+## Failed to build iwritenosx; make[2]: Entering directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/io'\n ## (cd ../../cxx/util && make mtest.o)\n-## make[3]: Entering directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/util'\n+## make[3]: Entering directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/util'\n ## depbase=`echo mtest.o | sed 's|[^/]*$|.deps/&|;s|\\.o$||'`;\\\n-## /usr/bin/mpicxx.openmpi -DHAVE_CONFIG_H -I. -I../../../cxx/util -I../../include -I../../include -I../../../include -Wdate-time -D_FORTIFY_SOURCE=2 -g -O2 -ffile-prefix-map=/build/1st/mpi-testsuite-3.2+dfsg=. -fstack-protector-strong -Wformat -Werror=format-security -MT mtest.o -MD -MP -MF $depbase.Tpo -c -o mtest.o ../../../cxx/util/mtest.cxx &&\\\n+## /usr/bin/mpicxx.openmpi -DHAVE_CONFIG_H -I. -I../../../cxx/util -I../../include -I../../include -I../../../include -Wdate-time -D_FORTIFY_SOURCE=2 -g -O2 -ffile-prefix-map=/build/mpi-testsuite-3.2+dfsg/2nd=. -fstack-protector-strong -Wformat -Werror=format-security -MT mtest.o -MD -MP -MF $depbase.Tpo -c -o mtest.o ../../../cxx/util/mtest.cxx &&\\\n ## mv -f $depbase.Tpo $depbase.Po\n ## ../../../cxx/util/mtest.cxx: In function 'void* MTestTypeContigInitRecv(MTestDatatype*)':\n ## ../../../cxx/util/mtest.cxx:189:42: warning: 'int MPI_Type_extent(MPI_Datatype, MPI_Aint*)' is deprecated: MPI_Type_extent is superseded by MPI_Type_get_extent in MPI-2.0 [-Wdeprecated-declarations]\n ## MPI_Type_extent( mtype->datatype, &size );\n ## ^\n ## In file included from ../../../cxx/util/mtest.cxx:7:\n ## /usr/lib/x86_64-linux-gnu/openmpi/include/mpi.h:1758:20: note: declared here\n@@ -1002352,30 +1002435,30 @@\n ## ../../../cxx/util/mtest.cxx:848:50: error: 'NULL_COPY_FN' is not a member of 'MPI::Win'\n ## mem_keyval = MPI::Win::Create_keyval( MPI::Win::NULL_COPY_FN,\n ## ^~~~~~~~~~~~\n ## ../../../cxx/util/mtest.cxx:849:22: error: 'NULL_DELETE_FN' is not a member of 'MPI::Win'\n ## MPI::Win::NULL_DELETE_FN, 0 );\n ## ^~~~~~~~~~~~~~\n ## make[3]: *** [Makefile:416: mtest.o] Error 1\n-## make[3]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/util'\n+## make[3]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/util'\n ## make[2]: *** [Makefile:989: ../../cxx/util/mtest.o] Error 2\n-## make[2]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/io'\n+## make[2]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/io'\n not ok 711 - ./cxx/io/iwriteshnosx \n ---\n Directory: ./cxx/io\n File: iwriteshnosx\n Num-procs: \n- Date: \"Sun Jan 5 01:33:48 2020\"\n+ Date: \"Mon Dec 3 17:05:37 2018\"\n ...\n ## Test output (expected 'No Errors'):\n-## Failed to build iwriteshnosx; make[2]: Entering directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/io'\n+## Failed to build iwriteshnosx; make[2]: Entering directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/io'\n ## (cd ../../cxx/util && make mtest.o)\n-## make[3]: Entering directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/util'\n+## make[3]: Entering directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/util'\n ## depbase=`echo mtest.o | sed 's|[^/]*$|.deps/&|;s|\\.o$||'`;\\\n-## /usr/bin/mpicxx.openmpi -DHAVE_CONFIG_H -I. -I../../../cxx/util -I../../include -I../../include -I../../../include -Wdate-time -D_FORTIFY_SOURCE=2 -g -O2 -ffile-prefix-map=/build/1st/mpi-testsuite-3.2+dfsg=. -fstack-protector-strong -Wformat -Werror=format-security -MT mtest.o -MD -MP -MF $depbase.Tpo -c -o mtest.o ../../../cxx/util/mtest.cxx &&\\\n+## /usr/bin/mpicxx.openmpi -DHAVE_CONFIG_H -I. -I../../../cxx/util -I../../include -I../../include -I../../../include -Wdate-time -D_FORTIFY_SOURCE=2 -g -O2 -ffile-prefix-map=/build/mpi-testsuite-3.2+dfsg/2nd=. -fstack-protector-strong -Wformat -Werror=format-security -MT mtest.o -MD -MP -MF $depbase.Tpo -c -o mtest.o ../../../cxx/util/mtest.cxx &&\\\n ## mv -f $depbase.Tpo $depbase.Po\n ## ../../../cxx/util/mtest.cxx: In function 'void* MTestTypeContigInitRecv(MTestDatatype*)':\n ## ../../../cxx/util/mtest.cxx:189:42: warning: 'int MPI_Type_extent(MPI_Datatype, MPI_Aint*)' is deprecated: MPI_Type_extent is superseded by MPI_Type_get_extent in MPI-2.0 [-Wdeprecated-declarations]\n ## MPI_Type_extent( mtype->datatype, &size );\n ## ^\n ## In file included from ../../../cxx/util/mtest.cxx:7:\n ## /usr/lib/x86_64-linux-gnu/openmpi/include/mpi.h:1758:20: note: declared here\n@@ -1002407,30 +1002490,30 @@\n ## ../../../cxx/util/mtest.cxx:848:50: error: 'NULL_COPY_FN' is not a member of 'MPI::Win'\n ## mem_keyval = MPI::Win::Create_keyval( MPI::Win::NULL_COPY_FN,\n ## ^~~~~~~~~~~~\n ## ../../../cxx/util/mtest.cxx:849:22: error: 'NULL_DELETE_FN' is not a member of 'MPI::Win'\n ## MPI::Win::NULL_DELETE_FN, 0 );\n ## ^~~~~~~~~~~~~~\n ## make[3]: *** [Makefile:416: mtest.o] Error 1\n-## make[3]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/util'\n+## make[3]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/util'\n ## make[2]: *** [Makefile:989: ../../cxx/util/mtest.o] Error 2\n-## make[2]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/io'\n+## make[2]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/io'\n not ok 712 - ./cxx/io/writenosx \n ---\n Directory: ./cxx/io\n File: writenosx\n Num-procs: \n- Date: \"Sun Jan 5 01:33:49 2020\"\n+ Date: \"Mon Dec 3 17:05:41 2018\"\n ...\n ## Test output (expected 'No Errors'):\n-## Failed to build writenosx; make[2]: Entering directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/io'\n+## Failed to build writenosx; make[2]: Entering directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/io'\n ## (cd ../../cxx/util && make mtest.o)\n-## make[3]: Entering directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/util'\n+## make[3]: Entering directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/util'\n ## depbase=`echo mtest.o | sed 's|[^/]*$|.deps/&|;s|\\.o$||'`;\\\n-## /usr/bin/mpicxx.openmpi -DHAVE_CONFIG_H -I. -I../../../cxx/util -I../../include -I../../include -I../../../include -Wdate-time -D_FORTIFY_SOURCE=2 -g -O2 -ffile-prefix-map=/build/1st/mpi-testsuite-3.2+dfsg=. -fstack-protector-strong -Wformat -Werror=format-security -MT mtest.o -MD -MP -MF $depbase.Tpo -c -o mtest.o ../../../cxx/util/mtest.cxx &&\\\n+## /usr/bin/mpicxx.openmpi -DHAVE_CONFIG_H -I. -I../../../cxx/util -I../../include -I../../include -I../../../include -Wdate-time -D_FORTIFY_SOURCE=2 -g -O2 -ffile-prefix-map=/build/mpi-testsuite-3.2+dfsg/2nd=. -fstack-protector-strong -Wformat -Werror=format-security -MT mtest.o -MD -MP -MF $depbase.Tpo -c -o mtest.o ../../../cxx/util/mtest.cxx &&\\\n ## mv -f $depbase.Tpo $depbase.Po\n ## ../../../cxx/util/mtest.cxx: In function 'void* MTestTypeContigInitRecv(MTestDatatype*)':\n ## ../../../cxx/util/mtest.cxx:189:42: warning: 'int MPI_Type_extent(MPI_Datatype, MPI_Aint*)' is deprecated: MPI_Type_extent is superseded by MPI_Type_get_extent in MPI-2.0 [-Wdeprecated-declarations]\n ## MPI_Type_extent( mtype->datatype, &size );\n ## ^\n ## In file included from ../../../cxx/util/mtest.cxx:7:\n ## /usr/lib/x86_64-linux-gnu/openmpi/include/mpi.h:1758:20: note: declared here\n@@ -1002462,30 +1002545,30 @@\n ## ../../../cxx/util/mtest.cxx:848:50: error: 'NULL_COPY_FN' is not a member of 'MPI::Win'\n ## mem_keyval = MPI::Win::Create_keyval( MPI::Win::NULL_COPY_FN,\n ## ^~~~~~~~~~~~\n ## ../../../cxx/util/mtest.cxx:849:22: error: 'NULL_DELETE_FN' is not a member of 'MPI::Win'\n ## MPI::Win::NULL_DELETE_FN, 0 );\n ## ^~~~~~~~~~~~~~\n ## make[3]: *** [Makefile:416: mtest.o] Error 1\n-## make[3]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/util'\n+## make[3]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/util'\n ## make[2]: *** [Makefile:989: ../../cxx/util/mtest.o] Error 2\n-## make[2]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/io'\n+## make[2]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/io'\n not ok 713 - ./cxx/io/writeatnosx \n ---\n Directory: ./cxx/io\n File: writeatnosx\n Num-procs: \n- Date: \"Sun Jan 5 01:33:50 2020\"\n+ Date: \"Mon Dec 3 17:05:45 2018\"\n ...\n ## Test output (expected 'No Errors'):\n-## Failed to build writeatnosx; make[2]: Entering directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/io'\n+## Failed to build writeatnosx; make[2]: Entering directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/io'\n ## (cd ../../cxx/util && make mtest.o)\n-## make[3]: Entering directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/util'\n+## make[3]: Entering directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/util'\n ## depbase=`echo mtest.o | sed 's|[^/]*$|.deps/&|;s|\\.o$||'`;\\\n-## /usr/bin/mpicxx.openmpi -DHAVE_CONFIG_H -I. -I../../../cxx/util -I../../include -I../../include -I../../../include -Wdate-time -D_FORTIFY_SOURCE=2 -g -O2 -ffile-prefix-map=/build/1st/mpi-testsuite-3.2+dfsg=. -fstack-protector-strong -Wformat -Werror=format-security -MT mtest.o -MD -MP -MF $depbase.Tpo -c -o mtest.o ../../../cxx/util/mtest.cxx &&\\\n+## /usr/bin/mpicxx.openmpi -DHAVE_CONFIG_H -I. -I../../../cxx/util -I../../include -I../../include -I../../../include -Wdate-time -D_FORTIFY_SOURCE=2 -g -O2 -ffile-prefix-map=/build/mpi-testsuite-3.2+dfsg/2nd=. -fstack-protector-strong -Wformat -Werror=format-security -MT mtest.o -MD -MP -MF $depbase.Tpo -c -o mtest.o ../../../cxx/util/mtest.cxx &&\\\n ## mv -f $depbase.Tpo $depbase.Po\n ## ../../../cxx/util/mtest.cxx: In function 'void* MTestTypeContigInitRecv(MTestDatatype*)':\n ## ../../../cxx/util/mtest.cxx:189:42: warning: 'int MPI_Type_extent(MPI_Datatype, MPI_Aint*)' is deprecated: MPI_Type_extent is superseded by MPI_Type_get_extent in MPI-2.0 [-Wdeprecated-declarations]\n ## MPI_Type_extent( mtype->datatype, &size );\n ## ^\n ## In file included from ../../../cxx/util/mtest.cxx:7:\n ## /usr/lib/x86_64-linux-gnu/openmpi/include/mpi.h:1758:20: note: declared here\n@@ -1002517,30 +1002600,30 @@\n ## ../../../cxx/util/mtest.cxx:848:50: error: 'NULL_COPY_FN' is not a member of 'MPI::Win'\n ## mem_keyval = MPI::Win::Create_keyval( MPI::Win::NULL_COPY_FN,\n ## ^~~~~~~~~~~~\n ## ../../../cxx/util/mtest.cxx:849:22: error: 'NULL_DELETE_FN' is not a member of 'MPI::Win'\n ## MPI::Win::NULL_DELETE_FN, 0 );\n ## ^~~~~~~~~~~~~~\n ## make[3]: *** [Makefile:416: mtest.o] Error 1\n-## make[3]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/util'\n+## make[3]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/util'\n ## make[2]: *** [Makefile:989: ../../cxx/util/mtest.o] Error 2\n-## make[2]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/io'\n+## make[2]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/io'\n not ok 714 - ./cxx/io/writeallnosx \n ---\n Directory: ./cxx/io\n File: writeallnosx\n Num-procs: \n- Date: \"Sun Jan 5 01:33:51 2020\"\n+ Date: \"Mon Dec 3 17:05:50 2018\"\n ...\n ## Test output (expected 'No Errors'):\n-## Failed to build writeallnosx; make[2]: Entering directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/io'\n+## Failed to build writeallnosx; make[2]: Entering directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/io'\n ## (cd ../../cxx/util && make mtest.o)\n-## make[3]: Entering directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/util'\n+## make[3]: Entering directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/util'\n ## depbase=`echo mtest.o | sed 's|[^/]*$|.deps/&|;s|\\.o$||'`;\\\n-## /usr/bin/mpicxx.openmpi -DHAVE_CONFIG_H -I. -I../../../cxx/util -I../../include -I../../include -I../../../include -Wdate-time -D_FORTIFY_SOURCE=2 -g -O2 -ffile-prefix-map=/build/1st/mpi-testsuite-3.2+dfsg=. -fstack-protector-strong -Wformat -Werror=format-security -MT mtest.o -MD -MP -MF $depbase.Tpo -c -o mtest.o ../../../cxx/util/mtest.cxx &&\\\n+## /usr/bin/mpicxx.openmpi -DHAVE_CONFIG_H -I. -I../../../cxx/util -I../../include -I../../include -I../../../include -Wdate-time -D_FORTIFY_SOURCE=2 -g -O2 -ffile-prefix-map=/build/mpi-testsuite-3.2+dfsg/2nd=. -fstack-protector-strong -Wformat -Werror=format-security -MT mtest.o -MD -MP -MF $depbase.Tpo -c -o mtest.o ../../../cxx/util/mtest.cxx &&\\\n ## mv -f $depbase.Tpo $depbase.Po\n ## ../../../cxx/util/mtest.cxx: In function 'void* MTestTypeContigInitRecv(MTestDatatype*)':\n ## ../../../cxx/util/mtest.cxx:189:42: warning: 'int MPI_Type_extent(MPI_Datatype, MPI_Aint*)' is deprecated: MPI_Type_extent is superseded by MPI_Type_get_extent in MPI-2.0 [-Wdeprecated-declarations]\n ## MPI_Type_extent( mtype->datatype, &size );\n ## ^\n ## In file included from ../../../cxx/util/mtest.cxx:7:\n ## /usr/lib/x86_64-linux-gnu/openmpi/include/mpi.h:1758:20: note: declared here\n@@ -1002572,30 +1002655,30 @@\n ## ../../../cxx/util/mtest.cxx:848:50: error: 'NULL_COPY_FN' is not a member of 'MPI::Win'\n ## mem_keyval = MPI::Win::Create_keyval( MPI::Win::NULL_COPY_FN,\n ## ^~~~~~~~~~~~\n ## ../../../cxx/util/mtest.cxx:849:22: error: 'NULL_DELETE_FN' is not a member of 'MPI::Win'\n ## MPI::Win::NULL_DELETE_FN, 0 );\n ## ^~~~~~~~~~~~~~\n ## make[3]: *** [Makefile:416: mtest.o] Error 1\n-## make[3]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/util'\n+## make[3]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/util'\n ## make[2]: *** [Makefile:989: ../../cxx/util/mtest.o] Error 2\n-## make[2]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/io'\n+## make[2]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/io'\n not ok 715 - ./cxx/io/writeshnosx \n ---\n Directory: ./cxx/io\n File: writeshnosx\n Num-procs: \n- Date: \"Sun Jan 5 01:33:53 2020\"\n+ Date: \"Mon Dec 3 17:05:54 2018\"\n ...\n ## Test output (expected 'No Errors'):\n-## Failed to build writeshnosx; make[2]: Entering directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/io'\n+## Failed to build writeshnosx; make[2]: Entering directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/io'\n ## (cd ../../cxx/util && make mtest.o)\n-## make[3]: Entering directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/util'\n+## make[3]: Entering directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/util'\n ## depbase=`echo mtest.o | sed 's|[^/]*$|.deps/&|;s|\\.o$||'`;\\\n-## /usr/bin/mpicxx.openmpi -DHAVE_CONFIG_H -I. -I../../../cxx/util -I../../include -I../../include -I../../../include -Wdate-time -D_FORTIFY_SOURCE=2 -g -O2 -ffile-prefix-map=/build/1st/mpi-testsuite-3.2+dfsg=. -fstack-protector-strong -Wformat -Werror=format-security -MT mtest.o -MD -MP -MF $depbase.Tpo -c -o mtest.o ../../../cxx/util/mtest.cxx &&\\\n+## /usr/bin/mpicxx.openmpi -DHAVE_CONFIG_H -I. -I../../../cxx/util -I../../include -I../../include -I../../../include -Wdate-time -D_FORTIFY_SOURCE=2 -g -O2 -ffile-prefix-map=/build/mpi-testsuite-3.2+dfsg/2nd=. -fstack-protector-strong -Wformat -Werror=format-security -MT mtest.o -MD -MP -MF $depbase.Tpo -c -o mtest.o ../../../cxx/util/mtest.cxx &&\\\n ## mv -f $depbase.Tpo $depbase.Po\n ## ../../../cxx/util/mtest.cxx: In function 'void* MTestTypeContigInitRecv(MTestDatatype*)':\n ## ../../../cxx/util/mtest.cxx:189:42: warning: 'int MPI_Type_extent(MPI_Datatype, MPI_Aint*)' is deprecated: MPI_Type_extent is superseded by MPI_Type_get_extent in MPI-2.0 [-Wdeprecated-declarations]\n ## MPI_Type_extent( mtype->datatype, &size );\n ## ^\n ## In file included from ../../../cxx/util/mtest.cxx:7:\n ## /usr/lib/x86_64-linux-gnu/openmpi/include/mpi.h:1758:20: note: declared here\n@@ -1002627,30 +1002710,30 @@\n ## ../../../cxx/util/mtest.cxx:848:50: error: 'NULL_COPY_FN' is not a member of 'MPI::Win'\n ## mem_keyval = MPI::Win::Create_keyval( MPI::Win::NULL_COPY_FN,\n ## ^~~~~~~~~~~~\n ## ../../../cxx/util/mtest.cxx:849:22: error: 'NULL_DELETE_FN' is not a member of 'MPI::Win'\n ## MPI::Win::NULL_DELETE_FN, 0 );\n ## ^~~~~~~~~~~~~~\n ## make[3]: *** [Makefile:416: mtest.o] Error 1\n-## make[3]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/util'\n+## make[3]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/util'\n ## make[2]: *** [Makefile:989: ../../cxx/util/mtest.o] Error 2\n-## make[2]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/io'\n+## make[2]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/io'\n not ok 716 - ./cxx/io/writeordnosx \n ---\n Directory: ./cxx/io\n File: writeordnosx\n Num-procs: \n- Date: \"Sun Jan 5 01:33:54 2020\"\n+ Date: \"Mon Dec 3 17:05:58 2018\"\n ...\n ## Test output (expected 'No Errors'):\n-## Failed to build writeordnosx; make[2]: Entering directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/io'\n+## Failed to build writeordnosx; make[2]: Entering directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/io'\n ## (cd ../../cxx/util && make mtest.o)\n-## make[3]: Entering directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/util'\n+## make[3]: Entering directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/util'\n ## depbase=`echo mtest.o | sed 's|[^/]*$|.deps/&|;s|\\.o$||'`;\\\n-## /usr/bin/mpicxx.openmpi -DHAVE_CONFIG_H -I. -I../../../cxx/util -I../../include -I../../include -I../../../include -Wdate-time -D_FORTIFY_SOURCE=2 -g -O2 -ffile-prefix-map=/build/1st/mpi-testsuite-3.2+dfsg=. -fstack-protector-strong -Wformat -Werror=format-security -MT mtest.o -MD -MP -MF $depbase.Tpo -c -o mtest.o ../../../cxx/util/mtest.cxx &&\\\n+## /usr/bin/mpicxx.openmpi -DHAVE_CONFIG_H -I. -I../../../cxx/util -I../../include -I../../include -I../../../include -Wdate-time -D_FORTIFY_SOURCE=2 -g -O2 -ffile-prefix-map=/build/mpi-testsuite-3.2+dfsg/2nd=. -fstack-protector-strong -Wformat -Werror=format-security -MT mtest.o -MD -MP -MF $depbase.Tpo -c -o mtest.o ../../../cxx/util/mtest.cxx &&\\\n ## mv -f $depbase.Tpo $depbase.Po\n ## ../../../cxx/util/mtest.cxx: In function 'void* MTestTypeContigInitRecv(MTestDatatype*)':\n ## ../../../cxx/util/mtest.cxx:189:42: warning: 'int MPI_Type_extent(MPI_Datatype, MPI_Aint*)' is deprecated: MPI_Type_extent is superseded by MPI_Type_get_extent in MPI-2.0 [-Wdeprecated-declarations]\n ## MPI_Type_extent( mtype->datatype, &size );\n ## ^\n ## In file included from ../../../cxx/util/mtest.cxx:7:\n ## /usr/lib/x86_64-linux-gnu/openmpi/include/mpi.h:1758:20: note: declared here\n@@ -1002682,30 +1002765,30 @@\n ## ../../../cxx/util/mtest.cxx:848:50: error: 'NULL_COPY_FN' is not a member of 'MPI::Win'\n ## mem_keyval = MPI::Win::Create_keyval( MPI::Win::NULL_COPY_FN,\n ## ^~~~~~~~~~~~\n ## ../../../cxx/util/mtest.cxx:849:22: error: 'NULL_DELETE_FN' is not a member of 'MPI::Win'\n ## MPI::Win::NULL_DELETE_FN, 0 );\n ## ^~~~~~~~~~~~~~\n ## make[3]: *** [Makefile:416: mtest.o] Error 1\n-## make[3]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/util'\n+## make[3]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/util'\n ## make[2]: *** [Makefile:989: ../../cxx/util/mtest.o] Error 2\n-## make[2]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/io'\n+## make[2]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/io'\n not ok 717 - ./cxx/io/writeatallnosx \n ---\n Directory: ./cxx/io\n File: writeatallnosx\n Num-procs: \n- Date: \"Sun Jan 5 01:33:55 2020\"\n+ Date: \"Mon Dec 3 17:06:03 2018\"\n ...\n ## Test output (expected 'No Errors'):\n-## Failed to build writeatallnosx; make[2]: Entering directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/io'\n+## Failed to build writeatallnosx; make[2]: Entering directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/io'\n ## (cd ../../cxx/util && make mtest.o)\n-## make[3]: Entering directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/util'\n+## make[3]: Entering directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/util'\n ## depbase=`echo mtest.o | sed 's|[^/]*$|.deps/&|;s|\\.o$||'`;\\\n-## /usr/bin/mpicxx.openmpi -DHAVE_CONFIG_H -I. -I../../../cxx/util -I../../include -I../../include -I../../../include -Wdate-time -D_FORTIFY_SOURCE=2 -g -O2 -ffile-prefix-map=/build/1st/mpi-testsuite-3.2+dfsg=. -fstack-protector-strong -Wformat -Werror=format-security -MT mtest.o -MD -MP -MF $depbase.Tpo -c -o mtest.o ../../../cxx/util/mtest.cxx &&\\\n+## /usr/bin/mpicxx.openmpi -DHAVE_CONFIG_H -I. -I../../../cxx/util -I../../include -I../../include -I../../../include -Wdate-time -D_FORTIFY_SOURCE=2 -g -O2 -ffile-prefix-map=/build/mpi-testsuite-3.2+dfsg/2nd=. -fstack-protector-strong -Wformat -Werror=format-security -MT mtest.o -MD -MP -MF $depbase.Tpo -c -o mtest.o ../../../cxx/util/mtest.cxx &&\\\n ## mv -f $depbase.Tpo $depbase.Po\n ## ../../../cxx/util/mtest.cxx: In function 'void* MTestTypeContigInitRecv(MTestDatatype*)':\n ## ../../../cxx/util/mtest.cxx:189:42: warning: 'int MPI_Type_extent(MPI_Datatype, MPI_Aint*)' is deprecated: MPI_Type_extent is superseded by MPI_Type_get_extent in MPI-2.0 [-Wdeprecated-declarations]\n ## MPI_Type_extent( mtype->datatype, &size );\n ## ^\n ## In file included from ../../../cxx/util/mtest.cxx:7:\n ## /usr/lib/x86_64-linux-gnu/openmpi/include/mpi.h:1758:20: note: declared here\n@@ -1002737,30 +1002820,30 @@\n ## ../../../cxx/util/mtest.cxx:848:50: error: 'NULL_COPY_FN' is not a member of 'MPI::Win'\n ## mem_keyval = MPI::Win::Create_keyval( MPI::Win::NULL_COPY_FN,\n ## ^~~~~~~~~~~~\n ## ../../../cxx/util/mtest.cxx:849:22: error: 'NULL_DELETE_FN' is not a member of 'MPI::Win'\n ## MPI::Win::NULL_DELETE_FN, 0 );\n ## ^~~~~~~~~~~~~~\n ## make[3]: *** [Makefile:416: mtest.o] Error 1\n-## make[3]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/util'\n+## make[3]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/util'\n ## make[2]: *** [Makefile:989: ../../cxx/util/mtest.o] Error 2\n-## make[2]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/io'\n+## make[2]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/io'\n not ok 718 - ./cxx/io/writeatallbenosx \n ---\n Directory: ./cxx/io\n File: writeatallbenosx\n Num-procs: \n- Date: \"Sun Jan 5 01:33:56 2020\"\n+ Date: \"Mon Dec 3 17:06:07 2018\"\n ...\n ## Test output (expected 'No Errors'):\n-## Failed to build writeatallbenosx; make[2]: Entering directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/io'\n+## Failed to build writeatallbenosx; make[2]: Entering directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/io'\n ## (cd ../../cxx/util && make mtest.o)\n-## make[3]: Entering directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/util'\n+## make[3]: Entering directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/util'\n ## depbase=`echo mtest.o | sed 's|[^/]*$|.deps/&|;s|\\.o$||'`;\\\n-## /usr/bin/mpicxx.openmpi -DHAVE_CONFIG_H -I. -I../../../cxx/util -I../../include -I../../include -I../../../include -Wdate-time -D_FORTIFY_SOURCE=2 -g -O2 -ffile-prefix-map=/build/1st/mpi-testsuite-3.2+dfsg=. -fstack-protector-strong -Wformat -Werror=format-security -MT mtest.o -MD -MP -MF $depbase.Tpo -c -o mtest.o ../../../cxx/util/mtest.cxx &&\\\n+## /usr/bin/mpicxx.openmpi -DHAVE_CONFIG_H -I. -I../../../cxx/util -I../../include -I../../include -I../../../include -Wdate-time -D_FORTIFY_SOURCE=2 -g -O2 -ffile-prefix-map=/build/mpi-testsuite-3.2+dfsg/2nd=. -fstack-protector-strong -Wformat -Werror=format-security -MT mtest.o -MD -MP -MF $depbase.Tpo -c -o mtest.o ../../../cxx/util/mtest.cxx &&\\\n ## mv -f $depbase.Tpo $depbase.Po\n ## ../../../cxx/util/mtest.cxx: In function 'void* MTestTypeContigInitRecv(MTestDatatype*)':\n ## ../../../cxx/util/mtest.cxx:189:42: warning: 'int MPI_Type_extent(MPI_Datatype, MPI_Aint*)' is deprecated: MPI_Type_extent is superseded by MPI_Type_get_extent in MPI-2.0 [-Wdeprecated-declarations]\n ## MPI_Type_extent( mtype->datatype, &size );\n ## ^\n ## In file included from ../../../cxx/util/mtest.cxx:7:\n ## /usr/lib/x86_64-linux-gnu/openmpi/include/mpi.h:1758:20: note: declared here\n@@ -1002792,30 +1002875,30 @@\n ## ../../../cxx/util/mtest.cxx:848:50: error: 'NULL_COPY_FN' is not a member of 'MPI::Win'\n ## mem_keyval = MPI::Win::Create_keyval( MPI::Win::NULL_COPY_FN,\n ## ^~~~~~~~~~~~\n ## ../../../cxx/util/mtest.cxx:849:22: error: 'NULL_DELETE_FN' is not a member of 'MPI::Win'\n ## MPI::Win::NULL_DELETE_FN, 0 );\n ## ^~~~~~~~~~~~~~\n ## make[3]: *** [Makefile:416: mtest.o] Error 1\n-## make[3]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/util'\n+## make[3]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/util'\n ## make[2]: *** [Makefile:989: ../../cxx/util/mtest.o] Error 2\n-## make[2]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/io'\n+## make[2]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/io'\n not ok 719 - ./cxx/io/writeallbenosx \n ---\n Directory: ./cxx/io\n File: writeallbenosx\n Num-procs: \n- Date: \"Sun Jan 5 01:33:57 2020\"\n+ Date: \"Mon Dec 3 17:06:11 2018\"\n ...\n ## Test output (expected 'No Errors'):\n-## Failed to build writeallbenosx; make[2]: Entering directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/io'\n+## Failed to build writeallbenosx; make[2]: Entering directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/io'\n ## (cd ../../cxx/util && make mtest.o)\n-## make[3]: Entering directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/util'\n+## make[3]: Entering directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/util'\n ## depbase=`echo mtest.o | sed 's|[^/]*$|.deps/&|;s|\\.o$||'`;\\\n-## /usr/bin/mpicxx.openmpi -DHAVE_CONFIG_H -I. -I../../../cxx/util -I../../include -I../../include -I../../../include -Wdate-time -D_FORTIFY_SOURCE=2 -g -O2 -ffile-prefix-map=/build/1st/mpi-testsuite-3.2+dfsg=. -fstack-protector-strong -Wformat -Werror=format-security -MT mtest.o -MD -MP -MF $depbase.Tpo -c -o mtest.o ../../../cxx/util/mtest.cxx &&\\\n+## /usr/bin/mpicxx.openmpi -DHAVE_CONFIG_H -I. -I../../../cxx/util -I../../include -I../../include -I../../../include -Wdate-time -D_FORTIFY_SOURCE=2 -g -O2 -ffile-prefix-map=/build/mpi-testsuite-3.2+dfsg/2nd=. -fstack-protector-strong -Wformat -Werror=format-security -MT mtest.o -MD -MP -MF $depbase.Tpo -c -o mtest.o ../../../cxx/util/mtest.cxx &&\\\n ## mv -f $depbase.Tpo $depbase.Po\n ## ../../../cxx/util/mtest.cxx: In function 'void* MTestTypeContigInitRecv(MTestDatatype*)':\n ## ../../../cxx/util/mtest.cxx:189:42: warning: 'int MPI_Type_extent(MPI_Datatype, MPI_Aint*)' is deprecated: MPI_Type_extent is superseded by MPI_Type_get_extent in MPI-2.0 [-Wdeprecated-declarations]\n ## MPI_Type_extent( mtype->datatype, &size );\n ## ^\n ## In file included from ../../../cxx/util/mtest.cxx:7:\n ## /usr/lib/x86_64-linux-gnu/openmpi/include/mpi.h:1758:20: note: declared here\n@@ -1002847,30 +1002930,30 @@\n ## ../../../cxx/util/mtest.cxx:848:50: error: 'NULL_COPY_FN' is not a member of 'MPI::Win'\n ## mem_keyval = MPI::Win::Create_keyval( MPI::Win::NULL_COPY_FN,\n ## ^~~~~~~~~~~~\n ## ../../../cxx/util/mtest.cxx:849:22: error: 'NULL_DELETE_FN' is not a member of 'MPI::Win'\n ## MPI::Win::NULL_DELETE_FN, 0 );\n ## ^~~~~~~~~~~~~~\n ## make[3]: *** [Makefile:416: mtest.o] Error 1\n-## make[3]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/util'\n+## make[3]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/util'\n ## make[2]: *** [Makefile:989: ../../cxx/util/mtest.o] Error 2\n-## make[2]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/io'\n+## make[2]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/io'\n not ok 720 - ./cxx/io/writeordbenosx \n ---\n Directory: ./cxx/io\n File: writeordbenosx\n Num-procs: \n- Date: \"Sun Jan 5 01:33:58 2020\"\n+ Date: \"Mon Dec 3 17:06:16 2018\"\n ...\n ## Test output (expected 'No Errors'):\n-## Failed to build writeordbenosx; make[2]: Entering directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/io'\n+## Failed to build writeordbenosx; make[2]: Entering directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/io'\n ## (cd ../../cxx/util && make mtest.o)\n-## make[3]: Entering directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/util'\n+## make[3]: Entering directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/util'\n ## depbase=`echo mtest.o | sed 's|[^/]*$|.deps/&|;s|\\.o$||'`;\\\n-## /usr/bin/mpicxx.openmpi -DHAVE_CONFIG_H -I. -I../../../cxx/util -I../../include -I../../include -I../../../include -Wdate-time -D_FORTIFY_SOURCE=2 -g -O2 -ffile-prefix-map=/build/1st/mpi-testsuite-3.2+dfsg=. -fstack-protector-strong -Wformat -Werror=format-security -MT mtest.o -MD -MP -MF $depbase.Tpo -c -o mtest.o ../../../cxx/util/mtest.cxx &&\\\n+## /usr/bin/mpicxx.openmpi -DHAVE_CONFIG_H -I. -I../../../cxx/util -I../../include -I../../include -I../../../include -Wdate-time -D_FORTIFY_SOURCE=2 -g -O2 -ffile-prefix-map=/build/mpi-testsuite-3.2+dfsg/2nd=. -fstack-protector-strong -Wformat -Werror=format-security -MT mtest.o -MD -MP -MF $depbase.Tpo -c -o mtest.o ../../../cxx/util/mtest.cxx &&\\\n ## mv -f $depbase.Tpo $depbase.Po\n ## ../../../cxx/util/mtest.cxx: In function 'void* MTestTypeContigInitRecv(MTestDatatype*)':\n ## ../../../cxx/util/mtest.cxx:189:42: warning: 'int MPI_Type_extent(MPI_Datatype, MPI_Aint*)' is deprecated: MPI_Type_extent is superseded by MPI_Type_get_extent in MPI-2.0 [-Wdeprecated-declarations]\n ## MPI_Type_extent( mtype->datatype, &size );\n ## ^\n ## In file included from ../../../cxx/util/mtest.cxx:7:\n ## /usr/lib/x86_64-linux-gnu/openmpi/include/mpi.h:1758:20: note: declared here\n@@ -1002902,30 +1002985,30 @@\n ## ../../../cxx/util/mtest.cxx:848:50: error: 'NULL_COPY_FN' is not a member of 'MPI::Win'\n ## mem_keyval = MPI::Win::Create_keyval( MPI::Win::NULL_COPY_FN,\n ## ^~~~~~~~~~~~\n ## ../../../cxx/util/mtest.cxx:849:22: error: 'NULL_DELETE_FN' is not a member of 'MPI::Win'\n ## MPI::Win::NULL_DELETE_FN, 0 );\n ## ^~~~~~~~~~~~~~\n ## make[3]: *** [Makefile:416: mtest.o] Error 1\n-## make[3]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/util'\n+## make[3]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/util'\n ## make[2]: *** [Makefile:989: ../../cxx/util/mtest.o] Error 2\n-## make[2]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/io'\n+## make[2]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/io'\n not ok 721 - ./cxx/io/fileerrx \n ---\n Directory: ./cxx/io\n File: fileerrx\n Num-procs: \n- Date: \"Sun Jan 5 01:33:59 2020\"\n+ Date: \"Mon Dec 3 17:06:21 2018\"\n ...\n ## Test output (expected 'No Errors'):\n-## Failed to build fileerrx; make[2]: Entering directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/io'\n+## Failed to build fileerrx; make[2]: Entering directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/io'\n ## (cd ../../cxx/util && make mtest.o)\n-## make[3]: Entering directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/util'\n+## make[3]: Entering directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/util'\n ## depbase=`echo mtest.o | sed 's|[^/]*$|.deps/&|;s|\\.o$||'`;\\\n-## /usr/bin/mpicxx.openmpi -DHAVE_CONFIG_H -I. -I../../../cxx/util -I../../include -I../../include -I../../../include -Wdate-time -D_FORTIFY_SOURCE=2 -g -O2 -ffile-prefix-map=/build/1st/mpi-testsuite-3.2+dfsg=. -fstack-protector-strong -Wformat -Werror=format-security -MT mtest.o -MD -MP -MF $depbase.Tpo -c -o mtest.o ../../../cxx/util/mtest.cxx &&\\\n+## /usr/bin/mpicxx.openmpi -DHAVE_CONFIG_H -I. -I../../../cxx/util -I../../include -I../../include -I../../../include -Wdate-time -D_FORTIFY_SOURCE=2 -g -O2 -ffile-prefix-map=/build/mpi-testsuite-3.2+dfsg/2nd=. -fstack-protector-strong -Wformat -Werror=format-security -MT mtest.o -MD -MP -MF $depbase.Tpo -c -o mtest.o ../../../cxx/util/mtest.cxx &&\\\n ## mv -f $depbase.Tpo $depbase.Po\n ## ../../../cxx/util/mtest.cxx: In function 'void* MTestTypeContigInitRecv(MTestDatatype*)':\n ## ../../../cxx/util/mtest.cxx:189:42: warning: 'int MPI_Type_extent(MPI_Datatype, MPI_Aint*)' is deprecated: MPI_Type_extent is superseded by MPI_Type_get_extent in MPI-2.0 [-Wdeprecated-declarations]\n ## MPI_Type_extent( mtype->datatype, &size );\n ## ^\n ## In file included from ../../../cxx/util/mtest.cxx:7:\n ## /usr/lib/x86_64-linux-gnu/openmpi/include/mpi.h:1758:20: note: declared here\n@@ -1002957,30 +1003040,30 @@\n ## ../../../cxx/util/mtest.cxx:848:50: error: 'NULL_COPY_FN' is not a member of 'MPI::Win'\n ## mem_keyval = MPI::Win::Create_keyval( MPI::Win::NULL_COPY_FN,\n ## ^~~~~~~~~~~~\n ## ../../../cxx/util/mtest.cxx:849:22: error: 'NULL_DELETE_FN' is not a member of 'MPI::Win'\n ## MPI::Win::NULL_DELETE_FN, 0 );\n ## ^~~~~~~~~~~~~~\n ## make[3]: *** [Makefile:416: mtest.o] Error 1\n-## make[3]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/util'\n+## make[3]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/util'\n ## make[2]: *** [Makefile:989: ../../cxx/util/mtest.o] Error 2\n-## make[2]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/io'\n+## make[2]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/io'\n not ok 722 - ./cxx/io/fileinfox \n ---\n Directory: ./cxx/io\n File: fileinfox\n Num-procs: \n- Date: \"Sun Jan 5 01:34:00 2020\"\n+ Date: \"Mon Dec 3 17:06:25 2018\"\n ...\n ## Test output (expected 'No Errors'):\n-## Failed to build fileinfox; make[2]: Entering directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/io'\n+## Failed to build fileinfox; make[2]: Entering directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/io'\n ## (cd ../../cxx/util && make mtest.o)\n-## make[3]: Entering directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/util'\n+## make[3]: Entering directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/util'\n ## depbase=`echo mtest.o | sed 's|[^/]*$|.deps/&|;s|\\.o$||'`;\\\n-## /usr/bin/mpicxx.openmpi -DHAVE_CONFIG_H -I. -I../../../cxx/util -I../../include -I../../include -I../../../include -Wdate-time -D_FORTIFY_SOURCE=2 -g -O2 -ffile-prefix-map=/build/1st/mpi-testsuite-3.2+dfsg=. -fstack-protector-strong -Wformat -Werror=format-security -MT mtest.o -MD -MP -MF $depbase.Tpo -c -o mtest.o ../../../cxx/util/mtest.cxx &&\\\n+## /usr/bin/mpicxx.openmpi -DHAVE_CONFIG_H -I. -I../../../cxx/util -I../../include -I../../include -I../../../include -Wdate-time -D_FORTIFY_SOURCE=2 -g -O2 -ffile-prefix-map=/build/mpi-testsuite-3.2+dfsg/2nd=. -fstack-protector-strong -Wformat -Werror=format-security -MT mtest.o -MD -MP -MF $depbase.Tpo -c -o mtest.o ../../../cxx/util/mtest.cxx &&\\\n ## mv -f $depbase.Tpo $depbase.Po\n ## ../../../cxx/util/mtest.cxx: In function 'void* MTestTypeContigInitRecv(MTestDatatype*)':\n ## ../../../cxx/util/mtest.cxx:189:42: warning: 'int MPI_Type_extent(MPI_Datatype, MPI_Aint*)' is deprecated: MPI_Type_extent is superseded by MPI_Type_get_extent in MPI-2.0 [-Wdeprecated-declarations]\n ## MPI_Type_extent( mtype->datatype, &size );\n ## ^\n ## In file included from ../../../cxx/util/mtest.cxx:7:\n ## /usr/lib/x86_64-linux-gnu/openmpi/include/mpi.h:1758:20: note: declared here\n@@ -1003012,30 +1003095,30 @@\n ## ../../../cxx/util/mtest.cxx:848:50: error: 'NULL_COPY_FN' is not a member of 'MPI::Win'\n ## mem_keyval = MPI::Win::Create_keyval( MPI::Win::NULL_COPY_FN,\n ## ^~~~~~~~~~~~\n ## ../../../cxx/util/mtest.cxx:849:22: error: 'NULL_DELETE_FN' is not a member of 'MPI::Win'\n ## MPI::Win::NULL_DELETE_FN, 0 );\n ## ^~~~~~~~~~~~~~\n ## make[3]: *** [Makefile:416: mtest.o] Error 1\n-## make[3]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/util'\n+## make[3]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/util'\n ## make[2]: *** [Makefile:989: ../../cxx/util/mtest.o] Error 2\n-## make[2]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/io'\n+## make[2]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/io'\n not ok 723 - ./cxx/io/filemiscx \n ---\n Directory: ./cxx/io\n File: filemiscx\n Num-procs: \n- Date: \"Sun Jan 5 01:34:02 2020\"\n+ Date: \"Mon Dec 3 17:06:30 2018\"\n ...\n ## Test output (expected 'No Errors'):\n-## Failed to build filemiscx; make[2]: Entering directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/io'\n+## Failed to build filemiscx; make[2]: Entering directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/io'\n ## (cd ../../cxx/util && make mtest.o)\n-## make[3]: Entering directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/util'\n+## make[3]: Entering directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/util'\n ## depbase=`echo mtest.o | sed 's|[^/]*$|.deps/&|;s|\\.o$||'`;\\\n-## /usr/bin/mpicxx.openmpi -DHAVE_CONFIG_H -I. -I../../../cxx/util -I../../include -I../../include -I../../../include -Wdate-time -D_FORTIFY_SOURCE=2 -g -O2 -ffile-prefix-map=/build/1st/mpi-testsuite-3.2+dfsg=. -fstack-protector-strong -Wformat -Werror=format-security -MT mtest.o -MD -MP -MF $depbase.Tpo -c -o mtest.o ../../../cxx/util/mtest.cxx &&\\\n+## /usr/bin/mpicxx.openmpi -DHAVE_CONFIG_H -I. -I../../../cxx/util -I../../include -I../../include -I../../../include -Wdate-time -D_FORTIFY_SOURCE=2 -g -O2 -ffile-prefix-map=/build/mpi-testsuite-3.2+dfsg/2nd=. -fstack-protector-strong -Wformat -Werror=format-security -MT mtest.o -MD -MP -MF $depbase.Tpo -c -o mtest.o ../../../cxx/util/mtest.cxx &&\\\n ## mv -f $depbase.Tpo $depbase.Po\n ## ../../../cxx/util/mtest.cxx: In function 'void* MTestTypeContigInitRecv(MTestDatatype*)':\n ## ../../../cxx/util/mtest.cxx:189:42: warning: 'int MPI_Type_extent(MPI_Datatype, MPI_Aint*)' is deprecated: MPI_Type_extent is superseded by MPI_Type_get_extent in MPI-2.0 [-Wdeprecated-declarations]\n ## MPI_Type_extent( mtype->datatype, &size );\n ## ^\n ## In file included from ../../../cxx/util/mtest.cxx:7:\n ## /usr/lib/x86_64-linux-gnu/openmpi/include/mpi.h:1758:20: note: declared here\n@@ -1003067,30 +1003150,30 @@\n ## ../../../cxx/util/mtest.cxx:848:50: error: 'NULL_COPY_FN' is not a member of 'MPI::Win'\n ## mem_keyval = MPI::Win::Create_keyval( MPI::Win::NULL_COPY_FN,\n ## ^~~~~~~~~~~~\n ## ../../../cxx/util/mtest.cxx:849:22: error: 'NULL_DELETE_FN' is not a member of 'MPI::Win'\n ## MPI::Win::NULL_DELETE_FN, 0 );\n ## ^~~~~~~~~~~~~~\n ## make[3]: *** [Makefile:416: mtest.o] Error 1\n-## make[3]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/util'\n+## make[3]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/util'\n ## make[2]: *** [Makefile:989: ../../cxx/util/mtest.o] Error 2\n-## make[2]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/io'\n+## make[2]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/io'\n not ok 724 - ./cxx/io/shpositionx \n ---\n Directory: ./cxx/io\n File: shpositionx\n Num-procs: \n- Date: \"Sun Jan 5 01:34:03 2020\"\n+ Date: \"Mon Dec 3 17:06:34 2018\"\n ...\n ## Test output (expected 'No Errors'):\n-## Failed to build shpositionx; make[2]: Entering directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/io'\n+## Failed to build shpositionx; make[2]: Entering directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/io'\n ## (cd ../../cxx/util && make mtest.o)\n-## make[3]: Entering directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/util'\n+## make[3]: Entering directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/util'\n ## depbase=`echo mtest.o | sed 's|[^/]*$|.deps/&|;s|\\.o$||'`;\\\n-## /usr/bin/mpicxx.openmpi -DHAVE_CONFIG_H -I. -I../../../cxx/util -I../../include -I../../include -I../../../include -Wdate-time -D_FORTIFY_SOURCE=2 -g -O2 -ffile-prefix-map=/build/1st/mpi-testsuite-3.2+dfsg=. -fstack-protector-strong -Wformat -Werror=format-security -MT mtest.o -MD -MP -MF $depbase.Tpo -c -o mtest.o ../../../cxx/util/mtest.cxx &&\\\n+## /usr/bin/mpicxx.openmpi -DHAVE_CONFIG_H -I. -I../../../cxx/util -I../../include -I../../include -I../../../include -Wdate-time -D_FORTIFY_SOURCE=2 -g -O2 -ffile-prefix-map=/build/mpi-testsuite-3.2+dfsg/2nd=. -fstack-protector-strong -Wformat -Werror=format-security -MT mtest.o -MD -MP -MF $depbase.Tpo -c -o mtest.o ../../../cxx/util/mtest.cxx &&\\\n ## mv -f $depbase.Tpo $depbase.Po\n ## ../../../cxx/util/mtest.cxx: In function 'void* MTestTypeContigInitRecv(MTestDatatype*)':\n ## ../../../cxx/util/mtest.cxx:189:42: warning: 'int MPI_Type_extent(MPI_Datatype, MPI_Aint*)' is deprecated: MPI_Type_extent is superseded by MPI_Type_get_extent in MPI-2.0 [-Wdeprecated-declarations]\n ## MPI_Type_extent( mtype->datatype, &size );\n ## ^\n ## In file included from ../../../cxx/util/mtest.cxx:7:\n ## /usr/lib/x86_64-linux-gnu/openmpi/include/mpi.h:1758:20: note: declared here\n@@ -1003122,30 +1003205,30 @@\n ## ../../../cxx/util/mtest.cxx:848:50: error: 'NULL_COPY_FN' is not a member of 'MPI::Win'\n ## mem_keyval = MPI::Win::Create_keyval( MPI::Win::NULL_COPY_FN,\n ## ^~~~~~~~~~~~\n ## ../../../cxx/util/mtest.cxx:849:22: error: 'NULL_DELETE_FN' is not a member of 'MPI::Win'\n ## MPI::Win::NULL_DELETE_FN, 0 );\n ## ^~~~~~~~~~~~~~\n ## make[3]: *** [Makefile:416: mtest.o] Error 1\n-## make[3]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/util'\n+## make[3]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/util'\n ## make[2]: *** [Makefile:989: ../../cxx/util/mtest.o] Error 2\n-## make[2]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/io'\n+## make[2]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/io'\n not ok 725 - ./cxx/io/seekavail \n ---\n Directory: ./cxx/io\n File: seekavail\n Num-procs: \n- Date: \"Sun Jan 5 01:34:04 2020\"\n+ Date: \"Mon Dec 3 17:06:37 2018\"\n ...\n ## Test output (expected 'No Errors'):\n-## Failed to build seekavail; make[2]: Entering directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/io'\n+## Failed to build seekavail; make[2]: Entering directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/io'\n ## (cd ../../cxx/util && make mtest.o)\n-## make[3]: Entering directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/util'\n+## make[3]: Entering directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/util'\n ## depbase=`echo mtest.o | sed 's|[^/]*$|.deps/&|;s|\\.o$||'`;\\\n-## /usr/bin/mpicxx.openmpi -DHAVE_CONFIG_H -I. -I../../../cxx/util -I../../include -I../../include -I../../../include -Wdate-time -D_FORTIFY_SOURCE=2 -g -O2 -ffile-prefix-map=/build/1st/mpi-testsuite-3.2+dfsg=. -fstack-protector-strong -Wformat -Werror=format-security -MT mtest.o -MD -MP -MF $depbase.Tpo -c -o mtest.o ../../../cxx/util/mtest.cxx &&\\\n+## /usr/bin/mpicxx.openmpi -DHAVE_CONFIG_H -I. -I../../../cxx/util -I../../include -I../../include -I../../../include -Wdate-time -D_FORTIFY_SOURCE=2 -g -O2 -ffile-prefix-map=/build/mpi-testsuite-3.2+dfsg/2nd=. -fstack-protector-strong -Wformat -Werror=format-security -MT mtest.o -MD -MP -MF $depbase.Tpo -c -o mtest.o ../../../cxx/util/mtest.cxx &&\\\n ## mv -f $depbase.Tpo $depbase.Po\n ## ../../../cxx/util/mtest.cxx: In function 'void* MTestTypeContigInitRecv(MTestDatatype*)':\n ## ../../../cxx/util/mtest.cxx:189:42: warning: 'int MPI_Type_extent(MPI_Datatype, MPI_Aint*)' is deprecated: MPI_Type_extent is superseded by MPI_Type_get_extent in MPI-2.0 [-Wdeprecated-declarations]\n ## MPI_Type_extent( mtype->datatype, &size );\n ## ^\n ## In file included from ../../../cxx/util/mtest.cxx:7:\n ## /usr/lib/x86_64-linux-gnu/openmpi/include/mpi.h:1758:20: note: declared here\n@@ -1003177,30 +1003260,30 @@\n ## ../../../cxx/util/mtest.cxx:848:50: error: 'NULL_COPY_FN' is not a member of 'MPI::Win'\n ## mem_keyval = MPI::Win::Create_keyval( MPI::Win::NULL_COPY_FN,\n ## ^~~~~~~~~~~~\n ## ../../../cxx/util/mtest.cxx:849:22: error: 'NULL_DELETE_FN' is not a member of 'MPI::Win'\n ## MPI::Win::NULL_DELETE_FN, 0 );\n ## ^~~~~~~~~~~~~~\n ## make[3]: *** [Makefile:416: mtest.o] Error 1\n-## make[3]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/util'\n+## make[3]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/util'\n ## make[2]: *** [Makefile:989: ../../cxx/util/mtest.o] Error 2\n-## make[2]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/io'\n+## make[2]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/io'\n not ok 726 - ./cxx/spawn/namepubx \n ---\n Directory: ./cxx/spawn\n File: namepubx\n Num-procs: \n- Date: \"Sun Jan 5 01:34:05 2020\"\n+ Date: \"Mon Dec 3 17:06:41 2018\"\n ...\n ## Test output (expected 'No Errors'):\n-## Failed to build namepubx; make[2]: Entering directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/spawn'\n+## Failed to build namepubx; make[2]: Entering directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/spawn'\n ## (cd ../../cxx/util && make mtest.o)\n-## make[3]: Entering directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/util'\n+## make[3]: Entering directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/util'\n ## depbase=`echo mtest.o | sed 's|[^/]*$|.deps/&|;s|\\.o$||'`;\\\n-## /usr/bin/mpicxx.openmpi -DHAVE_CONFIG_H -I. -I../../../cxx/util -I../../include -I../../include -I../../../include -Wdate-time -D_FORTIFY_SOURCE=2 -g -O2 -ffile-prefix-map=/build/1st/mpi-testsuite-3.2+dfsg=. -fstack-protector-strong -Wformat -Werror=format-security -MT mtest.o -MD -MP -MF $depbase.Tpo -c -o mtest.o ../../../cxx/util/mtest.cxx &&\\\n+## /usr/bin/mpicxx.openmpi -DHAVE_CONFIG_H -I. -I../../../cxx/util -I../../include -I../../include -I../../../include -Wdate-time -D_FORTIFY_SOURCE=2 -g -O2 -ffile-prefix-map=/build/mpi-testsuite-3.2+dfsg/2nd=. -fstack-protector-strong -Wformat -Werror=format-security -MT mtest.o -MD -MP -MF $depbase.Tpo -c -o mtest.o ../../../cxx/util/mtest.cxx &&\\\n ## mv -f $depbase.Tpo $depbase.Po\n ## ../../../cxx/util/mtest.cxx: In function 'void* MTestTypeContigInitRecv(MTestDatatype*)':\n ## ../../../cxx/util/mtest.cxx:189:42: warning: 'int MPI_Type_extent(MPI_Datatype, MPI_Aint*)' is deprecated: MPI_Type_extent is superseded by MPI_Type_get_extent in MPI-2.0 [-Wdeprecated-declarations]\n ## MPI_Type_extent( mtype->datatype, &size );\n ## ^\n ## In file included from ../../../cxx/util/mtest.cxx:7:\n ## /usr/lib/x86_64-linux-gnu/openmpi/include/mpi.h:1758:20: note: declared here\n@@ -1003232,30 +1003315,30 @@\n ## ../../../cxx/util/mtest.cxx:848:50: error: 'NULL_COPY_FN' is not a member of 'MPI::Win'\n ## mem_keyval = MPI::Win::Create_keyval( MPI::Win::NULL_COPY_FN,\n ## ^~~~~~~~~~~~\n ## ../../../cxx/util/mtest.cxx:849:22: error: 'NULL_DELETE_FN' is not a member of 'MPI::Win'\n ## MPI::Win::NULL_DELETE_FN, 0 );\n ## ^~~~~~~~~~~~~~\n ## make[3]: *** [Makefile:416: mtest.o] Error 1\n-## make[3]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/util'\n+## make[3]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/util'\n ## make[2]: *** [Makefile:673: ../../cxx/util/mtest.o] Error 2\n-## make[2]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/spawn'\n+## make[2]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/spawn'\n not ok 727 - ./cxx/spawn/spawnintrax \n ---\n Directory: ./cxx/spawn\n File: spawnintrax\n Num-procs: \n- Date: \"Sun Jan 5 01:34:06 2020\"\n+ Date: \"Mon Dec 3 17:06:46 2018\"\n ...\n ## Test output (expected 'No Errors'):\n-## Failed to build spawnintrax; make[2]: Entering directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/spawn'\n+## Failed to build spawnintrax; make[2]: Entering directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/spawn'\n ## (cd ../../cxx/util && make mtest.o)\n-## make[3]: Entering directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/util'\n+## make[3]: Entering directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/util'\n ## depbase=`echo mtest.o | sed 's|[^/]*$|.deps/&|;s|\\.o$||'`;\\\n-## /usr/bin/mpicxx.openmpi -DHAVE_CONFIG_H -I. -I../../../cxx/util -I../../include -I../../include -I../../../include -Wdate-time -D_FORTIFY_SOURCE=2 -g -O2 -ffile-prefix-map=/build/1st/mpi-testsuite-3.2+dfsg=. -fstack-protector-strong -Wformat -Werror=format-security -MT mtest.o -MD -MP -MF $depbase.Tpo -c -o mtest.o ../../../cxx/util/mtest.cxx &&\\\n+## /usr/bin/mpicxx.openmpi -DHAVE_CONFIG_H -I. -I../../../cxx/util -I../../include -I../../include -I../../../include -Wdate-time -D_FORTIFY_SOURCE=2 -g -O2 -ffile-prefix-map=/build/mpi-testsuite-3.2+dfsg/2nd=. -fstack-protector-strong -Wformat -Werror=format-security -MT mtest.o -MD -MP -MF $depbase.Tpo -c -o mtest.o ../../../cxx/util/mtest.cxx &&\\\n ## mv -f $depbase.Tpo $depbase.Po\n ## ../../../cxx/util/mtest.cxx: In function 'void* MTestTypeContigInitRecv(MTestDatatype*)':\n ## ../../../cxx/util/mtest.cxx:189:42: warning: 'int MPI_Type_extent(MPI_Datatype, MPI_Aint*)' is deprecated: MPI_Type_extent is superseded by MPI_Type_get_extent in MPI-2.0 [-Wdeprecated-declarations]\n ## MPI_Type_extent( mtype->datatype, &size );\n ## ^\n ## In file included from ../../../cxx/util/mtest.cxx:7:\n ## /usr/lib/x86_64-linux-gnu/openmpi/include/mpi.h:1758:20: note: declared here\n@@ -1003287,33 +1003370,33 @@\n ## ../../../cxx/util/mtest.cxx:848:50: error: 'NULL_COPY_FN' is not a member of 'MPI::Win'\n ## mem_keyval = MPI::Win::Create_keyval( MPI::Win::NULL_COPY_FN,\n ## ^~~~~~~~~~~~\n ## ../../../cxx/util/mtest.cxx:849:22: error: 'NULL_DELETE_FN' is not a member of 'MPI::Win'\n ## MPI::Win::NULL_DELETE_FN, 0 );\n ## ^~~~~~~~~~~~~~\n ## make[3]: *** [Makefile:416: mtest.o] Error 1\n-## make[3]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/util'\n+## make[3]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/util'\n ## make[2]: *** [Makefile:673: ../../cxx/util/mtest.o] Error 2\n-## make[2]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/spawn'\n+## make[2]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/spawn'\n not ok 728 - ./cxx/spawn/spawnintrax \n ---\n Directory: ./cxx/spawn\n File: spawnintrax\n Num-procs: \n- Date: \"Sun Jan 5 01:34:09 2020\"\n+ Date: \"Mon Dec 3 17:06:59 2018\"\n ...\n ## Test output (expected 'No Errors'):\n-## Failed to build spawnintrax; make[2]: Entering directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/spawn'\n+## Failed to build spawnintrax; make[2]: Entering directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/spawn'\n ## depbase=`echo spawnintrax.o | sed 's|[^/]*$|.deps/&|;s|\\.o$||'`;\\\n-## /usr/bin/mpicxx.openmpi -DHAVE_CONFIG_H -I. -I../../../cxx/spawn -I../../include -I../../include -I../../../include -Wdate-time -D_FORTIFY_SOURCE=2 -g -O2 -ffile-prefix-map=/build/1st/mpi-testsuite-3.2+dfsg=. -fstack-protector-strong -Wformat -Werror=format-security -MT spawnintrax.o -MD -MP -MF $depbase.Tpo -c -o spawnintrax.o ../../../cxx/spawn/spawnintrax.cxx &&\\\n+## /usr/bin/mpicxx.openmpi -DHAVE_CONFIG_H -I. -I../../../cxx/spawn -I../../include -I../../include -I../../../include -Wdate-time -D_FORTIFY_SOURCE=2 -g -O2 -ffile-prefix-map=/build/mpi-testsuite-3.2+dfsg/2nd=. -fstack-protector-strong -Wformat -Werror=format-security -MT spawnintrax.o -MD -MP -MF $depbase.Tpo -c -o spawnintrax.o ../../../cxx/spawn/spawnintrax.cxx &&\\\n ## mv -f $depbase.Tpo $depbase.Po\n ## (cd ../../cxx/util && make mtest.o)\n-## make[3]: Entering directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/util'\n+## make[3]: Entering directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/util'\n ## depbase=`echo mtest.o | sed 's|[^/]*$|.deps/&|;s|\\.o$||'`;\\\n-## /usr/bin/mpicxx.openmpi -DHAVE_CONFIG_H -I. -I../../../cxx/util -I../../include -I../../include -I../../../include -Wdate-time -D_FORTIFY_SOURCE=2 -g -O2 -ffile-prefix-map=/build/1st/mpi-testsuite-3.2+dfsg=. -fstack-protector-strong -Wformat -Werror=format-security -MT mtest.o -MD -MP -MF $depbase.Tpo -c -o mtest.o ../../../cxx/util/mtest.cxx &&\\\n+## /usr/bin/mpicxx.openmpi -DHAVE_CONFIG_H -I. -I../../../cxx/util -I../../include -I../../include -I../../../include -Wdate-time -D_FORTIFY_SOURCE=2 -g -O2 -ffile-prefix-map=/build/mpi-testsuite-3.2+dfsg/2nd=. -fstack-protector-strong -Wformat -Werror=format-security -MT mtest.o -MD -MP -MF $depbase.Tpo -c -o mtest.o ../../../cxx/util/mtest.cxx &&\\\n ## mv -f $depbase.Tpo $depbase.Po\n ## ../../../cxx/util/mtest.cxx: In function 'void* MTestTypeContigInitRecv(MTestDatatype*)':\n ## ../../../cxx/util/mtest.cxx:189:42: warning: 'int MPI_Type_extent(MPI_Datatype, MPI_Aint*)' is deprecated: MPI_Type_extent is superseded by MPI_Type_get_extent in MPI-2.0 [-Wdeprecated-declarations]\n ## MPI_Type_extent( mtype->datatype, &size );\n ## ^\n ## In file included from ../../../cxx/util/mtest.cxx:7:\n ## /usr/lib/x86_64-linux-gnu/openmpi/include/mpi.h:1758:20: note: declared here\n@@ -1003345,30 +1003428,30 @@\n ## ../../../cxx/util/mtest.cxx:848:50: error: 'NULL_COPY_FN' is not a member of 'MPI::Win'\n ## mem_keyval = MPI::Win::Create_keyval( MPI::Win::NULL_COPY_FN,\n ## ^~~~~~~~~~~~\n ## ../../../cxx/util/mtest.cxx:849:22: error: 'NULL_DELETE_FN' is not a member of 'MPI::Win'\n ## MPI::Win::NULL_DELETE_FN, 0 );\n ## ^~~~~~~~~~~~~~\n ## make[3]: *** [Makefile:416: mtest.o] Error 1\n-## make[3]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/util'\n+## make[3]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/util'\n ## make[2]: *** [Makefile:673: ../../cxx/util/mtest.o] Error 2\n-## make[2]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/spawn'\n+## make[2]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/spawn'\n not ok 729 - ./cxx/spawn/spawnargvx \n ---\n Directory: ./cxx/spawn\n File: spawnargvx\n Num-procs: \n- Date: \"Sun Jan 5 01:34:10 2020\"\n+ Date: \"Mon Dec 3 17:07:03 2018\"\n ...\n ## Test output (expected 'No Errors'):\n-## Failed to build spawnargvx; make[2]: Entering directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/spawn'\n+## Failed to build spawnargvx; make[2]: Entering directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/spawn'\n ## (cd ../../cxx/util && make mtest.o)\n-## make[3]: Entering directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/util'\n+## make[3]: Entering directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/util'\n ## depbase=`echo mtest.o | sed 's|[^/]*$|.deps/&|;s|\\.o$||'`;\\\n-## /usr/bin/mpicxx.openmpi -DHAVE_CONFIG_H -I. -I../../../cxx/util -I../../include -I../../include -I../../../include -Wdate-time -D_FORTIFY_SOURCE=2 -g -O2 -ffile-prefix-map=/build/1st/mpi-testsuite-3.2+dfsg=. -fstack-protector-strong -Wformat -Werror=format-security -MT mtest.o -MD -MP -MF $depbase.Tpo -c -o mtest.o ../../../cxx/util/mtest.cxx &&\\\n+## /usr/bin/mpicxx.openmpi -DHAVE_CONFIG_H -I. -I../../../cxx/util -I../../include -I../../include -I../../../include -Wdate-time -D_FORTIFY_SOURCE=2 -g -O2 -ffile-prefix-map=/build/mpi-testsuite-3.2+dfsg/2nd=. -fstack-protector-strong -Wformat -Werror=format-security -MT mtest.o -MD -MP -MF $depbase.Tpo -c -o mtest.o ../../../cxx/util/mtest.cxx &&\\\n ## mv -f $depbase.Tpo $depbase.Po\n ## ../../../cxx/util/mtest.cxx: In function 'void* MTestTypeContigInitRecv(MTestDatatype*)':\n ## ../../../cxx/util/mtest.cxx:189:42: warning: 'int MPI_Type_extent(MPI_Datatype, MPI_Aint*)' is deprecated: MPI_Type_extent is superseded by MPI_Type_get_extent in MPI-2.0 [-Wdeprecated-declarations]\n ## MPI_Type_extent( mtype->datatype, &size );\n ## ^\n ## In file included from ../../../cxx/util/mtest.cxx:7:\n ## /usr/lib/x86_64-linux-gnu/openmpi/include/mpi.h:1758:20: note: declared here\n@@ -1003400,30 +1003483,30 @@\n ## ../../../cxx/util/mtest.cxx:848:50: error: 'NULL_COPY_FN' is not a member of 'MPI::Win'\n ## mem_keyval = MPI::Win::Create_keyval( MPI::Win::NULL_COPY_FN,\n ## ^~~~~~~~~~~~\n ## ../../../cxx/util/mtest.cxx:849:22: error: 'NULL_DELETE_FN' is not a member of 'MPI::Win'\n ## MPI::Win::NULL_DELETE_FN, 0 );\n ## ^~~~~~~~~~~~~~\n ## make[3]: *** [Makefile:416: mtest.o] Error 1\n-## make[3]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/util'\n+## make[3]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/util'\n ## make[2]: *** [Makefile:673: ../../cxx/util/mtest.o] Error 2\n-## make[2]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/spawn'\n+## make[2]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/spawn'\n not ok 730 - ./cxx/spawn/selfconaccx \n ---\n Directory: ./cxx/spawn\n File: selfconaccx\n Num-procs: \n- Date: \"Sun Jan 5 01:34:12 2020\"\n+ Date: \"Mon Dec 3 17:07:06 2018\"\n ...\n ## Test output (expected 'No Errors'):\n-## Failed to build selfconaccx; make[2]: Entering directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/spawn'\n+## Failed to build selfconaccx; make[2]: Entering directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/spawn'\n ## (cd ../../cxx/util && make mtest.o)\n-## make[3]: Entering directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/util'\n+## make[3]: Entering directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/util'\n ## depbase=`echo mtest.o | sed 's|[^/]*$|.deps/&|;s|\\.o$||'`;\\\n-## /usr/bin/mpicxx.openmpi -DHAVE_CONFIG_H -I. -I../../../cxx/util -I../../include -I../../include -I../../../include -Wdate-time -D_FORTIFY_SOURCE=2 -g -O2 -ffile-prefix-map=/build/1st/mpi-testsuite-3.2+dfsg=. -fstack-protector-strong -Wformat -Werror=format-security -MT mtest.o -MD -MP -MF $depbase.Tpo -c -o mtest.o ../../../cxx/util/mtest.cxx &&\\\n+## /usr/bin/mpicxx.openmpi -DHAVE_CONFIG_H -I. -I../../../cxx/util -I../../include -I../../include -I../../../include -Wdate-time -D_FORTIFY_SOURCE=2 -g -O2 -ffile-prefix-map=/build/mpi-testsuite-3.2+dfsg/2nd=. -fstack-protector-strong -Wformat -Werror=format-security -MT mtest.o -MD -MP -MF $depbase.Tpo -c -o mtest.o ../../../cxx/util/mtest.cxx &&\\\n ## mv -f $depbase.Tpo $depbase.Po\n ## ../../../cxx/util/mtest.cxx: In function 'void* MTestTypeContigInitRecv(MTestDatatype*)':\n ## ../../../cxx/util/mtest.cxx:189:42: warning: 'int MPI_Type_extent(MPI_Datatype, MPI_Aint*)' is deprecated: MPI_Type_extent is superseded by MPI_Type_get_extent in MPI-2.0 [-Wdeprecated-declarations]\n ## MPI_Type_extent( mtype->datatype, &size );\n ## ^\n ## In file included from ../../../cxx/util/mtest.cxx:7:\n ## /usr/lib/x86_64-linux-gnu/openmpi/include/mpi.h:1758:20: note: declared here\n@@ -1003455,30 +1003538,30 @@\n ## ../../../cxx/util/mtest.cxx:848:50: error: 'NULL_COPY_FN' is not a member of 'MPI::Win'\n ## mem_keyval = MPI::Win::Create_keyval( MPI::Win::NULL_COPY_FN,\n ## ^~~~~~~~~~~~\n ## ../../../cxx/util/mtest.cxx:849:22: error: 'NULL_DELETE_FN' is not a member of 'MPI::Win'\n ## MPI::Win::NULL_DELETE_FN, 0 );\n ## ^~~~~~~~~~~~~~\n ## make[3]: *** [Makefile:416: mtest.o] Error 1\n-## make[3]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/util'\n+## make[3]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/util'\n ## make[2]: *** [Makefile:673: ../../cxx/util/mtest.o] Error 2\n-## make[2]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/spawn'\n+## make[2]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/spawn'\n not ok 731 - ./cxx/rma/winnamex \n ---\n Directory: ./cxx/rma\n File: winnamex\n Num-procs: \n- Date: \"Sun Jan 5 01:34:13 2020\"\n+ Date: \"Mon Dec 3 17:07:10 2018\"\n ...\n ## Test output (expected 'No Errors'):\n-## Failed to build winnamex; make[2]: Entering directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/rma'\n+## Failed to build winnamex; make[2]: Entering directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/rma'\n ## (cd ../../cxx/util && make mtest.o)\n-## make[3]: Entering directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/util'\n+## make[3]: Entering directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/util'\n ## depbase=`echo mtest.o | sed 's|[^/]*$|.deps/&|;s|\\.o$||'`;\\\n-## /usr/bin/mpicxx.openmpi -DHAVE_CONFIG_H -I. -I../../../cxx/util -I../../include -I../../include -I../../../include -Wdate-time -D_FORTIFY_SOURCE=2 -g -O2 -ffile-prefix-map=/build/1st/mpi-testsuite-3.2+dfsg=. -fstack-protector-strong -Wformat -Werror=format-security -MT mtest.o -MD -MP -MF $depbase.Tpo -c -o mtest.o ../../../cxx/util/mtest.cxx &&\\\n+## /usr/bin/mpicxx.openmpi -DHAVE_CONFIG_H -I. -I../../../cxx/util -I../../include -I../../include -I../../../include -Wdate-time -D_FORTIFY_SOURCE=2 -g -O2 -ffile-prefix-map=/build/mpi-testsuite-3.2+dfsg/2nd=. -fstack-protector-strong -Wformat -Werror=format-security -MT mtest.o -MD -MP -MF $depbase.Tpo -c -o mtest.o ../../../cxx/util/mtest.cxx &&\\\n ## mv -f $depbase.Tpo $depbase.Po\n ## ../../../cxx/util/mtest.cxx: In function 'void* MTestTypeContigInitRecv(MTestDatatype*)':\n ## ../../../cxx/util/mtest.cxx:189:42: warning: 'int MPI_Type_extent(MPI_Datatype, MPI_Aint*)' is deprecated: MPI_Type_extent is superseded by MPI_Type_get_extent in MPI-2.0 [-Wdeprecated-declarations]\n ## MPI_Type_extent( mtype->datatype, &size );\n ## ^\n ## In file included from ../../../cxx/util/mtest.cxx:7:\n ## /usr/lib/x86_64-linux-gnu/openmpi/include/mpi.h:1758:20: note: declared here\n@@ -1003510,30 +1003593,30 @@\n ## ../../../cxx/util/mtest.cxx:848:50: error: 'NULL_COPY_FN' is not a member of 'MPI::Win'\n ## mem_keyval = MPI::Win::Create_keyval( MPI::Win::NULL_COPY_FN,\n ## ^~~~~~~~~~~~\n ## ../../../cxx/util/mtest.cxx:849:22: error: 'NULL_DELETE_FN' is not a member of 'MPI::Win'\n ## MPI::Win::NULL_DELETE_FN, 0 );\n ## ^~~~~~~~~~~~~~\n ## make[3]: *** [Makefile:416: mtest.o] Error 1\n-## make[3]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/util'\n+## make[3]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/util'\n ## make[2]: *** [Makefile:700: ../../cxx/util/mtest.o] Error 2\n-## make[2]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/rma'\n+## make[2]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/rma'\n not ok 732 - ./cxx/rma/wincallx \n ---\n Directory: ./cxx/rma\n File: wincallx\n Num-procs: \n- Date: \"Sun Jan 5 01:34:14 2020\"\n+ Date: \"Mon Dec 3 17:07:14 2018\"\n ...\n ## Test output (expected 'No Errors'):\n-## Failed to build wincallx; make[2]: Entering directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/rma'\n+## Failed to build wincallx; make[2]: Entering directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/rma'\n ## (cd ../../cxx/util && make mtest.o)\n-## make[3]: Entering directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/util'\n+## make[3]: Entering directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/util'\n ## depbase=`echo mtest.o | sed 's|[^/]*$|.deps/&|;s|\\.o$||'`;\\\n-## /usr/bin/mpicxx.openmpi -DHAVE_CONFIG_H -I. -I../../../cxx/util -I../../include -I../../include -I../../../include -Wdate-time -D_FORTIFY_SOURCE=2 -g -O2 -ffile-prefix-map=/build/1st/mpi-testsuite-3.2+dfsg=. -fstack-protector-strong -Wformat -Werror=format-security -MT mtest.o -MD -MP -MF $depbase.Tpo -c -o mtest.o ../../../cxx/util/mtest.cxx &&\\\n+## /usr/bin/mpicxx.openmpi -DHAVE_CONFIG_H -I. -I../../../cxx/util -I../../include -I../../include -I../../../include -Wdate-time -D_FORTIFY_SOURCE=2 -g -O2 -ffile-prefix-map=/build/mpi-testsuite-3.2+dfsg/2nd=. -fstack-protector-strong -Wformat -Werror=format-security -MT mtest.o -MD -MP -MF $depbase.Tpo -c -o mtest.o ../../../cxx/util/mtest.cxx &&\\\n ## mv -f $depbase.Tpo $depbase.Po\n ## ../../../cxx/util/mtest.cxx: In function 'void* MTestTypeContigInitRecv(MTestDatatype*)':\n ## ../../../cxx/util/mtest.cxx:189:42: warning: 'int MPI_Type_extent(MPI_Datatype, MPI_Aint*)' is deprecated: MPI_Type_extent is superseded by MPI_Type_get_extent in MPI-2.0 [-Wdeprecated-declarations]\n ## MPI_Type_extent( mtype->datatype, &size );\n ## ^\n ## In file included from ../../../cxx/util/mtest.cxx:7:\n ## /usr/lib/x86_64-linux-gnu/openmpi/include/mpi.h:1758:20: note: declared here\n@@ -1003565,30 +1003648,30 @@\n ## ../../../cxx/util/mtest.cxx:848:50: error: 'NULL_COPY_FN' is not a member of 'MPI::Win'\n ## mem_keyval = MPI::Win::Create_keyval( MPI::Win::NULL_COPY_FN,\n ## ^~~~~~~~~~~~\n ## ../../../cxx/util/mtest.cxx:849:22: error: 'NULL_DELETE_FN' is not a member of 'MPI::Win'\n ## MPI::Win::NULL_DELETE_FN, 0 );\n ## ^~~~~~~~~~~~~~\n ## make[3]: *** [Makefile:416: mtest.o] Error 1\n-## make[3]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/util'\n+## make[3]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/util'\n ## make[2]: *** [Makefile:700: ../../cxx/util/mtest.o] Error 2\n-## make[2]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/rma'\n+## make[2]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/rma'\n not ok 733 - ./cxx/rma/getgroupx \n ---\n Directory: ./cxx/rma\n File: getgroupx\n Num-procs: \n- Date: \"Sun Jan 5 01:34:15 2020\"\n+ Date: \"Mon Dec 3 17:07:18 2018\"\n ...\n ## Test output (expected 'No Errors'):\n-## Failed to build getgroupx; make[2]: Entering directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/rma'\n+## Failed to build getgroupx; make[2]: Entering directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/rma'\n ## (cd ../../cxx/util && make mtest.o)\n-## make[3]: Entering directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/util'\n+## make[3]: Entering directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/util'\n ## depbase=`echo mtest.o | sed 's|[^/]*$|.deps/&|;s|\\.o$||'`;\\\n-## /usr/bin/mpicxx.openmpi -DHAVE_CONFIG_H -I. -I../../../cxx/util -I../../include -I../../include -I../../../include -Wdate-time -D_FORTIFY_SOURCE=2 -g -O2 -ffile-prefix-map=/build/1st/mpi-testsuite-3.2+dfsg=. -fstack-protector-strong -Wformat -Werror=format-security -MT mtest.o -MD -MP -MF $depbase.Tpo -c -o mtest.o ../../../cxx/util/mtest.cxx &&\\\n+## /usr/bin/mpicxx.openmpi -DHAVE_CONFIG_H -I. -I../../../cxx/util -I../../include -I../../include -I../../../include -Wdate-time -D_FORTIFY_SOURCE=2 -g -O2 -ffile-prefix-map=/build/mpi-testsuite-3.2+dfsg/2nd=. -fstack-protector-strong -Wformat -Werror=format-security -MT mtest.o -MD -MP -MF $depbase.Tpo -c -o mtest.o ../../../cxx/util/mtest.cxx &&\\\n ## mv -f $depbase.Tpo $depbase.Po\n ## ../../../cxx/util/mtest.cxx: In function 'void* MTestTypeContigInitRecv(MTestDatatype*)':\n ## ../../../cxx/util/mtest.cxx:189:42: warning: 'int MPI_Type_extent(MPI_Datatype, MPI_Aint*)' is deprecated: MPI_Type_extent is superseded by MPI_Type_get_extent in MPI-2.0 [-Wdeprecated-declarations]\n ## MPI_Type_extent( mtype->datatype, &size );\n ## ^\n ## In file included from ../../../cxx/util/mtest.cxx:7:\n ## /usr/lib/x86_64-linux-gnu/openmpi/include/mpi.h:1758:20: note: declared here\n@@ -1003620,30 +1003703,30 @@\n ## ../../../cxx/util/mtest.cxx:848:50: error: 'NULL_COPY_FN' is not a member of 'MPI::Win'\n ## mem_keyval = MPI::Win::Create_keyval( MPI::Win::NULL_COPY_FN,\n ## ^~~~~~~~~~~~\n ## ../../../cxx/util/mtest.cxx:849:22: error: 'NULL_DELETE_FN' is not a member of 'MPI::Win'\n ## MPI::Win::NULL_DELETE_FN, 0 );\n ## ^~~~~~~~~~~~~~\n ## make[3]: *** [Makefile:416: mtest.o] Error 1\n-## make[3]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/util'\n+## make[3]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/util'\n ## make[2]: *** [Makefile:700: ../../cxx/util/mtest.o] Error 2\n-## make[2]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/rma'\n+## make[2]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/rma'\n not ok 734 - ./cxx/rma/winfencex \n ---\n Directory: ./cxx/rma\n File: winfencex\n Num-procs: \n- Date: \"Sun Jan 5 01:34:16 2020\"\n+ Date: \"Mon Dec 3 17:07:22 2018\"\n ...\n ## Test output (expected 'No Errors'):\n-## Failed to build winfencex; make[2]: Entering directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/rma'\n+## Failed to build winfencex; make[2]: Entering directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/rma'\n ## (cd ../../cxx/util && make mtest.o)\n-## make[3]: Entering directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/util'\n+## make[3]: Entering directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/util'\n ## depbase=`echo mtest.o | sed 's|[^/]*$|.deps/&|;s|\\.o$||'`;\\\n-## /usr/bin/mpicxx.openmpi -DHAVE_CONFIG_H -I. -I../../../cxx/util -I../../include -I../../include -I../../../include -Wdate-time -D_FORTIFY_SOURCE=2 -g -O2 -ffile-prefix-map=/build/1st/mpi-testsuite-3.2+dfsg=. -fstack-protector-strong -Wformat -Werror=format-security -MT mtest.o -MD -MP -MF $depbase.Tpo -c -o mtest.o ../../../cxx/util/mtest.cxx &&\\\n+## /usr/bin/mpicxx.openmpi -DHAVE_CONFIG_H -I. -I../../../cxx/util -I../../include -I../../include -I../../../include -Wdate-time -D_FORTIFY_SOURCE=2 -g -O2 -ffile-prefix-map=/build/mpi-testsuite-3.2+dfsg/2nd=. -fstack-protector-strong -Wformat -Werror=format-security -MT mtest.o -MD -MP -MF $depbase.Tpo -c -o mtest.o ../../../cxx/util/mtest.cxx &&\\\n ## mv -f $depbase.Tpo $depbase.Po\n ## ../../../cxx/util/mtest.cxx: In function 'void* MTestTypeContigInitRecv(MTestDatatype*)':\n ## ../../../cxx/util/mtest.cxx:189:42: warning: 'int MPI_Type_extent(MPI_Datatype, MPI_Aint*)' is deprecated: MPI_Type_extent is superseded by MPI_Type_get_extent in MPI-2.0 [-Wdeprecated-declarations]\n ## MPI_Type_extent( mtype->datatype, &size );\n ## ^\n ## In file included from ../../../cxx/util/mtest.cxx:7:\n ## /usr/lib/x86_64-linux-gnu/openmpi/include/mpi.h:1758:20: note: declared here\n@@ -1003675,30 +1003758,30 @@\n ## ../../../cxx/util/mtest.cxx:848:50: error: 'NULL_COPY_FN' is not a member of 'MPI::Win'\n ## mem_keyval = MPI::Win::Create_keyval( MPI::Win::NULL_COPY_FN,\n ## ^~~~~~~~~~~~\n ## ../../../cxx/util/mtest.cxx:849:22: error: 'NULL_DELETE_FN' is not a member of 'MPI::Win'\n ## MPI::Win::NULL_DELETE_FN, 0 );\n ## ^~~~~~~~~~~~~~\n ## make[3]: *** [Makefile:416: mtest.o] Error 1\n-## make[3]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/util'\n+## make[3]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/util'\n ## make[2]: *** [Makefile:700: ../../cxx/util/mtest.o] Error 2\n-## make[2]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/rma'\n+## make[2]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/rma'\n not ok 735 - ./cxx/rma/winscale1x \n ---\n Directory: ./cxx/rma\n File: winscale1x\n Num-procs: \n- Date: \"Sun Jan 5 01:34:18 2020\"\n+ Date: \"Mon Dec 3 17:07:26 2018\"\n ...\n ## Test output (expected 'No Errors'):\n-## Failed to build winscale1x; make[2]: Entering directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/rma'\n+## Failed to build winscale1x; make[2]: Entering directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/rma'\n ## (cd ../../cxx/util && make mtest.o)\n-## make[3]: Entering directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/util'\n+## make[3]: Entering directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/util'\n ## depbase=`echo mtest.o | sed 's|[^/]*$|.deps/&|;s|\\.o$||'`;\\\n-## /usr/bin/mpicxx.openmpi -DHAVE_CONFIG_H -I. -I../../../cxx/util -I../../include -I../../include -I../../../include -Wdate-time -D_FORTIFY_SOURCE=2 -g -O2 -ffile-prefix-map=/build/1st/mpi-testsuite-3.2+dfsg=. -fstack-protector-strong -Wformat -Werror=format-security -MT mtest.o -MD -MP -MF $depbase.Tpo -c -o mtest.o ../../../cxx/util/mtest.cxx &&\\\n+## /usr/bin/mpicxx.openmpi -DHAVE_CONFIG_H -I. -I../../../cxx/util -I../../include -I../../include -I../../../include -Wdate-time -D_FORTIFY_SOURCE=2 -g -O2 -ffile-prefix-map=/build/mpi-testsuite-3.2+dfsg/2nd=. -fstack-protector-strong -Wformat -Werror=format-security -MT mtest.o -MD -MP -MF $depbase.Tpo -c -o mtest.o ../../../cxx/util/mtest.cxx &&\\\n ## mv -f $depbase.Tpo $depbase.Po\n ## ../../../cxx/util/mtest.cxx: In function 'void* MTestTypeContigInitRecv(MTestDatatype*)':\n ## ../../../cxx/util/mtest.cxx:189:42: warning: 'int MPI_Type_extent(MPI_Datatype, MPI_Aint*)' is deprecated: MPI_Type_extent is superseded by MPI_Type_get_extent in MPI-2.0 [-Wdeprecated-declarations]\n ## MPI_Type_extent( mtype->datatype, &size );\n ## ^\n ## In file included from ../../../cxx/util/mtest.cxx:7:\n ## /usr/lib/x86_64-linux-gnu/openmpi/include/mpi.h:1758:20: note: declared here\n@@ -1003730,30 +1003813,30 @@\n ## ../../../cxx/util/mtest.cxx:848:50: error: 'NULL_COPY_FN' is not a member of 'MPI::Win'\n ## mem_keyval = MPI::Win::Create_keyval( MPI::Win::NULL_COPY_FN,\n ## ^~~~~~~~~~~~\n ## ../../../cxx/util/mtest.cxx:849:22: error: 'NULL_DELETE_FN' is not a member of 'MPI::Win'\n ## MPI::Win::NULL_DELETE_FN, 0 );\n ## ^~~~~~~~~~~~~~\n ## make[3]: *** [Makefile:416: mtest.o] Error 1\n-## make[3]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/util'\n+## make[3]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/util'\n ## make[2]: *** [Makefile:700: ../../cxx/util/mtest.o] Error 2\n-## make[2]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/rma'\n+## make[2]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/rma'\n not ok 736 - ./cxx/rma/winscale2x \n ---\n Directory: ./cxx/rma\n File: winscale2x\n Num-procs: \n- Date: \"Sun Jan 5 01:34:19 2020\"\n+ Date: \"Mon Dec 3 17:07:30 2018\"\n ...\n ## Test output (expected 'No Errors'):\n-## Failed to build winscale2x; make[2]: Entering directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/rma'\n+## Failed to build winscale2x; make[2]: Entering directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/rma'\n ## (cd ../../cxx/util && make mtest.o)\n-## make[3]: Entering directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/util'\n+## make[3]: Entering directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/util'\n ## depbase=`echo mtest.o | sed 's|[^/]*$|.deps/&|;s|\\.o$||'`;\\\n-## /usr/bin/mpicxx.openmpi -DHAVE_CONFIG_H -I. -I../../../cxx/util -I../../include -I../../include -I../../../include -Wdate-time -D_FORTIFY_SOURCE=2 -g -O2 -ffile-prefix-map=/build/1st/mpi-testsuite-3.2+dfsg=. -fstack-protector-strong -Wformat -Werror=format-security -MT mtest.o -MD -MP -MF $depbase.Tpo -c -o mtest.o ../../../cxx/util/mtest.cxx &&\\\n+## /usr/bin/mpicxx.openmpi -DHAVE_CONFIG_H -I. -I../../../cxx/util -I../../include -I../../include -I../../../include -Wdate-time -D_FORTIFY_SOURCE=2 -g -O2 -ffile-prefix-map=/build/mpi-testsuite-3.2+dfsg/2nd=. -fstack-protector-strong -Wformat -Werror=format-security -MT mtest.o -MD -MP -MF $depbase.Tpo -c -o mtest.o ../../../cxx/util/mtest.cxx &&\\\n ## mv -f $depbase.Tpo $depbase.Po\n ## ../../../cxx/util/mtest.cxx: In function 'void* MTestTypeContigInitRecv(MTestDatatype*)':\n ## ../../../cxx/util/mtest.cxx:189:42: warning: 'int MPI_Type_extent(MPI_Datatype, MPI_Aint*)' is deprecated: MPI_Type_extent is superseded by MPI_Type_get_extent in MPI-2.0 [-Wdeprecated-declarations]\n ## MPI_Type_extent( mtype->datatype, &size );\n ## ^\n ## In file included from ../../../cxx/util/mtest.cxx:7:\n ## /usr/lib/x86_64-linux-gnu/openmpi/include/mpi.h:1758:20: note: declared here\n@@ -1003785,38 +1003868,38 @@\n ## ../../../cxx/util/mtest.cxx:848:50: error: 'NULL_COPY_FN' is not a member of 'MPI::Win'\n ## mem_keyval = MPI::Win::Create_keyval( MPI::Win::NULL_COPY_FN,\n ## ^~~~~~~~~~~~\n ## ../../../cxx/util/mtest.cxx:849:22: error: 'NULL_DELETE_FN' is not a member of 'MPI::Win'\n ## MPI::Win::NULL_DELETE_FN, 0 );\n ## ^~~~~~~~~~~~~~\n ## make[3]: *** [Makefile:416: mtest.o] Error 1\n-## make[3]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/util'\n+## make[3]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/util'\n ## make[2]: *** [Makefile:700: ../../cxx/util/mtest.o] Error 2\n-## make[2]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/rma'\n+## make[2]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/rma'\n not ok 737 - ./cxx/rma/fkeyvalwinx \n ---\n Directory: ./cxx/rma\n File: fkeyvalwinx\n Num-procs: \n- Date: \"Sun Jan 5 01:34:20 2020\"\n+ Date: \"Mon Dec 3 17:07:33 2018\"\n ...\n ## Test output (expected 'No Errors'):\n-## Failed to build fkeyvalwinx; make[2]: Entering directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/rma'\n+## Failed to build fkeyvalwinx; make[2]: Entering directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/rma'\n ## depbase=`echo fkeyvalwinx.o | sed 's|[^/]*$|.deps/&|;s|\\.o$||'`;\\\n-## /usr/bin/mpicxx.openmpi -DHAVE_CONFIG_H -I. -I../../../cxx/rma -I../../include -I../../include -I../../../include -Wdate-time -D_FORTIFY_SOURCE=2 -g -O2 -ffile-prefix-map=/build/1st/mpi-testsuite-3.2+dfsg=. -fstack-protector-strong -Wformat -Werror=format-security -MT fkeyvalwinx.o -MD -MP -MF $depbase.Tpo -c -o fkeyvalwinx.o ../../../cxx/rma/fkeyvalwinx.cxx &&\\\n+## /usr/bin/mpicxx.openmpi -DHAVE_CONFIG_H -I. -I../../../cxx/rma -I../../include -I../../include -I../../../include -Wdate-time -D_FORTIFY_SOURCE=2 -g -O2 -ffile-prefix-map=/build/mpi-testsuite-3.2+dfsg/2nd=. -fstack-protector-strong -Wformat -Werror=format-security -MT fkeyvalwinx.o -MD -MP -MF $depbase.Tpo -c -o fkeyvalwinx.o ../../../cxx/rma/fkeyvalwinx.cxx &&\\\n ## mv -f $depbase.Tpo $depbase.Po\n ## ../../../cxx/rma/fkeyvalwinx.cxx: In function 'int main(int, char**)':\n ## ../../../cxx/rma/fkeyvalwinx.cxx:73:13: error: 'NULL_COPY_FN' is not a member of 'MPI::Win'\n ## MPI::Win::NULL_COPY_FN, MPI::Win::NULL_DELETE_FN, (void *)0 );\n ## ^~~~~~~~~~~~\n ## ../../../cxx/rma/fkeyvalwinx.cxx:73:37: error: 'NULL_DELETE_FN' is not a member of 'MPI::Win'\n ## MPI::Win::NULL_COPY_FN, MPI::Win::NULL_DELETE_FN, (void *)0 );\n ## ^~~~~~~~~~~~~~\n ## make[2]: *** [Makefile:480: fkeyvalwinx.o] Error 1\n-## make[2]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/rma'\n+## make[2]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/rma'\n ok 738 - ./f90/attr/attrmpi1f90 1\n ok 739 - ./f90/attr/baseattrf90 1\n ok 740 - ./f90/attr/baseattr2f90 1\n ok 741 - ./f90/attr/commattrf90 1\n ok 742 - ./f90/attr/commattr2f90 1\n ok 743 - ./f90/attr/commattr3f90 1\n ok 744 - ./f90/attr/commattr4f90 1\n@@ -1003830,114 +1003913,130 @@\n ok 752 - ./f90/coll/exscanf90 2\n ok 753 - ./f90/coll/alltoallwf90 2\n not ok 754 - ./f90/coll/alltoallvf90 \n ---\n Directory: ./f90/coll\n File: alltoallvf90\n Num-procs: \n- Date: \"Sun Jan 5 01:34:37 2020\"\n+ Date: \"Mon Dec 3 17:08:03 2018\"\n ...\n ## Test output (expected 'No Errors'):\n-## Failed to build alltoallvf90; make[2]: Entering directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/f90/coll'\n-## /usr/bin/mpif90.openmpi -g -O2 -ffile-prefix-map=/build/1st/mpi-testsuite-3.2+dfsg=. -fstack-protector-strong -fcray-pointer -c -o alltoallvf90.o ../../../f90/coll/alltoallvf90.f90\n+## Failed to build alltoallvf90; make[2]: Entering directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/f90/coll'\n+## /usr/bin/mpif90.openmpi -g -O2 -ffile-prefix-map=/build/mpi-testsuite-3.2+dfsg/2nd=. -fstack-protector-strong -fcray-pointer -c -o alltoallvf90.o ../../../f90/coll/alltoallvf90.f90\n ## ../../../f90/coll/alltoallvf90.f90:50:64:\n ## \n ## & rbuf, rcounts, rdispls, rtypes, comm, ierr )\n ## 1\n ## Error: There is no specific subroutine for the generic 'mpi_alltoallv' at (1)\n ## ../../../f90/coll/alltoallvf90.f90:103:59:\n ## \n ## & rbuf, rcounts, rdispls, rtypes, comm, ierr )\n ## 1\n ## Error: There is no specific subroutine for the generic 'mpi_alltoallv' at (1)\n ## make[2]: *** [Makefile:552: alltoallvf90.o] Error 1\n-## make[2]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/f90/coll'\n+## make[2]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/f90/coll'\n ok 755 - ./f90/coll/inplacef90 2\n ok 756 - ./f90/coll/reducelocalf90 2\n ok 757 - ./f90/coll/redscatf90 2\n not ok 758 - ./f90/coll/split_typef90 \n ---\n Directory: ./f90/coll\n File: split_typef90\n Num-procs: \n- Date: \"Sun Jan 5 01:34:40 2020\"\n+ Date: \"Mon Dec 3 17:08:11 2018\"\n ...\n ## Test output (expected 'No Errors'):\n-## Failed to build split_typef90; make[2]: Entering directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/f90/coll'\n-## /usr/bin/mpif90.openmpi -g -O2 -ffile-prefix-map=/build/1st/mpi-testsuite-3.2+dfsg=. -fstack-protector-strong -fcray-pointer -c -o split_typef90.o ../../../f90/coll/split_typef90.f90\n+## Failed to build split_typef90; make[2]: Entering directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/f90/coll'\n+## /usr/bin/mpif90.openmpi -g -O2 -ffile-prefix-map=/build/mpi-testsuite-3.2+dfsg/2nd=. -fstack-protector-strong -fcray-pointer -c -o split_typef90.o ../../../f90/coll/split_typef90.f90\n ## ../../../f90/coll/split_typef90.f90:39:59:\n ## \n ## & rbuf, rcounts, rdispls, rtypes, newcomm, ierr )\n ## 1\n ## Error: There is no specific subroutine for the generic 'mpi_alltoallv' at (1)\n ## make[2]: *** [Makefile:552: split_typef90.o] Error 1\n-## make[2]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/f90/coll'\n+## make[2]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/f90/coll'\n ok 759 - ./f90/coll/nonblockingf90 2\n not ok 760 - ./f90/coll/vw_inplacef90 \n ---\n Directory: ./f90/coll\n File: vw_inplacef90\n Num-procs: \n- Date: \"Sun Jan 5 01:34:42 2020\"\n+ Date: \"Mon Dec 3 17:08:29 2018\"\n ...\n ## Test output (expected 'No Errors'):\n-## Failed to build vw_inplacef90; make[2]: Entering directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/f90/coll'\n-## /usr/bin/mpif90.openmpi -g -O2 -ffile-prefix-map=/build/1st/mpi-testsuite-3.2+dfsg=. -fstack-protector-strong -fcray-pointer -c -o vw_inplacef90.o ../../../f90/coll/vw_inplacef90.f90\n+## Failed to build vw_inplacef90; make[2]: Entering directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/f90/coll'\n+## /usr/bin/mpif90.openmpi -g -O2 -ffile-prefix-map=/build/mpi-testsuite-3.2+dfsg/2nd=. -fstack-protector-strong -fcray-pointer -c -o vw_inplacef90.o ../../../f90/coll/vw_inplacef90.f90\n ## ../../../f90/coll/vw_inplacef90.f90:63:40:\n ## \n ## & comm, ierr )\n ## 1\n ## Error: There is no specific subroutine for the generic 'mpi_alltoallv' at (1)\n ## ../../../f90/coll/vw_inplacef90.f90:92:40:\n ## \n ## & comm, ierr )\n ## 1\n ## Error: There is no specific subroutine for the generic 'mpi_alltoallw' at (1)\n ## make[2]: *** [Makefile:552: vw_inplacef90.o] Error 1\n-## make[2]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/f90/coll'\n+## make[2]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/f90/coll'\n ok 761 - ./f90/coll/red_scat_blockf90 2\n not ok 762 - ./f90/coll/nonblocking_inpf90 \n ---\n Directory: ./f90/coll\n File: nonblocking_inpf90\n Num-procs: \n- Date: \"Sun Jan 5 01:34:43 2020\"\n+ Date: \"Mon Dec 3 17:08:33 2018\"\n ...\n ## Test output (expected 'No Errors'):\n-## Failed to build nonblocking_inpf90; make[2]: Entering directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/f90/coll'\n-## /usr/bin/mpif90.openmpi -g -O2 -ffile-prefix-map=/build/1st/mpi-testsuite-3.2+dfsg=. -fstack-protector-strong -fcray-pointer -c -o nonblocking_inpf90.o ../../../f90/coll/nonblocking_inpf90.f90\n+## Failed to build nonblocking_inpf90; make[2]: Entering directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/f90/coll'\n+## /usr/bin/mpif90.openmpi -g -O2 -ffile-prefix-map=/build/mpi-testsuite-3.2+dfsg/2nd=. -fstack-protector-strong -fcray-pointer -c -o nonblocking_inpf90.o ../../../f90/coll/nonblocking_inpf90.f90\n ## ../../../f90/coll/nonblocking_inpf90.f90:58:47:\n ## \n ## & comm, req, ierr )\n ## 1\n ## Error: There is no specific subroutine for the generic 'mpi_ialltoallv' at (1)\n ## ../../../f90/coll/nonblocking_inpf90.f90:87:47:\n ## \n ## & comm, req, ierr )\n ## 1\n ## Error: There is no specific subroutine for the generic 'mpi_ialltoallw' at (1)\n ## make[2]: *** [Makefile:552: nonblocking_inpf90.o] Error 1\n-## make[2]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/f90/coll'\n+## make[2]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/f90/coll'\n ok 763 - ./f90/comm/commnamef90 2\n ok 764 - ./f90/comm/commerrf90 2\n ok 765 - ./f90/ext/c2f2cf90 1\n ok 766 - ./f90/ext/c2f90mult 1\n ok 767 - ./f90/ext/ctypesinf90 1\n ok 768 - ./f90/ext/allocmemf90 1\n ok 769 - ./f90/info/infotestf90 1\n ok 770 - ./f90/info/infotest2f90 1\n ok 771 - ./f90/init/baseenvf90 1\n-ok 772 - ./f90/io/iwriteatf90 2\n+not ok 772 - ./f90/io/iwriteatf90 2\n+ ---\n+ Directory: ./f90/io\n+ File: iwriteatf90\n+ Num-procs: 2\n+ Date: \"Mon Dec 3 17:11:54 2018\"\n+ ...\n+## Test output (expected 'No Errors'):\n+## --------------------------------------------------------------------------\n+## The user-provided time limit for job execution has been reached:\n+## \n+## Timeout: 180 seconds\n+## \n+## The job will now be aborted. Please check your code and/or\n+## adjust/remove the job execution time limit (as specified by --timeout\n+## command line option or MPIEXEC_TIMEOUT environment variable).\n+## --------------------------------------------------------------------------\n ok 773 - ./f90/io/iwritef90 2\n not ok 774 - ./f90/io/iwriteshf90 2\n ---\n Directory: ./f90/io\n File: iwriteshf90\n Num-procs: 2\n- Date: \"Sun Jan 5 01:40:59 2020\"\n+ Date: \"Mon Dec 3 17:16:23 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## --------------------------------------------------------------------------\n ## The user-provided time limit for job execution has been reached:\n ## \n ## Timeout: 180 seconds\n ## \n@@ -1003947,96 +1004046,127 @@\n ## --------------------------------------------------------------------------\n ok 775 - ./f90/io/writef90 2\n ok 776 - ./f90/io/writeatf90 2\n ok 777 - ./f90/io/writeallf90 2\n ok 778 - ./f90/io/writeshf90 2\n ok 779 - ./f90/io/writeordf90 2\n ok 780 - ./f90/io/writeatallf90 2\n-ok 781 - ./f90/io/writeatallbef90 2\n-ok 782 - ./f90/io/writeallbef90 2\n-not ok 783 - ./f90/io/writeordbef90 2\n+not ok 781 - ./f90/io/writeatallbef90 2\n ---\n Directory: ./f90/io\n- File: writeordbef90\n+ File: writeatallbef90\n Num-procs: 2\n- Date: \"Sun Jan 5 01:48:57 2020\"\n+ Date: \"Mon Dec 3 17:19:42 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## --------------------------------------------------------------------------\n ## The user-provided time limit for job execution has been reached:\n ## \n ## Timeout: 180 seconds\n ## \n ## The job will now be aborted. Please check your code and/or\n ## adjust/remove the job execution time limit (as specified by --timeout\n ## command line option or MPIEXEC_TIMEOUT environment variable).\n ## --------------------------------------------------------------------------\n-ok 784 - ./f90/io/fileerrf90 1\n-ok 785 - ./f90/io/fileinfof90 2\n-ok 786 - ./f90/io/shpositionf90 2\n-not ok 787 - ./f90/io/atomicityf90 2\n+not ok 782 - ./f90/io/writeallbef90 2\n ---\n Directory: ./f90/io\n- File: atomicityf90\n+ File: writeallbef90\n Num-procs: 2\n- Date: \"Sun Jan 5 01:49:07 2020\"\n+ Date: \"Mon Dec 3 17:22:43 2018\"\n ...\n ## Test output (expected 'No Errors'):\n-## (noncontig)Process 1 : readbuf( 1939 ) is 0 , should be 10\n ## --------------------------------------------------------------------------\n-## MPI_ABORT was invoked on rank 1 in communicator MPI_COMM_WORLD\n-## with errorcode 1.\n+## The user-provided time limit for job execution has been reached:\n ## \n-## NOTE: invoking MPI_ABORT causes Open MPI to kill all MPI processes.\n-## You may or may not see output from other processes, depending on\n-## exactly when Open MPI kills them.\n+## Timeout: 180 seconds\n+## \n+## The job will now be aborted. Please check your code and/or\n+## adjust/remove the job execution time limit (as specified by --timeout\n+## command line option or MPIEXEC_TIMEOUT environment variable).\n ## --------------------------------------------------------------------------\n+not ok 783 - ./f90/io/writeordbef90 2\n+ ---\n+ Directory: ./f90/io\n+ File: writeordbef90\n+ Num-procs: 2\n+ Date: \"Mon Dec 3 17:25:44 2018\"\n+ ...\n+## Test output (expected 'No Errors'):\n+## --------------------------------------------------------------------------\n+## The user-provided time limit for job execution has been reached:\n+## \n+## Timeout: 180 seconds\n+## \n+## The job will now be aborted. Please check your code and/or\n+## adjust/remove the job execution time limit (as specified by --timeout\n+## command line option or MPIEXEC_TIMEOUT environment variable).\n+## --------------------------------------------------------------------------\n+ok 784 - ./f90/io/fileerrf90 1\n+ok 785 - ./f90/io/fileinfof90 2\n+ok 786 - ./f90/io/shpositionf90 2\n+ok 787 - ./f90/io/atomicityf90 2\n not ok 788 - ./f90/io/miscfilef90 2\n ---\n Directory: ./f90/io\n File: miscfilef90\n Num-procs: 2\n- Date: \"Sun Jan 5 01:49:08 2020\"\n+ Date: \"Mon Dec 3 17:26:41 2018\"\n ...\n ## Test output (expected 'No Errors'):\n-## File pointer position = 275 , should be 20\n-## File pointer position = 275 , should be 20\n-## File pointer position in bytes = 2680 , should be 1000\n-## File pointer position in bytes = 2680 , should be 1000\n-## Found 4 errors\n+## [i-capture-the-hostname:00913] PMIX ERROR: PMIX TEMPORARILY UNAVAILABLE in file ../../../../../../src/mca/ptl/tcp/ptl_tcp.c at line 790\n+## [i-capture-the-hostname:00913] OPAL ERROR: Unreachable in file ext2x_client.c at line 109\n+## *** An error occurred in MPI_Init\n+## *** on a NULL communicator\n+## *** MPI_ERRORS_ARE_FATAL (processes in this communicator will now abort,\n+## *** and potentially your MPI job)\n+## [i-capture-the-hostname:00913] Local abort before MPI_INIT completed completed successfully, but am not able to aggregate error messages, and not able to guarantee that all other processes were killed!\n+## --------------------------------------------------------------------------\n+## Primary job terminated normally, but 1 process returned\n+## a non-zero exit code. Per user-direction, the job has been aborted.\n+## --------------------------------------------------------------------------\n+## [i-capture-the-hostname:00880] PMIX ERROR: UNREACHABLE in file ../../../../../../src/mca/ptl/tcp/ptl_tcp_component.c at line 1423\n+## [i-capture-the-hostname:00880] PMIX ERROR: UNREACHABLE in file ../../../../../../src/mca/ptl/tcp/ptl_tcp_component.c at line 1423\n+## --------------------------------------------------------------------------\n+## mpiexec.openmpi detected that one or more processes exited with non-zero status, thus causing\n+## the job to be terminated. The first process to do so was:\n+## \n+## Process name: [[36385,1],1]\n+## Exit code: 1\n+## --------------------------------------------------------------------------\n not ok 789 - ./f90/io/setviewcurf90 2\n ---\n Directory: ./f90/io\n File: setviewcurf90\n Num-procs: 2\n- Date: \"Sun Jan 5 01:49:10 2020\"\n+ Date: \"Mon Dec 3 17:27:03 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## 0 : buf(1) = -1\n+## Found 2 errors\n ## 1 : buf(1) = -1\n ## mca_fbtl_posix_preadv: error in (p)readv:Invalid argument\n ## mca_fbtl_posix_preadv: error in (p)readv:Invalid argument\n-## Found 2 errors\n ok 790 - ./f90/io/c2f2ciof90 1\n ok 791 - ./f90/io/c2f90multio 1\n ok 809 - ./f90/io/i_setviewcurf90 2 # SKIP requires MPI version 3.1\n ok 810 - ./f90/io/iwriteatallf90 2 # SKIP requires MPI version 3.1\n ok 792 - ./f90/misc/sizeof2 1\n ok 793 - ./f90/pt2pt/statusesf90 1\n ok 794 - ./f90/pt2pt/greqf90 1\n not ok 795 - ./f90/pt2pt/allpairf90 \n ---\n Directory: ./f90/pt2pt\n File: allpairf90\n Num-procs: \n- Date: \"Sun Jan 5 01:49:22 2020\"\n+ Date: \"Mon Dec 3 17:27:38 2018\"\n ...\n ## Test output (expected 'No Errors'):\n-## Failed to build allpairf90; make[2]: Entering directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/f90/pt2pt'\n-## /usr/bin/mpif90.openmpi -g -O2 -ffile-prefix-map=/build/1st/mpi-testsuite-3.2+dfsg=. -fstack-protector-strong -fcray-pointer -c -o allpairf90.o ../../../f90/pt2pt/allpairf90.f90\n+## Failed to build allpairf90; make[2]: Entering directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/f90/pt2pt'\n+## /usr/bin/mpif90.openmpi -g -O2 -ffile-prefix-map=/build/mpi-testsuite-3.2+dfsg/2nd=. -fstack-protector-strong -fcray-pointer -c -o allpairf90.o ../../../f90/pt2pt/allpairf90.f90\n ## ../../../f90/pt2pt/allpairf90.f90:392:64:\n ## \n ## call MPI_Waitany(2, requests, index, statuses, ierr)\n ## 1\n ## Error: There is no specific subroutine for the generic 'mpi_waitany' at (1)\n ## ../../../f90/pt2pt/allpairf90.f90:500:50:\n ## \n@@ -1004045,104 +1004175,104 @@\n ## Error: There is no specific subroutine for the generic 'mpi_testany' at (1)\n ## ../../../f90/pt2pt/allpairf90.f90:744:50:\n ## \n ## & statuses(1,1), ierr)\n ## 1\n ## Error: There is no specific subroutine for the generic 'mpi_testany' at (1)\n ## make[2]: *** [Makefile:462: allpairf90.o] Error 1\n-## make[2]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/f90/pt2pt'\n+## make[2]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/f90/pt2pt'\n not ok 796 - ./f90/pt2pt/mprobef90 2\n ---\n Directory: ./f90/pt2pt\n File: mprobef90\n Num-procs: 2\n- Date: \"Sun Jan 5 01:49:23 2020\"\n+ Date: \"Mon Dec 3 17:27:41 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## s1(MPI_ERROR) != MPI_ERR_DIMS at T4 Mprobe().\n-## s2(MPI_ERROR) != MPI_ERR_OTHER at T4 Mrecv().\n-## s1(MPI_ERROR) != MPI_ERR_DIMS at T5 Mprobe().\n-## s2(MPI_ERROR) != MPI_ERR_OTHER at T5 Imrecv().\n-## s1(MPI_ERROR) != MPI_ERR_DIMS at T6 Improbe().\n-## s2(MPI_ERROR) != MPI_ERR_OTHER at T6 Mrecv().\n-## s1(MPI_ERROR) != MPI_ERR_DIMS at T7 Improbe().\n-## s2(MPI_ERROR) != MPI_ERR_OTHER at T7 Imrecv().\n-## Found 24 errors\n ## s1(MPI_ERROR) != MPI_ERR_DIMS at T0 Mprobe().\n ## s2(MPI_ERROR) != MPI_ERR_OTHER at T0 Mrecv().\n ## s1(MPI_ERROR) != MPI_ERR_DIMS at T1 Mprobe().\n+## s2(MPI_ERROR) != MPI_ERR_OTHER at T4 Mrecv().\n ## s2(MPI_ERROR) != MPI_ERR_OTHER at T1 Imrecv().\n ## s1(MPI_ERROR) != MPI_ERR_DIMS at T2 Improbe().\n ## s2(MPI_ERROR) != MPI_ERR_OTHER at T2 Mrecv().\n ## s1(MPI_ERROR) != MPI_ERR_DIMS at T3 Improbe().\n ## s2(MPI_ERROR) != MPI_ERR_OTHER at T3 Imrecv().\n ## s1(MPI_ERROR) != MPI_ERR_DIMS at T4 Mprobe().\n ## s2(MPI_ERROR) != MPI_ERR_OTHER at T4 Mrecv().\n ## s1(MPI_ERROR) != MPI_ERR_DIMS at T5 Mprobe().\n ## s2(MPI_ERROR) != MPI_ERR_OTHER at T5 Imrecv().\n ## s1(MPI_ERROR) != MPI_ERR_DIMS at T6 Improbe().\n ## s2(MPI_ERROR) != MPI_ERR_OTHER at T6 Mrecv().\n ## s1(MPI_ERROR) != MPI_ERR_DIMS at T7 Improbe().\n ## s2(MPI_ERROR) != MPI_ERR_OTHER at T7 Imrecv().\n+## s1(MPI_ERROR) != MPI_ERR_DIMS at T5 Mprobe().\n+## s2(MPI_ERROR) != MPI_ERR_OTHER at T5 Imrecv().\n+## s1(MPI_ERROR) != MPI_ERR_DIMS at T6 Improbe().\n+## s2(MPI_ERROR) != MPI_ERR_OTHER at T6 Mrecv().\n+## s1(MPI_ERROR) != MPI_ERR_DIMS at T7 Improbe().\n+## s2(MPI_ERROR) != MPI_ERR_OTHER at T7 Imrecv().\n+## Found 24 errors\n ok 797 - ./f90/datatype/typenamef90 1\n ok 798 - ./f90/datatype/typename3f90 1\n not ok 799 - ./f90/datatype/typesnamef90 1\n ---\n Directory: ./f90/datatype\n File: typesnamef90\n Num-procs: 1\n- Date: \"Sun Jan 5 01:49:28 2020\"\n+ Date: \"Mon Dec 3 17:27:48 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## (type2) Expected length 0, got 17\n ## (type2) Datatype name is not all blank\n ## Found 2 errors\n not ok 800 - ./f90/datatype/typecntsf90 1\n ---\n Directory: ./f90/datatype\n File: typecntsf90\n Num-procs: 1\n- Date: \"Sun Jan 5 01:49:31 2020\"\n+ Date: \"Mon Dec 3 17:27:53 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## \n ## Program received signal SIGSEGV: Segmentation fault - invalid memory reference.\n ## \n ## Backtrace for this error:\n-## #0 0x7fc031c6589d in ???\n-## #1 0x7fc031c64ae3 in ???\n-## #2 0x7fc03189f8df in ???\n-## #3 0x7fc0319a463e in ???\n-## #4 0x7fc031118415 in ???\n-## #5 0x7fc031f1305e in ???\n-## #6 0x7fc031f44b26 in ???\n-## #7 0x7fc0320156d3 in ???\n-## #8 0x56423c6e3535 in explore_\n+## #0 0x7f33f0eb889d in ???\n+## #1 0x7f33f0eb7ae3 in ???\n+## #2 0x7f33f0af28df in ???\n+## #3 0x7f33f0bf4e90 in ???\n+## #4 0x7f33f036b415 in ???\n+## #5 0x7f33f116605e in ???\n+## #6 0x7f33f1197b26 in ???\n+## #7 0x7f33f12686d3 in ???\n+## #8 0x55868c871535 in explore_\n ## \tat ../../../f90/datatype/typecntsf90.f90:52\n-## #9 0x56423c6e327a in MAIN__\n+## #9 0x55868c87127a in MAIN__\n ## \tat ../../../f90/datatype/typecntsf90.f90:24\n-## #10 0x56423c6e327a in main\n+## #10 0x55868c87127a in main\n ## \tat ../../../f90/datatype/typecntsf90.f90:8\n ## --------------------------------------------------------------------------\n ## Primary job terminated normally, but 1 process returned\n ## a non-zero exit code. Per user-direction, the job has been aborted.\n ## --------------------------------------------------------------------------\n ## --------------------------------------------------------------------------\n-## mpiexec.openmpi noticed that process rank 0 with PID 0 on node profitbricks-build5-amd64 exited on signal 11 (Segmentation fault).\n+## mpiexec.openmpi noticed that process rank 0 with PID 0 on node i-capture-the-hostname exited on signal 11 (Segmentation fault).\n ## --------------------------------------------------------------------------\n ok 801 - ./f90/datatype/typem2f90 1\n ok 802 - ./f90/datatype/typesubf90 1\n ok 803 - ./f90/datatype/packef90 1\n ok 804 - ./f90/datatype/gaddressf90 1\n not ok 805 - ./f90/datatype/allctypesf90 1\n ---\n Directory: ./f90/datatype\n File: allctypesf90\n Num-procs: 1\n- Date: \"Sun Jan 5 01:49:38 2020\"\n+ Date: \"Mon Dec 3 17:28:05 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## For datatype MPI_C_FLOAT_COMPLEX found name MPI_C_COMPLEX\n ## Found 1 errors\n ok 806 - ./f90/datatype/hindex1f90 1\n ok 807 - ./f90/datatype/hindexed_blockf90 1\n ok 808 - ./f90/datatype/bottom 2\n@@ -1004153,15 +1004283,15 @@\n ok 813 - ./f90/datatype/kinds 2\n ok 814 - ./f90/datatype/trf90 1\n not ok 815 - ./f90/datatype/get_elem_d 2\n ---\n Directory: ./f90/datatype\n File: get_elem_d\n Num-procs: 2\n- Date: \"Sun Jan 5 01:49:53 2020\"\n+ Date: \"Mon Dec 3 17:28:27 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## 1 R1 Get_elements count=*** but should be 0\n ## 2 R1 Get_elements count=*** but should be 0\n ## 3 R1 Get_elements count=*** but should be 0\n ## 4 R1 Get_elements count=*** but should be 0\n ## 5 R1 Get_elements count=*** but should be 0\n@@ -1004190,72 +1004320,72 @@\n ## errs= 27\n ok 816 - ./f90/datatype/get_elem_u 2\n not ok 817 - ./f90/f90types/createf90types 1\n ---\n Directory: ./f90/f90types\n File: createf90types\n Num-procs: 1\n- Date: \"Sun Jan 5 01:49:58 2020\"\n+ Date: \"Mon Dec 3 17:28:33 2018\"\n ...\n ## Test output (expected 'No Errors'):\n-## [profitbricks-build5-amd64:21633] *** Process received signal ***\n-## [profitbricks-build5-amd64:21633] Signal: Segmentation fault (11)\n-## [profitbricks-build5-amd64:21633] Signal code: Address not mapped (1)\n-## [profitbricks-build5-amd64:21633] Failing at address: 0x11\n-## [profitbricks-build5-amd64:21633] [ 0] /lib/x86_64-linux-gnu/libpthread.so.0(+0x126b0)[0x7f5b7a86a6b0]\n-## [profitbricks-build5-amd64:21633] [ 1] /usr/lib/x86_64-linux-gnu/libmpi.so.40(MPI_Type_get_contents+0x139)[0x7f5b7a90cb19]\n-## [profitbricks-build5-amd64:21633] [ 2] ./createf90types(+0x38c1)[0x5571534ee8c1]\n-## [profitbricks-build5-amd64:21633] [ 3] ./createf90types(+0x35f9)[0x5571534ee5f9]\n-## [profitbricks-build5-amd64:21633] [ 4] /lib/x86_64-linux-gnu/libc.so.6(__libc_start_main+0xeb)[0x7f5b7a6b409b]\n-## [profitbricks-build5-amd64:21633] [ 5] ./createf90types(+0x373a)[0x5571534ee73a]\n-## [profitbricks-build5-amd64:21633] *** End of error message ***\n+## [i-capture-the-hostname:05855] *** Process received signal ***\n+## [i-capture-the-hostname:05855] Signal: Segmentation fault (11)\n+## [i-capture-the-hostname:05855] Signal code: Address not mapped (1)\n+## [i-capture-the-hostname:05855] Failing at address: 0x11\n+## [i-capture-the-hostname:05855] [ 0] /lib/x86_64-linux-gnu/libpthread.so.0(+0x126b0)[0x7fd86ab956b0]\n+## [i-capture-the-hostname:05855] [ 1] /usr/lib/x86_64-linux-gnu/libmpi.so.40(MPI_Type_get_contents+0x139)[0x7fd86ac37b19]\n+## [i-capture-the-hostname:05855] [ 2] ./createf90types(+0x38c1)[0x5653f4a9b8c1]\n+## [i-capture-the-hostname:05855] [ 3] ./createf90types(+0x35f9)[0x5653f4a9b5f9]\n+## [i-capture-the-hostname:05855] [ 4] /lib/x86_64-linux-gnu/libc.so.6(__libc_start_main+0xeb)[0x7fd86a9df09b]\n+## [i-capture-the-hostname:05855] [ 5] ./createf90types(+0x373a)[0x5653f4a9b73a]\n+## [i-capture-the-hostname:05855] *** End of error message ***\n ## --------------------------------------------------------------------------\n ## Primary job terminated normally, but 1 process returned\n ## a non-zero exit code. Per user-direction, the job has been aborted.\n ## --------------------------------------------------------------------------\n ## --------------------------------------------------------------------------\n-## mpiexec.openmpi noticed that process rank 0 with PID 0 on node profitbricks-build5-amd64 exited on signal 11 (Segmentation fault).\n+## mpiexec.openmpi noticed that process rank 0 with PID 0 on node i-capture-the-hostname exited on signal 11 (Segmentation fault).\n ## --------------------------------------------------------------------------\n not ok 818 - ./f90/f90types/createf90types 1\n ---\n Directory: ./f90/f90types\n File: createf90types\n Num-procs: 1\n- Date: \"Sun Jan 5 01:50:02 2020\"\n+ Date: \"Mon Dec 3 17:28:37 2018\"\n ...\n ## Test output (expected 'No Errors'):\n-## [profitbricks-build5-amd64:22139] *** Process received signal ***\n-## [profitbricks-build5-amd64:22139] Signal: Segmentation fault (11)\n-## [profitbricks-build5-amd64:22139] Signal code: Address not mapped (1)\n-## [profitbricks-build5-amd64:22139] Failing at address: 0x11\n-## [profitbricks-build5-amd64:22139] [ 0] /lib/x86_64-linux-gnu/libpthread.so.0(+0x126b0)[0x7f185f1aa6b0]\n-## [profitbricks-build5-amd64:22139] [ 1] /usr/lib/x86_64-linux-gnu/libmpi.so.40(MPI_Type_get_contents+0x139)[0x7f185f24cb19]\n-## [profitbricks-build5-amd64:22139] [ 2] ./createf90types(+0x38c1)[0x563f76edf8c1]\n-## [profitbricks-build5-amd64:22139] [ 3] ./createf90types(+0x35f9)[0x563f76edf5f9]\n-## [profitbricks-build5-amd64:22139] [ 4] /lib/x86_64-linux-gnu/libc.so.6(__libc_start_main+0xeb)[0x7f185eff409b]\n-## [profitbricks-build5-amd64:22139] [ 5] ./createf90types(+0x373a)[0x563f76edf73a]\n-## [profitbricks-build5-amd64:22139] *** End of error message ***\n+## [i-capture-the-hostname:05920] *** Process received signal ***\n+## [i-capture-the-hostname:05920] Signal: Segmentation fault (11)\n+## [i-capture-the-hostname:05920] Signal code: Address not mapped (1)\n+## [i-capture-the-hostname:05920] Failing at address: 0x11\n+## [i-capture-the-hostname:05920] [ 0] /lib/x86_64-linux-gnu/libpthread.so.0(+0x126b0)[0x7fa2ded4d6b0]\n+## [i-capture-the-hostname:05920] [ 1] /usr/lib/x86_64-linux-gnu/libmpi.so.40(MPI_Type_get_contents+0x139)[0x7fa2dedefb19]\n+## [i-capture-the-hostname:05920] [ 2] ./createf90types(+0x38c1)[0x561d264c18c1]\n+## [i-capture-the-hostname:05920] [ 3] ./createf90types(+0x35f9)[0x561d264c15f9]\n+## [i-capture-the-hostname:05920] [ 4] /lib/x86_64-linux-gnu/libc.so.6(__libc_start_main+0xeb)[0x7fa2deb9709b]\n+## [i-capture-the-hostname:05920] [ 5] ./createf90types(+0x373a)[0x561d264c173a]\n+## [i-capture-the-hostname:05920] *** End of error message ***\n ## --------------------------------------------------------------------------\n ## Primary job terminated normally, but 1 process returned\n ## a non-zero exit code. Per user-direction, the job has been aborted.\n ## --------------------------------------------------------------------------\n ## --------------------------------------------------------------------------\n-## mpiexec.openmpi noticed that process rank 0 with PID 0 on node profitbricks-build5-amd64 exited on signal 11 (Segmentation fault).\n+## mpiexec.openmpi noticed that process rank 0 with PID 0 on node i-capture-the-hostname exited on signal 11 (Segmentation fault).\n ## --------------------------------------------------------------------------\n ok 819 - ./f90/rma/winscale1f90 2\n ok 820 - ./f90/rma/winfencef90 2\n ok 821 - ./f90/rma/wingetf90 2\n ok 822 - ./f90/rma/winscale2f90 2\n ok 823 - ./f90/rma/winerrf90 1\n not ok 824 - ./f90/rma/winnamef90 1\n ---\n Directory: ./f90/rma\n File: winnamef90\n Num-procs: 1\n- Date: \"Sun Jan 5 01:50:10 2020\"\n+ Date: \"Mon Dec 3 17:28:51 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Did not get empty name from new window\n ## Found 1 errors\n ok 825 - ./f90/rma/wingroupf90 2\n ok 826 - ./f90/rma/winaccf90 2\n ok 827 - ./f90/rma/c2f2cwinf90 1\n@@ -1004265,406 +1004395,392 @@\n ok 850 - ./f90/rma/aintf90 2 # SKIP requires MPI version 3.1\n ok 831 - ./f90/spawn/namepubf90 2\n not ok 832 - ./f90/spawn/spawnf90 1\n ---\n Directory: ./f90/spawn\n File: spawnf90\n Num-procs: 1\n- Date: \"Sun Jan 5 01:50:22 2020\"\n+ Date: \"Mon Dec 3 17:29:12 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## --------------------------------------------------------------------------\n ## At least one pair of MPI processes are unable to reach each other for\n ## MPI communications. This means that no Open MPI device has indicated\n ## that it can be used to communicate between these processes. This is\n ## an error; Open MPI requires that all MPI processes be able to reach\n ## each other. This error can sometimes be the result of forgetting to\n ## specify the \"self\" BTL.\n ## \n-## Process 1 ([[39792,2],0]) is on host: profitbricks-build5-amd64\n-## Process 2 ([[39792,1],0]) is on host: profitbricks-build5-amd64\n+## Process 1 ([[37874,2],1]) is on host: i-capture-the-hostname\n+## Process 2 ([[37874,1],0]) is on host: i-capture-the-hostname\n ## BTLs attempted: vader self\n ## \n ## Your MPI job is now going to abort; sorry.\n ## --------------------------------------------------------------------------\n-## [profitbricks-build5-amd64:24549] [[39792,1],0] ORTE_ERROR_LOG: Unreachable in file dpm/dpm.c at line 433\n-## [profitbricks-build5-amd64:24613] [[39792,2],0] ORTE_ERROR_LOG: Unreachable in file dpm/dpm.c at line 433\n-## [profitbricks-build5-amd64:24615] [[39792,2],1] ORTE_ERROR_LOG: Unreachable in file dpm/dpm.c at line 433\n ## --------------------------------------------------------------------------\n ## It looks like MPI_INIT failed for some reason; your parallel process is\n ## likely to abort. There are many reasons that a parallel process can\n ## fail during MPI_INIT; some of which are due to configuration or environment\n ## problems. This failure appears to be an internal failure; here's some\n ## additional information (which may only be relevant to an Open MPI\n ## developer):\n ## \n ## ompi_dpm_dyn_init() failed\n ## --> Returned \"Unreachable\" (-12) instead of \"Success\" (0)\n ## --------------------------------------------------------------------------\n-## [profitbricks-build5-amd64:24549] *** An error occurred in MPI_Comm_spawn\n-## [profitbricks-build5-amd64:24549] *** reported by process [2607808513,0]\n-## [profitbricks-build5-amd64:24549] *** on communicator MPI_COMM_WORLD\n-## [profitbricks-build5-amd64:24549] *** MPI_ERR_INTERN: internal error\n-## [profitbricks-build5-amd64:24549] *** MPI_ERRORS_ARE_FATAL (processes in this communicator will now abort,\n-## [profitbricks-build5-amd64:24549] *** and potentially your MPI job)\n-## [profitbricks-build5-amd64:24613] *** An error occurred in MPI_Init\n-## [profitbricks-build5-amd64:24613] *** reported by process [2607808514,0]\n-## [profitbricks-build5-amd64:24613] *** on a NULL communicator\n-## [profitbricks-build5-amd64:24613] *** Unknown error\n-## [profitbricks-build5-amd64:24613] *** MPI_ERRORS_ARE_FATAL (processes in this communicator will now abort,\n-## [profitbricks-build5-amd64:24613] *** and potentially your MPI job)\n-## [profitbricks-build5-amd64:24482] 2 more processes have sent help message help-mca-bml-r2.txt / unreachable proc\n-## [profitbricks-build5-amd64:24482] Set MCA parameter \"orte_base_help_aggregate\" to 0 to see all help / error messages\n-## [profitbricks-build5-amd64:24482] 1 more process has sent help message help-mpi-runtime.txt / mpi_init:startup:internal-failure\n-## [profitbricks-build5-amd64:24482] 1 more process has sent help message help-mpi-errors.txt / mpi_errors_are_fatal unknown handle\n-## [warn] Epoll MOD(1) on fd 27 failed. Old events were 6; read change was 0 (none); write change was 2 (del); close change was 0 (none): Bad file descriptor\n+## [i-capture-the-hostname:07987] [[37874,2],0] ORTE_ERROR_LOG: Unreachable in file dpm/dpm.c at line 433\n+## [i-capture-the-hostname:07932] [[37874,1],0] ORTE_ERROR_LOG: Unreachable in file dpm/dpm.c at line 433\n+## [i-capture-the-hostname:07988] [[37874,2],1] ORTE_ERROR_LOG: Unreachable in file dpm/dpm.c at line 433\n+## [i-capture-the-hostname:07932] *** An error occurred in MPI_Comm_spawn\n+## [i-capture-the-hostname:07932] *** reported by process [2482110465,0]\n+## [i-capture-the-hostname:07932] *** on communicator MPI_COMM_WORLD\n+## [i-capture-the-hostname:07932] *** MPI_ERR_INTERN: internal error\n+## [i-capture-the-hostname:07932] *** MPI_ERRORS_ARE_FATAL (processes in this communicator will now abort,\n+## [i-capture-the-hostname:07932] *** and potentially your MPI job)\n+## [i-capture-the-hostname:07988] *** An error occurred in MPI_Init\n+## [i-capture-the-hostname:07988] *** reported by process [2482110466,1]\n+## [i-capture-the-hostname:07988] *** on a NULL communicator\n+## [i-capture-the-hostname:07988] *** Unknown error\n+## [i-capture-the-hostname:07988] *** MPI_ERRORS_ARE_FATAL (processes in this communicator will now abort,\n+## [i-capture-the-hostname:07988] *** and potentially your MPI job)\n+## [i-capture-the-hostname:07843] 2 more processes have sent help message help-mca-bml-r2.txt / unreachable proc\n+## [i-capture-the-hostname:07843] Set MCA parameter \"orte_base_help_aggregate\" to 0 to see all help / error messages\n+## [i-capture-the-hostname:07843] 1 more process has sent help message help-mpi-runtime.txt / mpi_init:startup:internal-failure\n+## [i-capture-the-hostname:07843] 1 more process has sent help message help-mpi-errors.txt / mpi_errors_are_fatal unknown handle\n not ok 833 - ./f90/spawn/spawnargvf90 1\n ---\n Directory: ./f90/spawn\n File: spawnargvf90\n Num-procs: 1\n- Date: \"Sun Jan 5 01:50:24 2020\"\n+ Date: \"Mon Dec 3 17:29:15 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## --------------------------------------------------------------------------\n ## At least one pair of MPI processes are unable to reach each other for\n ## MPI communications. This means that no Open MPI device has indicated\n ## that it can be used to communicate between these processes. This is\n ## an error; Open MPI requires that all MPI processes be able to reach\n ## each other. This error can sometimes be the result of forgetting to\n ## specify the \"self\" BTL.\n ## \n-## Process 1 ([[42070,2],0]) is on host: profitbricks-build5-amd64\n-## Process 2 ([[42070,1],0]) is on host: profitbricks-build5-amd64\n-## BTLs attempted: vader self\n+## Process 1 ([[44378,1],0]) is on host: i-capture-the-hostname\n+## Process 2 ([[44378,2],0]) is on host: i-capture-the-hostname\n+## BTLs attempted: self\n ## \n ## Your MPI job is now going to abort; sorry.\n ## --------------------------------------------------------------------------\n+## [i-capture-the-hostname:08249] [[44378,1],0] ORTE_ERROR_LOG: Unreachable in file dpm/dpm.c at line 433\n+## [i-capture-the-hostname:08249] *** An error occurred in MPI_Comm_spawn\n+## [i-capture-the-hostname:08249] *** reported by process [2908356609,0]\n+## [i-capture-the-hostname:08249] *** on communicator MPI_COMM_WORLD\n+## [i-capture-the-hostname:08249] *** MPI_ERR_INTERN: internal error\n+## [i-capture-the-hostname:08249] *** MPI_ERRORS_ARE_FATAL (processes in this communicator will now abort,\n+## [i-capture-the-hostname:08249] *** and potentially your MPI job)\n ## --------------------------------------------------------------------------\n ## It looks like MPI_INIT failed for some reason; your parallel process is\n ## likely to abort. There are many reasons that a parallel process can\n ## fail during MPI_INIT; some of which are due to configuration or environment\n ## problems. This failure appears to be an internal failure; here's some\n ## additional information (which may only be relevant to an Open MPI\n ## developer):\n ## \n ## ompi_dpm_dyn_init() failed\n ## --> Returned \"Unreachable\" (-12) instead of \"Success\" (0)\n ## --------------------------------------------------------------------------\n-## [profitbricks-build5-amd64:24799] [[42070,2],0] ORTE_ERROR_LOG: Unreachable in file dpm/dpm.c at line 433\n-## [profitbricks-build5-amd64:24799] *** An error occurred in MPI_Init\n-## [profitbricks-build5-amd64:24799] *** reported by process [2757099522,0]\n-## [profitbricks-build5-amd64:24799] *** on a NULL communicator\n-## [profitbricks-build5-amd64:24799] *** Unknown error\n-## [profitbricks-build5-amd64:24799] *** MPI_ERRORS_ARE_FATAL (processes in this communicator will now abort,\n-## [profitbricks-build5-amd64:24799] *** and potentially your MPI job)\n-## [profitbricks-build5-amd64:24752] [[42070,1],0] ORTE_ERROR_LOG: Unreachable in file dpm/dpm.c at line 433\n-## [profitbricks-build5-amd64:24800] [[42070,2],1] ORTE_ERROR_LOG: Unreachable in file dpm/dpm.c at line 433\n-## [profitbricks-build5-amd64:24752] *** An error occurred in MPI_Comm_spawn\n-## [profitbricks-build5-amd64:24752] *** reported by process [2757099521,0]\n-## [profitbricks-build5-amd64:24752] *** on communicator MPI_COMM_WORLD\n-## [profitbricks-build5-amd64:24752] *** MPI_ERR_INTERN: internal error\n-## [profitbricks-build5-amd64:24752] *** MPI_ERRORS_ARE_FATAL (processes in this communicator will now abort,\n-## [profitbricks-build5-amd64:24752] *** and potentially your MPI job)\n-## [profitbricks-build5-amd64:24708] 2 more processes have sent help message help-mca-bml-r2.txt / unreachable proc\n-## [profitbricks-build5-amd64:24708] Set MCA parameter \"orte_base_help_aggregate\" to 0 to see all help / error messages\n-## [profitbricks-build5-amd64:24708] 1 more process has sent help message help-mpi-runtime.txt / mpi_init:startup:internal-failure\n-## [profitbricks-build5-amd64:24708] 1 more process has sent help message help-mpi-errors.txt / mpi_errors_are_fatal unknown handle\n-## [warn] Epoll MOD(1) on fd 29 failed. Old events were 6; read change was 0 (none); write change was 2 (del); close change was 0 (none): Bad file descriptor\n+## [i-capture-the-hostname:08273] [[44378,2],0] ORTE_ERROR_LOG: Unreachable in file dpm/dpm.c at line 433\n+## [i-capture-the-hostname:08274] [[44378,2],1] ORTE_ERROR_LOG: Unreachable in file dpm/dpm.c at line 433\n+## [i-capture-the-hostname:08203] 2 more processes have sent help message help-mca-bml-r2.txt / unreachable proc\n+## [i-capture-the-hostname:08203] Set MCA parameter \"orte_base_help_aggregate\" to 0 to see all help / error messages\n+## [i-capture-the-hostname:08203] 1 more process has sent help message help-mpi-runtime.txt / mpi_init:startup:internal-failure\n ok 834 - ./f90/spawn/connaccf90 2\n not ok 835 - ./f90/spawn/spawnmultf90 1\n ---\n Directory: ./f90/spawn\n File: spawnmultf90\n Num-procs: 1\n- Date: \"Sun Jan 5 01:50:28 2020\"\n+ Date: \"Mon Dec 3 17:29:21 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## --------------------------------------------------------------------------\n ## At least one pair of MPI processes are unable to reach each other for\n ## MPI communications. This means that no Open MPI device has indicated\n ## that it can be used to communicate between these processes. This is\n ## an error; Open MPI requires that all MPI processes be able to reach\n ## each other. This error can sometimes be the result of forgetting to\n ## specify the \"self\" BTL.\n ## \n-## Process 1 ([[42300,1],0]) is on host: profitbricks-build5-amd64\n-## Process 2 ([[42300,2],0]) is on host: profitbricks-build5-amd64\n+## Process 1 ([[44498,1],0]) is on host: i-capture-the-hostname\n+## Process 2 ([[44498,2],0]) is on host: i-capture-the-hostname\n ## BTLs attempted: self\n ## \n ## Your MPI job is now going to abort; sorry.\n ## --------------------------------------------------------------------------\n-## [profitbricks-build5-amd64:25125] [[42300,1],0] ORTE_ERROR_LOG: Unreachable in file dpm/dpm.c at line 433\n-## [profitbricks-build5-amd64:25125] *** An error occurred in MPI_Comm_spawn_multiple\n-## [profitbricks-build5-amd64:25125] *** reported by process [2772172801,0]\n-## [profitbricks-build5-amd64:25125] *** on communicator MPI_COMM_WORLD\n-## [profitbricks-build5-amd64:25125] *** MPI_ERR_INTERN: internal error\n-## [profitbricks-build5-amd64:25125] *** MPI_ERRORS_ARE_FATAL (processes in this communicator will now abort,\n-## [profitbricks-build5-amd64:25125] *** and potentially your MPI job)\n+## [i-capture-the-hostname:08354] [[44498,1],0] ORTE_ERROR_LOG: Unreachable in file dpm/dpm.c at line 433\n+## [i-capture-the-hostname:08354] *** An error occurred in MPI_Comm_spawn_multiple\n+## [i-capture-the-hostname:08354] *** reported by process [2916220929,0]\n+## [i-capture-the-hostname:08354] *** on communicator MPI_COMM_WORLD\n+## [i-capture-the-hostname:08354] *** MPI_ERR_INTERN: internal error\n+## [i-capture-the-hostname:08354] *** MPI_ERRORS_ARE_FATAL (processes in this communicator will now abort,\n+## [i-capture-the-hostname:08354] *** and potentially your MPI job)\n ## --------------------------------------------------------------------------\n ## It looks like MPI_INIT failed for some reason; your parallel process is\n ## likely to abort. There are many reasons that a parallel process can\n ## fail during MPI_INIT; some of which are due to configuration or environment\n ## problems. This failure appears to be an internal failure; here's some\n ## additional information (which may only be relevant to an Open MPI\n ## developer):\n ## \n ## ompi_dpm_dyn_init() failed\n ## --> Returned \"Unreachable\" (-12) instead of \"Success\" (0)\n ## --------------------------------------------------------------------------\n-## [profitbricks-build5-amd64:25217] [[42300,2],1] ORTE_ERROR_LOG: Unreachable in file dpm/dpm.c at line 433\n-## [profitbricks-build5-amd64:25214] [[42300,2],0] ORTE_ERROR_LOG: Unreachable in file dpm/dpm.c at line 433\n-## [profitbricks-build5-amd64:25217] *** An error occurred in MPI_Init\n-## [profitbricks-build5-amd64:25217] *** reported by process [2772172802,1]\n-## [profitbricks-build5-amd64:25217] *** on a NULL communicator\n-## [profitbricks-build5-amd64:25217] *** Unknown error\n-## [profitbricks-build5-amd64:25217] *** MPI_ERRORS_ARE_FATAL (processes in this communicator will now abort,\n-## [profitbricks-build5-amd64:25217] *** and potentially your MPI job)\n-## [profitbricks-build5-amd64:25070] 2 more processes have sent help message help-mca-bml-r2.txt / unreachable proc\n-## [profitbricks-build5-amd64:25070] Set MCA parameter \"orte_base_help_aggregate\" to 0 to see all help / error messages\n-## [profitbricks-build5-amd64:25070] 1 more process has sent help message help-mpi-runtime.txt / mpi_init:startup:internal-failure\n-## [profitbricks-build5-amd64:25070] 1 more process has sent help message help-mpi-errors.txt / mpi_errors_are_fatal unknown handle\n+## [i-capture-the-hostname:08387] [[44498,2],1] ORTE_ERROR_LOG: Unreachable in file dpm/dpm.c at line 433\n+## [i-capture-the-hostname:08386] [[44498,2],0] ORTE_ERROR_LOG: Unreachable in file dpm/dpm.c at line 433\n+## [i-capture-the-hostname:08323] 2 more processes have sent help message help-mca-bml-r2.txt / unreachable proc\n+## [i-capture-the-hostname:08323] Set MCA parameter \"orte_base_help_aggregate\" to 0 to see all help / error messages\n+## [i-capture-the-hostname:08323] 1 more process has sent help message help-mpi-runtime.txt / mpi_init:startup:internal-failure\n not ok 836 - ./f90/spawn/spawnmult2f90 2\n ---\n Directory: ./f90/spawn\n File: spawnmult2f90\n Num-procs: 2\n- Date: \"Sun Jan 5 01:50:30 2020\"\n+ Date: \"Mon Dec 3 17:29:24 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## --------------------------------------------------------------------------\n ## At least one pair of MPI processes are unable to reach each other for\n ## MPI communications. This means that no Open MPI device has indicated\n ## that it can be used to communicate between these processes. This is\n ## an error; Open MPI requires that all MPI processes be able to reach\n ## each other. This error can sometimes be the result of forgetting to\n ## specify the \"self\" BTL.\n ## \n-## Process 1 ([[43006,1],0]) is on host: profitbricks-build5-amd64\n-## Process 2 ([[43006,2],0]) is on host: profitbricks-build5-amd64\n+## Process 1 ([[44099,1],1]) is on host: i-capture-the-hostname\n+## Process 2 ([[44099,2],0]) is on host: i-capture-the-hostname\n ## BTLs attempted: vader self\n ## \n ## Your MPI job is now going to abort; sorry.\n ## --------------------------------------------------------------------------\n-## [profitbricks-build5-amd64:25470] [[43006,1],0] ORTE_ERROR_LOG: Unreachable in file dpm/dpm.c at line 433\n-## [profitbricks-build5-amd64:25470] *** An error occurred in MPI_Comm_spawn_multiple\n-## [profitbricks-build5-amd64:25470] *** reported by process [2818441217,0]\n-## [profitbricks-build5-amd64:25470] *** on communicator MPI_COMM_WORLD\n-## [profitbricks-build5-amd64:25470] *** MPI_ERR_INTERN: internal error\n-## [profitbricks-build5-amd64:25470] *** MPI_ERRORS_ARE_FATAL (processes in this communicator will now abort,\n-## [profitbricks-build5-amd64:25470] *** and potentially your MPI job)\n-## [profitbricks-build5-amd64:25576] [[43006,2],1] ORTE_ERROR_LOG: Unreachable in file dpm/dpm.c at line 433\n-## [profitbricks-build5-amd64:25472] [[43006,1],1] ORTE_ERROR_LOG: Unreachable in file dpm/dpm.c at line 433\n-## [profitbricks-build5-amd64:25571] [[43006,2],0] ORTE_ERROR_LOG: Unreachable in file dpm/dpm.c at line 433\n+## [i-capture-the-hostname:08524] [[44099,1],1] ORTE_ERROR_LOG: Unreachable in file dpm/dpm.c at line 433\n+## [i-capture-the-hostname:08590] [[44099,2],0] ORTE_ERROR_LOG: Unreachable in file dpm/dpm.c at line 433\n+## [i-capture-the-hostname:08593] [[44099,2],1] ORTE_ERROR_LOG: Unreachable in file dpm/dpm.c at line 433\n ## --------------------------------------------------------------------------\n ## It looks like MPI_INIT failed for some reason; your parallel process is\n ## likely to abort. There are many reasons that a parallel process can\n ## fail during MPI_INIT; some of which are due to configuration or environment\n ## problems. This failure appears to be an internal failure; here's some\n ## additional information (which may only be relevant to an Open MPI\n ## developer):\n ## \n ## ompi_dpm_dyn_init() failed\n ## --> Returned \"Unreachable\" (-12) instead of \"Success\" (0)\n ## --------------------------------------------------------------------------\n-## [profitbricks-build5-amd64:25571] *** An error occurred in MPI_Init\n-## [profitbricks-build5-amd64:25571] *** reported by process [2818441218,0]\n-## [profitbricks-build5-amd64:25571] *** on a NULL communicator\n-## [profitbricks-build5-amd64:25571] *** Unknown error\n-## [profitbricks-build5-amd64:25571] *** MPI_ERRORS_ARE_FATAL (processes in this communicator will now abort,\n-## [profitbricks-build5-amd64:25571] *** and potentially your MPI job)\n-## [profitbricks-build5-amd64:25388] 3 more processes have sent help message help-mca-bml-r2.txt / unreachable proc\n-## [profitbricks-build5-amd64:25388] Set MCA parameter \"orte_base_help_aggregate\" to 0 to see all help / error messages\n-## [profitbricks-build5-amd64:25388] 1 more process has sent help message help-mpi-errors.txt / mpi_errors_are_fatal\n-## [profitbricks-build5-amd64:25388] 1 more process has sent help message help-mpi-runtime.txt / mpi_init:startup:internal-failure\n-## [profitbricks-build5-amd64:25388] 1 more process has sent help message help-mpi-errors.txt / mpi_errors_are_fatal unknown handle\n+## [i-capture-the-hostname:08524] *** An error occurred in MPI_Comm_spawn_multiple\n+## [i-capture-the-hostname:08524] *** reported by process [2890072065,1]\n+## [i-capture-the-hostname:08524] *** on communicator MPI_COMM_WORLD\n+## [i-capture-the-hostname:08524] *** MPI_ERR_INTERN: internal error\n+## [i-capture-the-hostname:08524] *** MPI_ERRORS_ARE_FATAL (processes in this communicator will now abort,\n+## [i-capture-the-hostname:08524] *** and potentially your MPI job)\n+## [i-capture-the-hostname:08519] [[44099,1],0] ORTE_ERROR_LOG: Unreachable in file dpm/dpm.c at line 433\n+## [i-capture-the-hostname:08590] *** An error occurred in MPI_Init\n+## [i-capture-the-hostname:08590] *** reported by process [2890072066,0]\n+## [i-capture-the-hostname:08590] *** on a NULL communicator\n+## [i-capture-the-hostname:08590] *** Unknown error\n+## [i-capture-the-hostname:08590] *** MPI_ERRORS_ARE_FATAL (processes in this communicator will now abort,\n+## [i-capture-the-hostname:08590] *** and potentially your MPI job)\n+## [i-capture-the-hostname:08466] 3 more processes have sent help message help-mca-bml-r2.txt / unreachable proc\n+## [i-capture-the-hostname:08466] Set MCA parameter \"orte_base_help_aggregate\" to 0 to see all help / error messages\n+## [i-capture-the-hostname:08466] 1 more process has sent help message help-mpi-runtime.txt / mpi_init:startup:internal-failure\n+## [i-capture-the-hostname:08466] 1 more process has sent help message help-mpi-errors.txt / mpi_errors_are_fatal\n+## [i-capture-the-hostname:08466] 1 more process has sent help message help-mpi-errors.txt / mpi_errors_are_fatal unknown handle\n not ok 837 - ./f90/spawn/spawnargvf03 1\n ---\n Directory: ./f90/spawn\n File: spawnargvf03\n Num-procs: 1\n- Date: \"Sun Jan 5 01:50:32 2020\"\n+ Date: \"Mon Dec 3 17:29:27 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## --------------------------------------------------------------------------\n ## At least one pair of MPI processes are unable to reach each other for\n ## MPI communications. This means that no Open MPI device has indicated\n ## that it can be used to communicate between these processes. This is\n ## an error; Open MPI requires that all MPI processes be able to reach\n ## each other. This error can sometimes be the result of forgetting to\n ## specify the \"self\" BTL.\n ## \n-## Process 1 ([[41128,2],1]) is on host: profitbricks-build5-amd64\n-## Process 2 ([[41128,1],0]) is on host: profitbricks-build5-amd64\n+## Process 1 ([[44881,2],0]) is on host: i-capture-the-hostname\n+## Process 2 ([[44881,1],0]) is on host: i-capture-the-hostname\n ## BTLs attempted: vader self\n ## \n ## Your MPI job is now going to abort; sorry.\n ## --------------------------------------------------------------------------\n-## [profitbricks-build5-amd64:25943] [[41128,2],1] ORTE_ERROR_LOG: Unreachable in file dpm/dpm.c at line 433\n-## [profitbricks-build5-amd64:25812] [[41128,1],0] ORTE_ERROR_LOG: Unreachable in file dpm/dpm.c at line 433\n+## [i-capture-the-hostname:08894] [[44881,2],0] ORTE_ERROR_LOG: Unreachable in file dpm/dpm.c at line 433\n ## --------------------------------------------------------------------------\n ## It looks like MPI_INIT failed for some reason; your parallel process is\n ## likely to abort. There are many reasons that a parallel process can\n ## fail during MPI_INIT; some of which are due to configuration or environment\n ## problems. This failure appears to be an internal failure; here's some\n ## additional information (which may only be relevant to an Open MPI\n ## developer):\n ## \n ## ompi_dpm_dyn_init() failed\n ## --> Returned \"Unreachable\" (-12) instead of \"Success\" (0)\n ## --------------------------------------------------------------------------\n-## [profitbricks-build5-amd64:25812] *** An error occurred in MPI_Comm_spawn\n-## [profitbricks-build5-amd64:25812] *** reported by process [2695364609,0]\n-## [profitbricks-build5-amd64:25812] *** on communicator MPI_COMM_WORLD\n-## [profitbricks-build5-amd64:25812] *** MPI_ERR_INTERN: internal error\n-## [profitbricks-build5-amd64:25812] *** MPI_ERRORS_ARE_FATAL (processes in this communicator will now abort,\n-## [profitbricks-build5-amd64:25812] *** and potentially your MPI job)\n-## [profitbricks-build5-amd64:25937] [[41128,2],0] ORTE_ERROR_LOG: Unreachable in file dpm/dpm.c at line 433\n-## [profitbricks-build5-amd64:25937] *** An error occurred in MPI_Init\n-## [profitbricks-build5-amd64:25937] *** reported by process [2695364610,0]\n-## [profitbricks-build5-amd64:25937] *** on a NULL communicator\n-## [profitbricks-build5-amd64:25937] *** Unknown error\n-## [profitbricks-build5-amd64:25937] *** MPI_ERRORS_ARE_FATAL (processes in this communicator will now abort,\n-## [profitbricks-build5-amd64:25937] *** and potentially your MPI job)\n-## [profitbricks-build5-amd64:25722] 2 more processes have sent help message help-mca-bml-r2.txt / unreachable proc\n-## [profitbricks-build5-amd64:25722] Set MCA parameter \"orte_base_help_aggregate\" to 0 to see all help / error messages\n-## [profitbricks-build5-amd64:25722] 1 more process has sent help message help-mpi-runtime.txt / mpi_init:startup:internal-failure\n-## [profitbricks-build5-amd64:25722] 1 more process has sent help message help-mpi-errors.txt / mpi_errors_are_fatal unknown handle\n+## [i-capture-the-hostname:08792] [[44881,1],0] ORTE_ERROR_LOG: Unreachable in file dpm/dpm.c at line 433\n+## [i-capture-the-hostname:08900] [[44881,2],1] ORTE_ERROR_LOG: Unreachable in file dpm/dpm.c at line 433\n+## [i-capture-the-hostname:08792] *** An error occurred in MPI_Comm_spawn\n+## [i-capture-the-hostname:08792] *** reported by process [2941321217,0]\n+## [i-capture-the-hostname:08792] *** on communicator MPI_COMM_WORLD\n+## [i-capture-the-hostname:08792] *** MPI_ERR_INTERN: internal error\n+## [i-capture-the-hostname:08792] *** MPI_ERRORS_ARE_FATAL (processes in this communicator will now abort,\n+## [i-capture-the-hostname:08792] *** and potentially your MPI job)\n+## [i-capture-the-hostname:08894] *** An error occurred in MPI_Init\n+## [i-capture-the-hostname:08894] *** reported by process [2941321218,0]\n+## [i-capture-the-hostname:08894] *** on a NULL communicator\n+## [i-capture-the-hostname:08894] *** Unknown error\n+## [i-capture-the-hostname:08894] *** MPI_ERRORS_ARE_FATAL (processes in this communicator will now abort,\n+## [i-capture-the-hostname:08894] *** and potentially your MPI job)\n+## [i-capture-the-hostname:08704] 2 more processes have sent help message help-mca-bml-r2.txt / unreachable proc\n+## [i-capture-the-hostname:08704] Set MCA parameter \"orte_base_help_aggregate\" to 0 to see all help / error messages\n+## [i-capture-the-hostname:08704] 1 more process has sent help message help-mpi-runtime.txt / mpi_init:startup:internal-failure\n+## [i-capture-the-hostname:08704] 1 more process has sent help message help-mpi-errors.txt / mpi_errors_are_fatal unknown handle\n+## [warn] Epoll MOD(1) on fd 25 failed. Old events were 6; read change was 0 (none); write change was 2 (del); close change was 0 (none): Bad file descriptor\n not ok 838 - ./f90/spawn/spawnmultf03 1\n ---\n Directory: ./f90/spawn\n File: spawnmultf03\n Num-procs: 1\n- Date: \"Sun Jan 5 01:50:34 2020\"\n+ Date: \"Mon Dec 3 17:29:31 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## --------------------------------------------------------------------------\n ## At least one pair of MPI processes are unable to reach each other for\n ## MPI communications. This means that no Open MPI device has indicated\n ## that it can be used to communicate between these processes. This is\n ## an error; Open MPI requires that all MPI processes be able to reach\n ## each other. This error can sometimes be the result of forgetting to\n ## specify the \"self\" BTL.\n ## \n-## Process 1 ([[41220,1],0]) is on host: profitbricks-build5-amd64\n-## Process 2 ([[41220,2],0]) is on host: profitbricks-build5-amd64\n-## BTLs attempted: self\n+## Process 1 ([[44617,2],1]) is on host: i-capture-the-hostname\n+## Process 2 ([[44617,1],0]) is on host: i-capture-the-hostname\n+## BTLs attempted: vader self\n ## \n ## Your MPI job is now going to abort; sorry.\n ## --------------------------------------------------------------------------\n-## [profitbricks-build5-amd64:26121] [[41220,1],0] ORTE_ERROR_LOG: Unreachable in file dpm/dpm.c at line 433\n-## [profitbricks-build5-amd64:26121] *** An error occurred in MPI_Comm_spawn_multiple\n-## [profitbricks-build5-amd64:26121] *** reported by process [2701393921,0]\n-## [profitbricks-build5-amd64:26121] *** on communicator MPI_COMM_WORLD\n-## [profitbricks-build5-amd64:26121] *** MPI_ERR_INTERN: internal error\n-## [profitbricks-build5-amd64:26121] *** MPI_ERRORS_ARE_FATAL (processes in this communicator will now abort,\n-## [profitbricks-build5-amd64:26121] *** and potentially your MPI job)\n+## [i-capture-the-hostname:09053] [[44617,2],0] ORTE_ERROR_LOG: Unreachable in file dpm/dpm.c at line 433\n+## [i-capture-the-hostname:09054] [[44617,2],1] ORTE_ERROR_LOG: Unreachable in file dpm/dpm.c at line 433\n+## [i-capture-the-hostname:09025] [[44617,1],0] ORTE_ERROR_LOG: Unreachable in file dpm/dpm.c at line 433\n ## --------------------------------------------------------------------------\n ## It looks like MPI_INIT failed for some reason; your parallel process is\n ## likely to abort. There are many reasons that a parallel process can\n ## fail during MPI_INIT; some of which are due to configuration or environment\n ## problems. This failure appears to be an internal failure; here's some\n ## additional information (which may only be relevant to an Open MPI\n ## developer):\n ## \n ## ompi_dpm_dyn_init() failed\n ## --> Returned \"Unreachable\" (-12) instead of \"Success\" (0)\n ## --------------------------------------------------------------------------\n-## [profitbricks-build5-amd64:26209] [[41220,2],1] ORTE_ERROR_LOG: Unreachable in file dpm/dpm.c at line 433\n-## [profitbricks-build5-amd64:26203] [[41220,2],0] ORTE_ERROR_LOG: Unreachable in file dpm/dpm.c at line 433\n-## [profitbricks-build5-amd64:26209] *** An error occurred in MPI_Init\n-## [profitbricks-build5-amd64:26209] *** reported by process [2701393922,1]\n-## [profitbricks-build5-amd64:26209] *** on a NULL communicator\n-## [profitbricks-build5-amd64:26209] *** Unknown error\n-## [profitbricks-build5-amd64:26209] *** MPI_ERRORS_ARE_FATAL (processes in this communicator will now abort,\n-## [profitbricks-build5-amd64:26209] *** and potentially your MPI job)\n-## [profitbricks-build5-amd64:26070] 2 more processes have sent help message help-mca-bml-r2.txt / unreachable proc\n-## [profitbricks-build5-amd64:26070] Set MCA parameter \"orte_base_help_aggregate\" to 0 to see all help / error messages\n-## [profitbricks-build5-amd64:26070] 1 more process has sent help message help-mpi-runtime.txt / mpi_init:startup:internal-failure\n-## [profitbricks-build5-amd64:26070] 1 more process has sent help message help-mpi-errors.txt / mpi_errors_are_fatal unknown handle\n+## [i-capture-the-hostname:09025] *** An error occurred in MPI_Comm_spawn_multiple\n+## [i-capture-the-hostname:09025] *** reported by process [2924019713,0]\n+## [i-capture-the-hostname:09025] *** on communicator MPI_COMM_WORLD\n+## [i-capture-the-hostname:09025] *** MPI_ERR_INTERN: internal error\n+## [i-capture-the-hostname:09025] *** MPI_ERRORS_ARE_FATAL (processes in this communicator will now abort,\n+## [i-capture-the-hostname:09025] *** and potentially your MPI job)\n+## [i-capture-the-hostname:09053] *** An error occurred in MPI_Init\n+## [i-capture-the-hostname:09053] *** reported by process [2924019714,0]\n+## [i-capture-the-hostname:09053] *** on a NULL communicator\n+## [i-capture-the-hostname:09053] *** Unknown error\n+## [i-capture-the-hostname:09053] *** MPI_ERRORS_ARE_FATAL (processes in this communicator will now abort,\n+## [i-capture-the-hostname:09053] *** and potentially your MPI job)\n+## [i-capture-the-hostname:08984] 2 more processes have sent help message help-mca-bml-r2.txt / unreachable proc\n+## [i-capture-the-hostname:08984] Set MCA parameter \"orte_base_help_aggregate\" to 0 to see all help / error messages\n+## [i-capture-the-hostname:08984] 1 more process has sent help message help-mpi-runtime.txt / mpi_init:startup:internal-failure\n+## [i-capture-the-hostname:08984] 1 more process has sent help message help-mpi-errors.txt / mpi_errors_are_fatal unknown handle\n+## [warn] Epoll MOD(1) on fd 29 failed. Old events were 6; read change was 0 (none); write change was 2 (del); close change was 0 (none): Bad file descriptor\n ok 839 - ./f90/timer/wtimef90 1\n ok 840 - ./f90/topo/cartcrf90 2\n not ok 841 - ./f90/topo/dgraph_wgtf90 2\n ---\n Directory: ./f90/topo\n File: dgraph_wgtf90\n Num-procs: 2\n- Date: \"Sun Jan 5 01:50:39 2020\"\n+ Date: \"Mon Dec 3 17:29:39 2018\"\n ...\n ## Test output (expected 'No Errors'):\n-## Error: Cannot partition 16 elements in 17 parts\n+## Error: Cannot partition 15 elements in 2 parts\n ## \n ## Program received signal SIGSEGV: Segmentation fault - invalid memory reference.\n ## \n ## Backtrace for this error:\n-## #0 0x7f06aef7589d in ???\n-## #1 0x7f06aef74ae3 in ???\n-## #2 0x7f06aebaf8df in ???\n-## #3 0x7f06ac0ede8b in ???\n-## #4 0x7f06ac0ee422 in ???\n-## #5 0x7f06ac0ee7c1 in ???\n-## #6 0x7f06ac0eb28c in ???\n-## #7 0x7f06ac0e19e8 in ???\n-## #8 0x7f06af2313cc in ???\n-## #9 0x7f06af31cbca in ???\n-## #10 0x561cf1e2c9e6 in dgraph_unwgt\n+## #0 0x7fe32072889d in ???\n+## #1 0x7fe320727ae3 in ???\n+## #2 0x7fe3203628df in ???\n+## #3 0x7fe31cfa0e8b in ???\n+## #4 0x7fe31cfa1422 in ???\n+## #5 0x7fe31cfa17c1 in ???\n+## #6 0x7fe31cf9e28c in ???\n+## #7 0x7fe31cf949e8 in ???\n+## #8 0x7fe3209e43cc in ???\n+## #9 0x7fe320acfbca in ???\n+## #10 0x56497adc79e6 in dgraph_unwgt\n ## \tat ../../../f90/topo/dgraph_wgtf90.f90:155\n-## #11 0x561cf1e2c1fe in main\n+## #11 0x56497adc71fe in main\n ## \tat ../../../f90/topo/dgraph_wgtf90.f90:127\n ## --------------------------------------------------------------------------\n ## Primary job terminated normally, but 1 process returned\n ## a non-zero exit code. Per user-direction, the job has been aborted.\n ## --------------------------------------------------------------------------\n ## --------------------------------------------------------------------------\n-## mpiexec.openmpi noticed that process rank 0 with PID 0 on node profitbricks-build5-amd64 exited on signal 11 (Segmentation fault).\n+## mpiexec.openmpi noticed that process rank 0 with PID 0 on node i-capture-the-hostname exited on signal 11 (Segmentation fault).\n ## --------------------------------------------------------------------------\n not ok 842 - ./f90/topo/dgraph_unwgtf90 2\n ---\n Directory: ./f90/topo\n File: dgraph_unwgtf90\n Num-procs: 2\n- Date: \"Sun Jan 5 01:50:42 2020\"\n+ Date: \"Mon Dec 3 17:29:42 2018\"\n ...\n ## Test output (expected 'No Errors'):\n-## Error: Cannot partition 16 elements in 17 parts\n+## Error: Cannot partition 15 elements in 2 parts\n ## \n ## Program received signal SIGSEGV: Segmentation fault - invalid memory reference.\n ## \n ## Backtrace for this error:\n-## #0 0x7f961bc2589d in ???\n-## #1 0x7f961bc24ae3 in ???\n-## #2 0x7f961b85f8df in ???\n-## #3 0x7f961849de8b in ???\n-## #4 0x7f961849e422 in ???\n-## #5 0x7f961849e7c1 in ???\n-## #6 0x7f961849b28c in ???\n-## #7 0x7f96184919e8 in ???\n-## #8 0x7f961bee13cc in ???\n-## #9 0x7f961bfccbca in ???\n-## #10 0x55a829f7a8dd in dgraph_unwgt\n+## #0 0x7fc31c3b089d in ???\n+## #1 0x7fc31c3afae3 in ???\n+## #2 0x7fc31bfea8df in ???\n+## #3 0x7fc318c28e8b in ???\n+## #4 0x7fc318c29422 in ???\n+## #5 0x7fc318c297c1 in ???\n+## #6 0x7fc318c2628c in ???\n+## #7 0x7fc318c1c9e8 in ???\n+## #8 0x7fc31c66c3cc in ???\n+## #9 0x7fc31c757bca in ???\n+## #10 0x55b5eed118dd in dgraph_unwgt\n ## \tat ../../../f90/topo/dgraph_unwgtf90.f90:139\n-## #11 0x55a829f7a1fe in main\n+## #11 0x55b5eed111fe in main\n ## \tat ../../../f90/topo/dgraph_unwgtf90.f90:114\n ## --------------------------------------------------------------------------\n ## Primary job terminated normally, but 1 process returned\n ## a non-zero exit code. Per user-direction, the job has been aborted.\n ## --------------------------------------------------------------------------\n ## --------------------------------------------------------------------------\n-## mpiexec.openmpi noticed that process rank 0 with PID 0 on node profitbricks-build5-amd64 exited on signal 11 (Segmentation fault).\n+## mpiexec.openmpi noticed that process rank 0 with PID 0 on node i-capture-the-hostname exited on signal 11 (Segmentation fault).\n ## --------------------------------------------------------------------------\n not ok 843 - ./f90/profile/profile1f90 \n ---\n Directory: ./f90/profile\n File: profile1f90\n Num-procs: \n- Date: \"Sun Jan 5 01:50:43 2020\"\n+ Date: \"Mon Dec 3 17:29:46 2018\"\n ...\n ## Test output (expected 'No Errors'):\n-## Failed to build profile1f90; make[2]: Entering directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/f90/profile'\n-## /usr/bin/mpif90.openmpi -g -O2 -ffile-prefix-map=/build/1st/mpi-testsuite-3.2+dfsg=. -fstack-protector-strong -fcray-pointer -c -o profile1f90.o ../../../f90/profile/profile1f90.f90\n+## Failed to build profile1f90; make[2]: Entering directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/f90/profile'\n+## /usr/bin/mpif90.openmpi -g -O2 -ffile-prefix-map=/build/mpi-testsuite-3.2+dfsg/2nd=. -fstack-protector-strong -fcray-pointer -c -o profile1f90.o ../../../f90/profile/profile1f90.f90\n ## ../../../f90/profile/profile1f90.f90:68:11:\n ## \n ## use mpi\n ## 1\n ## Error: 'mpi_send' of module 'mpi', imported at (1), is also the name of the current program unit\n ## ../../../f90/profile/profile1f90.f90:68:11:\n ## \n@@ -1004678,59 +1004794,59 @@\n ## Error: 'mpi_recv' of module 'mpi', imported at (1), is also the name of the current program unit\n ## ../../../f90/profile/profile1f90.f90:82:11:\n ## \n ## use mpi\n ## 1\n ## Error: Name 'mpi_recv' at (1) is an ambiguous reference to 'mpi_recv' from current program unit\n ## make[2]: *** [Makefile:434: profile1f90.o] Error 1\n-## make[2]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/f90/profile'\n+## make[2]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/f90/profile'\n not ok 844 - ./errors/attr/keyvalmis 1\n ---\n Directory: ./errors/attr\n File: keyvalmis\n Num-procs: 1\n- Date: \"Sun Jan 5 01:50:45 2020\"\n+ Date: \"Mon Dec 3 17:29:49 2018\"\n ...\n ## Test output (expected 'No Errors'):\n-## [profitbricks-build5-amd64:27846] *** An error occurred in MPI_Win_set_attr\n-## [profitbricks-build5-amd64:27846] *** reported by process [2829385729,0]\n-## [profitbricks-build5-amd64:27846] *** on win pt2pt window 3\n-## [profitbricks-build5-amd64:27846] *** MPI_ERR_OTHER: known error not in list\n-## [profitbricks-build5-amd64:27846] *** MPI_ERRORS_ARE_FATAL (processes in this win will now abort,\n-## [profitbricks-build5-amd64:27846] *** and potentially your MPI job)\n+## [i-capture-the-hostname:09931] *** An error occurred in MPI_Win_set_attr\n+## [i-capture-the-hostname:09931] *** reported by process [2885484545,0]\n+## [i-capture-the-hostname:09931] *** on win pt2pt window 3\n+## [i-capture-the-hostname:09931] *** MPI_ERR_OTHER: known error not in list\n+## [i-capture-the-hostname:09931] *** MPI_ERRORS_ARE_FATAL (processes in this win will now abort,\n+## [i-capture-the-hostname:09931] *** and potentially your MPI job)\n not ok 845 - ./errors/coll/noalias 2\n ---\n Directory: ./errors/coll\n File: noalias\n Num-procs: 2\n- Date: \"Sun Jan 5 01:50:47 2020\"\n+ Date: \"Mon Dec 3 17:29:52 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Did not detect aliased arguments in MPI_Allreduce\n ## Did not detect aliased arguments in MPI_Gather\n ## Did not detect aliased arguments in MPI_Scatter\n ## Found 4 errors\n not ok 846 - ./errors/coll/nb_noalias 2\n ---\n Directory: ./errors/coll\n File: nb_noalias\n Num-procs: 2\n- Date: \"Sun Jan 5 01:50:48 2020\"\n+ Date: \"Mon Dec 3 17:29:54 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Did not detect aliased arguments in MPI_Allreduce\n ## Did not detect aliased arguments in MPI_Gather\n ## Did not detect aliased arguments in MPI_Scatter\n ## Found 4 errors\n not ok 847 - ./errors/coll/noalias2 2\n ---\n Directory: ./errors/coll\n File: noalias2\n Num-procs: 2\n- Date: \"Sun Jan 5 01:53:49 2020\"\n+ Date: \"Mon Dec 3 17:32:56 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## --------------------------------------------------------------------------\n ## The user-provided time limit for job execution has been reached:\n ## \n ## Timeout: 180 seconds\n ## \n@@ -1004739,337 +1004855,346 @@\n ## command line option or MPIEXEC_TIMEOUT environment variable).\n ## --------------------------------------------------------------------------\n not ok 848 - ./errors/coll/noalias3 2\n ---\n Directory: ./errors/coll\n File: noalias3\n Num-procs: 2\n- Date: \"Sun Jan 5 01:53:52 2020\"\n+ Date: \"Mon Dec 3 17:33:00 2018\"\n ...\n ## Test output (expected 'No Errors'):\n-## [profitbricks-build5-amd64:10279] *** Process received signal ***\n-## [profitbricks-build5-amd64:10279] Signal: Segmentation fault (11)\n-## [profitbricks-build5-amd64:10279] Signal code: (128)\n-## [profitbricks-build5-amd64:10279] Failing at address: (nil)\n-## [profitbricks-build5-amd64:10279] [ 0] /lib/x86_64-linux-gnu/libpthread.so.0(+0x126b0)[0x7f4bf959a6b0]\n-## [profitbricks-build5-amd64:10279] [ 1] /usr/lib/x86_64-linux-gnu/libmpi.so.40(PMPI_Ialltoallw+0x117)[0x7f4bf9613287]\n-## [profitbricks-build5-amd64:10279] [ 2] ./noalias3(+0x388f)[0x55d8cc02688f]\n-## [profitbricks-build5-amd64:10279] [ 3] /lib/x86_64-linux-gnu/libc.so.6(__libc_start_main+0xeb)[0x7f4bf93e409b]\n-## [profitbricks-build5-amd64:10279] [ 4] ./noalias3(+0x3d9a)[0x55d8cc026d9a]\n-## [profitbricks-build5-amd64:10279] *** End of error message ***\n+## [i-capture-the-hostname:16299] *** Process received signal ***\n+## [i-capture-the-hostname:16299] Signal: Segmentation fault (11)\n+## [i-capture-the-hostname:16299] Signal code: (128)\n+## [i-capture-the-hostname:16299] Failing at address: (nil)\n+## [i-capture-the-hostname:16298] *** Process received signal ***\n+## [i-capture-the-hostname:16298] Signal: Segmentation fault (11)\n+## [i-capture-the-hostname:16298] Signal code: (128)\n+## [i-capture-the-hostname:16298] Failing at address: (nil)\n+## [i-capture-the-hostname:16298] [ 0] /lib/x86_64-linux-gnu/libpthread.so.0(+0x126b0)[0x7f9c822f56b0]\n+## [i-capture-the-hostname:16298] [ 1] /usr/lib/x86_64-linux-gnu/libmpi.so.40(PMPI_Ialltoallw+0x117)[0x7f9c8236e287]\n+## [i-capture-the-hostname:16298] [ 2] ./noalias3(+0x388f)[0x5616c4ad488f]\n+## [i-capture-the-hostname:16298] [ 3] /lib/x86_64-linux-gnu/libc.so.6(__libc_start_main+0xeb)[0x7f9c8213f09b]\n+## [i-capture-the-hostname:16298] [ 4] ./noalias3(+0x3d9a)[0x5616c4ad4d9a]\n+## [i-capture-the-hostname:16298] *** End of error message ***\n+## [i-capture-the-hostname:16299] [ 0] /lib/x86_64-linux-gnu/libpthread.so.0(+0x126b0)[0x7f18c167d6b0]\n+## [i-capture-the-hostname:16299] [ 1] /usr/lib/x86_64-linux-gnu/libmpi.so.40(PMPI_Ialltoallw+0x117)[0x7f18c16f6287]\n+## [i-capture-the-hostname:16299] [ 2] ./noalias3(+0x388f)[0x55840866d88f]\n+## [i-capture-the-hostname:16299] [ 3] /lib/x86_64-linux-gnu/libc.so.6(__libc_start_main+0xeb)[0x7f18c14c709b]\n+## [i-capture-the-hostname:16299] [ 4] ./noalias3(+0x3d9a)[0x55840866dd9a]\n+## [i-capture-the-hostname:16299] *** End of error message ***\n ## --------------------------------------------------------------------------\n ## Primary job terminated normally, but 1 process returned\n ## a non-zero exit code. Per user-direction, the job has been aborted.\n ## --------------------------------------------------------------------------\n-## [profitbricks-build5-amd64:10300] *** Process received signal ***\n-## [profitbricks-build5-amd64:10300] Signal: Segmentation fault (11)\n-## [profitbricks-build5-amd64:10300] Signal code: (128)\n-## [profitbricks-build5-amd64:10300] Failing at address: (nil)\n-## [profitbricks-build5-amd64:10300] [ 0] /lib/x86_64-linux-gnu/libpthread.so.0(+0x126b0)[0x7f7eed4ba6b0]\n-## [profitbricks-build5-amd64:10300] [ 1] /usr/lib/x86_64-linux-gnu/libmpi.so.40(PMPI_Ialltoallw+0x117)[0x7f7eed533287]\n-## [profitbricks-build5-amd64:10300] [ 2] ./noalias3(+0x388f)[0x562e132d888f]\n-## [profitbricks-build5-amd64:10300] [ 3] /lib/x86_64-linux-gnu/libc.so.6(__libc_start_main+0xeb)[0x7f7eed30409b]\n-## [profitbricks-build5-amd64:10300] [ 4] ./noalias3(+0x3d9a)[0x562e132d8d9a]\n-## [profitbricks-build5-amd64:10300] *** End of error message ***\n ## --------------------------------------------------------------------------\n-## mpiexec.openmpi noticed that process rank 0 with PID 0 on node profitbricks-build5-amd64 exited on signal 11 (Segmentation fault).\n+## mpiexec.openmpi noticed that process rank 1 with PID 0 on node i-capture-the-hostname exited on signal 11 (Segmentation fault).\n ## --------------------------------------------------------------------------\n not ok 849 - ./errors/coll/rerr 2\n ---\n Directory: ./errors/coll\n File: rerr\n Num-procs: 2\n- Date: \"Sun Jan 5 01:53:54 2020\"\n+ Date: \"Mon Dec 3 17:33:02 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Did not detect invalid type/op pair (byte,max) in Allreduce\n-## Found 2 errors\n ## Did not detect invalid type/op pair (byte,max) in Allreduce\n+## Found 2 errors\n not ok 850 - ./errors/coll/nb_rerr 2\n ---\n Directory: ./errors/coll\n File: nb_rerr\n Num-procs: 2\n- Date: \"Sun Jan 5 01:53:55 2020\"\n+ Date: \"Mon Dec 3 17:33:04 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Did not detect invalid type/op pair (byte,max) in Allreduce\n ## Did not detect invalid type/op pair (byte,max) in Allreduce\n ## Found 2 errors\n not ok 851 - ./errors/coll/reduce_local 1\n ---\n Directory: ./errors/coll\n File: reduce_local\n Num-procs: 1\n- Date: \"Sun Jan 5 01:53:58 2020\"\n+ Date: \"Mon Dec 3 17:33:08 2018\"\n ...\n ## Test output (expected 'No Errors'):\n-## [profitbricks-build5-amd64:17027] *** Process received signal ***\n-## [profitbricks-build5-amd64:17027] Signal: Segmentation fault (11)\n-## [profitbricks-build5-amd64:17027] Signal code: Address not mapped (1)\n-## [profitbricks-build5-amd64:17027] Failing at address: 0x1\n-## [profitbricks-build5-amd64:17027] [ 0] /lib/x86_64-linux-gnu/libpthread.so.0(+0x126b0)[0x7f06717fa6b0]\n-## [profitbricks-build5-amd64:17027] [ 1] /usr/lib/x86_64-linux-gnu/libmpi.so.40(+0xaf2c0)[0x7f06718bf2c0]\n-## [profitbricks-build5-amd64:17027] [ 2] /usr/lib/x86_64-linux-gnu/libmpi.so.40(mca_coll_base_reduce_local+0xba)[0x7f06718b5bda]\n-## [profitbricks-build5-amd64:17027] [ 3] /usr/lib/x86_64-linux-gnu/libmpi.so.40(MPI_Reduce_local+0x131)[0x7f0671893291]\n-## [profitbricks-build5-amd64:17027] [ 4] ./reduce_local(+0x25ef)[0x5611902995ef]\n-## [profitbricks-build5-amd64:17027] [ 5] /lib/x86_64-linux-gnu/libc.so.6(__libc_start_main+0xeb)[0x7f067164409b]\n-## [profitbricks-build5-amd64:17027] [ 6] ./reduce_local(+0x27ca)[0x5611902997ca]\n-## [profitbricks-build5-amd64:17027] *** End of error message ***\n+## [i-capture-the-hostname:16439] *** Process received signal ***\n+## [i-capture-the-hostname:16439] Signal: Segmentation fault (11)\n+## [i-capture-the-hostname:16439] Signal code: Address not mapped (1)\n+## [i-capture-the-hostname:16439] Failing at address: 0x1\n+## [i-capture-the-hostname:16439] [ 0] /lib/x86_64-linux-gnu/libpthread.so.0(+0x126b0)[0x7f52c1ce56b0]\n+## [i-capture-the-hostname:16439] [ 1] /usr/lib/x86_64-linux-gnu/libmpi.so.40(+0xaf2c0)[0x7f52c1daa2c0]\n+## [i-capture-the-hostname:16439] [ 2] /usr/lib/x86_64-linux-gnu/libmpi.so.40(mca_coll_base_reduce_local+0xba)[0x7f52c1da0bda]\n+## [i-capture-the-hostname:16439] [ 3] /usr/lib/x86_64-linux-gnu/libmpi.so.40(MPI_Reduce_local+0x131)[0x7f52c1d7e291]\n+## [i-capture-the-hostname:16439] [ 4] ./reduce_local(+0x25ef)[0x55d790b705ef]\n+## [i-capture-the-hostname:16439] [ 5] /lib/x86_64-linux-gnu/libc.so.6(__libc_start_main+0xeb)[0x7f52c1b2f09b]\n+## [i-capture-the-hostname:16439] [ 6] ./reduce_local(+0x27ca)[0x55d790b707ca]\n+## [i-capture-the-hostname:16439] *** End of error message ***\n ## --------------------------------------------------------------------------\n ## Primary job terminated normally, but 1 process returned\n ## a non-zero exit code. Per user-direction, the job has been aborted.\n ## --------------------------------------------------------------------------\n ## --------------------------------------------------------------------------\n-## mpiexec.openmpi noticed that process rank 0 with PID 0 on node profitbricks-build5-amd64 exited on signal 11 (Segmentation fault).\n+## mpiexec.openmpi noticed that process rank 0 with PID 0 on node i-capture-the-hostname exited on signal 11 (Segmentation fault).\n ## --------------------------------------------------------------------------\n not ok 852 - ./errors/coll/bcastlength 2\n ---\n Directory: ./errors/coll\n File: bcastlength\n Num-procs: 2\n- Date: \"Sun Jan 5 01:53:59 2020\"\n+ Date: \"Mon Dec 3 17:33:11 2018\"\n ...\n ## Test output (expected 'No Errors'):\n-## Did not detect mismatched length (long) on process 1\n ## Found 1 errors\n+## Did not detect mismatched length (long) on process 1\n not ok 853 - ./errors/coll/ibcastlength 2\n ---\n Directory: ./errors/coll\n File: ibcastlength\n Num-procs: 2\n- Date: \"Sun Jan 5 01:54:00 2020\"\n+ Date: \"Mon Dec 3 17:33:13 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Did not detect mismatched length (long) on process 1\n-## MPI Error in MPI_Testall() (req 0 = 15)\n ## Found 1 errors\n+## MPI Error in MPI_Testall() (req 0 = 15)\n not ok 854 - ./errors/comm/cfree 2\n ---\n Directory: ./errors/comm\n File: cfree\n Num-procs: 2\n- Date: \"Sun Jan 5 01:54:04 2020\"\n+ Date: \"Mon Dec 3 17:33:17 2018\"\n ...\n ## Test output (expected 'No Errors'):\n-## [profitbricks-build5-amd64:31465] *** Process received signal ***\n-## [profitbricks-build5-amd64:31465] Signal: Segmentation fault (11)\n-## [profitbricks-build5-amd64:31465] Signal code: Address not mapped (1)\n-## [profitbricks-build5-amd64:31465] Failing at address: 0x10\n-## [profitbricks-build5-amd64:31488] *** Process received signal ***\n-## [profitbricks-build5-amd64:31488] Signal: Segmentation fault (11)\n-## [profitbricks-build5-amd64:31488] Signal code: Address not mapped (1)\n-## [profitbricks-build5-amd64:31488] Failing at address: 0x10\n-## [profitbricks-build5-amd64:31488] [ 0] /lib/x86_64-linux-gnu/libpthread.so.0(+0x126b0)[0x7fad54f426b0]\n-## [profitbricks-build5-amd64:31488] [ 1] /usr/lib/x86_64-linux-gnu/libmpi.so.40(MPI_Barrier+0x91)[0x7fad54fbb961]\n-## [profitbricks-build5-amd64:31488] [ 2] ./cfree(+0x255e)[0x564800fc455e]\n-## [profitbricks-build5-amd64:31488] [ 3] /lib/x86_64-linux-gnu/libc.so.6(__libc_start_main+0xeb)[0x7fad54d8c09b]\n-## [profitbricks-build5-amd64:31488] [ 4] ./cfree(+0x261a)[0x564800fc461a]\n-## [profitbricks-build5-amd64:31488] *** End of error message ***\n-## [profitbricks-build5-amd64:31465] [ 0] /lib/x86_64-linux-gnu/libpthread.so.0(+0x126b0)[0x7f0bd9eba6b0]\n-## [profitbricks-build5-amd64:31465] [ 1] /usr/lib/x86_64-linux-gnu/libmpi.so.40(MPI_Barrier+0x91)[0x7f0bd9f33961]\n-## [profitbricks-build5-amd64:31465] [ 2] ./cfree(+0x255e)[0x5623ec9e255e]\n-## [profitbricks-build5-amd64:31465] [ 3] /lib/x86_64-linux-gnu/libc.so.6(__libc_start_main+0xeb)[0x7f0bd9d0409b]\n-## [profitbricks-build5-amd64:31465] [ 4] ./cfree(+0x261a)[0x5623ec9e261a]\n-## [profitbricks-build5-amd64:31465] *** End of error message ***\n+## [i-capture-the-hostname:16661] *** Process received signal ***\n+## [i-capture-the-hostname:16661] Signal: Segmentation fault (11)\n+## [i-capture-the-hostname:16661] Signal code: Address not mapped (1)\n+## [i-capture-the-hostname:16661] Failing at address: 0x10\n+## [i-capture-the-hostname:16661] [ 0] /lib/x86_64-linux-gnu/libpthread.so.0(+0x126b0)[0x7f6612cd56b0]\n+## [i-capture-the-hostname:16661] [ 1] /usr/lib/x86_64-linux-gnu/libmpi.so.40(MPI_Barrier+0x91)[0x7f6612d4e961]\n+## [i-capture-the-hostname:16661] [ 2] ./cfree(+0x255e)[0x55cdbcc4d55e]\n+## [i-capture-the-hostname:16661] [ 3] /lib/x86_64-linux-gnu/libc.so.6(__libc_start_main+0xeb)[0x7f6612b1f09b]\n+## [i-capture-the-hostname:16661] [ 4] ./cfree(+0x261a)[0x55cdbcc4d61a]\n+## [i-capture-the-hostname:16661] *** End of error message ***\n+## [i-capture-the-hostname:16662] *** Process received signal ***\n+## [i-capture-the-hostname:16662] Signal: Segmentation fault (11)\n+## [i-capture-the-hostname:16662] Signal code: Address not mapped (1)\n+## [i-capture-the-hostname:16662] Failing at address: 0x10\n ## --------------------------------------------------------------------------\n ## Primary job terminated normally, but 1 process returned\n ## a non-zero exit code. Per user-direction, the job has been aborted.\n ## --------------------------------------------------------------------------\n+## [i-capture-the-hostname:16662] [ 0] /lib/x86_64-linux-gnu/libpthread.so.0(+0x126b0)[0x7fb4186656b0]\n+## [i-capture-the-hostname:16662] [ 1] /usr/lib/x86_64-linux-gnu/libmpi.so.40(MPI_Barrier+0x91)[0x7fb4186de961]\n+## [i-capture-the-hostname:16662] [ 2] ./cfree(+0x255e)[0x55a7f6cbf55e]\n+## [i-capture-the-hostname:16662] [ 3] /lib/x86_64-linux-gnu/libc.so.6(__libc_start_main+0xeb)[0x7fb4184af09b]\n+## [i-capture-the-hostname:16662] [ 4] ./cfree(+0x261a)[0x55a7f6cbf61a]\n+## [i-capture-the-hostname:16662] *** End of error message ***\n ## --------------------------------------------------------------------------\n-## mpiexec.openmpi noticed that process rank 0 with PID 0 on node profitbricks-build5-amd64 exited on signal 11 (Segmentation fault).\n+## mpiexec.openmpi noticed that process rank 0 with PID 0 on node i-capture-the-hostname exited on signal 11 (Segmentation fault).\n ## --------------------------------------------------------------------------\n not ok 855 - ./errors/comm/ccreate1 2\n ---\n Directory: ./errors/comm\n File: ccreate1\n Num-procs: 2\n- Date: \"Sun Jan 5 01:54:05 2020\"\n+ Date: \"Mon Dec 3 17:33:19 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Found 1 errors\n ## Did not detect group of high ranks in low comm\n not ok 856 - ./errors/comm/userdup 2\n ---\n Directory: ./errors/comm\n File: userdup\n Num-procs: 2\n- Date: \"Sun Jan 5 01:54:06 2020\"\n+ Date: \"Mon Dec 3 17:33:21 2018\"\n ...\n ## Test output (expected 'No Errors'):\n-## Found 2 errors\n ## Unexpected error class = 14, expected user-defined class 93\n ## Unexpected error class = 14, expected user-defined class 93\n+## Found 2 errors\n ok 857 - ./errors/comm/manysplit 2\n ok 878 - ./errors/comm/too_many_comms 2 # SKIP non-strict test, strict MPI mode requested\n ok 879 - ./errors/comm/too_many_icomms 2 # SKIP non-strict test, strict MPI mode requested\n ok 880 - ./errors/comm/too_many_comms2 2 # SKIP non-strict test, strict MPI mode requested\n ok 881 - ./errors/comm/too_many_comms3 2 # SKIP non-strict test, strict MPI mode requested\n ok 882 - ./errors/comm/too_many_icomms2 2 # SKIP non-strict test, strict MPI mode requested\n-ok 858 - ./errors/group/gerr 1\n+not ok 858 - ./errors/group/gerr 1\n+ ---\n+ Directory: ./errors/group\n+ File: gerr\n+ Num-procs: 1\n+ Date: \"Mon Dec 3 17:33:28 2018\"\n+ ...\n+## Test output (expected 'No Errors'):\n+## Did not detect invalid handle (comm) in group_incl\n+## Found 1 errors\n ok 859 - ./errors/pt2pt/proberank 1\n ok 860 - ./errors/pt2pt/truncmsg1 2\n not ok 861 - ./errors/pt2pt/truncmsg2 2\n ---\n Directory: ./errors/pt2pt\n File: truncmsg2\n Num-procs: 2\n- Date: \"Sun Jan 5 01:54:14 2020\"\n+ Date: \"Mon Dec 3 17:33:37 2018\"\n ...\n ## Test output (expected 'No Errors'):\n-## [profitbricks-build5-amd64:60615] *** Process received signal ***\n-## [profitbricks-build5-amd64:60615] Signal: Segmentation fault (11)\n-## [profitbricks-build5-amd64:60615] Signal code: Address not mapped (1)\n-## [profitbricks-build5-amd64:60615] Failing at address: 0x3017ea34\n-## [profitbricks-build5-amd64:60615] [ 0] /lib/x86_64-linux-gnu/libpthread.so.0(+0x126b0)[0x7fb68d1fa6b0]\n-## [profitbricks-build5-amd64:60615] [ 1] /usr/lib/x86_64-linux-gnu/libmpi.so.40(ompi_datatype_create_indexed+0x1f)[0x7fb68d26b15f]\n-## [profitbricks-build5-amd64:60615] [ 2] /usr/lib/x86_64-linux-gnu/libmpi.so.40(MPI_Type_indexed+0xd2)[0x7fb68d29d592]\n-## [profitbricks-build5-amd64:60615] [ 3] ./truncmsg2(+0x264d)[0x55a13017364d]\n-## [profitbricks-build5-amd64:60615] [ 4] /lib/x86_64-linux-gnu/libc.so.6(__libc_start_main+0xeb)[0x7fb68d04409b]\n-## [profitbricks-build5-amd64:60615] [ 5] ./truncmsg2(+0x28ba)[0x55a1301738ba]\n-## [profitbricks-build5-amd64:60615] *** End of error message ***\n-## [profitbricks-build5-amd64:60635] *** Process received signal ***\n-## [profitbricks-build5-amd64:60635] Signal: Segmentation fault (11)\n-## [profitbricks-build5-amd64:60635] Signal code: Address not mapped (1)\n-## [profitbricks-build5-amd64:60635] Failing at address: 0xffffffffeba84a34\n-## [profitbricks-build5-amd64:60635] [ 0] /lib/x86_64-linux-gnu/libpthread.so.0(+0x126b0)[0x7f567a86a6b0]\n-## [profitbricks-build5-amd64:60635] [ 1] /usr/lib/x86_64-linux-gnu/libmpi.so.40(ompi_datatype_create_indexed+0x1f)[0x7f567a8db15f]\n-## [profitbricks-build5-amd64:60635] [ 2] /usr/lib/x86_64-linux-gnu/libmpi.so.40(MPI_Type_indexed+0xd2)[0x7f567a90d592]\n-## [profitbricks-build5-amd64:60635] [ 3] ./truncmsg2(+0x264d)[0x561beba7964d]\n-## [profitbricks-build5-amd64:60635] [ 4] /lib/x86_64-linux-gnu/libc.so.6(__libc_start_main+0xeb)[0x7f567a6b409b]\n-## [profitbricks-build5-amd64:60635] [ 5] ./truncmsg2(+0x28ba)[0x561beba798ba]\n-## [profitbricks-build5-amd64:60635] *** End of error message ***\n+## [i-capture-the-hostname:17676] *** Process received signal ***\n+## [i-capture-the-hostname:17676] Signal: Segmentation fault (11)\n+## [i-capture-the-hostname:17676] Signal code: Address not mapped (1)\n+## [i-capture-the-hostname:17676] Failing at address: 0xffffffff907a5a34\n+## [i-capture-the-hostname:17676] [ 0] /lib/x86_64-linux-gnu/libpthread.so.0(+0x126b0)[0x7f0cac9356b0]\n+## [i-capture-the-hostname:17676] [ 1] /usr/lib/x86_64-linux-gnu/libmpi.so.40(ompi_datatype_create_indexed+0x1f)[0x7f0cac9a615f]\n+## [i-capture-the-hostname:17676] [ 2] /usr/lib/x86_64-linux-gnu/libmpi.so.40(MPI_Type_indexed+0xd2)[0x7f0cac9d8592]\n+## [i-capture-the-hostname:17676] [ 3] ./truncmsg2(+0x264d)[0x55c09079a64d]\n+## [i-capture-the-hostname:17676] [ 4] /lib/x86_64-linux-gnu/libc.so.6(__libc_start_main+0xeb)[0x7f0cac77f09b]\n+## [i-capture-the-hostname:17676] [ 5] ./truncmsg2(+0x28ba)[0x55c09079a8ba]\n+## [i-capture-the-hostname:17676] *** End of error message ***\n+## [i-capture-the-hostname:17679] *** Process received signal ***\n+## [i-capture-the-hostname:17679] Signal: Segmentation fault (11)\n+## [i-capture-the-hostname:17679] Signal code: Address not mapped (1)\n+## [i-capture-the-hostname:17679] Failing at address: 0x3fcf1a34\n+## [i-capture-the-hostname:17679] [ 0] /lib/x86_64-linux-gnu/libpthread.so.0(+0x126b0)[0x7f8bec2b56b0]\n+## [i-capture-the-hostname:17679] [ 1] /usr/lib/x86_64-linux-gnu/libmpi.so.40(ompi_datatype_create_indexed+0x1f)[0x7f8bec32615f]\n+## [i-capture-the-hostname:17679] [ 2] /usr/lib/x86_64-linux-gnu/libmpi.so.40(MPI_Type_indexed+0xd2)[0x7f8bec358592]\n+## [i-capture-the-hostname:17679] [ 3] ./truncmsg2(+0x264d)[0x555b3fce664d]\n+## [i-capture-the-hostname:17679] [ 4] /lib/x86_64-linux-gnu/libc.so.6(__libc_start_main+0xeb)[0x7f8bec0ff09b]\n+## [i-capture-the-hostname:17679] [ 5] ./truncmsg2(+0x28ba)[0x555b3fce68ba]\n+## [i-capture-the-hostname:17679] *** End of error message ***\n ## --------------------------------------------------------------------------\n ## Primary job terminated normally, but 1 process returned\n ## a non-zero exit code. Per user-direction, the job has been aborted.\n ## --------------------------------------------------------------------------\n ## --------------------------------------------------------------------------\n-## mpiexec.openmpi noticed that process rank 0 with PID 0 on node profitbricks-build5-amd64 exited on signal 11 (Segmentation fault).\n+## mpiexec.openmpi noticed that process rank 0 with PID 0 on node i-capture-the-hostname exited on signal 11 (Segmentation fault).\n ## --------------------------------------------------------------------------\n ok 862 - ./errors/pt2pt/errinstatts 2\n ok 863 - ./errors/pt2pt/errinstatta 2\n ok 864 - ./errors/pt2pt/errinstatws 2\n ok 865 - ./errors/pt2pt/errinstatwa 2\n not ok 866 - ./errors/topo/cartsmall 2\n ---\n Directory: ./errors/topo\n File: cartsmall\n Num-procs: 2\n- Date: \"Sun Jan 5 01:54:19 2020\"\n+ Date: \"Mon Dec 3 17:33:48 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Expected a null comm from cart create\n ## Expected a null comm from cart create\n ## Found 2 errors\n ok 867 - ./errors/rma/winerr 2\n ok 868 - ./errors/rma/winerr2 2\n not ok 869 - ./errors/rma/cas_type_check 2\n ---\n Directory: ./errors/rma\n File: cas_type_check\n Num-procs: 2\n- Date: \"Sun Jan 5 01:54:30 2020\"\n+ Date: \"Mon Dec 3 17:34:29 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## cas_type_check: ../../../errors/rma/cas_type_check.c:73: main: Assertion `err_class == 3' failed.\n-## [profitbricks-build5-amd64:23319] *** Process received signal ***\n-## [profitbricks-build5-amd64:23319] Signal: Aborted (6)\n-## [profitbricks-build5-amd64:23319] Signal code: (-6)\n-## [profitbricks-build5-amd64:23319] [ 0] /lib/x86_64-linux-gnu/libpthread.so.0(+0x126b0)[0x7f7e950426b0]\n-## [profitbricks-build5-amd64:23319] [ 1] /lib/x86_64-linux-gnu/libc.so.6(gsignal+0x10b)[0x7f7e94e9f85b]\n-## [profitbricks-build5-amd64:23319] [ 2] /lib/x86_64-linux-gnu/libc.so.6(abort+0x121)[0x7f7e94e8a535]\n-## [profitbricks-build5-amd64:23319] [ 3] /lib/x86_64-linux-gnu/libc.so.6(+0x2240f)[0x7f7e94e8a40f]\n-## [profitbricks-build5-amd64:23319] [ 4] /lib/x86_64-linux-gnu/libc.so.6(+0x300a2)[0x7f7e94e980a2]\n-## [profitbricks-build5-amd64:23319] [ 5] ./cas_type_check(+0x4722)[0x55d869d03722]\n-## [profitbricks-build5-amd64:23319] [ 6] /lib/x86_64-linux-gnu/libc.so.6(__libc_start_main+0xeb)[0x7f7e94e8c09b]\n-## [profitbricks-build5-amd64:23319] [ 7] ./cas_type_check(+0x47fa)[0x55d869d037fa]\n-## [profitbricks-build5-amd64:23319] *** End of error message ***\n+## [i-capture-the-hostname:18545] *** Process received signal ***\n+## [i-capture-the-hostname:18545] Signal: Aborted (6)\n+## [i-capture-the-hostname:18545] Signal code: (-6)\n ## cas_type_check: ../../../errors/rma/cas_type_check.c:73: main: Assertion `err_class == 3' failed.\n-## [profitbricks-build5-amd64:23305] *** Process received signal ***\n-## [profitbricks-build5-amd64:23305] Signal: Aborted (6)\n-## [profitbricks-build5-amd64:23305] Signal code: (-6)\n-## [profitbricks-build5-amd64:23305] [ 0] /lib/x86_64-linux-gnu/libpthread.so.0(+0x126b0)[0x7f5f36bc26b0]\n-## [profitbricks-build5-amd64:23305] [ 1] /lib/x86_64-linux-gnu/libc.so.6(gsignal+0x10b)[0x7f5f36a1f85b]\n-## [profitbricks-build5-amd64:23305] [ 2] /lib/x86_64-linux-gnu/libc.so.6(abort+0x121)[0x7f5f36a0a535]\n-## [profitbricks-build5-amd64:23305] [ 3] /lib/x86_64-linux-gnu/libc.so.6(+0x2240f)[0x7f5f36a0a40f]\n-## [profitbricks-build5-amd64:23305] [ 4] /lib/x86_64-linux-gnu/libc.so.6(+0x300a2)[0x7f5f36a180a2]\n-## [profitbricks-build5-amd64:23305] [ 5] ./cas_type_check(+0x4722)[0x5557f00f0722]\n-## [profitbricks-build5-amd64:23305] [ 6] /lib/x86_64-linux-gnu/libc.so.6(__libc_start_main+0xeb)[0x7f5f36a0c09b]\n-## [profitbricks-build5-amd64:23305] [ 7] ./cas_type_check(+0x47fa)[0x5557f00f07fa]\n-## [profitbricks-build5-amd64:23305] *** End of error message ***\n+## [i-capture-the-hostname:18544] *** Process received signal ***\n+## [i-capture-the-hostname:18544] Signal: Aborted (6)\n+## [i-capture-the-hostname:18544] Signal code: (-6)\n+## [i-capture-the-hostname:18545] [ 0] /lib/x86_64-linux-gnu/libpthread.so.0(+0x126b0)[0x7ff4f84856b0]\n+## [i-capture-the-hostname:18545] [ 1] /lib/x86_64-linux-gnu/libc.so.6(gsignal+0x10b)[0x7ff4f82e285b]\n+## [i-capture-the-hostname:18545] [ 2] /lib/x86_64-linux-gnu/libc.so.6(abort+0x121)[0x7ff4f82cd535]\n+## [i-capture-the-hostname:18545] [ 3] /lib/x86_64-linux-gnu/libc.so.6(+0x2240f)[0x7ff4f82cd40f]\n+## [i-capture-the-hostname:18545] [ 4] /lib/x86_64-linux-gnu/libc.so.6(+0x300a2)[0x7ff4f82db0a2]\n+## [i-capture-the-hostname:18545] [ 5] ./cas_type_check(+0x4722)[0x5579237e3722]\n+## [i-capture-the-hostname:18545] [ 6] /lib/x86_64-linux-gnu/libc.so.6(__libc_start_main+0xeb)[0x7ff4f82cf09b]\n+## [i-capture-the-hostname:18545] [ 7] ./cas_type_check(+0x47fa)[0x5579237e37fa]\n+## [i-capture-the-hostname:18545] *** End of error message ***\n+## [i-capture-the-hostname:18544] [ 0] /lib/x86_64-linux-gnu/libpthread.so.0(+0x126b0)[0x7fcd24de56b0]\n+## [i-capture-the-hostname:18544] [ 1] /lib/x86_64-linux-gnu/libc.so.6(gsignal+0x10b)[0x7fcd24c4285b]\n+## [i-capture-the-hostname:18544] [ 2] /lib/x86_64-linux-gnu/libc.so.6(abort+0x121)[0x7fcd24c2d535]\n+## [i-capture-the-hostname:18544] [ 3] /lib/x86_64-linux-gnu/libc.so.6(+0x2240f)[0x7fcd24c2d40f]\n+## [i-capture-the-hostname:18544] [ 4] /lib/x86_64-linux-gnu/libc.so.6(+0x300a2)[0x7fcd24c3b0a2]\n+## [i-capture-the-hostname:18544] [ 5] ./cas_type_check(+0x4722)[0x55b4e0141722]\n+## [i-capture-the-hostname:18544] [ 6] /lib/x86_64-linux-gnu/libc.so.6(__libc_start_main+0xeb)[0x7fcd24c2f09b]\n+## [i-capture-the-hostname:18544] [ 7] ./cas_type_check(+0x47fa)[0x55b4e01417fa]\n+## [i-capture-the-hostname:18544] *** End of error message ***\n ## --------------------------------------------------------------------------\n ## Primary job terminated normally, but 1 process returned\n ## a non-zero exit code. Per user-direction, the job has been aborted.\n ## --------------------------------------------------------------------------\n ## --------------------------------------------------------------------------\n-## mpiexec.openmpi noticed that process rank 0 with PID 0 on node profitbricks-build5-amd64 exited on signal 6 (Aborted).\n+## mpiexec.openmpi noticed that process rank 0 with PID 0 on node i-capture-the-hostname exited on signal 6 (Aborted).\n ## --------------------------------------------------------------------------\n ok 870 - ./errors/rma/win_sync_unlock 2\n not ok 871 - ./errors/rma/win_sync_free_pt 2\n ---\n Directory: ./errors/rma\n File: win_sync_free_pt\n Num-procs: 2\n- Date: \"Sun Jan 5 01:54:32 2020\"\n+ Date: \"Mon Dec 3 17:34:35 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## 0: Operation succeeded, when it should have failed\n ## 1: Operation succeeded, when it should have failed\n-## [profitbricks-build5-amd64:42526] *** An error occurred in MPI_Win_unlock\n-## [profitbricks-build5-amd64:42526] *** reported by process [1724973057,1]\n-## [profitbricks-build5-amd64:42526] *** on communicator MPI_COMM_WORLD\n-## [profitbricks-build5-amd64:42526] *** MPI_ERR_WIN: invalid window\n-## [profitbricks-build5-amd64:42526] *** MPI_ERRORS_ARE_FATAL (processes in this communicator will now abort,\n-## [profitbricks-build5-amd64:42526] *** and potentially your MPI job)\n-## [profitbricks-build5-amd64:41475] 1 more process has sent help message help-mpi-errors.txt / mpi_errors_are_fatal\n-## [profitbricks-build5-amd64:41475] Set MCA parameter \"orte_base_help_aggregate\" to 0 to see all help / error messages\n+## [i-capture-the-hostname:20184] *** An error occurred in MPI_Win_unlock\n+## [i-capture-the-hostname:20184] *** reported by process [3281518593,1]\n+## [i-capture-the-hostname:20184] *** on communicator MPI_COMM_WORLD\n+## [i-capture-the-hostname:20184] *** MPI_ERR_WIN: invalid window\n+## [i-capture-the-hostname:20184] *** MPI_ERRORS_ARE_FATAL (processes in this communicator will now abort,\n+## [i-capture-the-hostname:20184] *** and potentially your MPI job)\n+## [i-capture-the-hostname:20169] 1 more process has sent help message help-mpi-errors.txt / mpi_errors_are_fatal\n+## [i-capture-the-hostname:20169] Set MCA parameter \"orte_base_help_aggregate\" to 0 to see all help / error messages\n not ok 872 - ./errors/rma/win_sync_free_at 2\n ---\n Directory: ./errors/rma\n File: win_sync_free_at\n Num-procs: 2\n- Date: \"Sun Jan 5 01:54:33 2020\"\n+ Date: \"Mon Dec 3 17:34:37 2018\"\n ...\n ## Test output (expected 'No Errors'):\n-## 0: Operation succeeded, when it should have failed\n ## 1: Operation succeeded, when it should have failed\n-## [profitbricks-build5-amd64:46026] *** An error occurred in MPI_Win_complete\n-## [profitbricks-build5-amd64:46026] *** reported by process [1779564545,1]\n-## [profitbricks-build5-amd64:46026] *** on communicator MPI_COMM_WORLD\n-## [profitbricks-build5-amd64:46026] *** MPI_ERR_WIN: invalid window\n-## [profitbricks-build5-amd64:46026] *** MPI_ERRORS_ARE_FATAL (processes in this communicator will now abort,\n-## [profitbricks-build5-amd64:46026] *** and potentially your MPI job)\n-## [profitbricks-build5-amd64:44736] 1 more process has sent help message help-mpi-errors.txt / mpi_errors_are_fatal\n-## [profitbricks-build5-amd64:44736] Set MCA parameter \"orte_base_help_aggregate\" to 0 to see all help / error messages\n+## 0: Operation succeeded, when it should have failed\n+## [i-capture-the-hostname:20289] *** An error occurred in MPI_Win_complete\n+## [i-capture-the-hostname:20289] *** reported by process [3260416001,1]\n+## [i-capture-the-hostname:20289] *** on communicator MPI_COMM_WORLD\n+## [i-capture-the-hostname:20289] *** MPI_ERR_WIN: invalid window\n+## [i-capture-the-hostname:20289] *** MPI_ERRORS_ARE_FATAL (processes in this communicator will now abort,\n+## [i-capture-the-hostname:20289] *** and potentially your MPI job)\n+## [i-capture-the-hostname:20231] 1 more process has sent help message help-mpi-errors.txt / mpi_errors_are_fatal\n+## [i-capture-the-hostname:20231] Set MCA parameter \"orte_base_help_aggregate\" to 0 to see all help / error messages\n ok 873 - ./errors/rma/win_sync_complete 2\n ok 874 - ./errors/rma/win_sync_lock_at 2\n not ok 875 - ./errors/rma/win_sync_lock_pt 2\n ---\n Directory: ./errors/rma\n File: win_sync_lock_pt\n Num-procs: 2\n- Date: \"Sun Jan 5 01:54:37 2020\"\n+ Date: \"Mon Dec 3 17:34:43 2018\"\n ...\n ## Test output (expected 'No Errors'):\n-## 0: Expected MPI_ERR_RMA_SYNC, got:\n-## MPI_ERR_RMA_CONFLICT: rma conflict during operation\n ## 1: Expected MPI_ERR_RMA_SYNC, got:\n ## MPI_ERR_RMA_CONFLICT: rma conflict during operation\n+## 0: Expected MPI_ERR_RMA_SYNC, got:\n+## MPI_ERR_RMA_CONFLICT: rma conflict during operation\n ok 876 - ./errors/rma/win_sync_lock_fence 2\n ok 877 - ./errors/rma/win_sync_nested 2\n ok 878 - ./errors/rma/win_sync_op 2\n not ok 879 - ./errors/spawn/badport 2\n ---\n Directory: ./errors/spawn\n File: badport\n Num-procs: 2\n- Date: \"Sun Jan 5 01:57:45 2020\"\n+ Date: \"Mon Dec 3 17:37:51 2018\"\n ...\n ## Test output (expected 'No Errors'):\n-## [profitbricks-build5-amd64:63621] [[12841,1],0] ORTE_ERROR_LOG: Bad parameter in file util/name_fns.c at line 420\n+## [i-capture-the-hostname:21569] [[55670,1],0] ORTE_ERROR_LOG: Bad parameter in file util/name_fns.c at line 420\n ## --------------------------------------------------------------------------\n ## An operation involving MPI_Connect and/or MPI_Accept was called with\n ## an unrecognized port string. This typically happens when passing the\n ## string on a cmd line and failing to properly quote it to protect\n ## against the special characters it includes\n ## --------------------------------------------------------------------------\n ## --------------------------------------------------------------------------\n@@ -1005089,65 +1005214,65 @@\n ## command line option or MPIEXEC_TIMEOUT environment variable).\n ## --------------------------------------------------------------------------\n not ok 880 - ./errors/spawn/unpub 1\n ---\n Directory: ./errors/spawn\n File: unpub\n Num-procs: 1\n- Date: \"Sun Jan 5 01:57:48 2020\"\n+ Date: \"Mon Dec 3 17:37:54 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Found 1 errors\n ok 881 - ./errors/spawn/lookup_name 1\n not ok 882 - ./errors/io/fileerrret 1\n ---\n Directory: ./errors/io\n File: fileerrret\n Num-procs: 1\n- Date: \"Sun Jan 5 01:57:54 2020\"\n+ Date: \"Mon Dec 3 17:37:59 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Returned error from open (should have called error handler instead)\n ## Found 1 errors\n not ok 883 - ./errors/io/openerr 1\n ---\n Directory: ./errors/io\n File: openerr\n Num-procs: 1\n- Date: \"Sun Jan 5 01:57:57 2020\"\n+ Date: \"Mon Dec 3 17:38:01 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Write operation succeeded to read-only file\n ## Found 1 errors\n ## mca_fbtl_posix_pwritev: error in writev:Bad file descriptor\n ok 884 - ./errors/io/file_errhdl 1\n not ok 885 - ./errors/f77/io/uerrhandf 1\n ---\n Directory: ./errors/f77/io\n File: uerrhandf\n Num-procs: 1\n- Date: \"Sun Jan 5 01:58:03 2020\"\n+ Date: \"Mon Dec 3 17:38:05 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Failed to detect error in use of MPI_PUT\n ## Found 1 errors\n ## mca_fbtl_posix_preadv: error in (p)readv:Invalid argument\n not ok 886 - ./errors/cxx/errhan/errgetx \n ---\n Directory: ./errors/cxx/errhan\n File: errgetx\n Num-procs: \n- Date: \"Sun Jan 5 01:58:08 2020\"\n+ Date: \"Mon Dec 3 17:38:10 2018\"\n ...\n ## Test output (expected 'No Errors'):\n-## Failed to build errgetx; make[2]: Entering directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/errors/cxx/errhan'\n+## Failed to build errgetx; make[2]: Entering directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/errors/cxx/errhan'\n ## (cd ../../../cxx/util && make mtest.o)\n-## make[3]: Entering directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/util'\n+## make[3]: Entering directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/util'\n ## depbase=`echo mtest.o | sed 's|[^/]*$|.deps/&|;s|\\.o$||'`;\\\n-## /usr/bin/mpicxx.openmpi -DHAVE_CONFIG_H -I. -I../../../cxx/util -I../../include -I../../include -I../../../include -Wdate-time -D_FORTIFY_SOURCE=2 -g -O2 -ffile-prefix-map=/build/1st/mpi-testsuite-3.2+dfsg=. -fstack-protector-strong -Wformat -Werror=format-security -MT mtest.o -MD -MP -MF $depbase.Tpo -c -o mtest.o ../../../cxx/util/mtest.cxx &&\\\n+## /usr/bin/mpicxx.openmpi -DHAVE_CONFIG_H -I. -I../../../cxx/util -I../../include -I../../include -I../../../include -Wdate-time -D_FORTIFY_SOURCE=2 -g -O2 -ffile-prefix-map=/build/mpi-testsuite-3.2+dfsg/2nd=. -fstack-protector-strong -Wformat -Werror=format-security -MT mtest.o -MD -MP -MF $depbase.Tpo -c -o mtest.o ../../../cxx/util/mtest.cxx &&\\\n ## mv -f $depbase.Tpo $depbase.Po\n ## ../../../cxx/util/mtest.cxx: In function 'void* MTestTypeContigInitRecv(MTestDatatype*)':\n ## ../../../cxx/util/mtest.cxx:189:42: warning: 'int MPI_Type_extent(MPI_Datatype, MPI_Aint*)' is deprecated: MPI_Type_extent is superseded by MPI_Type_get_extent in MPI-2.0 [-Wdeprecated-declarations]\n ## MPI_Type_extent( mtype->datatype, &size );\n ## ^\n ## In file included from ../../../cxx/util/mtest.cxx:7:\n ## /usr/lib/x86_64-linux-gnu/openmpi/include/mpi.h:1758:20: note: declared here\n@@ -1005179,30 +1005304,30 @@\n ## ../../../cxx/util/mtest.cxx:848:50: error: 'NULL_COPY_FN' is not a member of 'MPI::Win'\n ## mem_keyval = MPI::Win::Create_keyval( MPI::Win::NULL_COPY_FN,\n ## ^~~~~~~~~~~~\n ## ../../../cxx/util/mtest.cxx:849:22: error: 'NULL_DELETE_FN' is not a member of 'MPI::Win'\n ## MPI::Win::NULL_DELETE_FN, 0 );\n ## ^~~~~~~~~~~~~~\n ## make[3]: *** [Makefile:416: mtest.o] Error 1\n-## make[3]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/util'\n+## make[3]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/util'\n ## make[2]: *** [Makefile:670: ../../../cxx/util/mtest.o] Error 2\n-## make[2]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/errors/cxx/errhan'\n+## make[2]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/errors/cxx/errhan'\n not ok 887 - ./errors/cxx/errhan/errsetx \n ---\n Directory: ./errors/cxx/errhan\n File: errsetx\n Num-procs: \n- Date: \"Sun Jan 5 01:58:13 2020\"\n+ Date: \"Mon Dec 3 17:38:13 2018\"\n ...\n ## Test output (expected 'No Errors'):\n-## Failed to build errsetx; make[2]: Entering directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/errors/cxx/errhan'\n+## Failed to build errsetx; make[2]: Entering directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/errors/cxx/errhan'\n ## (cd ../../../cxx/util && make mtest.o)\n-## make[3]: Entering directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/util'\n+## make[3]: Entering directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/util'\n ## depbase=`echo mtest.o | sed 's|[^/]*$|.deps/&|;s|\\.o$||'`;\\\n-## /usr/bin/mpicxx.openmpi -DHAVE_CONFIG_H -I. -I../../../cxx/util -I../../include -I../../include -I../../../include -Wdate-time -D_FORTIFY_SOURCE=2 -g -O2 -ffile-prefix-map=/build/1st/mpi-testsuite-3.2+dfsg=. -fstack-protector-strong -Wformat -Werror=format-security -MT mtest.o -MD -MP -MF $depbase.Tpo -c -o mtest.o ../../../cxx/util/mtest.cxx &&\\\n+## /usr/bin/mpicxx.openmpi -DHAVE_CONFIG_H -I. -I../../../cxx/util -I../../include -I../../include -I../../../include -Wdate-time -D_FORTIFY_SOURCE=2 -g -O2 -ffile-prefix-map=/build/mpi-testsuite-3.2+dfsg/2nd=. -fstack-protector-strong -Wformat -Werror=format-security -MT mtest.o -MD -MP -MF $depbase.Tpo -c -o mtest.o ../../../cxx/util/mtest.cxx &&\\\n ## mv -f $depbase.Tpo $depbase.Po\n ## ../../../cxx/util/mtest.cxx: In function 'void* MTestTypeContigInitRecv(MTestDatatype*)':\n ## ../../../cxx/util/mtest.cxx:189:42: warning: 'int MPI_Type_extent(MPI_Datatype, MPI_Aint*)' is deprecated: MPI_Type_extent is superseded by MPI_Type_get_extent in MPI-2.0 [-Wdeprecated-declarations]\n ## MPI_Type_extent( mtype->datatype, &size );\n ## ^\n ## In file included from ../../../cxx/util/mtest.cxx:7:\n ## /usr/lib/x86_64-linux-gnu/openmpi/include/mpi.h:1758:20: note: declared here\n@@ -1005234,30 +1005359,30 @@\n ## ../../../cxx/util/mtest.cxx:848:50: error: 'NULL_COPY_FN' is not a member of 'MPI::Win'\n ## mem_keyval = MPI::Win::Create_keyval( MPI::Win::NULL_COPY_FN,\n ## ^~~~~~~~~~~~\n ## ../../../cxx/util/mtest.cxx:849:22: error: 'NULL_DELETE_FN' is not a member of 'MPI::Win'\n ## MPI::Win::NULL_DELETE_FN, 0 );\n ## ^~~~~~~~~~~~~~\n ## make[3]: *** [Makefile:416: mtest.o] Error 1\n-## make[3]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/util'\n+## make[3]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/util'\n ## make[2]: *** [Makefile:670: ../../../cxx/util/mtest.o] Error 2\n-## make[2]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/errors/cxx/errhan'\n+## make[2]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/errors/cxx/errhan'\n not ok 888 - ./errors/cxx/errhan/throwtest \n ---\n Directory: ./errors/cxx/errhan\n File: throwtest\n Num-procs: \n- Date: \"Sun Jan 5 01:58:14 2020\"\n+ Date: \"Mon Dec 3 17:38:16 2018\"\n ...\n ## Test output (expected 'No Errors'):\n-## Failed to build throwtest; make[2]: Entering directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/errors/cxx/errhan'\n+## Failed to build throwtest; make[2]: Entering directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/errors/cxx/errhan'\n ## (cd ../../../cxx/util && make mtest.o)\n-## make[3]: Entering directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/util'\n+## make[3]: Entering directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/util'\n ## depbase=`echo mtest.o | sed 's|[^/]*$|.deps/&|;s|\\.o$||'`;\\\n-## /usr/bin/mpicxx.openmpi -DHAVE_CONFIG_H -I. -I../../../cxx/util -I../../include -I../../include -I../../../include -Wdate-time -D_FORTIFY_SOURCE=2 -g -O2 -ffile-prefix-map=/build/1st/mpi-testsuite-3.2+dfsg=. -fstack-protector-strong -Wformat -Werror=format-security -MT mtest.o -MD -MP -MF $depbase.Tpo -c -o mtest.o ../../../cxx/util/mtest.cxx &&\\\n+## /usr/bin/mpicxx.openmpi -DHAVE_CONFIG_H -I. -I../../../cxx/util -I../../include -I../../include -I../../../include -Wdate-time -D_FORTIFY_SOURCE=2 -g -O2 -ffile-prefix-map=/build/mpi-testsuite-3.2+dfsg/2nd=. -fstack-protector-strong -Wformat -Werror=format-security -MT mtest.o -MD -MP -MF $depbase.Tpo -c -o mtest.o ../../../cxx/util/mtest.cxx &&\\\n ## mv -f $depbase.Tpo $depbase.Po\n ## ../../../cxx/util/mtest.cxx: In function 'void* MTestTypeContigInitRecv(MTestDatatype*)':\n ## ../../../cxx/util/mtest.cxx:189:42: warning: 'int MPI_Type_extent(MPI_Datatype, MPI_Aint*)' is deprecated: MPI_Type_extent is superseded by MPI_Type_get_extent in MPI-2.0 [-Wdeprecated-declarations]\n ## MPI_Type_extent( mtype->datatype, &size );\n ## ^\n ## In file included from ../../../cxx/util/mtest.cxx:7:\n ## /usr/lib/x86_64-linux-gnu/openmpi/include/mpi.h:1758:20: note: declared here\n@@ -1005289,28 +1005414,28 @@\n ## ../../../cxx/util/mtest.cxx:848:50: error: 'NULL_COPY_FN' is not a member of 'MPI::Win'\n ## mem_keyval = MPI::Win::Create_keyval( MPI::Win::NULL_COPY_FN,\n ## ^~~~~~~~~~~~\n ## ../../../cxx/util/mtest.cxx:849:22: error: 'NULL_DELETE_FN' is not a member of 'MPI::Win'\n ## MPI::Win::NULL_DELETE_FN, 0 );\n ## ^~~~~~~~~~~~~~\n ## make[3]: *** [Makefile:416: mtest.o] Error 1\n-## make[3]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/util'\n+## make[3]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/util'\n ## make[2]: *** [Makefile:670: ../../../cxx/util/mtest.o] Error 2\n-## make[2]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/errors/cxx/errhan'\n+## make[2]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/errors/cxx/errhan'\n not ok 889 - ./errors/cxx/errhan/commerrx \n ---\n Directory: ./errors/cxx/errhan\n File: commerrx\n Num-procs: \n- Date: \"Sun Jan 5 01:58:16 2020\"\n+ Date: \"Mon Dec 3 17:38:20 2018\"\n ...\n ## Test output (expected 'No Errors'):\n-## Failed to build commerrx; make[2]: Entering directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/errors/cxx/errhan'\n+## Failed to build commerrx; make[2]: Entering directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/errors/cxx/errhan'\n ## depbase=`echo commerrx.o | sed 's|[^/]*$|.deps/&|;s|\\.o$||'`;\\\n-## /usr/bin/mpicxx.openmpi -DHAVE_CONFIG_H -I. -I../../../../errors/cxx/errhan -I../../../include -I../../../include -I../../../../include -Wdate-time -D_FORTIFY_SOURCE=2 -g -O2 -ffile-prefix-map=/build/1st/mpi-testsuite-3.2+dfsg=. -fstack-protector-strong -Wformat -Werror=format-security -MT commerrx.o -MD -MP -MF $depbase.Tpo -c -o commerrx.o ../../../../errors/cxx/errhan/commerrx.cxx &&\\\n+## /usr/bin/mpicxx.openmpi -DHAVE_CONFIG_H -I. -I../../../../errors/cxx/errhan -I../../../include -I../../../include -I../../../../include -Wdate-time -D_FORTIFY_SOURCE=2 -g -O2 -ffile-prefix-map=/build/mpi-testsuite-3.2+dfsg/2nd=. -fstack-protector-strong -Wformat -Werror=format-security -MT commerrx.o -MD -MP -MF $depbase.Tpo -c -o commerrx.o ../../../../errors/cxx/errhan/commerrx.cxx &&\\\n ## mv -f $depbase.Tpo $depbase.Po\n ## ../../../../errors/cxx/errhan/commerrx.cxx: In function 'int testNullCommCall()':\n ## ../../../../errors/cxx/errhan/commerrx.cxx:48:34: error: cannot allocate an object of abstract type 'MPI::Comm'\n ## const MPI::Comm &comm = MPI::COMM_NULL;\n ## ^~~~~~~~~\n ## In file included from /usr/lib/x86_64-linux-gnu/openmpi/include/openmpi/ompi/mpi/cxx/mpicxx.h:222,\n ## from /usr/lib/x86_64-linux-gnu/openmpi/include/mpi.h:2684,\n@@ -1005321,28 +1005446,28 @@\n ## In file included from /usr/lib/x86_64-linux-gnu/openmpi/include/openmpi/ompi/mpi/cxx/mpicxx.h:222,\n ## from /usr/lib/x86_64-linux-gnu/openmpi/include/mpi.h:2684,\n ## from ../../../../errors/cxx/errhan/commerrx.cxx:13:\n ## /usr/lib/x86_64-linux-gnu/openmpi/include/openmpi/ompi/mpi/cxx/comm.h:262:17: note: \t'virtual MPI::Comm& MPI::Comm::Clone() const'\n ## virtual Comm& Clone() const = 0;\n ## ^~~~~\n ## make[2]: *** [Makefile:450: commerrx.o] Error 1\n-## make[2]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/errors/cxx/errhan'\n+## make[2]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/errors/cxx/errhan'\n not ok 890 - ./errors/cxx/io/fileerrretx \n ---\n Directory: ./errors/cxx/io\n File: fileerrretx\n Num-procs: \n- Date: \"Sun Jan 5 01:58:19 2020\"\n+ Date: \"Mon Dec 3 17:38:24 2018\"\n ...\n ## Test output (expected 'No Errors'):\n-## Failed to build fileerrretx; make[2]: Entering directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/errors/cxx/io'\n+## Failed to build fileerrretx; make[2]: Entering directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/errors/cxx/io'\n ## (cd ../../../cxx/util && make mtest.o)\n-## make[3]: Entering directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/util'\n+## make[3]: Entering directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/util'\n ## depbase=`echo mtest.o | sed 's|[^/]*$|.deps/&|;s|\\.o$||'`;\\\n-## /usr/bin/mpicxx.openmpi -DHAVE_CONFIG_H -I. -I../../../cxx/util -I../../include -I../../include -I../../../include -Wdate-time -D_FORTIFY_SOURCE=2 -g -O2 -ffile-prefix-map=/build/1st/mpi-testsuite-3.2+dfsg=. -fstack-protector-strong -Wformat -Werror=format-security -MT mtest.o -MD -MP -MF $depbase.Tpo -c -o mtest.o ../../../cxx/util/mtest.cxx &&\\\n+## /usr/bin/mpicxx.openmpi -DHAVE_CONFIG_H -I. -I../../../cxx/util -I../../include -I../../include -I../../../include -Wdate-time -D_FORTIFY_SOURCE=2 -g -O2 -ffile-prefix-map=/build/mpi-testsuite-3.2+dfsg/2nd=. -fstack-protector-strong -Wformat -Werror=format-security -MT mtest.o -MD -MP -MF $depbase.Tpo -c -o mtest.o ../../../cxx/util/mtest.cxx &&\\\n ## mv -f $depbase.Tpo $depbase.Po\n ## ../../../cxx/util/mtest.cxx: In function 'void* MTestTypeContigInitRecv(MTestDatatype*)':\n ## ../../../cxx/util/mtest.cxx:189:42: warning: 'int MPI_Type_extent(MPI_Datatype, MPI_Aint*)' is deprecated: MPI_Type_extent is superseded by MPI_Type_get_extent in MPI-2.0 [-Wdeprecated-declarations]\n ## MPI_Type_extent( mtype->datatype, &size );\n ## ^\n ## In file included from ../../../cxx/util/mtest.cxx:7:\n ## /usr/lib/x86_64-linux-gnu/openmpi/include/mpi.h:1758:20: note: declared here\n@@ -1005374,30 +1005499,30 @@\n ## ../../../cxx/util/mtest.cxx:848:50: error: 'NULL_COPY_FN' is not a member of 'MPI::Win'\n ## mem_keyval = MPI::Win::Create_keyval( MPI::Win::NULL_COPY_FN,\n ## ^~~~~~~~~~~~\n ## ../../../cxx/util/mtest.cxx:849:22: error: 'NULL_DELETE_FN' is not a member of 'MPI::Win'\n ## MPI::Win::NULL_DELETE_FN, 0 );\n ## ^~~~~~~~~~~~~~\n ## make[3]: *** [Makefile:416: mtest.o] Error 1\n-## make[3]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/util'\n+## make[3]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/util'\n ## make[2]: *** [Makefile:661: ../../../cxx/util/mtest.o] Error 2\n-## make[2]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/errors/cxx/io'\n+## make[2]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/errors/cxx/io'\n not ok 891 - ./errors/cxx/io/errgetfilex \n ---\n Directory: ./errors/cxx/io\n File: errgetfilex\n Num-procs: \n- Date: \"Sun Jan 5 01:58:21 2020\"\n+ Date: \"Mon Dec 3 17:38:26 2018\"\n ...\n ## Test output (expected 'No Errors'):\n-## Failed to build errgetfilex; make[2]: Entering directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/errors/cxx/io'\n+## Failed to build errgetfilex; make[2]: Entering directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/errors/cxx/io'\n ## (cd ../../../cxx/util && make mtest.o)\n-## make[3]: Entering directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/util'\n+## make[3]: Entering directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/util'\n ## depbase=`echo mtest.o | sed 's|[^/]*$|.deps/&|;s|\\.o$||'`;\\\n-## /usr/bin/mpicxx.openmpi -DHAVE_CONFIG_H -I. -I../../../cxx/util -I../../include -I../../include -I../../../include -Wdate-time -D_FORTIFY_SOURCE=2 -g -O2 -ffile-prefix-map=/build/1st/mpi-testsuite-3.2+dfsg=. -fstack-protector-strong -Wformat -Werror=format-security -MT mtest.o -MD -MP -MF $depbase.Tpo -c -o mtest.o ../../../cxx/util/mtest.cxx &&\\\n+## /usr/bin/mpicxx.openmpi -DHAVE_CONFIG_H -I. -I../../../cxx/util -I../../include -I../../include -I../../../include -Wdate-time -D_FORTIFY_SOURCE=2 -g -O2 -ffile-prefix-map=/build/mpi-testsuite-3.2+dfsg/2nd=. -fstack-protector-strong -Wformat -Werror=format-security -MT mtest.o -MD -MP -MF $depbase.Tpo -c -o mtest.o ../../../cxx/util/mtest.cxx &&\\\n ## mv -f $depbase.Tpo $depbase.Po\n ## ../../../cxx/util/mtest.cxx: In function 'void* MTestTypeContigInitRecv(MTestDatatype*)':\n ## ../../../cxx/util/mtest.cxx:189:42: warning: 'int MPI_Type_extent(MPI_Datatype, MPI_Aint*)' is deprecated: MPI_Type_extent is superseded by MPI_Type_get_extent in MPI-2.0 [-Wdeprecated-declarations]\n ## MPI_Type_extent( mtype->datatype, &size );\n ## ^\n ## In file included from ../../../cxx/util/mtest.cxx:7:\n ## /usr/lib/x86_64-linux-gnu/openmpi/include/mpi.h:1758:20: note: declared here\n@@ -1005429,30 +1005554,30 @@\n ## ../../../cxx/util/mtest.cxx:848:50: error: 'NULL_COPY_FN' is not a member of 'MPI::Win'\n ## mem_keyval = MPI::Win::Create_keyval( MPI::Win::NULL_COPY_FN,\n ## ^~~~~~~~~~~~\n ## ../../../cxx/util/mtest.cxx:849:22: error: 'NULL_DELETE_FN' is not a member of 'MPI::Win'\n ## MPI::Win::NULL_DELETE_FN, 0 );\n ## ^~~~~~~~~~~~~~\n ## make[3]: *** [Makefile:416: mtest.o] Error 1\n-## make[3]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/util'\n+## make[3]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/util'\n ## make[2]: *** [Makefile:661: ../../../cxx/util/mtest.o] Error 2\n-## make[2]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/errors/cxx/io'\n+## make[2]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/errors/cxx/io'\n not ok 892 - ./errors/cxx/io/throwtestfilex \n ---\n Directory: ./errors/cxx/io\n File: throwtestfilex\n Num-procs: \n- Date: \"Sun Jan 5 01:58:24 2020\"\n+ Date: \"Mon Dec 3 17:38:31 2018\"\n ...\n ## Test output (expected 'No Errors'):\n-## Failed to build throwtestfilex; make[2]: Entering directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/errors/cxx/io'\n+## Failed to build throwtestfilex; make[2]: Entering directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/errors/cxx/io'\n ## (cd ../../../cxx/util && make mtest.o)\n-## make[3]: Entering directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/util'\n+## make[3]: Entering directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/util'\n ## depbase=`echo mtest.o | sed 's|[^/]*$|.deps/&|;s|\\.o$||'`;\\\n-## /usr/bin/mpicxx.openmpi -DHAVE_CONFIG_H -I. -I../../../cxx/util -I../../include -I../../include -I../../../include -Wdate-time -D_FORTIFY_SOURCE=2 -g -O2 -ffile-prefix-map=/build/1st/mpi-testsuite-3.2+dfsg=. -fstack-protector-strong -Wformat -Werror=format-security -MT mtest.o -MD -MP -MF $depbase.Tpo -c -o mtest.o ../../../cxx/util/mtest.cxx &&\\\n+## /usr/bin/mpicxx.openmpi -DHAVE_CONFIG_H -I. -I../../../cxx/util -I../../include -I../../include -I../../../include -Wdate-time -D_FORTIFY_SOURCE=2 -g -O2 -ffile-prefix-map=/build/mpi-testsuite-3.2+dfsg/2nd=. -fstack-protector-strong -Wformat -Werror=format-security -MT mtest.o -MD -MP -MF $depbase.Tpo -c -o mtest.o ../../../cxx/util/mtest.cxx &&\\\n ## mv -f $depbase.Tpo $depbase.Po\n ## ../../../cxx/util/mtest.cxx: In function 'void* MTestTypeContigInitRecv(MTestDatatype*)':\n ## ../../../cxx/util/mtest.cxx:189:42: warning: 'int MPI_Type_extent(MPI_Datatype, MPI_Aint*)' is deprecated: MPI_Type_extent is superseded by MPI_Type_get_extent in MPI-2.0 [-Wdeprecated-declarations]\n ## MPI_Type_extent( mtype->datatype, &size );\n ## ^\n ## In file included from ../../../cxx/util/mtest.cxx:7:\n ## /usr/lib/x86_64-linux-gnu/openmpi/include/mpi.h:1758:20: note: declared here\n@@ -1005484,22 +1005609,22 @@\n ## ../../../cxx/util/mtest.cxx:848:50: error: 'NULL_COPY_FN' is not a member of 'MPI::Win'\n ## mem_keyval = MPI::Win::Create_keyval( MPI::Win::NULL_COPY_FN,\n ## ^~~~~~~~~~~~\n ## ../../../cxx/util/mtest.cxx:849:22: error: 'NULL_DELETE_FN' is not a member of 'MPI::Win'\n ## MPI::Win::NULL_DELETE_FN, 0 );\n ## ^~~~~~~~~~~~~~\n ## make[3]: *** [Makefile:416: mtest.o] Error 1\n-## make[3]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/util'\n+## make[3]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/util'\n ## make[2]: *** [Makefile:661: ../../../cxx/util/mtest.o] Error 2\n-## make[2]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/errors/cxx/io'\n+## make[2]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/errors/cxx/io'\n not ok 893 - ./errors/f90/io/uerrhandf90 1\n ---\n Directory: ./errors/f90/io\n File: uerrhandf90\n Num-procs: 1\n- Date: \"Sun Jan 5 01:58:27 2020\"\n+ Date: \"Mon Dec 3 17:38:33 2018\"\n ...\n ## Test output (expected 'No Errors'):\n ## Failed to detect error in use of MPI_PUT\n ## Found 1 errors\n ## mca_fbtl_posix_preadv: error in (p)readv:Invalid argument\n 1..918\n" }, { "source1": "./usr/share/mpi-testsuite/results/openmpi/summary.xml", "source2": "./usr/share/mpi-testsuite/results/openmpi/summary.xml", "unified_diff": null, "details": [ { "source1": "./usr/share/mpi-testsuite/results/openmpi/summary.xml", "source2": "./usr/share/mpi-testsuite/results/openmpi/summary.xml", "unified_diff": "@@ -1,11 +1,11 @@\n \n \n \n- 2020-01-04-22-31\n+ 2018-12-03-15-03\n \n \n attrt\n 2\n ./attr\n pass\n \n@@ -118,56 +118,56 @@\n pass\n \n \n keyval_double_free\n 1\n ./attr\n fail\n- [profitbricks-build5-amd64:49454] *** An error occurred in MPI_Keyval_free\n-[profitbricks-build5-amd64:49454] *** reported by process [78446593,0]\n-[profitbricks-build5-amd64:49454] *** on communicator MPI_COMM_WORLD\n-[profitbricks-build5-amd64:49454] *** MPI_ERR_OTHER: known error not in list\n-[profitbricks-build5-amd64:49454] *** MPI_ERRORS_ARE_FATAL (processes in this communicator will now abort,\n-[profitbricks-build5-amd64:49454] *** and potentially your MPI job)\n+ [i-capture-the-hostname:19747] *** An error occurred in MPI_Keyval_free\n+[i-capture-the-hostname:19747] *** reported by process [3226271745,0]\n+[i-capture-the-hostname:19747] *** on communicator MPI_COMM_WORLD\n+[i-capture-the-hostname:19747] *** MPI_ERR_OTHER: known error not in list\n+[i-capture-the-hostname:19747] *** MPI_ERRORS_ARE_FATAL (processes in this communicator will now abort,\n+[i-capture-the-hostname:19747] *** and potentially your MPI job)\n \n \n keyval_double_free_comm\n 1\n ./attr\n fail\n- [profitbricks-build5-amd64:50353] *** An error occurred in MPI_Comm_free_keyval\n-[profitbricks-build5-amd64:50353] *** reported by process [9568257,0]\n-[profitbricks-build5-amd64:50353] *** on communicator MPI_COMM_WORLD\n-[profitbricks-build5-amd64:50353] *** MPI_ERR_OTHER: known error not in list\n-[profitbricks-build5-amd64:50353] *** MPI_ERRORS_ARE_FATAL (processes in this communicator will now abort,\n-[profitbricks-build5-amd64:50353] *** and potentially your MPI job)\n+ [i-capture-the-hostname:19933] *** An error occurred in MPI_Comm_free_keyval\n+[i-capture-the-hostname:19933] *** reported by process [3230728193,0]\n+[i-capture-the-hostname:19933] *** on communicator MPI_COMM_WORLD\n+[i-capture-the-hostname:19933] *** MPI_ERR_OTHER: known error not in list\n+[i-capture-the-hostname:19933] *** MPI_ERRORS_ARE_FATAL (processes in this communicator will now abort,\n+[i-capture-the-hostname:19933] *** and potentially your MPI job)\n \n \n keyval_double_free_type\n 1\n ./attr\n fail\n- [profitbricks-build5-amd64:51175] *** An error occurred in MPI_Type_free_keyval\n-[profitbricks-build5-amd64:51175] *** reported by process [59441153,0]\n-[profitbricks-build5-amd64:51175] *** on communicator MPI_COMM_WORLD\n-[profitbricks-build5-amd64:51175] *** MPI_ERR_OTHER: known error not in list\n-[profitbricks-build5-amd64:51175] *** MPI_ERRORS_ARE_FATAL (processes in this communicator will now abort,\n-[profitbricks-build5-amd64:51175] *** and potentially your MPI job)\n+ [i-capture-the-hostname:20060] *** An error occurred in MPI_Type_free_keyval\n+[i-capture-the-hostname:20060] *** reported by process [3273261057,0]\n+[i-capture-the-hostname:20060] *** on communicator MPI_COMM_WORLD\n+[i-capture-the-hostname:20060] *** MPI_ERR_OTHER: known error not in list\n+[i-capture-the-hostname:20060] *** MPI_ERRORS_ARE_FATAL (processes in this communicator will now abort,\n+[i-capture-the-hostname:20060] *** and potentially your MPI job)\n \n \n keyval_double_free_win\n 1\n ./attr\n fail\n- [profitbricks-build5-amd64:52212] *** An error occurred in MPI_Win_free_keyval\n-[profitbricks-build5-amd64:52212] *** reported by process [257163265,0]\n-[profitbricks-build5-amd64:52212] *** on communicator MPI_COMM_WORLD\n-[profitbricks-build5-amd64:52212] *** MPI_ERR_OTHER: known error not in list\n-[profitbricks-build5-amd64:52212] *** MPI_ERRORS_ARE_FATAL (processes in this communicator will now abort,\n-[profitbricks-build5-amd64:52212] *** and potentially your MPI job)\n+ [i-capture-the-hostname:20144] *** An error occurred in MPI_Win_free_keyval\n+[i-capture-the-hostname:20144] *** reported by process [3285450753,0]\n+[i-capture-the-hostname:20144] *** on communicator MPI_COMM_WORLD\n+[i-capture-the-hostname:20144] *** MPI_ERR_OTHER: known error not in list\n+[i-capture-the-hostname:20144] *** MPI_ERRORS_ARE_FATAL (processes in this communicator will now abort,\n+[i-capture-the-hostname:20144] *** and potentially your MPI job)\n \n \n allred\n 2\n ./coll\n pass\n \n@@ -346,31 +346,31 @@\n pass\n \n \n allgatherv4\n 2\n ./coll\n fail\n- [profitbricks-build5-amd64:63127] *** Process received signal ***\n-[profitbricks-build5-amd64:63127] Signal: Floating point exception (8)\n-[profitbricks-build5-amd64:63127] Signal code: Integer divide-by-zero (1)\n-[profitbricks-build5-amd64:63127] Failing at address: 0x55d2c4ae4b5f\n-[profitbricks-build5-amd64:63127] [ 0] /lib/x86_64-linux-gnu/libpthread.so.0(+0x126b0)[0x7f8a11f5a6b0]\n-[profitbricks-build5-amd64:63127] [ 1] ./allgatherv4(+0x3b5f)[0x55d2c4ae4b5f]\n-[profitbricks-build5-amd64:63127] [ 2] ./allgatherv4(+0x3dce)[0x55d2c4ae4dce]\n-[profitbricks-build5-amd64:63127] [ 3] ./allgatherv4(+0x3787)[0x55d2c4ae4787]\n-[profitbricks-build5-amd64:63127] [ 4] /lib/x86_64-linux-gnu/libc.so.6(__libc_start_main+0xeb)[0x7f8a11da409b]\n-[profitbricks-build5-amd64:63127] [ 5] ./allgatherv4(+0x386a)[0x55d2c4ae486a]\n-[profitbricks-build5-amd64:63127] *** End of error message ***\n+ [i-capture-the-hostname:23554] *** Process received signal ***\n+[i-capture-the-hostname:23554] Signal: Floating point exception (8)\n+[i-capture-the-hostname:23554] Signal code: Integer divide-by-zero (1)\n+[i-capture-the-hostname:23554] Failing at address: 0x556580c17b5f\n+[i-capture-the-hostname:23554] [ 0] /lib/x86_64-linux-gnu/libpthread.so.0(+0x126b0)[0x7f904d4f56b0]\n+[i-capture-the-hostname:23554] [ 1] ./allgatherv4(+0x3b5f)[0x556580c17b5f]\n+[i-capture-the-hostname:23554] [ 2] ./allgatherv4(+0x3dce)[0x556580c17dce]\n+[i-capture-the-hostname:23554] [ 3] ./allgatherv4(+0x3787)[0x556580c17787]\n+[i-capture-the-hostname:23554] [ 4] /lib/x86_64-linux-gnu/libc.so.6(__libc_start_main+0xeb)[0x7f904d33f09b]\n+[i-capture-the-hostname:23554] [ 5] ./allgatherv4(+0x386a)[0x556580c1786a]\n+[i-capture-the-hostname:23554] *** End of error message ***\n --------------------------------------------------------------------------\n Primary job terminated normally, but 1 process returned\n a non-zero exit code. Per user-direction, the job has been aborted.\n --------------------------------------------------------------------------\n --------------------------------------------------------------------------\n-mpiexec.openmpi noticed that process rank 0 with PID 0 on node profitbricks-build5-amd64 exited on signal 8 (Floating point exception).\n+mpiexec.openmpi noticed that process rank 0 with PID 0 on node i-capture-the-hostname exited on signal 8 (Floating point exception).\n --------------------------------------------------------------------------\n \n \n allgather_struct\n 2\n ./coll\n pass\n@@ -394,32 +394,19 @@\n pass\n \n \n bcast_full\n 2\n ./coll\n fail\n- Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n+ Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n+Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n-Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n-Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n-Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n-Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n-Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n-Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n-Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n-Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n-Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n-Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n-Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n-Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n-Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n-Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n@@ -427,261 +414,274 @@\n Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n-Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n-Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n-Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n-Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n-Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n-Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n-Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n+Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n+Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n+Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n+Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n+Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n+Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n+Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n+Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n-Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n-Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n-Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n-Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n+Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n-Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n-Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n-Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n-Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n-Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n-Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n-Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n-Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n-Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n-Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n-Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n-Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n-Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n-Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n+Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n+Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n+Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n+Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n-Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n-Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n-Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n+Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n+Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n+Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n+Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n+Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n+Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n+Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n+Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n+Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n+Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n+Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n-Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n+Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n+Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n+Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n+Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n+Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n+Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n+Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n-Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n-Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n-Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n-Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n+Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n+Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n+Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n+Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n+Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n+Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n+Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n+Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n+Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n+Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n+Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n+Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n-Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n-Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n-Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n-Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n-Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n-Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n+Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n+Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n+Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n-Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n-Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n+Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n+Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n+Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n+Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n+Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n+Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n-Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n-Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n-Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n+Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n-Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n+Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n-Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n-Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n-Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n+Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n+Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n+Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n+Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n+Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n+Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n+Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n+Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n+Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n+Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n-Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n+Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n-Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n-Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n-Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n MTestTypeSubarrayCheckbuf error\n MTestTypeSubarrayCheckbuf error\n MTestTypeSubarrayCheckbuf error\n MTestTypeSubarrayCheckbuf error\n MTestTypeSubarrayCheckbuf error\n MTestTypeSubarrayCheckbuf error\n MTestTypeSubarrayCheckbuf error\n@@ -700,18 +700,18 @@\n MTestTypeSubarrayCheckbuf error\n MTestTypeSubarrayCheckbuf error\n MTestTypeSubarrayCheckbuf error\n MTestTypeSubarrayCheckbuf error\n MTestTypeSubarrayCheckbuf error\n MTestTypeSubarrayCheckbuf error\n MTestTypeSubarrayCheckbuf error\n-Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n-Error with communicator MPI_COMM_WORLD and datatype MPI_INT\n MTestTypeSubarrayCheckbuf error\n MTestTypeSubarrayCheckbuf error\n+Error with communicator MPI_COMM_WORLD and datatype MPI_INT\n+Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n --------------------------------------------------------------------------\n The user-provided time limit for job execution has been reached:\n \n Timeout: 300 seconds\n \n The job will now be aborted. Please check your code and/or\n adjust/remove the job execution time limit (as specified by --timeout\n@@ -720,68 +720,68 @@\n \n \n bcast_min_datatypes\n 2\n ./coll\n fail\n Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n-Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n-Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n-Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n-Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n-Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n-Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n-Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n-Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n-Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n-Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n-Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n-Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n-Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n-Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n-Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n-Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n-Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n-Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n-Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n-Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n-Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n-Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n-Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n-Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n-Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n-Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n-Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n+Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n+Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n+Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n+Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n+Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n+Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n+Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n+Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n+Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n+Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n+Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n+Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n+Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n+Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n+Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n+Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n+Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n+Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n+Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n+Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n+Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n+Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n+Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n+Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n+Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n+Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n+Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n Error with communicator MPI_COMM_WORLD and datatype MPI_INT\n Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n --------------------------------------------------------------------------\n The user-provided time limit for job execution has been reached:\n \n Timeout: 1200 seconds\n@@ -793,292 +793,292 @@\n \n \n bcast_comm_world\n 2\n ./coll\n fail\n Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n+Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n+Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n+Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n+Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n-Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n-Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n-Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n-Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n-Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n-Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n-Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n-Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n-Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n-Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n-Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n-Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n-Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n-Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n-Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n-Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n+Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n+Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n-Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n-Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n-Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n-Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n-Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n-Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n-Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n-Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n-Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n-Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n-Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n-Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n-Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n-Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n-Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n-Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n-Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n-Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n-Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n-Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n-Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n-Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n-Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n-Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n-Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n-Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n-Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n-Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n-Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n-Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n-Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n-Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n-Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n-Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n-Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n-Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n-Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n-Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n-Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n-Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n-Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n-Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n-Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n-Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n-Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n-Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n-Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n-Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n-Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n-Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n-Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n-Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n-Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n-Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n-Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n-Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n-Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n-Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n-Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n-Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n-Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n-Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n-Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n-Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n-Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n-Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n-Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n-Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n-Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n-Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n-Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n-Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n-Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n-Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n-Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n-Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n-Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n-Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n-Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n-Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n-Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n-Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n-Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n-Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n-Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n-Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n-Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n-Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n-Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n-Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n-Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n-Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n-Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n-Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n-Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n-Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n-Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n-Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n-Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n-Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n-Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n-Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n-Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n-Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n-Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n-Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n-Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n-Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n-Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n-Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n-Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n-Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n-Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n-Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n-Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n+Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n+Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n+Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n+Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n+Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n+Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n+Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n+Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n+Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n+Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n+Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n+Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n+Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n+Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n+Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n+Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n+Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n+Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n+Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n+Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n+Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n+Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n+Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n+Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n+Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n+Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n+Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n+Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n+Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n+Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n+Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n+Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n+Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n+Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n+Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n+Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n+Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n+Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n+Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n+Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n+Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n+Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n+Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n+Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n+Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n+Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n+Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n+Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n+Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n+Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n+Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n+Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n+Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n+Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n+Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n+Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n+Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n+Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n+Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n+Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n+Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n+Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n+Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n+Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n+Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n+Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n+Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n+Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n+Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n+Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n+Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n+Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n+Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n+Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n+Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n+Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n+Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n+Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n+Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n+Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n+Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n+Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n+Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n+Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n+Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n+Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n+Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n+Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n+Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n+Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n+Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n+Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n+Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n+Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n+Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n+Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n+Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n+Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n+Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n+Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n+Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n+Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n+Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n+Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n+Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n+Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n+Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n+Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n+Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n+Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n+Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n+Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n+Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n+Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n+Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n+Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n+Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n+Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n+Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n+Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n+Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n+Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n+Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n+Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n+Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n+Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n-Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n-Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n+Error with communicator MPI_COMM_WORLD and datatype MPI_DOUBLE\n Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n MTestTypeSubarrayCheckbuf error\n MTestTypeSubarrayCheckbuf error\n MTestTypeSubarrayCheckbuf error\n MTestTypeSubarrayCheckbuf error\n MTestTypeSubarrayCheckbuf error\n MTestTypeSubarrayCheckbuf error\n@@ -1100,16 +1100,16 @@\n MTestTypeSubarrayCheckbuf error\n MTestTypeSubarrayCheckbuf error\n MTestTypeSubarrayCheckbuf error\n MTestTypeSubarrayCheckbuf error\n MTestTypeSubarrayCheckbuf error\n MTestTypeSubarrayCheckbuf error\n MTestTypeSubarrayCheckbuf error\n-Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n Error with communicator MPI_COMM_WORLD and datatype MPI_INT\n+Error class 15 (MPI_ERR_TRUNCATE: message truncated)\n --------------------------------------------------------------------------\n The user-provided time limit for job execution has been reached:\n \n Timeout: 1200 seconds\n \n The job will now be aborted. Please check your code and/or\n adjust/remove the job execution time limit (as specified by --timeout\n@@ -1651,315 +1651,315 @@\n pass\n \n \n nonblocking\n 2\n ./coll\n fail\n- [profitbricks-build5-amd64:30536] *** Process received signal ***\n-[profitbricks-build5-amd64:30536] Signal: Segmentation fault (11)\n-[profitbricks-build5-amd64:30536] Signal code: (128)\n-[profitbricks-build5-amd64:30536] Failing at address: (nil)\n-[profitbricks-build5-amd64:30536] [ 0] /lib/x86_64-linux-gnu/libpthread.so.0(+0x126b0)[0x7f22e27ca6b0]\n-[profitbricks-build5-amd64:30536] [ 1] /usr/lib/x86_64-linux-gnu/libmpi.so.40(PMPI_Ialltoallw+0x117)[0x7f22e2843287]\n-[profitbricks-build5-amd64:30536] [ 2] ./nonblocking(+0x3b85)[0x559a420b0b85]\n-[profitbricks-build5-amd64:30536] [ 3] /lib/x86_64-linux-gnu/libc.so.6(__libc_start_main+0xeb)[0x7f22e261409b]\n-[profitbricks-build5-amd64:30536] [ 4] ./nonblocking(+0x415a)[0x559a420b115a]\n-[profitbricks-build5-amd64:30536] *** End of error message ***\n-[profitbricks-build5-amd64:30537] *** Process received signal ***\n-[profitbricks-build5-amd64:30537] Signal: Segmentation fault (11)\n-[profitbricks-build5-amd64:30537] Signal code: (128)\n-[profitbricks-build5-amd64:30537] Failing at address: (nil)\n+ [i-capture-the-hostname:37607] *** Process received signal ***\n+[i-capture-the-hostname:37607] Signal: Segmentation fault (11)\n+[i-capture-the-hostname:37607] Signal code: (128)\n+[i-capture-the-hostname:37607] Failing at address: (nil)\n+[i-capture-the-hostname:37609] *** Process received signal ***\n+[i-capture-the-hostname:37609] Signal: Segmentation fault (11)\n+[i-capture-the-hostname:37609] Signal code: (128)\n+[i-capture-the-hostname:37609] Failing at address: (nil)\n+[i-capture-the-hostname:37607] [ 0] /lib/x86_64-linux-gnu/libpthread.so.0(+0x126b0)[0x7f60f867d6b0]\n+[i-capture-the-hostname:37607] [ 1] /usr/lib/x86_64-linux-gnu/libmpi.so.40(PMPI_Ialltoallw+0x117)[0x7f60f86f6287]\n+[i-capture-the-hostname:37607] [ 2] ./nonblocking(+0x3b85)[0x55d5d13c4b85]\n+[i-capture-the-hostname:37607] [ 3] /lib/x86_64-linux-gnu/libc.so.6(__libc_start_main+0xeb)[0x7f60f84c709b]\n+[i-capture-the-hostname:37607] [ 4] ./nonblocking(+0x415a)[0x55d5d13c515a]\n+[i-capture-the-hostname:37607] *** End of error message ***\n+[i-capture-the-hostname:37609] [ 0] /lib/x86_64-linux-gnu/libpthread.so.0(+0x126b0)[0x7f6fb1d956b0]\n+[i-capture-the-hostname:37609] [ 1] /usr/lib/x86_64-linux-gnu/libmpi.so.40(PMPI_Ialltoallw+0x117)[0x7f6fb1e0e287]\n+[i-capture-the-hostname:37609] [ 2] ./nonblocking(+0x3b85)[0x55a43c14fb85]\n+[i-capture-the-hostname:37609] [ 3] /lib/x86_64-linux-gnu/libc.so.6(__libc_start_main+0xeb)[0x7f6fb1bdf09b]\n+[i-capture-the-hostname:37609] [ 4] ./nonblocking(+0x415a)[0x55a43c15015a]\n+[i-capture-the-hostname:37609] *** End of error message ***\n --------------------------------------------------------------------------\n Primary job terminated normally, but 1 process returned\n a non-zero exit code. Per user-direction, the job has been aborted.\n --------------------------------------------------------------------------\n-[profitbricks-build5-amd64:30537] [ 0] /lib/x86_64-linux-gnu/libpthread.so.0(+0x126b0)[0x7fb16074a6b0]\n-[profitbricks-build5-amd64:30537] [ 1] /usr/lib/x86_64-linux-gnu/libmpi.so.40(PMPI_Ialltoallw+0x117)[0x7fb1607c3287]\n-[profitbricks-build5-amd64:30537] [ 2] ./nonblocking(+0x3b85)[0x5602f3c53b85]\n-[profitbricks-build5-amd64:30537] [ 3] /lib/x86_64-linux-gnu/libc.so.6(__libc_start_main+0xeb)[0x7fb16059409b]\n-[profitbricks-build5-amd64:30537] [ 4] ./nonblocking(+0x415a)[0x5602f3c5415a]\n-[profitbricks-build5-amd64:30537] *** End of error message ***\n --------------------------------------------------------------------------\n-mpiexec.openmpi noticed that process rank 0 with PID 0 on node profitbricks-build5-amd64 exited on signal 11 (Segmentation fault).\n+mpiexec.openmpi noticed that process rank 0 with PID 0 on node i-capture-the-hostname exited on signal 11 (Segmentation fault).\n --------------------------------------------------------------------------\n \n \n nonblocking\n 2\n ./coll\n fail\n- [profitbricks-build5-amd64:30785] *** Process received signal ***\n-[profitbricks-build5-amd64:30785] Signal: Segmentation fault (11)\n-[profitbricks-build5-amd64:30785] Signal code: (128)\n-[profitbricks-build5-amd64:30785] Failing at address: (nil)\n-[profitbricks-build5-amd64:30783] *** Process received signal ***\n-[profitbricks-build5-amd64:30783] Signal: Segmentation fault (11)\n-[profitbricks-build5-amd64:30783] Signal code: (128)\n-[profitbricks-build5-amd64:30783] Failing at address: (nil)\n-[profitbricks-build5-amd64:30783] [ 0] /lib/x86_64-linux-gnu/libpthread.so.0(+0x126b0)[0x7fd0c2caa6b0]\n-[profitbricks-build5-amd64:30783] [ 1] /usr/lib/x86_64-linux-gnu/libmpi.so.40(PMPI_Ialltoallw+0x117)[0x7fd0c2d23287]\n-[profitbricks-build5-amd64:30783] [ 2] ./nonblocking(+0x3b85)[0x55e59ba0eb85]\n-[profitbricks-build5-amd64:30783] [ 3] /lib/x86_64-linux-gnu/libc.so.6(__libc_start_main+0xeb)[0x7fd0c2af409b]\n-[profitbricks-build5-amd64:30783] [ 4] ./nonblocking(+0x415a)[0x55e59ba0f15a]\n-[profitbricks-build5-amd64:30783] *** End of error message ***\n+ [i-capture-the-hostname:37754] *** Process received signal ***\n+[i-capture-the-hostname:37754] Signal: Segmentation fault (11)\n+[i-capture-the-hostname:37754] Signal code: (128)\n+[i-capture-the-hostname:37754] Failing at address: (nil)\n+[i-capture-the-hostname:37755] *** Process received signal ***\n+[i-capture-the-hostname:37755] Signal: Segmentation fault (11)\n+[i-capture-the-hostname:37755] Signal code: (128)\n+[i-capture-the-hostname:37755] Failing at address: (nil)\n+[i-capture-the-hostname:37754] [ 0] /lib/x86_64-linux-gnu/libpthread.so.0(+0x126b0)[0x7fadeb1a56b0]\n+[i-capture-the-hostname:37754] [ 1] /usr/lib/x86_64-linux-gnu/libmpi.so.40(PMPI_Ialltoallw+0x117)[0x7fadeb21e287]\n+[i-capture-the-hostname:37754] [ 2] ./nonblocking(+0x3b85)[0x565461425b85]\n+[i-capture-the-hostname:37754] [ 3] /lib/x86_64-linux-gnu/libc.so.6(__libc_start_main+0xeb)[0x7fadeafef09b]\n+[i-capture-the-hostname:37754] [ 4] ./nonblocking(+0x415a)[0x56546142615a]\n+[i-capture-the-hostname:37754] *** End of error message ***\n+[i-capture-the-hostname:37755] [ 0] /lib/x86_64-linux-gnu/libpthread.so.0(+0x126b0)[0x7f05f0ef56b0]\n+[i-capture-the-hostname:37755] [ 1] /usr/lib/x86_64-linux-gnu/libmpi.so.40(PMPI_Ialltoallw+0x117)[0x7f05f0f6e287]\n+[i-capture-the-hostname:37755] [ 2] ./nonblocking(+0x3b85)[0x559ff9929b85]\n+[i-capture-the-hostname:37755] [ 3] /lib/x86_64-linux-gnu/libc.so.6(__libc_start_main+0xeb)[0x7f05f0d3f09b]\n+[i-capture-the-hostname:37755] [ 4] ./nonblocking(+0x415a)[0x559ff992a15a]\n+[i-capture-the-hostname:37755] *** End of error message ***\n --------------------------------------------------------------------------\n Primary job terminated normally, but 1 process returned\n a non-zero exit code. Per user-direction, the job has been aborted.\n --------------------------------------------------------------------------\n-[profitbricks-build5-amd64:30785] [ 0] /lib/x86_64-linux-gnu/libpthread.so.0(+0x126b0)[0x7fde328326b0]\n-[profitbricks-build5-amd64:30785] [ 1] /usr/lib/x86_64-linux-gnu/libmpi.so.40(PMPI_Ialltoallw+0x117)[0x7fde328ab287]\n-[profitbricks-build5-amd64:30785] [ 2] ./nonblocking(+0x3b85)[0x55df699ffb85]\n-[profitbricks-build5-amd64:30785] [ 3] /lib/x86_64-linux-gnu/libc.so.6(__libc_start_main+0xeb)[0x7fde3267c09b]\n-[profitbricks-build5-amd64:30785] [ 4] ./nonblocking(+0x415a)[0x55df69a0015a]\n-[profitbricks-build5-amd64:30785] *** End of error message ***\n --------------------------------------------------------------------------\n-mpiexec.openmpi noticed that process rank 0 with PID 0 on node profitbricks-build5-amd64 exited on signal 11 (Segmentation fault).\n+mpiexec.openmpi noticed that process rank 0 with PID 0 on node i-capture-the-hostname exited on signal 11 (Segmentation fault).\n --------------------------------------------------------------------------\n \n \n nonblocking\n 2\n ./coll\n fail\n- [profitbricks-build5-amd64:30846] *** Process received signal ***\n-[profitbricks-build5-amd64:30846] Signal: Segmentation fault (11)\n-[profitbricks-build5-amd64:30846] Signal code: (128)\n-[profitbricks-build5-amd64:30846] Failing at address: (nil)\n-[profitbricks-build5-amd64:30847] *** Process received signal ***\n-[profitbricks-build5-amd64:30847] Signal: Segmentation fault (11)\n-[profitbricks-build5-amd64:30847] Signal code: (128)\n-[profitbricks-build5-amd64:30847] Failing at address: (nil)\n-[profitbricks-build5-amd64:30846] [ 0] /lib/x86_64-linux-gnu/libpthread.so.0(+0x126b0)[0x7f4cb65826b0]\n-[profitbricks-build5-amd64:30846] [ 1] /usr/lib/x86_64-linux-gnu/libmpi.so.40(PMPI_Ialltoallw+0x117)[0x7f4cb65fb287]\n-[profitbricks-build5-amd64:30846] [ 2] ./nonblocking(+0x3b85)[0x55e34c213b85]\n-[profitbricks-build5-amd64:30846] [ 3] /lib/x86_64-linux-gnu/libc.so.6(__libc_start_main+0xeb)[0x7f4cb63cc09b]\n-[profitbricks-build5-amd64:30846] [ 4] ./nonblocking(+0x415a)[0x55e34c21415a]\n-[profitbricks-build5-amd64:30846] *** End of error message ***\n-[profitbricks-build5-amd64:30847] [ 0] /lib/x86_64-linux-gnu/libpthread.so.0(+0x126b0)[0x7f0c8d5da6b0]\n-[profitbricks-build5-amd64:30847] [ 1] /usr/lib/x86_64-linux-gnu/libmpi.so.40(PMPI_Ialltoallw+0x117)[0x7f0c8d653287]\n-[profitbricks-build5-amd64:30847] [ 2] ./nonblocking(+0x3b85)[0x55886d5e5b85]\n-[profitbricks-build5-amd64:30847] [ 3] /lib/x86_64-linux-gnu/libc.so.6(__libc_start_main+0xeb)[0x7f0c8d42409b]\n-[profitbricks-build5-amd64:30847] [ 4] ./nonblocking(+0x415a)[0x55886d5e615a]\n-[profitbricks-build5-amd64:30847] *** End of error message ***\n+ [i-capture-the-hostname:37870] *** Process received signal ***\n+[i-capture-the-hostname:37870] Signal: Segmentation fault (11)\n+[i-capture-the-hostname:37870] Signal code: (128)\n+[i-capture-the-hostname:37870] Failing at address: (nil)\n+[i-capture-the-hostname:37871] *** Process received signal ***\n+[i-capture-the-hostname:37871] Signal: Segmentation fault (11)\n+[i-capture-the-hostname:37871] Signal code: (128)\n+[i-capture-the-hostname:37871] Failing at address: (nil)\n+[i-capture-the-hostname:37870] [ 0] /lib/x86_64-linux-gnu/libpthread.so.0(+0x126b0)[0x7f392c84d6b0]\n+[i-capture-the-hostname:37870] [ 1] [i-capture-the-hostname:37871] [ 0] /lib/x86_64-linux-gnu/libpthread.so.0(+0x126b0)[0x7fb44a9956b0]\n+[i-capture-the-hostname:37871] [ 1] /usr/lib/x86_64-linux-gnu/libmpi.so.40(PMPI_Ialltoallw+0x117)[0x7fb44aa0e287]\n+[i-capture-the-hostname:37871] [ 2] ./nonblocking(+0x3b85)[0x55e08bbeab85]\n+[i-capture-the-hostname:37871] [ 3] /usr/lib/x86_64-linux-gnu/libmpi.so.40(PMPI_Ialltoallw+0x117)[0x7f392c8c6287]\n+[i-capture-the-hostname:37870] [ 2] ./nonblocking(+0x3b85)[0x555cf20f5b85]\n+[i-capture-the-hostname:37870] [ 3] /lib/x86_64-linux-gnu/libc.so.6(__libc_start_main+0xeb)[0x7fb44a7df09b]\n+[i-capture-the-hostname:37871] [ 4] ./nonblocking(+0x415a)[0x55e08bbeb15a]\n+[i-capture-the-hostname:37871] *** End of error message ***\n+/lib/x86_64-linux-gnu/libc.so.6(__libc_start_main+0xeb)[0x7f392c69709b]\n+[i-capture-the-hostname:37870] [ 4] ./nonblocking(+0x415a)[0x555cf20f615a]\n+[i-capture-the-hostname:37870] *** End of error message ***\n --------------------------------------------------------------------------\n Primary job terminated normally, but 1 process returned\n a non-zero exit code. Per user-direction, the job has been aborted.\n --------------------------------------------------------------------------\n --------------------------------------------------------------------------\n-mpiexec.openmpi noticed that process rank 0 with PID 0 on node profitbricks-build5-amd64 exited on signal 11 (Segmentation fault).\n+mpiexec.openmpi noticed that process rank 0 with PID 0 on node i-capture-the-hostname exited on signal 11 (Segmentation fault).\n --------------------------------------------------------------------------\n \n \n nonblocking2\n 1\n ./coll\n fail\n- [profitbricks-build5-amd64:30894] *** Process received signal ***\n-[profitbricks-build5-amd64:30894] Signal: Segmentation fault (11)\n-[profitbricks-build5-amd64:30894] Signal code: Address not mapped (1)\n-[profitbricks-build5-amd64:30894] Failing at address: 0x7fc404afc450\n-[profitbricks-build5-amd64:30894] [ 0] /lib/x86_64-linux-gnu/libpthread.so.0(+0x126b0)[0x7fc4632b26b0]\n-[profitbricks-build5-amd64:30894] [ 1] /usr/lib/x86_64-linux-gnu/libmpi.so.40(PMPI_Ialltoallw+0x117)[0x7fc46332b287]\n-[profitbricks-build5-amd64:30894] [ 2] ./nonblocking2(+0x4dc3)[0x55aa04aefdc3]\n-[profitbricks-build5-amd64:30894] [ 3] /lib/x86_64-linux-gnu/libc.so.6(__libc_start_main+0xeb)[0x7fc4630fc09b]\n-[profitbricks-build5-amd64:30894] [ 4] ./nonblocking2(+0x50ca)[0x55aa04af00ca]\n-[profitbricks-build5-amd64:30894] *** End of error message ***\n+ [i-capture-the-hostname:37895] *** Process received signal ***\n+[i-capture-the-hostname:37895] Signal: Segmentation fault (11)\n+[i-capture-the-hostname:37895] Signal code: Address not mapped (1)\n+[i-capture-the-hostname:37895] Failing at address: 0x7fe7ac466450\n+[i-capture-the-hostname:37895] [ 0] /lib/x86_64-linux-gnu/libpthread.so.0(+0x126b0)[0x7fe73f7056b0]\n+[i-capture-the-hostname:37895] [ 1] /usr/lib/x86_64-linux-gnu/libmpi.so.40(PMPI_Ialltoallw+0x117)[0x7fe73f77e287]\n+[i-capture-the-hostname:37895] [ 2] ./nonblocking2(+0x4dc3)[0x558fac459dc3]\n+[i-capture-the-hostname:37895] [ 3] /lib/x86_64-linux-gnu/libc.so.6(__libc_start_main+0xeb)[0x7fe73f54f09b]\n+[i-capture-the-hostname:37895] [ 4] ./nonblocking2(+0x50ca)[0x558fac45a0ca]\n+[i-capture-the-hostname:37895] *** End of error message ***\n --------------------------------------------------------------------------\n Primary job terminated normally, but 1 process returned\n a non-zero exit code. Per user-direction, the job has been aborted.\n --------------------------------------------------------------------------\n --------------------------------------------------------------------------\n-mpiexec.openmpi noticed that process rank 0 with PID 0 on node profitbricks-build5-amd64 exited on signal 11 (Segmentation fault).\n+mpiexec.openmpi noticed that process rank 0 with PID 0 on node i-capture-the-hostname exited on signal 11 (Segmentation fault).\n --------------------------------------------------------------------------\n \n \n nonblocking2\n 2\n ./coll\n fail\n- [profitbricks-build5-amd64:31326] *** Process received signal ***\n-[profitbricks-build5-amd64:31326] Signal: Segmentation fault (11)\n-[profitbricks-build5-amd64:31326] Signal code: Address not mapped (1)\n-[profitbricks-build5-amd64:31326] Failing at address: (nil)\n-[profitbricks-build5-amd64:31326] [ 0] /lib/x86_64-linux-gnu/libpthread.so.0(+0x126b0)[0x7f73b474a6b0]\n-[profitbricks-build5-amd64:31326] *** End of error message ***\n+ [i-capture-the-hostname:37937] *** Process received signal ***\n+[i-capture-the-hostname:37937] Signal: Segmentation fault (11)\n+[i-capture-the-hostname:37937] Signal code: Address not mapped (1)\n+[i-capture-the-hostname:37937] Failing at address: (nil)\n+[i-capture-the-hostname:37937] [ 0] /lib/x86_64-linux-gnu/libpthread.so.0(+0x126b0)[0x7fb16f00d6b0]\n+[i-capture-the-hostname:37937] *** End of error message ***\n --------------------------------------------------------------------------\n Primary job terminated normally, but 1 process returned\n a non-zero exit code. Per user-direction, the job has been aborted.\n --------------------------------------------------------------------------\n --------------------------------------------------------------------------\n-mpiexec.openmpi noticed that process rank 0 with PID 0 on node profitbricks-build5-amd64 exited on signal 11 (Segmentation fault).\n+mpiexec.openmpi noticed that process rank 0 with PID 0 on node i-capture-the-hostname exited on signal 11 (Segmentation fault).\n --------------------------------------------------------------------------\n \n \n nonblocking2\n 2\n ./coll\n fail\n- [profitbricks-build5-amd64:31879] *** Process received signal ***\n-[profitbricks-build5-amd64:31879] Signal: Segmentation fault (11)\n-[profitbricks-build5-amd64:31879] Signal code: Address not mapped (1)\n-[profitbricks-build5-amd64:31879] Failing at address: (nil)\n-[profitbricks-build5-amd64:31879] [ 0] /lib/x86_64-linux-gnu/libpthread.so.0(+0x126b0)[0x7f5e5d35a6b0]\n-[profitbricks-build5-amd64:31879] *** End of error message ***\n+ [i-capture-the-hostname:38074] *** Process received signal ***\n+[i-capture-the-hostname:38074] Signal: Segmentation fault (11)\n+[i-capture-the-hostname:38074] Signal code: Address not mapped (1)\n+[i-capture-the-hostname:38074] Failing at address: (nil)\n+[i-capture-the-hostname:38074] [ 0] /lib/x86_64-linux-gnu/libpthread.so.0(+0x126b0)[0x7f28e79956b0]\n+[i-capture-the-hostname:38074] *** End of error message ***\n --------------------------------------------------------------------------\n Primary job terminated normally, but 1 process returned\n a non-zero exit code. Per user-direction, the job has been aborted.\n --------------------------------------------------------------------------\n --------------------------------------------------------------------------\n-mpiexec.openmpi noticed that process rank 0 with PID 0 on node profitbricks-build5-amd64 exited on signal 11 (Segmentation fault).\n+mpiexec.openmpi noticed that process rank 0 with PID 0 on node i-capture-the-hostname exited on signal 11 (Segmentation fault).\n --------------------------------------------------------------------------\n \n \n nonblocking2\n 2\n ./coll\n fail\n- [profitbricks-build5-amd64:32249] *** Process received signal ***\n-[profitbricks-build5-amd64:32249] Signal: Segmentation fault (11)\n-[profitbricks-build5-amd64:32249] Signal code: Address not mapped (1)\n-[profitbricks-build5-amd64:32249] Failing at address: (nil)\n-[profitbricks-build5-amd64:32249] [ 0] /lib/x86_64-linux-gnu/libpthread.so.0(+0x126b0)[0x7f739c3ea6b0]\n-[profitbricks-build5-amd64:32249] *** End of error message ***\n+ [i-capture-the-hostname:38185] *** Process received signal ***\n+[i-capture-the-hostname:38185] Signal: Segmentation fault (11)\n+[i-capture-the-hostname:38185] Signal code: Address not mapped (1)\n+[i-capture-the-hostname:38185] Failing at address: (nil)\n+[i-capture-the-hostname:38185] [ 0] /lib/x86_64-linux-gnu/libpthread.so.0(+0x126b0)[0x7fe1b2c656b0]\n+[i-capture-the-hostname:38185] *** End of error message ***\n --------------------------------------------------------------------------\n Primary job terminated normally, but 1 process returned\n a non-zero exit code. Per user-direction, the job has been aborted.\n --------------------------------------------------------------------------\n --------------------------------------------------------------------------\n-mpiexec.openmpi noticed that process rank 0 with PID 0 on node profitbricks-build5-amd64 exited on signal 11 (Segmentation fault).\n+mpiexec.openmpi noticed that process rank 0 with PID 0 on node i-capture-the-hostname exited on signal 11 (Segmentation fault).\n --------------------------------------------------------------------------\n \n \n nonblocking3\n 1\n ./coll\n fail\n- [profitbricks-build5-amd64:32604] *** Process received signal ***\n-[profitbricks-build5-amd64:32604] Signal: Segmentation fault (11)\n-[profitbricks-build5-amd64:32604] Signal code: Address not mapped (1)\n-[profitbricks-build5-amd64:32604] Failing at address: 0x4137d590\n-[profitbricks-build5-amd64:32604] [ 0] /lib/x86_64-linux-gnu/libpthread.so.0(+0x126b0)[0x7fd8e51026b0]\n-[profitbricks-build5-amd64:32604] [ 1] /usr/lib/x86_64-linux-gnu/libmpi.so.40(PMPI_Ialltoallw+0x117)[0x7fd8e517b287]\n-[profitbricks-build5-amd64:32604] [ 2] ./nonblocking3(+0x5d84)[0x557941370d84]\n-[profitbricks-build5-amd64:32604] [ 3] ./nonblocking3(+0x3820)[0x55794136e820]\n-[profitbricks-build5-amd64:32604] [ 4] /lib/x86_64-linux-gnu/libc.so.6(__libc_start_main+0xeb)[0x7fd8e4f4c09b]\n-[profitbricks-build5-amd64:32604] [ 5] ./nonblocking3(+0x50ba)[0x5579413700ba]\n-[profitbricks-build5-amd64:32604] *** End of error message ***\n+ [i-capture-the-hostname:38277] *** Process received signal ***\n+[i-capture-the-hostname:38277] Signal: Segmentation fault (11)\n+[i-capture-the-hostname:38277] Signal code: Address not mapped (1)\n+[i-capture-the-hostname:38277] Failing at address: 0x60fc6590\n+[i-capture-the-hostname:38277] [ 0] /lib/x86_64-linux-gnu/libpthread.so.0(+0x126b0)[0x7fd76add56b0]\n+[i-capture-the-hostname:38277] [ 1] /usr/lib/x86_64-linux-gnu/libmpi.so.40(PMPI_Ialltoallw+0x117)[0x7fd76ae4e287]\n+[i-capture-the-hostname:38277] [ 2] ./nonblocking3(+0x5d84)[0x55ec60fb9d84]\n+[i-capture-the-hostname:38277] [ 3] ./nonblocking3(+0x3820)[0x55ec60fb7820]\n+[i-capture-the-hostname:38277] [ 4] /lib/x86_64-linux-gnu/libc.so.6(__libc_start_main+0xeb)[0x7fd76ac1f09b]\n+[i-capture-the-hostname:38277] [ 5] ./nonblocking3(+0x50ba)[0x55ec60fb90ba]\n+[i-capture-the-hostname:38277] *** End of error message ***\n --------------------------------------------------------------------------\n Primary job terminated normally, but 1 process returned\n a non-zero exit code. Per user-direction, the job has been aborted.\n --------------------------------------------------------------------------\n --------------------------------------------------------------------------\n-mpiexec.openmpi noticed that process rank 0 with PID 0 on node profitbricks-build5-amd64 exited on signal 11 (Segmentation fault).\n+mpiexec.openmpi noticed that process rank 0 with PID 0 on node i-capture-the-hostname exited on signal 11 (Segmentation fault).\n --------------------------------------------------------------------------\n \n \n nonblocking3\n 2\n ./coll\n fail\n- [profitbricks-build5-amd64:32946] *** Process received signal ***\n-[profitbricks-build5-amd64:32946] Signal: Segmentation fault (11)\n-[profitbricks-build5-amd64:32946] Signal code: (128)\n-[profitbricks-build5-amd64:32946] Failing at address: (nil)\n-[profitbricks-build5-amd64:32946] [ 0] /lib/x86_64-linux-gnu/libpthread.so.0(+0x126b0)[0x7f9c909ea6b0]\n-[profitbricks-build5-amd64:32946] [ 1] /usr/lib/x86_64-linux-gnu/libmpi.so.40(PMPI_Ialltoallw+0x117)[0x7f9c90a63287]\n-[profitbricks-build5-amd64:32946] [ 2] ./nonblocking3(+0x5d84)[0x55c0a8161d84]\n-[profitbricks-build5-amd64:32946] [ 3] ./nonblocking3(+0x3820)[0x55c0a815f820]\n-[profitbricks-build5-amd64:32946] [ 4] /lib/x86_64-linux-gnu/libc.so.6(__libc_start_main+0xeb)[0x7f9c9083409b]\n-[profitbricks-build5-amd64:32946] [ 5] ./nonblocking3(+0x50ba)[0x55c0a81610ba]\n-[profitbricks-build5-amd64:32946] *** End of error message ***\n-[profitbricks-build5-amd64:32947] *** Process received signal ***\n-[profitbricks-build5-amd64:32947] Signal: Segmentation fault (11)\n-[profitbricks-build5-amd64:32947] Signal code: (128)\n-[profitbricks-build5-amd64:32947] Failing at address: (nil)\n-[profitbricks-build5-amd64:32947] [ 0] /lib/x86_64-linux-gnu/libpthread.so.0(+0x126b0)[0x7f54dea026b0]\n-[profitbricks-build5-amd64:32947] [ 1] /usr/lib/x86_64-linux-gnu/libmpi.so.40(PMPI_Ialltoallw+0x117)[0x7f54dea7b287]\n-[profitbricks-build5-amd64:32947] [ 2] ./nonblocking3(+0x5d84)[0x55b22304cd84]\n-[profitbricks-build5-amd64:32947] [ 3] ./nonblocking3(+0x3820)[0x55b22304a820]\n-[profitbricks-build5-amd64:32947] [ 4] /lib/x86_64-linux-gnu/libc.so.6(__libc_start_main+0xeb)[0x7f54de84c09b]\n-[profitbricks-build5-amd64:32947] [ 5] ./nonblocking3(+0x50ba)[0x55b22304c0ba]\n-[profitbricks-build5-amd64:32947] *** End of error message ***\n+ [i-capture-the-hostname:38299] *** Process received signal ***\n+[i-capture-the-hostname:38299] Signal: Segmentation fault (11)\n+[i-capture-the-hostname:38299] Signal code: (128)\n+[i-capture-the-hostname:38299] Failing at address: (nil)\n+[i-capture-the-hostname:38299] [ 0] /lib/x86_64-linux-gnu/libpthread.so.0(+0x126b0)[0x7fc1859cd6b0]\n+[i-capture-the-hostname:38299] [ 1] /usr/lib/x86_64-linux-gnu/libmpi.so.40(PMPI_Ialltoallw+0x117)[0x7fc185a46287]\n+[i-capture-the-hostname:38299] [ 2] ./nonblocking3(+0x5d84)[0x55b7dbc50d84]\n+[i-capture-the-hostname:38299] [ 3] ./nonblocking3(+0x3820)[0x55b7dbc4e820]\n+[i-capture-the-hostname:38299] [ 4] /lib/x86_64-linux-gnu/libc.so.6(__libc_start_main+0xeb)[0x7fc18581709b]\n+[i-capture-the-hostname:38299] [ 5] ./nonblocking3(+0x50ba)[0x55b7dbc500ba]\n+[i-capture-the-hostname:38299] *** End of error message ***\n+[i-capture-the-hostname:38300] *** Process received signal ***\n+[i-capture-the-hostname:38300] Signal: Segmentation fault (11)\n+[i-capture-the-hostname:38300] Signal code: (128)\n+[i-capture-the-hostname:38300] Failing at address: (nil)\n+[i-capture-the-hostname:38300] [ 0] /lib/x86_64-linux-gnu/libpthread.so.0(+0x126b0)[0x7f60b80c56b0]\n+[i-capture-the-hostname:38300] [ 1] /usr/lib/x86_64-linux-gnu/libmpi.so.40(PMPI_Ialltoallw+0x117)[0x7f60b813e287]\n+[i-capture-the-hostname:38300] [ 2] ./nonblocking3(+0x5d84)[0x5600d2373d84]\n+[i-capture-the-hostname:38300] [ 3] ./nonblocking3(+0x3820)[0x5600d2371820]\n+[i-capture-the-hostname:38300] [ 4] /lib/x86_64-linux-gnu/libc.so.6(__libc_start_main+0xeb)[0x7f60b7f0f09b]\n+[i-capture-the-hostname:38300] [ 5] ./nonblocking3(+0x50ba)[0x5600d23730ba]\n+[i-capture-the-hostname:38300] *** End of error message ***\n --------------------------------------------------------------------------\n Primary job terminated normally, but 1 process returned\n a non-zero exit code. Per user-direction, the job has been aborted.\n --------------------------------------------------------------------------\n --------------------------------------------------------------------------\n-mpiexec.openmpi noticed that process rank 0 with PID 0 on node profitbricks-build5-amd64 exited on signal 11 (Segmentation fault).\n+mpiexec.openmpi noticed that process rank 0 with PID 0 on node i-capture-the-hostname exited on signal 11 (Segmentation fault).\n --------------------------------------------------------------------------\n \n \n nonblocking3\n 2\n ./coll\n fail\n- [profitbricks-build5-amd64:33277] *** Process received signal ***\n-[profitbricks-build5-amd64:33277] Signal: Segmentation fault (11)\n-[profitbricks-build5-amd64:33277] Signal code: (128)\n-[profitbricks-build5-amd64:33277] Failing at address: (nil)\n-[profitbricks-build5-amd64:33282] *** Process received signal ***\n-[profitbricks-build5-amd64:33282] Signal: Segmentation fault (11)\n-[profitbricks-build5-amd64:33282] Signal code: (128)\n-[profitbricks-build5-amd64:33282] Failing at address: (nil)\n-[profitbricks-build5-amd64:33277] [ 0] /lib/x86_64-linux-gnu/libpthread.so.0(+0x126b0)[0x7fa10ee726b0]\n-[profitbricks-build5-amd64:33277] [ 1] /usr/lib/x86_64-linux-gnu/libmpi.so.40(PMPI_Ialltoallw+0x117)[0x7fa10eeeb287]\n-[profitbricks-build5-amd64:33277] [ 2] ./nonblocking3(+0x5d84)[0x560e67595d84]\n-[profitbricks-build5-amd64:33277] [ 3] ./nonblocking3(+0x3820)[0x560e67593820]\n-[profitbricks-build5-amd64:33277] [ 4] /lib/x86_64-linux-gnu/libc.so.6(__libc_start_main+0xeb)[0x7fa10ecbc09b]\n-[profitbricks-build5-amd64:33277] [ 5] ./nonblocking3(+0x50ba)[0x560e675950ba]\n-[profitbricks-build5-amd64:33277] *** End of error message ***\n-[profitbricks-build5-amd64:33282] [ 0] /lib/x86_64-linux-gnu/libpthread.so.0(+0x126b0)[0x7f57161326b0]\n-[profitbricks-build5-amd64:33282] [ 1] /usr/lib/x86_64-linux-gnu/libmpi.so.40(PMPI_Ialltoallw+0x117)[0x7f57161ab287]\n-[profitbricks-build5-amd64:33282] [ 2] ./nonblocking3(+0x5d84)[0x55f7faef6d84]\n-[profitbricks-build5-amd64:33282] [ 3] ./nonblocking3(+0x3820)[0x55f7faef4820]\n-[profitbricks-build5-amd64:33282] [ 4] /lib/x86_64-linux-gnu/libc.so.6(__libc_start_main+0xeb)[0x7f5715f7c09b]\n-[profitbricks-build5-amd64:33282] [ 5] ./nonblocking3(+0x50ba)[0x55f7faef60ba]\n-[profitbricks-build5-amd64:33282] *** End of error message ***\n+ [i-capture-the-hostname:38313] *** Process received signal ***\n+[i-capture-the-hostname:38313] Signal: Segmentation fault (11)\n+[i-capture-the-hostname:38313] Signal code: (128)\n+[i-capture-the-hostname:38313] Failing at address: (nil)\n+[i-capture-the-hostname:38314] *** Process received signal ***\n+[i-capture-the-hostname:38314] Signal: Segmentation fault (11)\n+[i-capture-the-hostname:38314] Signal code: (128)\n+[i-capture-the-hostname:38314] Failing at address: (nil)\n+[i-capture-the-hostname:38313] [ 0] /lib/x86_64-linux-gnu/libpthread.so.0(+0x126b0)[0x7fd4ef1cd6b0]\n+[i-capture-the-hostname:38313] [ 1] /usr/lib/x86_64-linux-gnu/libmpi.so.40(PMPI_Ialltoallw+0x117)[0x7fd4ef246287]\n+[i-capture-the-hostname:38313] [ 2] ./nonblocking3(+0x5d84)[0x55bbd8de9d84]\n+[i-capture-the-hostname:38313] [ 3] ./nonblocking3(+0x3820)[0x55bbd8de7820]\n+[i-capture-the-hostname:38313] [ 4] /lib/x86_64-linux-gnu/libc.so.6(__libc_start_main+0xeb)[0x7fd4ef01709b]\n+[i-capture-the-hostname:38313] [ 5] ./nonblocking3(+0x50ba)[0x55bbd8de90ba]\n+[i-capture-the-hostname:38313] *** End of error message ***\n+[i-capture-the-hostname:38314] [ 0] /lib/x86_64-linux-gnu/libpthread.so.0(+0x126b0)[0x7f87432956b0]\n+[i-capture-the-hostname:38314] [ 1] /usr/lib/x86_64-linux-gnu/libmpi.so.40(PMPI_Ialltoallw+0x117)[0x7f874330e287]\n+[i-capture-the-hostname:38314] [ 2] ./nonblocking3(+0x5d84)[0x562cb341bd84]\n+[i-capture-the-hostname:38314] [ 3] ./nonblocking3(+0x3820)[0x562cb3419820]\n+[i-capture-the-hostname:38314] [ 4] /lib/x86_64-linux-gnu/libc.so.6(__libc_start_main+0xeb)[0x7f87430df09b]\n+[i-capture-the-hostname:38314] [ 5] ./nonblocking3(+0x50ba)[0x562cb341b0ba]\n+[i-capture-the-hostname:38314] *** End of error message ***\n --------------------------------------------------------------------------\n Primary job terminated normally, but 1 process returned\n a non-zero exit code. Per user-direction, the job has been aborted.\n --------------------------------------------------------------------------\n --------------------------------------------------------------------------\n-mpiexec.openmpi noticed that process rank 0 with PID 0 on node profitbricks-build5-amd64 exited on signal 11 (Segmentation fault).\n+mpiexec.openmpi noticed that process rank 0 with PID 0 on node i-capture-the-hostname exited on signal 11 (Segmentation fault).\n --------------------------------------------------------------------------\n \n \n nonblocking3\n 2\n ./coll\n fail\n- [profitbricks-build5-amd64:33779] *** Process received signal ***\n-[profitbricks-build5-amd64:33779] Signal: Segmentation fault (11)\n-[profitbricks-build5-amd64:33779] Signal code: (128)\n-[profitbricks-build5-amd64:33779] Failing at address: (nil)\n-[profitbricks-build5-amd64:33782] *** Process received signal ***\n-[profitbricks-build5-amd64:33782] Signal: Segmentation fault (11)\n-[profitbricks-build5-amd64:33782] Signal code: (128)\n-[profitbricks-build5-amd64:33782] Failing at address: (nil)\n-[profitbricks-build5-amd64:33779] [ 0] /lib/x86_64-linux-gnu/libpthread.so.0(+0x126b0)[0x7f994932a6b0]\n-[profitbricks-build5-amd64:33779] [ 1] /usr/lib/x86_64-linux-gnu/libmpi.so.40(PMPI_Ialltoallw+0x117)[0x7f99493a3287]\n-[profitbricks-build5-amd64:33779] [ 2] ./nonblocking3(+0x5d84)[0x56021c71bd84]\n-[profitbricks-build5-amd64:33779] [ 3] ./nonblocking3(+0x3820)[0x56021c719820]\n-[profitbricks-build5-amd64:33779] [ 4] /lib/x86_64-linux-gnu/libc.so.6(__libc_start_main+0xeb)[0x7f994917409b]\n-[profitbricks-build5-amd64:33779] [ 5] ./nonblocking3(+0x50ba)[0x56021c71b0ba]\n-[profitbricks-build5-amd64:33779] *** End of error message ***\n-[profitbricks-build5-amd64:33782] [ 0] /lib/x86_64-linux-gnu/libpthread.so.0(+0x126b0)[0x7fe0b826a6b0]\n-[profitbricks-build5-amd64:33782] [ 1] /usr/lib/x86_64-linux-gnu/libmpi.so.40(PMPI_Ialltoallw+0x117)[0x7fe0b82e3287]\n-[profitbricks-build5-amd64:33782] [ 2] ./nonblocking3(+0x5d84)[0x55da0ae00d84]\n-[profitbricks-build5-amd64:33782] [ 3] ./nonblocking3(+0x3820)[0x55da0adfe820]\n-[profitbricks-build5-amd64:33782] [ 4] /lib/x86_64-linux-gnu/libc.so.6(__libc_start_main+0xeb)[0x7fe0b80b409b]\n-[profitbricks-build5-amd64:33782] [ 5] ./nonblocking3(+0x50ba)[0x55da0ae000ba]\n-[profitbricks-build5-amd64:33782] *** End of error message ***\n+ [i-capture-the-hostname:38338] *** Process received signal ***\n+[i-capture-the-hostname:38338] Signal: Segmentation fault (11)\n+[i-capture-the-hostname:38338] Signal code: (128)\n+[i-capture-the-hostname:38338] Failing at address: (nil)\n+[i-capture-the-hostname:38338] [ 0] /lib/x86_64-linux-gnu/libpthread.so.0(+0x126b0)[0x7f6100d556b0]\n+[i-capture-the-hostname:38338] [ 1] /usr/lib/x86_64-linux-gnu/libmpi.so.40(PMPI_Ialltoallw+0x117)[0x7f6100dce287]\n+[i-capture-the-hostname:38338] [ 2] ./nonblocking3(+0x5d84)[0x55867f79ed84]\n+[i-capture-the-hostname:38338] [ 3] ./nonblocking3(+0x3820)[0x55867f79c820]\n+[i-capture-the-hostname:38338] [ 4] /lib/x86_64-linux-gnu/libc.so.6(__libc_start_main+0xeb)[0x7f6100b9f09b]\n+[i-capture-the-hostname:38338] [ 5] ./nonblocking3(+0x50ba)[0x55867f79e0ba]\n+[i-capture-the-hostname:38338] *** End of error message ***\n+[i-capture-the-hostname:38339] *** Process received signal ***\n+[i-capture-the-hostname:38339] Signal: Segmentation fault (11)\n+[i-capture-the-hostname:38339] Signal code: (128)\n+[i-capture-the-hostname:38339] Failing at address: (nil)\n+[i-capture-the-hostname:38339] [ 0] /lib/x86_64-linux-gnu/libpthread.so.0(+0x126b0)[0x7f0be281d6b0]\n+[i-capture-the-hostname:38339] [ 1] /usr/lib/x86_64-linux-gnu/libmpi.so.40(PMPI_Ialltoallw+0x117)[0x7f0be2896287]\n+[i-capture-the-hostname:38339] [ 2] ./nonblocking3(+0x5d84)[0x555ef259ed84]\n+[i-capture-the-hostname:38339] [ 3] ./nonblocking3(+0x3820)[0x555ef259c820]\n+[i-capture-the-hostname:38339] [ 4] /lib/x86_64-linux-gnu/libc.so.6(__libc_start_main+0xeb)[0x7f0be266709b]\n+[i-capture-the-hostname:38339] [ 5] ./nonblocking3(+0x50ba)[0x555ef259e0ba]\n+[i-capture-the-hostname:38339] *** End of error message ***\n --------------------------------------------------------------------------\n Primary job terminated normally, but 1 process returned\n a non-zero exit code. Per user-direction, the job has been aborted.\n --------------------------------------------------------------------------\n --------------------------------------------------------------------------\n-mpiexec.openmpi noticed that process rank 0 with PID 0 on node profitbricks-build5-amd64 exited on signal 11 (Segmentation fault).\n+mpiexec.openmpi noticed that process rank 0 with PID 0 on node i-capture-the-hostname exited on signal 11 (Segmentation fault).\n --------------------------------------------------------------------------\n \n \n iallred\n 2\n ./coll\n pass\n@@ -2097,22 +2097,22 @@\n pass\n \n \n commcreate1\n 2\n ./comm\n fail\n- [profitbricks-build5-amd64:55812] *** An error occurred in MPI_Group_range_incl\n-[profitbricks-build5-amd64:55812] *** reported by process [497090561,0]\n-[profitbricks-build5-amd64:55812] *** on communicator MPI_COMM_WORLD\n-[profitbricks-build5-amd64:55812] *** MPI_ERR_RANK: invalid rank\n-[profitbricks-build5-amd64:55812] *** MPI_ERRORS_ARE_FATAL (processes in this communicator will now abort,\n-[profitbricks-build5-amd64:55812] *** and potentially your MPI job)\n-[profitbricks-build5-amd64:55667] 1 more process has sent help message help-mpi-errors.txt / mpi_errors_are_fatal\n-[profitbricks-build5-amd64:55667] Set MCA parameter "orte_base_help_aggregate" to 0 to see all help / error messages\n+ [i-capture-the-hostname:39200] *** An error occurred in MPI_Group_range_incl\n+[i-capture-the-hostname:39200] *** reported by process [340393985,0]\n+[i-capture-the-hostname:39200] *** on communicator MPI_COMM_WORLD\n+[i-capture-the-hostname:39200] *** MPI_ERR_RANK: invalid rank\n+[i-capture-the-hostname:39200] *** MPI_ERRORS_ARE_FATAL (processes in this communicator will now abort,\n+[i-capture-the-hostname:39200] *** and potentially your MPI job)\n+[i-capture-the-hostname:39195] 1 more process has sent help message help-mpi-errors.txt / mpi_errors_are_fatal\n+[i-capture-the-hostname:39195] Set MCA parameter "orte_base_help_aggregate" to 0 to see all help / error messages\n \n \n commname\n 2\n ./comm\n pass\n \n@@ -2133,16 +2133,16 @@\n MPI_ABORT was invoked on rank 0 in communicator MPI_COMM_WORLD\n with errorcode 1.\n \n NOTE: invoking MPI_ABORT causes Open MPI to kill all MPI processes.\n You may or may not see output from other processes, depending on\n exactly when Open MPI kills them.\n --------------------------------------------------------------------------\n-[profitbricks-build5-amd64:57101] 1 more process has sent help message help-mpi-api.txt / mpi-abort\n-[profitbricks-build5-amd64:57101] Set MCA parameter "orte_base_help_aggregate" to 0 to see all help / error messages\n+[i-capture-the-hostname:39256] 1 more process has sent help message help-mpi-api.txt / mpi-abort\n+[i-capture-the-hostname:39256] Set MCA parameter "orte_base_help_aggregate" to 0 to see all help / error messages\n \n \n icgroup\n 2\n ./comm\n pass\n \n@@ -2163,16 +2163,16 @@\n MPI_ABORT was invoked on rank 1 in communicator MPI_COMM_WORLD\n with errorcode 1.\n \n NOTE: invoking MPI_ABORT causes Open MPI to kill all MPI processes.\n You may or may not see output from other processes, depending on\n exactly when Open MPI kills them.\n --------------------------------------------------------------------------\n-[profitbricks-build5-amd64:58296] 1 more process has sent help message help-mpi-api.txt / mpi-abort\n-[profitbricks-build5-amd64:58296] Set MCA parameter "orte_base_help_aggregate" to 0 to see all help / error messages\n+[i-capture-the-hostname:39298] 1 more process has sent help message help-mpi-api.txt / mpi-abort\n+[i-capture-the-hostname:39298] Set MCA parameter "orte_base_help_aggregate" to 0 to see all help / error messages\n \n \n iccreate\n 2\n ./comm\n fail\n This test requires at least 4 processes\n@@ -2181,16 +2181,16 @@\n MPI_ABORT was invoked on rank 1 in communicator MPI_COMM_WORLD\n with errorcode 1.\n \n NOTE: invoking MPI_ABORT causes Open MPI to kill all MPI processes.\n You may or may not see output from other processes, depending on\n exactly when Open MPI kills them.\n --------------------------------------------------------------------------\n-[profitbricks-build5-amd64:58904] 1 more process has sent help message help-mpi-api.txt / mpi-abort\n-[profitbricks-build5-amd64:58904] Set MCA parameter "orte_base_help_aggregate" to 0 to see all help / error messages\n+[i-capture-the-hostname:39312] 1 more process has sent help message help-mpi-api.txt / mpi-abort\n+[i-capture-the-hostname:39312] Set MCA parameter "orte_base_help_aggregate" to 0 to see all help / error messages\n \n \n ctxalloc\n 2\n ./comm\n pass\n \n@@ -2215,16 +2215,16 @@\n MPI_ABORT was invoked on rank 0 in communicator MPI_COMM_WORLD\n with errorcode 1.\n \n NOTE: invoking MPI_ABORT causes Open MPI to kill all MPI processes.\n You may or may not see output from other processes, depending on\n exactly when Open MPI kills them.\n --------------------------------------------------------------------------\n-[profitbricks-build5-amd64:63276] 1 more process has sent help message help-mpi-api.txt / mpi-abort\n-[profitbricks-build5-amd64:63276] Set MCA parameter "orte_base_help_aggregate" to 0 to see all help / error messages\n+[i-capture-the-hostname:39425] 1 more process has sent help message help-mpi-api.txt / mpi-abort\n+[i-capture-the-hostname:39425] Set MCA parameter "orte_base_help_aggregate" to 0 to see all help / error messages\n \n \n cmsplit2\n 2\n ./comm\n pass\n \n@@ -2445,18 +2445,18 @@\n pass\n \n \n typename\n 1\n ./datatype\n fail\n- Expected MPI_C_FLOAT_COMPLEX but got MPI_C_COMPLEX\n+ Found 3 errors\n+Expected MPI_C_FLOAT_COMPLEX but got MPI_C_COMPLEX\n Expected MPI_LONG_LONG but got MPI_LONG_LONG_INT\n-MPI_INTEGER16 is not available\n- Found 3 errors\n+MPI_INTEGER16 is not available\n \n \n typefree\n 1\n ./datatype\n pass\n \n@@ -2600,18 +2600,18 @@\n pass\n \n \n struct-no-real-types\n 1\n ./datatype\n fail\n- error: extent is 9223372036854775798 but should be -10 in no_real_types_test()\n+ Found 1 errors\n+error: extent is 9223372036854775798 but should be -10 in no_real_types_test()\n type map is { (LB,10) }, so UB is 0 and extent is ub-lb\n-1 errors in blockindexed test.\n- Found 1 errors\n+1 errors in blockindexed test.\n \n \n struct-empty-el\n 1\n ./datatype\n pass\n \n@@ -2807,62 +2807,62 @@\n pass\n \n \n @groupcreate\n \n ./group\n fail\n- Failed to build @groupcreate; make[2]: Entering directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/group'\n+ Failed to build @groupcreate; make[2]: Entering directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/group'\n make[2]: *** No rule to make target '@groupcreate'. Stop.\n-make[2]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/group'\n+make[2]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/group'\n \n \n @grouptest\n \n ./group\n fail\n- Failed to build @grouptest; make[2]: Entering directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/group'\n+ Failed to build @grouptest; make[2]: Entering directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/group'\n make[2]: *** No rule to make target '@grouptest'. Stop.\n-make[2]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/group'\n+make[2]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/group'\n \n \n @grouptest2\n \n ./group\n fail\n- Failed to build @grouptest2; make[2]: Entering directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/group'\n+ Failed to build @grouptest2; make[2]: Entering directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/group'\n make[2]: *** No rule to make target '@grouptest2'. Stop.\n-make[2]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/group'\n+make[2]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/group'\n \n \n @groupnullincl\n \n ./group\n fail\n- Failed to build @groupnullincl; make[2]: Entering directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/group'\n+ Failed to build @groupnullincl; make[2]: Entering directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/group'\n make[2]: *** No rule to make target '@groupnullincl'. Stop.\n-make[2]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/group'\n+make[2]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/group'\n \n \n @gtranks\n \n ./group\n fail\n- Failed to build @gtranks; make[2]: Entering directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/group'\n+ Failed to build @gtranks; make[2]: Entering directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/group'\n make[2]: *** No rule to make target '@gtranks'. Stop.\n-make[2]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/group'\n+make[2]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/group'\n \n \n @gtranksperf\n \n ./group\n fail\n- Failed to build @gtranksperf; make[2]: Entering directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/group'\n+ Failed to build @gtranksperf; make[2]: Entering directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/group'\n make[2]: *** No rule to make target '@gtranksperf'. Stop.\n-make[2]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/group'\n+make[2]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/group'\n \n \n infodup\n 1\n ./info\n pass\n \n@@ -2972,47 +2972,47 @@\n pass\n \n \n mpi_t_str\n 1\n ./mpi_t\n fail\n- check failed: ((strlen(desc) + 1) == min(desc_len, STR_SZ)), line 88\n+ found 1464 errors\n check failed: ((strlen(desc) + 1) == min(desc_len, STR_SZ)), line 88\n check failed: ((strlen(desc) + 1) == min(desc_len, STR_SZ)), line 88\n check failed: ((strlen(desc) + 1) == min(desc_len, STR_SZ)), line 88\n check failed: ((strlen(desc) + 1) == min(desc_len, STR_SZ)), line 88\n check failed: ((strlen(desc) + 1) == min(desc_len, STR_SZ)), line 88\n check failed: ((strlen(desc) + 1) == min(desc_len, STR_SZ)), line 88\n check failed: ((strlen(desc) + 1) == min(desc_len, STR_SZ)), line 88\n check failed: ((strlen(desc) + 1) == min(desc_len, STR_SZ)), line 88\n-found 1464 errors\n+check failed: ((strlen(desc) + 1) == min(desc_len, STR_SZ)), line 88\n \n \n mpit_vars\n 1\n ./mpi_t\n fail\n- [profitbricks-build5-amd64:45361] *** Process received signal ***\n-[profitbricks-build5-amd64:45361] Signal: Segmentation fault (11)\n-[profitbricks-build5-amd64:45361] Signal code: Address not mapped (1)\n-[profitbricks-build5-amd64:45361] Failing at address: (nil)\n-[profitbricks-build5-amd64:45361] [ 0] /lib/x86_64-linux-gnu/libpthread.so.0(+0x126b0)[0x7fe71f54a6b0]\n-[profitbricks-build5-amd64:45361] [ 1] /usr/lib/x86_64-linux-gnu/libmpi.so.40(PMPI_T_cvar_read+0x4e)[0x7fe71f5f485e]\n-[profitbricks-build5-amd64:45361] [ 2] ./mpit_vars(+0x457f)[0x55dbdd7d757f]\n-[profitbricks-build5-amd64:45361] [ 3] ./mpit_vars(+0x36b5)[0x55dbdd7d66b5]\n-[profitbricks-build5-amd64:45361] [ 4] /lib/x86_64-linux-gnu/libc.so.6(__libc_start_main+0xeb)[0x7fe71f39409b]\n-[profitbricks-build5-amd64:45361] [ 5] ./mpit_vars(+0x377a)[0x55dbdd7d677a]\n-[profitbricks-build5-amd64:45361] *** End of error message ***\n+ [i-capture-the-hostname:49703] *** Process received signal ***\n+[i-capture-the-hostname:49703] Signal: Segmentation fault (11)\n+[i-capture-the-hostname:49703] Signal code: Address not mapped (1)\n+[i-capture-the-hostname:49703] Failing at address: (nil)\n+[i-capture-the-hostname:49703] [ 0] /lib/x86_64-linux-gnu/libpthread.so.0(+0x126b0)[0x7f0df58356b0]\n+[i-capture-the-hostname:49703] [ 1] /usr/lib/x86_64-linux-gnu/libmpi.so.40(PMPI_T_cvar_read+0x4e)[0x7f0df58df85e]\n+[i-capture-the-hostname:49703] [ 2] ./mpit_vars(+0x457f)[0x5590e91d957f]\n+[i-capture-the-hostname:49703] [ 3] ./mpit_vars(+0x36b5)[0x5590e91d86b5]\n+[i-capture-the-hostname:49703] [ 4] /lib/x86_64-linux-gnu/libc.so.6(__libc_start_main+0xeb)[0x7f0df567f09b]\n+[i-capture-the-hostname:49703] [ 5] ./mpit_vars(+0x377a)[0x5590e91d877a]\n+[i-capture-the-hostname:49703] *** End of error message ***\n --------------------------------------------------------------------------\n Primary job terminated normally, but 1 process returned\n a non-zero exit code. Per user-direction, the job has been aborted.\n --------------------------------------------------------------------------\n --------------------------------------------------------------------------\n-mpiexec.openmpi noticed that process rank 0 with PID 0 on node profitbricks-build5-amd64 exited on signal 11 (Segmentation fault).\n+mpiexec.openmpi noticed that process rank 0 with PID 0 on node i-capture-the-hostname exited on signal 11 (Segmentation fault).\n --------------------------------------------------------------------------\n \n \n cvarwrite\n 1\n ./mpi_t\n pass\n@@ -3317,23 +3317,23 @@\n adlb_mimic1\n 2\n ./rma\n fail\n This test requires at least 3 processes\n This test requires at least 3 processes\n --------------------------------------------------------------------------\n-MPI_ABORT was invoked on rank 1 in communicator MPI_COMM_WORLD\n+MPI_ABORT was invoked on rank 0 in communicator MPI_COMM_WORLD\n with errorcode 1.\n \n NOTE: invoking MPI_ABORT causes Open MPI to kill all MPI processes.\n You may or may not see output from other processes, depending on\n exactly when Open MPI kills them.\n --------------------------------------------------------------------------\n-[profitbricks-build5-amd64:38388] 1 more process has sent help message help-mpi-api.txt / mpi-abort\n-[profitbricks-build5-amd64:38388] Set MCA parameter "orte_base_help_aggregate" to 0 to see all help / error messages\n+[i-capture-the-hostname:33131] 1 more process has sent help message help-mpi-api.txt / mpi-abort\n+[i-capture-the-hostname:33131] Set MCA parameter "orte_base_help_aggregate" to 0 to see all help / error messages\n \n \n accfence2\n 2\n ./rma\n pass\n \n@@ -3450,16 +3450,16 @@\n MPI_ABORT was invoked on rank 1 in communicator MPI_COMM_WORLD\n with errorcode 1.\n \n NOTE: invoking MPI_ABORT causes Open MPI to kill all MPI processes.\n You may or may not see output from other processes, depending on\n exactly when Open MPI kills them.\n --------------------------------------------------------------------------\n-[profitbricks-build5-amd64:51493] 1 more process has sent help message help-mpi-api.txt / mpi-abort\n-[profitbricks-build5-amd64:51493] Set MCA parameter "orte_base_help_aggregate" to 0 to see all help / error messages\n+[i-capture-the-hostname:36578] 1 more process has sent help message help-mpi-api.txt / mpi-abort\n+[i-capture-the-hostname:36578] Set MCA parameter "orte_base_help_aggregate" to 0 to see all help / error messages\n \n \n lockcontention2\n 2\n ./rma\n pass\n \n@@ -3473,39 +3473,43 @@\n lockcontention3\n 2\n ./rma\n fail\n This test requires at least 3 processes\n This test requires at least 3 processes\n --------------------------------------------------------------------------\n-MPI_ABORT was invoked on rank 1 in communicator MPI_COMM_WORLD\n+MPI_ABORT was invoked on rank 0 in communicator MPI_COMM_WORLD\n with errorcode 1.\n \n NOTE: invoking MPI_ABORT causes Open MPI to kill all MPI processes.\n You may or may not see output from other processes, depending on\n exactly when Open MPI kills them.\n --------------------------------------------------------------------------\n-[profitbricks-build5-amd64:52424] 1 more process has sent help message help-mpi-api.txt / mpi-abort\n-[profitbricks-build5-amd64:52424] Set MCA parameter "orte_base_help_aggregate" to 0 to see all help / error messages\n+[i-capture-the-hostname:37006] 1 more process has sent help message help-mpi-api.txt / mpi-abort\n+[i-capture-the-hostname:37006] Set MCA parameter "orte_base_help_aggregate" to 0 to see all help / error messages\n \n \n lockopts\n 2\n ./rma\n pass\n \n \n lock_dt\n 2\n ./rma\n fail\n- malloc(): corrupted top size\n-[profitbricks-build5-amd64:53184] *** Process received signal ***\n-[profitbricks-build5-amd64:53184] Signal: Aborted (6)\n-[profitbricks-build5-amd64:53184] Signal code: (-6)\n+ [i-capture-the-hostname:37260] *** Process received signal ***\n+[i-capture-the-hostname:37260] Signal: Segmentation fault (11)\n+[i-capture-the-hostname:37260] Signal code: Address not mapped (1)\n+[i-capture-the-hostname:37260] Failing at address: 0x55d831d1e0c0\n+malloc(): corrupted top size\n+[i-capture-the-hostname:37260] *** Process received signal ***\n+[i-capture-the-hostname:37260] Signal: Aborted (6)\n+[i-capture-the-hostname:37260] Signal code: (-6)\n --------------------------------------------------------------------------\n The user-provided time limit for job execution has been reached:\n \n Timeout: 180 seconds\n \n The job will now be aborted. Please check your code and/or\n adjust/remove the job execution time limit (as specified by --timeout\n@@ -3513,211 +3517,228 @@\n --------------------------------------------------------------------------\n \n \n lock_dt_flush\n 2\n ./rma\n fail\n- [profitbricks-build5-amd64:51145] *** Process received signal ***\n-[profitbricks-build5-amd64:51145] Signal: Segmentation fault (11)\n-[profitbricks-build5-amd64:51145] Signal code: Address not mapped (1)\n-[profitbricks-build5-amd64:51145] Failing at address: 0x55bcbd5cb3e0\n-[profitbricks-build5-amd64:51145] [ 0] /lib/x86_64-linux-gnu/libpthread.so.0(+0x126b0)[0x7f55c0dba6b0]\n-[profitbricks-build5-amd64:51145] [ 1] /lib/x86_64-linux-gnu/libc.so.6(+0x13ad67)[0x7f55c0d1ad67]\n-[profitbricks-build5-amd64:51145] [ 2] /usr/lib/x86_64-linux-gnu/libopen-pal.so.40(opal_generic_simple_unpack+0x4e6)[0x7f55c0ab4596]\n-[profitbricks-build5-amd64:51145] [ 3] /usr/lib/x86_64-linux-gnu/openmpi/lib/openmpi3/mca_pml_ob1.so(mca_pml_ob1_recv_request_progress_frag+0x1c2)[0x7f55be1e8b62]\n-[profitbricks-build5-amd64:51145] [ 4] /usr/lib/x86_64-linux-gnu/openmpi/lib/openmpi3/mca_btl_vader.so(mca_btl_vader_poll_handle_frag+0x8f)[0x7f55be2246bf]\n-[profitbricks-build5-amd64:51145] [ 5] /usr/lib/x86_64-linux-gnu/openmpi/lib/openmpi3/mca_btl_vader.so(+0x4b8a)[0x7f55be224b8a]\n-[profitbricks-build5-amd64:51145] [ 6] /usr/lib/x86_64-linux-gnu/libopen-pal.so.40(opal_progress+0x2c)[0x7f55c0a98e0c]\n-[profitbricks-build5-amd64:51145] [ 7] /usr/lib/x86_64-linux-gnu/libopen-pal.so.40(ompi_sync_wait_mt+0x115)[0x7f55c0a9f835]\n-[profitbricks-build5-amd64:51145] [ 8] /usr/lib/x86_64-linux-gnu/libmpi.so.40(ompi_request_default_wait+0x1d9)[0x7f55c0e1a9b9]\n-[profitbricks-build5-amd64:51145] [ 9] /usr/lib/x86_64-linux-gnu/libmpi.so.40(ompi_coll_base_barrier_intra_two_procs+0xea)[0x7f55c0e77c8a]\n-[profitbricks-build5-amd64:51145] [10] /usr/lib/x86_64-linux-gnu/libmpi.so.40(MPI_Barrier+0xa8)[0x7f55c0e33978]\n-[profitbricks-build5-amd64:51145] [11] ./lock_dt_flush(+0x386b)[0x55bcbbc8386b]\n-[profitbricks-build5-amd64:51145] [12] /lib/x86_64-linux-gnu/libc.so.6(__libc_start_main+0xeb)[0x7f55c0c0409b]\n-[profitbricks-build5-amd64:51145] [13] ./lock_dt_flush(+0x393a)[0x55bcbbc8393a]\n-[profitbricks-build5-amd64:51145] *** End of error message ***\n+ [i-capture-the-hostname:50873] *** Process received signal ***\n+[i-capture-the-hostname:50873] Signal: Segmentation fault (11)\n+[i-capture-the-hostname:50873] Signal code: Address not mapped (1)\n+[i-capture-the-hostname:50873] Failing at address: 0x562662628260\n+[i-capture-the-hostname:50873] [ 0] /lib/x86_64-linux-gnu/libpthread.so.0(+0x126b0)[0x7f386cd7d6b0]\n+[i-capture-the-hostname:50873] [ 1] /lib/x86_64-linux-gnu/libc.so.6(+0x13ad67)[0x7f386ccddd67]\n+[i-capture-the-hostname:50873] [ 2] /usr/lib/x86_64-linux-gnu/libopen-pal.so.40(opal_generic_simple_unpack+0x4e6)[0x7f386ca77596]\n+[i-capture-the-hostname:50873] [ 3] /usr/lib/x86_64-linux-gnu/openmpi/lib/openmpi3/mca_pml_ob1.so(mca_pml_ob1_recv_request_progress_frag+0x1c2)[0x7f386a1abb62]\n+[i-capture-the-hostname:50873] [ 4] /usr/lib/x86_64-linux-gnu/openmpi/lib/openmpi3/mca_btl_vader.so(mca_btl_vader_poll_handle_frag+0x8f)[0x7f386a1e76bf]\n+[i-capture-the-hostname:50873] [ 5] /usr/lib/x86_64-linux-gnu/openmpi/lib/openmpi3/mca_btl_vader.so(+0x4b8a)[0x7f386a1e7b8a]\n+[i-capture-the-hostname:50873] [ 6] /usr/lib/x86_64-linux-gnu/libopen-pal.so.40(opal_progress+0x2c)[0x7f386ca5be0c]\n+[i-capture-the-hostname:50873] [ 7] /usr/lib/x86_64-linux-gnu/libopen-pal.so.40(ompi_sync_wait_mt+0x115)[0x7f386ca62835]\n+[i-capture-the-hostname:50873] [ 8] /usr/lib/x86_64-linux-gnu/libmpi.so.40(ompi_request_default_wait+0x1d9)[0x7f386cddd9b9]\n+[i-capture-the-hostname:50873] [ 9] /usr/lib/x86_64-linux-gnu/libmpi.so.40(ompi_coll_base_barrier_intra_two_procs+0xea)[0x7f386ce3ac8a]\n+[i-capture-the-hostname:50873] [10] /usr/lib/x86_64-linux-gnu/libmpi.so.40(MPI_Barrier+0xa8)[0x7f386cdf6978]\n+[i-capture-the-hostname:50873] [11] ./lock_dt_flush(+0x386b)[0x5626603df86b]\n+[i-capture-the-hostname:50873] [12] /lib/x86_64-linux-gnu/libc.so.6(__libc_start_main+0xeb)[0x7f386cbc709b]\n+[i-capture-the-hostname:50873] [13] ./lock_dt_flush(+0x393a)[0x5626603df93a]\n+[i-capture-the-hostname:50873] *** End of error message ***\n --------------------------------------------------------------------------\n Primary job terminated normally, but 1 process returned\n a non-zero exit code. Per user-direction, the job has been aborted.\n --------------------------------------------------------------------------\n --------------------------------------------------------------------------\n-mpiexec.openmpi noticed that process rank 1 with PID 0 on node profitbricks-build5-amd64 exited on signal 11 (Segmentation fault).\n+mpiexec.openmpi noticed that process rank 1 with PID 0 on node i-capture-the-hostname exited on signal 11 (Segmentation fault).\n --------------------------------------------------------------------------\n \n \n lock_dt_flushlocal\n 2\n ./rma\n fail\n- [profitbricks-build5-amd64:51804] *** Process received signal ***\n-[profitbricks-build5-amd64:51804] Signal: Segmentation fault (11)\n-[profitbricks-build5-amd64:51804] Signal code: Address not mapped (1)\n-[profitbricks-build5-amd64:51804] Failing at address: 0x560c5b483030\n-malloc(): corrupted top size\n-[profitbricks-build5-amd64:51804] *** Process received signal ***\n-[profitbricks-build5-amd64:51804] Signal: Aborted (6)\n-[profitbricks-build5-amd64:51804] Signal code: (-6)\n+ [i-capture-the-hostname:50905] *** Process received signal ***\n+[i-capture-the-hostname:50905] Signal: Segmentation fault (11)\n+[i-capture-the-hostname:50905] Signal code: Address not mapped (1)\n+[i-capture-the-hostname:50905] Failing at address: 0x55ab28e87050\n+[i-capture-the-hostname:50905] [ 0] /lib/x86_64-linux-gnu/libpthread.so.0(+0x126b0)[0x7f68e594d6b0]\n+[i-capture-the-hostname:50905] [ 1] /lib/x86_64-linux-gnu/libc.so.6(+0x13ad67)[0x7f68e58add67]\n+[i-capture-the-hostname:50905] [ 2] /usr/lib/x86_64-linux-gnu/libopen-pal.so.40(opal_generic_simple_unpack+0x4e6)[0x7f68e5647596]\n+[i-capture-the-hostname:50905] [ 3] /usr/lib/x86_64-linux-gnu/openmpi/lib/openmpi3/mca_pml_ob1.so(mca_pml_ob1_recv_request_progress_frag+0x1c2)[0x7f68e2d7bb62]\n+[i-capture-the-hostname:50905] [ 4] /usr/lib/x86_64-linux-gnu/openmpi/lib/openmpi3/mca_btl_vader.so(mca_btl_vader_poll_handle_frag+0x8f)[0x7f68e2db76bf]\n+[i-capture-the-hostname:50905] [ 5] /usr/lib/x86_64-linux-gnu/openmpi/lib/openmpi3/mca_btl_vader.so(+0x4b8a)[0x7f68e2db7b8a]\n+[i-capture-the-hostname:50905] [ 6] /usr/lib/x86_64-linux-gnu/libopen-pal.so.40(opal_progress+0x2c)[0x7f68e562be0c]\n+[i-capture-the-hostname:50905] [ 7] /usr/lib/x86_64-linux-gnu/libopen-pal.so.40(ompi_sync_wait_mt+0x115)[0x7f68e5632835]\n+[i-capture-the-hostname:50905] [ 8] /usr/lib/x86_64-linux-gnu/libmpi.so.40(ompi_request_default_wait+0x1d9)[0x7f68e59ad9b9]\n+[i-capture-the-hostname:50905] [ 9] /usr/lib/x86_64-linux-gnu/libmpi.so.40(ompi_coll_base_barrier_intra_two_procs+0xea)[0x7f68e5a0ac8a]\n+[i-capture-the-hostname:50905] [10] /usr/lib/x86_64-linux-gnu/libmpi.so.40(MPI_Barrier+0xa8)[0x7f68e59c6978]\n+[i-capture-the-hostname:50905] [11] ./lock_dt_flushlocal(+0x38be)[0x55ab27eee8be]\n+[i-capture-the-hostname:50905] [12] /lib/x86_64-linux-gnu/libc.so.6(__libc_start_main+0xeb)[0x7f68e579709b]\n+[i-capture-the-hostname:50905] [13] ./lock_dt_flushlocal(+0x398a)[0x55ab27eee98a]\n+[i-capture-the-hostname:50905] *** End of error message ***\n --------------------------------------------------------------------------\n-The user-provided time limit for job execution has been reached:\n-\n- Timeout: 180 seconds\n-\n-The job will now be aborted. Please check your code and/or\n-adjust/remove the job execution time limit (as specified by --timeout\n-command line option or MPIEXEC_TIMEOUT environment variable).\n+Primary job terminated normally, but 1 process returned\n+a non-zero exit code. Per user-direction, the job has been aborted.\n+--------------------------------------------------------------------------\n+--------------------------------------------------------------------------\n+mpiexec.openmpi noticed that process rank 1 with PID 0 on node i-capture-the-hostname exited on signal 11 (Segmentation fault).\n --------------------------------------------------------------------------\n \n \n lockall_dt\n 2\n ./rma\n fail\n+ [i-capture-the-hostname:51015] *** Process received signal ***\n+[i-capture-the-hostname:51015] Signal: Segmentation fault (11)\n+[i-capture-the-hostname:51015] Signal code: Address not mapped (1)\n+[i-capture-the-hostname:51015] Failing at address: 0x56121ca8c290\n+[i-capture-the-hostname:51015] [ 0] /lib/x86_64-linux-gnu/libpthread.so.0(+0x126b0)[0x7f00c40656b0]\n+[i-capture-the-hostname:51015] [ 1] /lib/x86_64-linux-gnu/libc.so.6(+0x13ad67)[0x7f00c3fc5d67]\n+[i-capture-the-hostname:51015] [ 2] /usr/lib/x86_64-linux-gnu/libopen-pal.so.40(opal_generic_simple_unpack+0x4e6)[0x7f00c3d5f596]\n+[i-capture-the-hostname:51015] [ 3] /usr/lib/x86_64-linux-gnu/openmpi/lib/openmpi3/mca_pml_ob1.so(mca_pml_ob1_recv_request_progress_frag+0x1c2)[0x7f00c1493b62]\n+[i-capture-the-hostname:51015] [ 4] /usr/lib/x86_64-linux-gnu/openmpi/lib/openmpi3/mca_btl_vader.so(mca_btl_vader_poll_handle_frag+0x8f)[0x7f00c14cf6bf]\n+[i-capture-the-hostname:51015] [ 5] /usr/lib/x86_64-linux-gnu/openmpi/lib/openmpi3/mca_btl_vader.so(+0x4b8a)[0x7f00c14cfb8a]\n+[i-capture-the-hostname:51015] [ 6] /usr/lib/x86_64-linux-gnu/libopen-pal.so.40(opal_progress+0x2c)[0x7f00c3d43e0c]\n+[i-capture-the-hostname:51015] [ 7] /usr/lib/x86_64-linux-gnu/libopen-pal.so.40(ompi_sync_wait_mt+0x115)[0x7f00c3d4a835]\n+[i-capture-the-hostname:51015] [ 8] /usr/lib/x86_64-linux-gnu/libmpi.so.40(ompi_request_default_wait+0x1d9)[0x7f00c40c59b9]\n+[i-capture-the-hostname:51015] [ 9] /usr/lib/x86_64-linux-gnu/libmpi.so.40(ompi_coll_base_barrier_intra_two_procs+0xea)[0x7f00c4122c8a]\n+[i-capture-the-hostname:51015] [10] /usr/lib/x86_64-linux-gnu/libmpi.so.40(MPI_Barrier+0xa8)[0x7f00c40de978]\n+[i-capture-the-hostname:51015] [11] ./lockall_dt(+0x368c)[0x56121c1cd68c]\n+[i-capture-the-hostname:51015] [12] /lib/x86_64-linux-gnu/libc.so.6(__libc_start_main+0xeb)[0x7f00c3eaf09b]\n+[i-capture-the-hostname:51015] [13] ./lockall_dt(+0x392a)[0x56121c1cd92a]\n+[i-capture-the-hostname:51015] *** End of error message ***\n+--------------------------------------------------------------------------\n+Primary job terminated normally, but 1 process returned\n+a non-zero exit code. Per user-direction, the job has been aborted.\n+--------------------------------------------------------------------------\n+--------------------------------------------------------------------------\n+mpiexec.openmpi noticed that process rank 1 with PID 0 on node i-capture-the-hostname exited on signal 11 (Segmentation fault).\n+--------------------------------------------------------------------------\n+ \n+ \n+ lockall_dt_flush\n+ 2\n+ ./rma\n+ fail\n malloc(): corrupted top size\n-[profitbricks-build5-amd64:59257] *** Process received signal ***\n-[profitbricks-build5-amd64:59257] Signal: Aborted (6)\n-[profitbricks-build5-amd64:59257] Signal code: (-6)\n+[i-capture-the-hostname:51104] *** Process received signal ***\n+[i-capture-the-hostname:51104] Signal: Aborted (6)\n+[i-capture-the-hostname:51104] Signal code: (-6)\n --------------------------------------------------------------------------\n The user-provided time limit for job execution has been reached:\n \n Timeout: 240 seconds\n \n The job will now be aborted. Please check your code and/or\n adjust/remove the job execution time limit (as specified by --timeout\n command line option or MPIEXEC_TIMEOUT environment variable).\n --------------------------------------------------------------------------\n \n \n- lockall_dt_flush\n+ lockall_dt_flushall\n 2\n ./rma\n fail\n- [profitbricks-build5-amd64:41357] *** Process received signal ***\n-[profitbricks-build5-amd64:41357] Signal: Segmentation fault (11)\n-[profitbricks-build5-amd64:41357] Signal code: Address not mapped (1)\n-[profitbricks-build5-amd64:41357] Failing at address: 0x5572fd3763e0\n-[profitbricks-build5-amd64:41357] [ 0] /lib/x86_64-linux-gnu/libpthread.so.0(+0x126b0)[0x7f966dc826b0]\n-[profitbricks-build5-amd64:41357] [ 1] /lib/x86_64-linux-gnu/libc.so.6(+0x13ad67)[0x7f966dbe2d67]\n-[profitbricks-build5-amd64:41357] [ 2] /usr/lib/x86_64-linux-gnu/libopen-pal.so.40(opal_generic_simple_unpack+0x4e6)[0x7f966d97c596]\n-[profitbricks-build5-amd64:41357] [ 3] /usr/lib/x86_64-linux-gnu/openmpi/lib/openmpi3/mca_pml_ob1.so(mca_pml_ob1_recv_request_progress_frag+0x1c2)[0x7f966b0b0b62]\n-[profitbricks-build5-amd64:41357] [ 4] /usr/lib/x86_64-linux-gnu/openmpi/lib/openmpi3/mca_btl_vader.so(mca_btl_vader_poll_handle_frag+0x8f)[0x7f966b0ec6bf]\n-[profitbricks-build5-amd64:41357] [ 5] /usr/lib/x86_64-linux-gnu/openmpi/lib/openmpi3/mca_btl_vader.so(+0x4b8a)[0x7f966b0ecb8a]\n-[profitbricks-build5-amd64:41357] [ 6] /usr/lib/x86_64-linux-gnu/libopen-pal.so.40(opal_progress+0x2c)[0x7f966d960e0c]\n-[profitbricks-build5-amd64:41357] [ 7] /usr/lib/x86_64-linux-gnu/libopen-pal.so.40(ompi_sync_wait_mt+0x115)[0x7f966d967835]\n-[profitbricks-build5-amd64:41357] [ 8] /usr/lib/x86_64-linux-gnu/libmpi.so.40(ompi_request_default_wait+0x1d9)[0x7f966dce29b9]\n-[profitbricks-build5-amd64:41357] [ 9] /usr/lib/x86_64-linux-gnu/libmpi.so.40(ompi_coll_base_barrier_intra_two_procs+0xea)[0x7f966dd3fc8a]\n-[profitbricks-build5-amd64:41357] [10] /usr/lib/x86_64-linux-gnu/libmpi.so.40(MPI_Barrier+0xa8)[0x7f966dcfb978]\n-[profitbricks-build5-amd64:41357] [11] ./lockall_dt_flush(+0x36a6)[0x5572fbda66a6]\n-[profitbricks-build5-amd64:41357] [12] /lib/x86_64-linux-gnu/libc.so.6(__libc_start_main+0xeb)[0x7f966dacc09b]\n-[profitbricks-build5-amd64:41357] [13] ./lockall_dt_flush(+0x396a)[0x5572fbda696a]\n-[profitbricks-build5-amd64:41357] *** End of error message ***\n+ [i-capture-the-hostname:63672] *** Process received signal ***\n+[i-capture-the-hostname:63672] Signal: Segmentation fault (11)\n+[i-capture-the-hostname:63672] Signal code: Address not mapped (1)\n+[i-capture-the-hostname:63672] Failing at address: 0x558ffefe7260\n+[i-capture-the-hostname:63672] [ 0] /lib/x86_64-linux-gnu/libpthread.so.0(+0x126b0)[0x7f704fd9d6b0]\n+[i-capture-the-hostname:63672] [ 1] /lib/x86_64-linux-gnu/libc.so.6(+0x13ad67)[0x7f704fcfdd67]\n+[i-capture-the-hostname:63672] [ 2] /usr/lib/x86_64-linux-gnu/libopen-pal.so.40(opal_generic_simple_unpack+0x4e6)[0x7f704fa97596]\n+[i-capture-the-hostname:63672] [ 3] /usr/lib/x86_64-linux-gnu/openmpi/lib/openmpi3/mca_pml_ob1.so(mca_pml_ob1_recv_request_progress_frag+0x1c2)[0x7f704d1cbb62]\n+[i-capture-the-hostname:63672] [ 4] /usr/lib/x86_64-linux-gnu/openmpi/lib/openmpi3/mca_btl_vader.so(mca_btl_vader_poll_handle_frag+0x8f)[0x7f704d2076bf]\n+[i-capture-the-hostname:63672] [ 5] /usr/lib/x86_64-linux-gnu/openmpi/lib/openmpi3/mca_btl_vader.so(+0x4b8a)[0x7f704d207b8a]\n+[i-capture-the-hostname:63672] [ 6] /usr/lib/x86_64-linux-gnu/libopen-pal.so.40(opal_progress+0x2c)[0x7f704fa7be0c]\n+[i-capture-the-hostname:63672] [ 7] /usr/lib/x86_64-linux-gnu/libopen-pal.so.40(ompi_sync_wait_mt+0x115)[0x7f704fa82835]\n+[i-capture-the-hostname:63672] [ 8] /usr/lib/x86_64-linux-gnu/libmpi.so.40(ompi_request_default_wait+0x1d9)[0x7f704fdfd9b9]\n+[i-capture-the-hostname:63672] [ 9] /usr/lib/x86_64-linux-gnu/libmpi.so.40(ompi_coll_base_barrier_intra_two_procs+0xea)[0x7f704fe5ac8a]\n+[i-capture-the-hostname:63672] [10] /usr/lib/x86_64-linux-gnu/libmpi.so.40(MPI_Barrier+0xa8)[0x7f704fe16978]\n+[i-capture-the-hostname:63672] [11] ./lockall_dt_flushall(+0x36a6)[0x558ffe64e6a6]\n+[i-capture-the-hostname:63672] [12] /lib/x86_64-linux-gnu/libc.so.6(__libc_start_main+0xeb)[0x7f704fbe709b]\n+[i-capture-the-hostname:63672] [13] ./lockall_dt_flushall(+0x396a)[0x558ffe64e96a]\n+[i-capture-the-hostname:63672] *** End of error message ***\n --------------------------------------------------------------------------\n Primary job terminated normally, but 1 process returned\n a non-zero exit code. Per user-direction, the job has been aborted.\n --------------------------------------------------------------------------\n --------------------------------------------------------------------------\n-mpiexec.openmpi noticed that process rank 1 with PID 0 on node profitbricks-build5-amd64 exited on signal 11 (Segmentation fault).\n+mpiexec.openmpi noticed that process rank 1 with PID 0 on node i-capture-the-hostname exited on signal 11 (Segmentation fault).\n --------------------------------------------------------------------------\n \n \n- lockall_dt_flushall\n+ lockall_dt_flushlocal\n 2\n ./rma\n fail\n- malloc(): corrupted top size\n-[profitbricks-build5-amd64:41800] *** Process received signal ***\n-[profitbricks-build5-amd64:41800] Signal: Aborted (6)\n-[profitbricks-build5-amd64:41800] Signal code: (-6)\n+ [i-capture-the-hostname:63776] *** Process received signal ***\n+[i-capture-the-hostname:63776] Signal: Segmentation fault (11)\n+[i-capture-the-hostname:63776] Signal code: Address not mapped (1)\n+[i-capture-the-hostname:63776] Failing at address: 0x55b1515b30e0\n+malloc(): corrupted top size\n+[i-capture-the-hostname:63776] *** Process received signal ***\n+[i-capture-the-hostname:63776] Signal: Aborted (6)\n+[i-capture-the-hostname:63776] Signal code: (-6)\n --------------------------------------------------------------------------\n The user-provided time limit for job execution has been reached:\n \n Timeout: 240 seconds\n \n The job will now be aborted. Please check your code and/or\n adjust/remove the job execution time limit (as specified by --timeout\n command line option or MPIEXEC_TIMEOUT environment variable).\n --------------------------------------------------------------------------\n \n \n- lockall_dt_flushlocal\n+ lockall_dt_flushlocalall\n 2\n ./rma\n fail\n malloc(): corrupted top size\n-[profitbricks-build5-amd64:54159] *** Process received signal ***\n-[profitbricks-build5-amd64:54159] Signal: Aborted (6)\n-[profitbricks-build5-amd64:54159] Signal code: (-6)\n+[i-capture-the-hostname:14144] *** Process received signal ***\n+[i-capture-the-hostname:14144] Signal: Aborted (6)\n+[i-capture-the-hostname:14144] Signal code: (-6)\n --------------------------------------------------------------------------\n The user-provided time limit for job execution has been reached:\n \n Timeout: 240 seconds\n \n The job will now be aborted. Please check your code and/or\n adjust/remove the job execution time limit (as specified by --timeout\n command line option or MPIEXEC_TIMEOUT environment variable).\n --------------------------------------------------------------------------\n \n \n- lockall_dt_flushlocalall\n+ lock_contention_dt\n 2\n ./rma\n fail\n- [profitbricks-build5-amd64:09247] *** Process received signal ***\n-[profitbricks-build5-amd64:09247] Signal: Segmentation fault (11)\n-[profitbricks-build5-amd64:09247] Signal code: Address not mapped (1)\n-[profitbricks-build5-amd64:09247] Failing at address: 0x564fe1409190\n-[profitbricks-build5-amd64:09247] [ 0] /lib/x86_64-linux-gnu/libpthread.so.0(+0x126b0)[0x7f1500d4a6b0]\n-[profitbricks-build5-amd64:09247] [ 1] /lib/x86_64-linux-gnu/libc.so.6(+0x13acc1)[0x7f1500caacc1]\n-[profitbricks-build5-amd64:09247] [ 2] /usr/lib/x86_64-linux-gnu/libopen-pal.so.40(opal_generic_simple_unpack+0x4e6)[0x7f1500a44596]\n-[profitbricks-build5-amd64:09247] [ 3] /usr/lib/x86_64-linux-gnu/openmpi/lib/openmpi3/mca_pml_ob1.so(mca_pml_ob1_recv_request_progress_frag+0x1c2)[0x7f14fe178b62]\n-[profitbricks-build5-amd64:09247] [ 4] /usr/lib/x86_64-linux-gnu/openmpi/lib/openmpi3/mca_btl_vader.so(mca_btl_vader_poll_handle_frag+0x8f)[0x7f14fe1b46bf]\n-[profitbricks-build5-amd64:09247] [ 5] /usr/lib/x86_64-linux-gnu/openmpi/lib/openmpi3/mca_btl_vader.so(+0x4b8a)[0x7f14fe1b4b8a]\n-[profitbricks-build5-amd64:09247] [ 6] /usr/lib/x86_64-linux-gnu/libopen-pal.so.40(opal_progress+0x2c)[0x7f1500a28e0c]\n-[profitbricks-build5-amd64:09247] [ 7] /usr/lib/x86_64-linux-gnu/libopen-pal.so.40(ompi_sync_wait_mt+0x115)[0x7f1500a2f835]\n-[profitbricks-build5-amd64:09247] [ 8] /usr/lib/x86_64-linux-gnu/libmpi.so.40(ompi_request_default_wait+0x1d9)[0x7f1500daa9b9]\n-[profitbricks-build5-amd64:09247] [ 9] /usr/lib/x86_64-linux-gnu/libmpi.so.40(ompi_coll_base_barrier_intra_two_procs+0xea)[0x7f1500e07c8a]\n-[profitbricks-build5-amd64:09247] [10] /usr/lib/x86_64-linux-gnu/libmpi.so.40(MPI_Barrier+0xa8)[0x7f1500dc3978]\n-[profitbricks-build5-amd64:09247] [11] ./lockall_dt_flushlocalall(+0x36ac)[0x564fdf2226ac]\n-[profitbricks-build5-amd64:09247] [12] /lib/x86_64-linux-gnu/libc.so.6(__libc_start_main+0xeb)[0x7f1500b9409b]\n-[profitbricks-build5-amd64:09247] [13] ./lockall_dt_flushlocalall(+0x39ca)[0x564fdf2229ca]\n-[profitbricks-build5-amd64:09247] *** End of error message ***\n+ [i-capture-the-hostname:40682] *** Process received signal ***\n+[i-capture-the-hostname:40682] Signal: Segmentation fault (11)\n+[i-capture-the-hostname:40682] Signal code: Address not mapped (1)\n+[i-capture-the-hostname:40682] Failing at address: 0x555b11725290\n+[i-capture-the-hostname:40682] [ 0] /lib/x86_64-linux-gnu/libpthread.so.0(+0x126b0)[0x7fca9e9356b0]\n+[i-capture-the-hostname:40682] [ 1] /lib/x86_64-linux-gnu/libc.so.6(+0x13ad67)[0x7fca9e895d67]\n+[i-capture-the-hostname:40682] [ 2] /usr/lib/x86_64-linux-gnu/libopen-pal.so.40(opal_generic_simple_unpack+0x4e6)[0x7fca9e62f596]\n+[i-capture-the-hostname:40682] [ 3] /usr/lib/x86_64-linux-gnu/openmpi/lib/openmpi3/mca_pml_ob1.so(mca_pml_ob1_recv_request_progress_frag+0x1c2)[0x7fca9c163b62]\n+[i-capture-the-hostname:40682] [ 4] /usr/lib/x86_64-linux-gnu/openmpi/lib/openmpi3/mca_btl_vader.so(mca_btl_vader_poll_handle_frag+0x8f)[0x7fca9c19f6bf]\n+[i-capture-the-hostname:40682] [ 5] /usr/lib/x86_64-linux-gnu/openmpi/lib/openmpi3/mca_btl_vader.so(+0x4b8a)[0x7fca9c19fb8a]\n+[i-capture-the-hostname:40682] [ 6] /usr/lib/x86_64-linux-gnu/libopen-pal.so.40(opal_progress+0x2c)[0x7fca9e613e0c]\n+[i-capture-the-hostname:40682] [ 7] /usr/lib/x86_64-linux-gnu/libopen-pal.so.40(ompi_sync_wait_mt+0x115)[0x7fca9e61a835]\n+[i-capture-the-hostname:40682] [ 8] /usr/lib/x86_64-linux-gnu/libmpi.so.40(ompi_request_default_wait+0x1d9)[0x7fca9e9959b9]\n+[i-capture-the-hostname:40682] [ 9] /usr/lib/x86_64-linux-gnu/libmpi.so.40(ompi_coll_base_barrier_intra_two_procs+0xea)[0x7fca9e9f2c8a]\n+[i-capture-the-hostname:40682] [10] /usr/lib/x86_64-linux-gnu/libmpi.so.40(MPI_Barrier+0xa8)[0x7fca9e9ae978]\n+[i-capture-the-hostname:40682] [11] ./lock_contention_dt(+0x2824)[0x555b1057c824]\n+[i-capture-the-hostname:40682] [12] /lib/x86_64-linux-gnu/libc.so.6(__libc_start_main+0xeb)[0x7fca9e77f09b]\n+[i-capture-the-hostname:40682] [13] ./lock_contention_dt(+0x28fa)[0x555b1057c8fa]\n+[i-capture-the-hostname:40682] *** End of error message ***\n --------------------------------------------------------------------------\n Primary job terminated normally, but 1 process returned\n a non-zero exit code. Per user-direction, the job has been aborted.\n --------------------------------------------------------------------------\n --------------------------------------------------------------------------\n-mpiexec.openmpi noticed that process rank 1 with PID 0 on node profitbricks-build5-amd64 exited on signal 11 (Segmentation fault).\n---------------------------------------------------------------------------\n- \n- \n- lock_contention_dt\n- 2\n- ./rma\n- fail\n- Data expected = ff but got p[1,0] = df\n-Data expected = fe but got p[1,1] = de\n-Data expected = fd but got p[1,2] = dd\n-Data expected = fc but got p[1,3] = dc\n-Data expected = fb but got p[1,4] = db\n-Data expected = fa but got p[1,5] = da\n-Data expected = f9 but got p[1,6] = d9\n-Data expected = f8 but got p[1,7] = d8\n-Data expected = f7 but got p[1,8] = d7\n-free(): invalid next size (normal)\n-[profitbricks-build5-amd64:10369] *** Process received signal ***\n-[profitbricks-build5-amd64:10369] Signal: Aborted (6)\n-[profitbricks-build5-amd64:10369] Signal code: (-6)\n---------------------------------------------------------------------------\n-The user-provided time limit for job execution has been reached:\n-\n- Timeout: 240 seconds\n-\n-The job will now be aborted. Please check your code and/or\n-adjust/remove the job execution time limit (as specified by --timeout\n-command line option or MPIEXEC_TIMEOUT environment variable).\n+mpiexec.openmpi noticed that process rank 1 with PID 0 on node i-capture-the-hostname exited on signal 11 (Segmentation fault).\n --------------------------------------------------------------------------\n \n \n transpose4\n 2\n ./rma\n pass\n@@ -4204,18 +4225,18 @@\n pass\n \n \n win_info\n 2\n ./rma\n fail\n- 0: no_locks is not defined\n-0: no_locks is not defined\n+ 1: no_locks is not defined\n 1: no_locks is not defined\n-1: no_locks is not defined\n+0: no_locks is not defined\n+0: no_locks is not defined\n \n \n linked_list_lockall\n 2\n ./rma\n pass\n \n@@ -4295,20 +4316,20 @@\n --------------------------------------------------------------------------\n \n \n badrma\n 2\n ./rma\n fail\n- [profitbricks-build5-amd64:23555] *** An error occurred in MPI_Accumulate\n-[profitbricks-build5-amd64:23555] *** reported by process [2679439361,0]\n-[profitbricks-build5-amd64:23555] *** on win pt2pt window 3\n-[profitbricks-build5-amd64:23555] *** MPI_ERR_ARG: invalid argument of some other kind\n-[profitbricks-build5-amd64:23555] *** MPI_ERRORS_ARE_FATAL (processes in this win will now abort,\n-[profitbricks-build5-amd64:23555] *** and potentially your MPI job)\n+ [i-capture-the-hostname:60307] *** An error occurred in MPI_Accumulate\n+[i-capture-the-hostname:60307] *** reported by process [1713504257,0]\n+[i-capture-the-hostname:60307] *** on win pt2pt window 3\n+[i-capture-the-hostname:60307] *** MPI_ERR_ARG: invalid argument of some other kind\n+[i-capture-the-hostname:60307] *** MPI_ERRORS_ARE_FATAL (processes in this win will now abort,\n+[i-capture-the-hostname:60307] *** and potentially your MPI job)\n \n \n acc-loc\n 2\n ./rma\n pass\n \n@@ -4325,25 +4346,25 @@\n pass\n \n \n win_shared_put_flush_get\n 2\n ./rma\n fail\n- --------------------------------------------------------------------------\n-MPI_ABORT was invoked on rank 1 in communicator MPI_COMM_WORLD\n+ Error: must be run with four processes\n+--------------------------------------------------------------------------\n+MPI_ABORT was invoked on rank 0 in communicator MPI_COMM_WORLD\n with errorcode 1.\n \n NOTE: invoking MPI_ABORT causes Open MPI to kill all MPI processes.\n You may or may not see output from other processes, depending on\n exactly when Open MPI kills them.\n --------------------------------------------------------------------------\n-Error: must be run with four processes\n-[profitbricks-build5-amd64:24810] 1 more process has sent help message help-mpi-api.txt / mpi-abort\n-[profitbricks-build5-amd64:24810] Set MCA parameter "orte_base_help_aggregate" to 0 to see all help / error messages\n+[i-capture-the-hostname:60531] 1 more process has sent help message help-mpi-api.txt / mpi-abort\n+[i-capture-the-hostname:60531] Set MCA parameter "orte_base_help_aggregate" to 0 to see all help / error messages\n \n \n get-struct\n 2\n ./rma\n pass\n \n@@ -999427,33 +999448,33 @@\n MPI_ABORT was invoked on rank 1 in communicator MPI_COMM_WORLD\n with errorcode 1.\n \n NOTE: invoking MPI_ABORT causes Open MPI to kill all MPI processes.\n You may or may not see output from other processes, depending on\n exactly when Open MPI kills them.\n --------------------------------------------------------------------------\n-[profitbricks-build5-amd64:29563] 1 more process has sent help message help-mpi-api.txt / mpi-abort\n-[profitbricks-build5-amd64:29563] Set MCA parameter "orte_base_help_aggregate" to 0 to see all help / error messages\n+[i-capture-the-hostname:45288] 1 more process has sent help message help-mpi-api.txt / mpi-abort\n+[i-capture-the-hostname:45288] Set MCA parameter "orte_base_help_aggregate" to 0 to see all help / error messages\n \n \n win_shared_acc_flush_load\n 2\n ./rma\n fail\n Error: must be run with three processes\n --------------------------------------------------------------------------\n-MPI_ABORT was invoked on rank 0 in communicator MPI_COMM_WORLD\n+MPI_ABORT was invoked on rank 1 in communicator MPI_COMM_WORLD\n with errorcode 1.\n \n NOTE: invoking MPI_ABORT causes Open MPI to kill all MPI processes.\n You may or may not see output from other processes, depending on\n exactly when Open MPI kills them.\n --------------------------------------------------------------------------\n-[profitbricks-build5-amd64:29815] 1 more process has sent help message help-mpi-api.txt / mpi-abort\n-[profitbricks-build5-amd64:29815] Set MCA parameter "orte_base_help_aggregate" to 0 to see all help / error messages\n+[i-capture-the-hostname:45419] 1 more process has sent help message help-mpi-api.txt / mpi-abort\n+[i-capture-the-hostname:45419] Set MCA parameter "orte_base_help_aggregate" to 0 to see all help / error messages\n \n \n win_shared_gacc_flush_load\n 2\n ./rma\n fail\n Error: must be run with three processes\n@@ -999461,33 +999482,33 @@\n MPI_ABORT was invoked on rank 1 in communicator MPI_COMM_WORLD\n with errorcode 1.\n \n NOTE: invoking MPI_ABORT causes Open MPI to kill all MPI processes.\n You may or may not see output from other processes, depending on\n exactly when Open MPI kills them.\n --------------------------------------------------------------------------\n-[profitbricks-build5-amd64:29966] 1 more process has sent help message help-mpi-api.txt / mpi-abort\n-[profitbricks-build5-amd64:29966] Set MCA parameter "orte_base_help_aggregate" to 0 to see all help / error messages\n+[i-capture-the-hostname:45518] 1 more process has sent help message help-mpi-api.txt / mpi-abort\n+[i-capture-the-hostname:45518] Set MCA parameter "orte_base_help_aggregate" to 0 to see all help / error messages\n \n \n win_shared_fop_flush_load\n 2\n ./rma\n fail\n- --------------------------------------------------------------------------\n+ Error: must be run with three processes\n+--------------------------------------------------------------------------\n MPI_ABORT was invoked on rank 1 in communicator MPI_COMM_WORLD\n with errorcode 1.\n \n NOTE: invoking MPI_ABORT causes Open MPI to kill all MPI processes.\n You may or may not see output from other processes, depending on\n exactly when Open MPI kills them.\n --------------------------------------------------------------------------\n-Error: must be run with three processes\n-[profitbricks-build5-amd64:30327] 1 more process has sent help message help-mpi-api.txt / mpi-abort\n-[profitbricks-build5-amd64:30327] Set MCA parameter "orte_base_help_aggregate" to 0 to see all help / error messages\n+[i-capture-the-hostname:45615] 1 more process has sent help message help-mpi-api.txt / mpi-abort\n+[i-capture-the-hostname:45615] Set MCA parameter "orte_base_help_aggregate" to 0 to see all help / error messages\n \n \n win_shared_cas_flush_load\n 2\n ./rma\n fail\n Error: must be run with three processes\n@@ -999495,33 +999516,33 @@\n MPI_ABORT was invoked on rank 1 in communicator MPI_COMM_WORLD\n with errorcode 1.\n \n NOTE: invoking MPI_ABORT causes Open MPI to kill all MPI processes.\n You may or may not see output from other processes, depending on\n exactly when Open MPI kills them.\n --------------------------------------------------------------------------\n-[profitbricks-build5-amd64:30683] 1 more process has sent help message help-mpi-api.txt / mpi-abort\n-[profitbricks-build5-amd64:30683] Set MCA parameter "orte_base_help_aggregate" to 0 to see all help / error messages\n+[i-capture-the-hostname:45755] 1 more process has sent help message help-mpi-api.txt / mpi-abort\n+[i-capture-the-hostname:45755] Set MCA parameter "orte_base_help_aggregate" to 0 to see all help / error messages\n \n \n put_flush_get\n 2\n ./rma\n fail\n Error: must be run with three processes\n --------------------------------------------------------------------------\n-MPI_ABORT was invoked on rank 0 in communicator MPI_COMM_WORLD\n+MPI_ABORT was invoked on rank 1 in communicator MPI_COMM_WORLD\n with errorcode 1.\n \n NOTE: invoking MPI_ABORT causes Open MPI to kill all MPI processes.\n You may or may not see output from other processes, depending on\n exactly when Open MPI kills them.\n --------------------------------------------------------------------------\n-[profitbricks-build5-amd64:30917] 1 more process has sent help message help-mpi-api.txt / mpi-abort\n-[profitbricks-build5-amd64:30917] Set MCA parameter "orte_base_help_aggregate" to 0 to see all help / error messages\n+[i-capture-the-hostname:45895] 1 more process has sent help message help-mpi-api.txt / mpi-abort\n+[i-capture-the-hostname:45895] Set MCA parameter "orte_base_help_aggregate" to 0 to see all help / error messages\n \n \n acc_flush_get\n 2\n ./rma\n fail\n Error: must be run with three processes\n@@ -999529,67 +999550,67 @@\n MPI_ABORT was invoked on rank 1 in communicator MPI_COMM_WORLD\n with errorcode 1.\n \n NOTE: invoking MPI_ABORT causes Open MPI to kill all MPI processes.\n You may or may not see output from other processes, depending on\n exactly when Open MPI kills them.\n --------------------------------------------------------------------------\n-[profitbricks-build5-amd64:31120] 1 more process has sent help message help-mpi-api.txt / mpi-abort\n-[profitbricks-build5-amd64:31120] Set MCA parameter "orte_base_help_aggregate" to 0 to see all help / error messages\n+[i-capture-the-hostname:45926] 1 more process has sent help message help-mpi-api.txt / mpi-abort\n+[i-capture-the-hostname:45926] Set MCA parameter "orte_base_help_aggregate" to 0 to see all help / error messages\n \n \n gacc_flush_get\n 2\n ./rma\n fail\n- --------------------------------------------------------------------------\n-MPI_ABORT was invoked on rank 0 in communicator MPI_COMM_WORLD\n+ Error: must be run with three processes\n+--------------------------------------------------------------------------\n+MPI_ABORT was invoked on rank 1 in communicator MPI_COMM_WORLD\n with errorcode 1.\n \n NOTE: invoking MPI_ABORT causes Open MPI to kill all MPI processes.\n You may or may not see output from other processes, depending on\n exactly when Open MPI kills them.\n --------------------------------------------------------------------------\n-Error: must be run with three processes\n-[profitbricks-build5-amd64:31342] 1 more process has sent help message help-mpi-api.txt / mpi-abort\n-[profitbricks-build5-amd64:31342] Set MCA parameter "orte_base_help_aggregate" to 0 to see all help / error messages\n+[i-capture-the-hostname:45957] 1 more process has sent help message help-mpi-api.txt / mpi-abort\n+[i-capture-the-hostname:45957] Set MCA parameter "orte_base_help_aggregate" to 0 to see all help / error messages\n \n \n fop_flush_get\n 2\n ./rma\n fail\n- --------------------------------------------------------------------------\n+ Error: must be run with three processes\n+--------------------------------------------------------------------------\n MPI_ABORT was invoked on rank 1 in communicator MPI_COMM_WORLD\n with errorcode 1.\n \n NOTE: invoking MPI_ABORT causes Open MPI to kill all MPI processes.\n You may or may not see output from other processes, depending on\n exactly when Open MPI kills them.\n --------------------------------------------------------------------------\n-Error: must be run with three processes\n-[profitbricks-build5-amd64:31706] 1 more process has sent help message help-mpi-api.txt / mpi-abort\n-[profitbricks-build5-amd64:31706] Set MCA parameter "orte_base_help_aggregate" to 0 to see all help / error messages\n+[i-capture-the-hostname:45988] 1 more process has sent help message help-mpi-api.txt / mpi-abort\n+[i-capture-the-hostname:45988] Set MCA parameter "orte_base_help_aggregate" to 0 to see all help / error messages\n \n \n cas_flush_get\n 2\n ./rma\n fail\n Error: must be run with three processes\n --------------------------------------------------------------------------\n-MPI_ABORT was invoked on rank 0 in communicator MPI_COMM_WORLD\n+MPI_ABORT was invoked on rank 1 in communicator MPI_COMM_WORLD\n with errorcode 1.\n \n NOTE: invoking MPI_ABORT causes Open MPI to kill all MPI processes.\n You may or may not see output from other processes, depending on\n exactly when Open MPI kills them.\n --------------------------------------------------------------------------\n-[profitbricks-build5-amd64:31908] 1 more process has sent help message help-mpi-api.txt / mpi-abort\n-[profitbricks-build5-amd64:31908] Set MCA parameter "orte_base_help_aggregate" to 0 to see all help / error messages\n+[i-capture-the-hostname:46013] 1 more process has sent help message help-mpi-api.txt / mpi-abort\n+[i-capture-the-hostname:46013] Set MCA parameter "orte_base_help_aggregate" to 0 to see all help / error messages\n \n \n namepub\n 2\n ./spawn\n pass\n \n@@ -999602,93 +999623,91 @@\n At least one pair of MPI processes are unable to reach each other for\n MPI communications. This means that no Open MPI device has indicated\n that it can be used to communicate between these processes. This is\n an error; Open MPI requires that all MPI processes be able to reach\n each other. This error can sometimes be the result of forgetting to\n specify the "self" BTL.\n \n- Process 1 ([[47742,2],1]) is on host: profitbricks-build5-amd64\n- Process 2 ([[47742,1],0]) is on host: profitbricks-build5-amd64\n+ Process 1 ([[14664,2],0]) is on host: i-capture-the-hostname\n+ Process 2 ([[14664,1],0]) is on host: i-capture-the-hostname\n BTLs attempted: vader self\n \n Your MPI job is now going to abort; sorry.\n --------------------------------------------------------------------------\n+[i-capture-the-hostname:46135] [[14664,2],0] ORTE_ERROR_LOG: Unreachable in file dpm/dpm.c at line 433\n+[i-capture-the-hostname:46136] [[14664,2],1] ORTE_ERROR_LOG: Unreachable in file dpm/dpm.c at line 433\n+[i-capture-the-hostname:46122] [[14664,1],0] ORTE_ERROR_LOG: Unreachable in file dpm/dpm.c at line 433\n+[i-capture-the-hostname:46122] *** An error occurred in MPI_Comm_spawn\n+[i-capture-the-hostname:46122] *** reported by process [961019905,0]\n+[i-capture-the-hostname:46122] *** on communicator MPI_COMM_WORLD\n+[i-capture-the-hostname:46122] *** MPI_ERR_INTERN: internal error\n+[i-capture-the-hostname:46122] *** MPI_ERRORS_ARE_FATAL (processes in this communicator will now abort,\n+[i-capture-the-hostname:46122] *** and potentially your MPI job)\n --------------------------------------------------------------------------\n It looks like MPI_INIT failed for some reason; your parallel process is\n likely to abort. There are many reasons that a parallel process can\n fail during MPI_INIT; some of which are due to configuration or environment\n problems. This failure appears to be an internal failure; here's some\n additional information (which may only be relevant to an Open MPI\n developer):\n \n ompi_dpm_dyn_init() failed\n --> Returned "Unreachable" (-12) instead of "Success" (0)\n --------------------------------------------------------------------------\n-[profitbricks-build5-amd64:32597] [[47742,2],1] ORTE_ERROR_LOG: Unreachable in file dpm/dpm.c at line 433\n-[profitbricks-build5-amd64:32597] *** An error occurred in MPI_Init_thread\n-[profitbricks-build5-amd64:32597] *** reported by process [3128819714,1]\n-[profitbricks-build5-amd64:32597] *** on a NULL communicator\n-[profitbricks-build5-amd64:32597] *** Unknown error\n-[profitbricks-build5-amd64:32597] *** MPI_ERRORS_ARE_FATAL (processes in this communicator will now abort,\n-[profitbricks-build5-amd64:32597] *** and potentially your MPI job)\n-[profitbricks-build5-amd64:32594] [[47742,2],0] ORTE_ERROR_LOG: Unreachable in file dpm/dpm.c at line 433\n-[profitbricks-build5-amd64:32477] [[47742,1],0] ORTE_ERROR_LOG: Unreachable in file dpm/dpm.c at line 433\n-[profitbricks-build5-amd64:32477] *** An error occurred in MPI_Comm_spawn\n-[profitbricks-build5-amd64:32477] *** reported by process [3128819713,0]\n-[profitbricks-build5-amd64:32477] *** on communicator MPI_COMM_WORLD\n-[profitbricks-build5-amd64:32477] *** MPI_ERR_INTERN: internal error\n-[profitbricks-build5-amd64:32477] *** MPI_ERRORS_ARE_FATAL (processes in this communicator will now abort,\n-[profitbricks-build5-amd64:32477] *** and potentially your MPI job)\n-[profitbricks-build5-amd64:32428] 2 more processes have sent help message help-mca-bml-r2.txt / unreachable proc\n-[profitbricks-build5-amd64:32428] Set MCA parameter "orte_base_help_aggregate" to 0 to see all help / error messages\n-[profitbricks-build5-amd64:32428] 1 more process has sent help message help-mpi-runtime.txt / mpi_init:startup:internal-failure\n-[profitbricks-build5-amd64:32428] 1 more process has sent help message help-mpi-errors.txt / mpi_errors_are_fatal unknown handle\n+[i-capture-the-hostname:46135] *** An error occurred in MPI_Init_thread\n+[i-capture-the-hostname:46135] *** reported by process [961019906,0]\n+[i-capture-the-hostname:46135] *** on a NULL communicator\n+[i-capture-the-hostname:46135] *** Unknown error\n+[i-capture-the-hostname:46135] *** MPI_ERRORS_ARE_FATAL (processes in this communicator will now abort,\n+[i-capture-the-hostname:46135] *** and potentially your MPI job)\n+[i-capture-the-hostname:46105] 2 more processes have sent help message help-mca-bml-r2.txt / unreachable proc\n+[i-capture-the-hostname:46105] Set MCA parameter "orte_base_help_aggregate" to 0 to see all help / error messages\n+[i-capture-the-hostname:46105] 1 more process has sent help message help-mpi-runtime.txt / mpi_init:startup:internal-failure\n+[i-capture-the-hostname:46105] 1 more process has sent help message help-mpi-errors.txt / mpi_errors_are_fatal unknown handle\n \n \n spawn2\n 1\n ./spawn\n fail\n --------------------------------------------------------------------------\n At least one pair of MPI processes are unable to reach each other for\n MPI communications. This means that no Open MPI device has indicated\n that it can be used to communicate between these processes. This is\n an error; Open MPI requires that all MPI processes be able to reach\n each other. This error can sometimes be the result of forgetting to\n specify the "self" BTL.\n \n- Process 1 ([[17614,1],0]) is on host: profitbricks-build5-amd64\n- Process 2 ([[17614,2],0]) is on host: profitbricks-build5-amd64\n- BTLs attempted: self\n+ Process 1 ([[14622,2],0]) is on host: i-capture-the-hostname\n+ Process 2 ([[14622,1],0]) is on host: i-capture-the-hostname\n+ BTLs attempted: vader self\n \n Your MPI job is now going to abort; sorry.\n --------------------------------------------------------------------------\n-[profitbricks-build5-amd64:32912] [[17614,1],0] ORTE_ERROR_LOG: Unreachable in file dpm/dpm.c at line 433\n-[profitbricks-build5-amd64:32912] *** An error occurred in MPI_Comm_spawn\n-[profitbricks-build5-amd64:32912] *** reported by process [1154351105,0]\n-[profitbricks-build5-amd64:32912] *** on communicator MPI_COMM_WORLD\n-[profitbricks-build5-amd64:32912] *** MPI_ERR_INTERN: internal error\n-[profitbricks-build5-amd64:32912] *** MPI_ERRORS_ARE_FATAL (processes in this communicator will now abort,\n-[profitbricks-build5-amd64:32912] *** and potentially your MPI job)\n+[i-capture-the-hostname:46183] [[14622,2],0] ORTE_ERROR_LOG: Unreachable in file dpm/dpm.c at line 433\n --------------------------------------------------------------------------\n It looks like MPI_INIT failed for some reason; your parallel process is\n likely to abort. There are many reasons that a parallel process can\n fail during MPI_INIT; some of which are due to configuration or environment\n problems. This failure appears to be an internal failure; here's some\n additional information (which may only be relevant to an Open MPI\n developer):\n \n ompi_dpm_dyn_init() failed\n --> Returned "Unreachable" (-12) instead of "Success" (0)\n --------------------------------------------------------------------------\n-[profitbricks-build5-amd64:33075] [[17614,2],1] ORTE_ERROR_LOG: Unreachable in file dpm/dpm.c at line 433\n-[profitbricks-build5-amd64:33069] [[17614,2],0] ORTE_ERROR_LOG: Unreachable in file dpm/dpm.c at line 433\n-[profitbricks-build5-amd64:32796] 2 more processes have sent help message help-mca-bml-r2.txt / unreachable proc\n-[profitbricks-build5-amd64:32796] Set MCA parameter "orte_base_help_aggregate" to 0 to see all help / error messages\n-[profitbricks-build5-amd64:32796] 1 more process has sent help message help-mpi-runtime.txt / mpi_init:startup:internal-failure\n+[i-capture-the-hostname:46183] *** An error occurred in MPI_Init_thread\n+[i-capture-the-hostname:46183] *** reported by process [958267394,0]\n+[i-capture-the-hostname:46183] *** on a NULL communicator\n+[i-capture-the-hostname:46183] *** Unknown error\n+[i-capture-the-hostname:46183] *** MPI_ERRORS_ARE_FATAL (processes in this communicator will now abort,\n+[i-capture-the-hostname:46183] *** and potentially your MPI job)\n+[i-capture-the-hostname:46164] [[14622,1],0] ORTE_ERROR_LOG: Unreachable in file dpm/dpm.c at line 433\n+[i-capture-the-hostname:46159] 2 more processes have sent help message help-mca-bml-r2.txt / unreachable proc\n+[i-capture-the-hostname:46159] Set MCA parameter "orte_base_help_aggregate" to 0 to see all help / error messages\n \n \n spawninfo1\n 1\n ./spawn\n fail\n --------------------------------------------------------------------------\n@@ -999696,15 +999715,15 @@\n did not launch the job. This error was first reported for process\n rank 0; it may have occurred for other processes as well.\n \n NOTE: A common cause for this error is misspelling a mpiexec.openmpi command\n line parameter option (remember that mpiexec.openmpi interprets the first\n unrecognized command line token as the executable).\n \n-Node: profitbricks-build5-amd64\n+Node: i-capture-the-hostname\n Executable: spawninfo1\n --------------------------------------------------------------------------\n 2 total processes failed to start\n \n \n spawnminfo1\n 1\n@@ -999715,15 +999734,15 @@\n did not launch the job. This error was first reported for process\n rank 0; it may have occurred for other processes as well.\n \n NOTE: A common cause for this error is misspelling a mpiexec.openmpi command\n line parameter option (remember that mpiexec.openmpi interprets the first\n unrecognized command line token as the executable).\n \n-Node: profitbricks-build5-amd64\n+Node: i-capture-the-hostname\n Executable: spawnminfo1\n --------------------------------------------------------------------------\n \n \n spawnintra\n 1\n ./spawn\n@@ -999732,350 +999751,315 @@\n At least one pair of MPI processes are unable to reach each other for\n MPI communications. This means that no Open MPI device has indicated\n that it can be used to communicate between these processes. This is\n an error; Open MPI requires that all MPI processes be able to reach\n each other. This error can sometimes be the result of forgetting to\n specify the "self" BTL.\n \n- Process 1 ([[17123,2],1]) is on host: profitbricks-build5-amd64\n- Process 2 ([[17123,1],0]) is on host: profitbricks-build5-amd64\n- BTLs attempted: vader self\n+ Process 1 ([[14779,1],0]) is on host: i-capture-the-hostname\n+ Process 2 ([[14779,2],0]) is on host: i-capture-the-hostname\n+ BTLs attempted: self\n \n Your MPI job is now going to abort; sorry.\n --------------------------------------------------------------------------\n-[profitbricks-build5-amd64:34436] *** An error occurred in MPI_Comm_spawn\n-[profitbricks-build5-amd64:34436] *** reported by process [1122172929,0]\n-[profitbricks-build5-amd64:34436] *** on communicator MPI_COMM_WORLD\n-[profitbricks-build5-amd64:34436] *** MPI_ERR_INTERN: internal error\n-[profitbricks-build5-amd64:34436] *** MPI_ERRORS_ARE_FATAL (processes in this communicator will now abort,\n-[profitbricks-build5-amd64:34436] *** and potentially your MPI job)\n+[i-capture-the-hostname:46331] [[14779,1],0] ORTE_ERROR_LOG: Unreachable in file dpm/dpm.c at line 433\n+[i-capture-the-hostname:46331] *** An error occurred in MPI_Comm_spawn\n+[i-capture-the-hostname:46331] *** reported by process [968556545,0]\n+[i-capture-the-hostname:46331] *** on communicator MPI_COMM_WORLD\n+[i-capture-the-hostname:46331] *** MPI_ERR_INTERN: internal error\n+[i-capture-the-hostname:46331] *** MPI_ERRORS_ARE_FATAL (processes in this communicator will now abort,\n+[i-capture-the-hostname:46331] *** and potentially your MPI job)\n+[i-capture-the-hostname:46344] [[14779,2],1] ORTE_ERROR_LOG: Unreachable in file dpm/dpm.c at line 433\n+[i-capture-the-hostname:46342] [[14779,2],0] ORTE_ERROR_LOG: Unreachable in file dpm/dpm.c at line 433\n --------------------------------------------------------------------------\n It looks like MPI_INIT failed for some reason; your parallel process is\n likely to abort. There are many reasons that a parallel process can\n fail during MPI_INIT; some of which are due to configuration or environment\n problems. This failure appears to be an internal failure; here's some\n additional information (which may only be relevant to an Open MPI\n developer):\n \n ompi_dpm_dyn_init() failed\n --> Returned "Unreachable" (-12) instead of "Success" (0)\n --------------------------------------------------------------------------\n-[profitbricks-build5-amd64:34436] [[17123,1],0] ORTE_ERROR_LOG: Unreachable in file dpm/dpm.c at line 433\n-[profitbricks-build5-amd64:34586] [[17123,2],0] ORTE_ERROR_LOG: Unreachable in file dpm/dpm.c at line 433\n-[profitbricks-build5-amd64:34588] [[17123,2],1] ORTE_ERROR_LOG: Unreachable in file dpm/dpm.c at line 433\n-[profitbricks-build5-amd64:34586] *** An error occurred in MPI_Init_thread\n-[profitbricks-build5-amd64:34586] *** reported by process [1122172930,0]\n-[profitbricks-build5-amd64:34586] *** on a NULL communicator\n-[profitbricks-build5-amd64:34586] *** Unknown error\n-[profitbricks-build5-amd64:34586] *** MPI_ERRORS_ARE_FATAL (processes in this communicator will now abort,\n-[profitbricks-build5-amd64:34586] *** and potentially your MPI job)\n-[profitbricks-build5-amd64:34353] 2 more processes have sent help message help-mca-bml-r2.txt / unreachable proc\n-[profitbricks-build5-amd64:34353] Set MCA parameter "orte_base_help_aggregate" to 0 to see all help / error messages\n-[profitbricks-build5-amd64:34353] 1 more process has sent help message help-mpi-runtime.txt / mpi_init:startup:internal-failure\n-[profitbricks-build5-amd64:34353] 1 more process has sent help message help-mpi-errors.txt / mpi_errors_are_fatal unknown handle\n+[i-capture-the-hostname:46344] *** An error occurred in MPI_Init_thread\n+[i-capture-the-hostname:46344] *** reported by process [968556546,1]\n+[i-capture-the-hostname:46344] *** on a NULL communicator\n+[i-capture-the-hostname:46344] *** Unknown error\n+[i-capture-the-hostname:46344] *** MPI_ERRORS_ARE_FATAL (processes in this communicator will now abort,\n+[i-capture-the-hostname:46344] *** and potentially your MPI job)\n+[i-capture-the-hostname:46314] 2 more processes have sent help message help-mca-bml-r2.txt / unreachable proc\n+[i-capture-the-hostname:46314] Set MCA parameter "orte_base_help_aggregate" to 0 to see all help / error messages\n+[i-capture-the-hostname:46314] 1 more process has sent help message help-mpi-runtime.txt / mpi_init:startup:internal-failure\n+[i-capture-the-hostname:46314] 1 more process has sent help message help-mpi-errors.txt / mpi_errors_are_fatal unknown handle\n \n \n spawnintra\n 2\n ./spawn\n fail\n --------------------------------------------------------------------------\n At least one pair of MPI processes are unable to reach each other for\n MPI communications. This means that no Open MPI device has indicated\n that it can be used to communicate between these processes. This is\n an error; Open MPI requires that all MPI processes be able to reach\n each other. This error can sometimes be the result of forgetting to\n specify the "self" BTL.\n \n- Process 1 ([[17183,1],1]) is on host: profitbricks-build5-amd64\n- Process 2 ([[17183,2],0]) is on host: profitbricks-build5-amd64\n+ Process 1 ([[14412,1],1]) is on host: i-capture-the-hostname\n+ Process 2 ([[14412,2],0]) is on host: i-capture-the-hostname\n BTLs attempted: vader self\n \n Your MPI job is now going to abort; sorry.\n --------------------------------------------------------------------------\n-[profitbricks-build5-amd64:34850] *** An error occurred in MPI_Comm_spawn\n-[profitbricks-build5-amd64:34850] *** reported by process [1126105089,1]\n-[profitbricks-build5-amd64:34850] *** on communicator MPI_COMM_WORLD\n-[profitbricks-build5-amd64:34850] *** MPI_ERR_INTERN: internal error\n-[profitbricks-build5-amd64:34850] *** MPI_ERRORS_ARE_FATAL (processes in this communicator will now abort,\n-[profitbricks-build5-amd64:34850] *** and potentially your MPI job)\n+[i-capture-the-hostname:46382] [[14412,1],1] ORTE_ERROR_LOG: Unreachable in file dpm/dpm.c at line 433\n+[i-capture-the-hostname:46399] [[14412,2],1] ORTE_ERROR_LOG: Unreachable in file dpm/dpm.c at line 433\n+[i-capture-the-hostname:46381] [[14412,1],0] ORTE_ERROR_LOG: Unreachable in file dpm/dpm.c at line 433\n+[i-capture-the-hostname:46381] *** An error occurred in MPI_Comm_spawn\n+[i-capture-the-hostname:46381] *** reported by process [944504833,0]\n+[i-capture-the-hostname:46381] *** on communicator MPI_COMM_WORLD\n+[i-capture-the-hostname:46381] *** MPI_ERR_INTERN: internal error\n+[i-capture-the-hostname:46381] *** MPI_ERRORS_ARE_FATAL (processes in this communicator will now abort,\n+[i-capture-the-hostname:46381] *** and potentially your MPI job)\n --------------------------------------------------------------------------\n It looks like MPI_INIT failed for some reason; your parallel process is\n likely to abort. There are many reasons that a parallel process can\n fail during MPI_INIT; some of which are due to configuration or environment\n problems. This failure appears to be an internal failure; here's some\n additional information (which may only be relevant to an Open MPI\n developer):\n \n ompi_dpm_dyn_init() failed\n --> Returned "Unreachable" (-12) instead of "Success" (0)\n --------------------------------------------------------------------------\n-[profitbricks-build5-amd64:34974] [[17183,2],1] ORTE_ERROR_LOG: Unreachable in file dpm/dpm.c at line 433\n-[profitbricks-build5-amd64:34850] [[17183,1],1] ORTE_ERROR_LOG: Unreachable in file dpm/dpm.c at line 433\n-[profitbricks-build5-amd64:34849] [[17183,1],0] ORTE_ERROR_LOG: Unreachable in file dpm/dpm.c at line 433\n-[profitbricks-build5-amd64:34966] [[17183,2],0] ORTE_ERROR_LOG: Unreachable in file dpm/dpm.c at line 433\n-[profitbricks-build5-amd64:34974] *** An error occurred in MPI_Init_thread\n-[profitbricks-build5-amd64:34974] *** reported by process [1126105090,1]\n-[profitbricks-build5-amd64:34974] *** on a NULL communicator\n-[profitbricks-build5-amd64:34974] *** Unknown error\n-[profitbricks-build5-amd64:34974] *** MPI_ERRORS_ARE_FATAL (processes in this communicator will now abort,\n-[profitbricks-build5-amd64:34974] *** and potentially your MPI job)\n-[profitbricks-build5-amd64:34765] 3 more processes have sent help message help-mca-bml-r2.txt / unreachable proc\n-[profitbricks-build5-amd64:34765] Set MCA parameter "orte_base_help_aggregate" to 0 to see all help / error messages\n-[profitbricks-build5-amd64:34765] 1 more process has sent help message help-mpi-errors.txt / mpi_errors_are_fatal\n-[profitbricks-build5-amd64:34765] 1 more process has sent help message help-mpi-runtime.txt / mpi_init:startup:internal-failure\n-[profitbricks-build5-amd64:34765] 1 more process has sent help message help-mpi-errors.txt / mpi_errors_are_fatal unknown handle\n+[i-capture-the-hostname:46398] [[14412,2],0] ORTE_ERROR_LOG: Unreachable in file dpm/dpm.c at line 433\n+[i-capture-the-hostname:46399] *** An error occurred in MPI_Init_thread\n+[i-capture-the-hostname:46399] *** reported by process [944504834,1]\n+[i-capture-the-hostname:46399] *** on a NULL communicator\n+[i-capture-the-hostname:46399] *** Unknown error\n+[i-capture-the-hostname:46399] *** MPI_ERRORS_ARE_FATAL (processes in this communicator will now abort,\n+[i-capture-the-hostname:46399] *** and potentially your MPI job)\n+[i-capture-the-hostname:46365] 3 more processes have sent help message help-mca-bml-r2.txt / unreachable proc\n+[i-capture-the-hostname:46365] Set MCA parameter "orte_base_help_aggregate" to 0 to see all help / error messages\n+[i-capture-the-hostname:46365] 1 more process has sent help message help-mpi-errors.txt / mpi_errors_are_fatal\n+[i-capture-the-hostname:46365] 1 more process has sent help message help-mpi-runtime.txt / mpi_init:startup:internal-failure\n+[i-capture-the-hostname:46365] 1 more process has sent help message help-mpi-errors.txt / mpi_errors_are_fatal unknown handle\n \n \n spawnargv\n 1\n ./spawn\n fail\n --------------------------------------------------------------------------\n At least one pair of MPI processes are unable to reach each other for\n MPI communications. This means that no Open MPI device has indicated\n that it can be used to communicate between these processes. This is\n an error; Open MPI requires that all MPI processes be able to reach\n each other. This error can sometimes be the result of forgetting to\n specify the "self" BTL.\n \n- Process 1 ([[19493,2],0]) is on host: profitbricks-build5-amd64\n- Process 2 ([[19493,1],0]) is on host: profitbricks-build5-amd64\n- BTLs attempted: vader self\n+ Process 1 ([[14338,1],0]) is on host: i-capture-the-hostname\n+ Process 2 ([[14338,2],0]) is on host: i-capture-the-hostname\n+ BTLs attempted: self\n \n Your MPI job is now going to abort; sorry.\n --------------------------------------------------------------------------\n+[i-capture-the-hostname:46424] [[14338,1],0] ORTE_ERROR_LOG: Unreachable in file dpm/dpm.c at line 433\n+[i-capture-the-hostname:46424] *** An error occurred in MPI_Comm_spawn\n+[i-capture-the-hostname:46424] *** reported by process [939655169,0]\n+[i-capture-the-hostname:46424] *** on communicator MPI_COMM_WORLD\n+[i-capture-the-hostname:46424] *** MPI_ERR_INTERN: internal error\n+[i-capture-the-hostname:46424] *** MPI_ERRORS_ARE_FATAL (processes in this communicator will now abort,\n+[i-capture-the-hostname:46424] *** and potentially your MPI job)\n+[i-capture-the-hostname:46444] [[14338,2],0] ORTE_ERROR_LOG: Unreachable in file dpm/dpm.c at line 433\n+[i-capture-the-hostname:46445] [[14338,2],1] ORTE_ERROR_LOG: Unreachable in file dpm/dpm.c at line 433\n --------------------------------------------------------------------------\n It looks like MPI_INIT failed for some reason; your parallel process is\n likely to abort. There are many reasons that a parallel process can\n fail during MPI_INIT; some of which are due to configuration or environment\n problems. This failure appears to be an internal failure; here's some\n additional information (which may only be relevant to an Open MPI\n developer):\n \n ompi_dpm_dyn_init() failed\n --> Returned "Unreachable" (-12) instead of "Success" (0)\n --------------------------------------------------------------------------\n-[profitbricks-build5-amd64:35244] [[19493,2],0] ORTE_ERROR_LOG: Unreachable in file dpm/dpm.c at line 433\n-[profitbricks-build5-amd64:35244] *** An error occurred in MPI_Init_thread\n-[profitbricks-build5-amd64:35244] *** reported by process [1277493250,0]\n-[profitbricks-build5-amd64:35244] *** on a NULL communicator\n-[profitbricks-build5-amd64:35244] *** Unknown error\n-[profitbricks-build5-amd64:35244] *** MPI_ERRORS_ARE_FATAL (processes in this communicator will now abort,\n-[profitbricks-build5-amd64:35244] *** and potentially your MPI job)\n-[profitbricks-build5-amd64:35095] [[19493,1],0] ORTE_ERROR_LOG: Unreachable in file dpm/dpm.c at line 433\n-[profitbricks-build5-amd64:35095] *** An error occurred in MPI_Comm_spawn\n-[profitbricks-build5-amd64:35095] *** reported by process [1277493249,0]\n-[profitbricks-build5-amd64:35095] *** on communicator MPI_COMM_WORLD\n-[profitbricks-build5-amd64:35095] *** MPI_ERR_INTERN: internal error\n-[profitbricks-build5-amd64:35095] *** MPI_ERRORS_ARE_FATAL (processes in this communicator will now abort,\n-[profitbricks-build5-amd64:35095] *** and potentially your MPI job)\n-[profitbricks-build5-amd64:35246] [[19493,2],1] ORTE_ERROR_LOG: Unreachable in file dpm/dpm.c at line 433\n-[profitbricks-build5-amd64:35063] 2 more processes have sent help message help-mca-bml-r2.txt / unreachable proc\n-[profitbricks-build5-amd64:35063] Set MCA parameter "orte_base_help_aggregate" to 0 to see all help / error messages\n-[profitbricks-build5-amd64:35063] 1 more process has sent help message help-mpi-runtime.txt / mpi_init:startup:internal-failure\n+[i-capture-the-hostname:46444] *** An error occurred in MPI_Init_thread\n+[i-capture-the-hostname:46444] *** reported by process [939655170,0]\n+[i-capture-the-hostname:46444] *** on a NULL communicator\n+[i-capture-the-hostname:46444] *** Unknown error\n+[i-capture-the-hostname:46444] *** MPI_ERRORS_ARE_FATAL (processes in this communicator will now abort,\n+[i-capture-the-hostname:46444] *** and potentially your MPI job)\n+[i-capture-the-hostname:46419] 2 more processes have sent help message help-mca-bml-r2.txt / unreachable proc\n+[i-capture-the-hostname:46419] Set MCA parameter "orte_base_help_aggregate" to 0 to see all help / error messages\n+[i-capture-the-hostname:46419] 1 more process has sent help message help-mpi-runtime.txt / mpi_init:startup:internal-failure\n+[i-capture-the-hostname:46419] 1 more process has sent help message help-mpi-errors.txt / mpi_errors_are_fatal unknown handle\n \n \n spawnmanyarg\n 1\n ./spawn\n fail\n --------------------------------------------------------------------------\n At least one pair of MPI processes are unable to reach each other for\n MPI communications. This means that no Open MPI device has indicated\n that it can be used to communicate between these processes. This is\n an error; Open MPI requires that all MPI processes be able to reach\n each other. This error can sometimes be the result of forgetting to\n specify the "self" BTL.\n \n- Process 1 ([[20005,2],1]) is on host: profitbricks-build5-amd64\n- Process 2 ([[20005,1],0]) is on host: profitbricks-build5-amd64\n+ Process 1 ([[14530,2],0]) is on host: i-capture-the-hostname\n+ Process 2 ([[14530,1],0]) is on host: i-capture-the-hostname\n BTLs attempted: vader self\n \n Your MPI job is now going to abort; sorry.\n --------------------------------------------------------------------------\n+[i-capture-the-hostname:46571] [[14530,2],0] ORTE_ERROR_LOG: Unreachable in file dpm/dpm.c at line 433\n+[i-capture-the-hostname:46575] [[14530,2],1] ORTE_ERROR_LOG: Unreachable in file dpm/dpm.c at line 433\n --------------------------------------------------------------------------\n It looks like MPI_INIT failed for some reason; your parallel process is\n likely to abort. There are many reasons that a parallel process can\n fail during MPI_INIT; some of which are due to configuration or environment\n problems. This failure appears to be an internal failure; here's some\n additional information (which may only be relevant to an Open MPI\n developer):\n \n ompi_dpm_dyn_init() failed\n --> Returned "Unreachable" (-12) instead of "Success" (0)\n --------------------------------------------------------------------------\n-[profitbricks-build5-amd64:35871] [[20005,2],1] ORTE_ERROR_LOG: Unreachable in file dpm/dpm.c at line 433\n-[profitbricks-build5-amd64:35866] [[20005,2],0] ORTE_ERROR_LOG: Unreachable in file dpm/dpm.c at line 433\n-[profitbricks-build5-amd64:35871] *** An error occurred in MPI_Init_thread\n-[profitbricks-build5-amd64:35871] *** reported by process [1311047682,1]\n-[profitbricks-build5-amd64:35871] *** on a NULL communicator\n-[profitbricks-build5-amd64:35871] *** Unknown error\n-[profitbricks-build5-amd64:35871] *** MPI_ERRORS_ARE_FATAL (processes in this communicator will now abort,\n-[profitbricks-build5-amd64:35871] *** and potentially your MPI job)\n-[profitbricks-build5-amd64:35672] [[20005,1],0] ORTE_ERROR_LOG: Unreachable in file dpm/dpm.c at line 433\n-[profitbricks-build5-amd64:35672] *** An error occurred in MPI_Comm_spawn\n-[profitbricks-build5-amd64:35672] *** reported by process [1311047681,0]\n-[profitbricks-build5-amd64:35672] *** on communicator MPI_COMM_WORLD\n-[profitbricks-build5-amd64:35672] *** MPI_ERR_INTERN: internal error\n-[profitbricks-build5-amd64:35672] *** MPI_ERRORS_ARE_FATAL (processes in this communicator will now abort,\n-[profitbricks-build5-amd64:35672] *** and potentially your MPI job)\n-[profitbricks-build5-amd64:35575] 2 more processes have sent help message help-mca-bml-r2.txt / unreachable proc\n-[profitbricks-build5-amd64:35575] Set MCA parameter "orte_base_help_aggregate" to 0 to see all help / error messages\n-[profitbricks-build5-amd64:35575] 1 more process has sent help message help-mpi-runtime.txt / mpi_init:startup:internal-failure\n-[profitbricks-build5-amd64:35575] 1 more process has sent help message help-mpi-errors.txt / mpi_errors_are_fatal unknown handle\n+[i-capture-the-hostname:46575] *** An error occurred in MPI_Init_thread\n+[i-capture-the-hostname:46575] *** reported by process [952238082,1]\n+[i-capture-the-hostname:46575] *** on a NULL communicator\n+[i-capture-the-hostname:46575] *** Unknown error\n+[i-capture-the-hostname:46575] *** MPI_ERRORS_ARE_FATAL (processes in this communicator will now abort,\n+[i-capture-the-hostname:46575] *** and potentially your MPI job)\n+[i-capture-the-hostname:46483] 2 more processes have sent help message help-mca-bml-r2.txt / unreachable proc\n+[i-capture-the-hostname:46483] Set MCA parameter "orte_base_help_aggregate" to 0 to see all help / error messages\n+[i-capture-the-hostname:46483] 1 more process has sent help message help-mpi-runtime.txt / mpi_init:startup:internal-failure\n+[i-capture-the-hostname:46483] 1 more process has sent help message help-mpi-errors.txt / mpi_errors_are_fatal unknown handle\n \n \n spawnmult2\n 2\n ./spawn\n fail\n --------------------------------------------------------------------------\n At least one pair of MPI processes are unable to reach each other for\n MPI communications. This means that no Open MPI device has indicated\n that it can be used to communicate between these processes. This is\n an error; Open MPI requires that all MPI processes be able to reach\n each other. This error can sometimes be the result of forgetting to\n specify the "self" BTL.\n \n- Process 1 ([[18925,1],0]) is on host: profitbricks-build5-amd64\n- Process 2 ([[18925,2],0]) is on host: profitbricks-build5-amd64\n+ Process 1 ([[15263,1],0]) is on host: i-capture-the-hostname\n+ Process 2 ([[15263,2],0]) is on host: i-capture-the-hostname\n BTLs attempted: vader self\n \n Your MPI job is now going to abort; sorry.\n --------------------------------------------------------------------------\n-[profitbricks-build5-amd64:36366] [[18925,2],1] ORTE_ERROR_LOG: Unreachable in file dpm/dpm.c at line 433\n-[profitbricks-build5-amd64:36278] [[18925,1],0] ORTE_ERROR_LOG: Unreachable in file dpm/dpm.c at line 433\n---------------------------------------------------------------------------\n-It looks like MPI_INIT failed for some reason; your parallel process is\n-likely to abort. There are many reasons that a parallel process can\n-fail during MPI_INIT; some of which are due to configuration or environment\n-problems. This failure appears to be an internal failure; here's some\n-additional information (which may only be relevant to an Open MPI\n-developer):\n-\n- ompi_dpm_dyn_init() failed\n- --> Returned "Unreachable" (-12) instead of "Success" (0)\n---------------------------------------------------------------------------\n-[profitbricks-build5-amd64:36278] *** An error occurred in MPI_Comm_spawn_multiple\n-[profitbricks-build5-amd64:36278] *** reported by process [1240268801,0]\n-[profitbricks-build5-amd64:36278] *** on communicator MPI_COMM_WORLD\n-[profitbricks-build5-amd64:36278] *** MPI_ERR_INTERN: internal error\n-[profitbricks-build5-amd64:36278] *** MPI_ERRORS_ARE_FATAL (processes in this communicator will now abort,\n-[profitbricks-build5-amd64:36278] *** and potentially your MPI job)\n-[profitbricks-build5-amd64:36283] [[18925,1],1] ORTE_ERROR_LOG: Unreachable in file dpm/dpm.c at line 433\n-[profitbricks-build5-amd64:36365] [[18925,2],0] ORTE_ERROR_LOG: Unreachable in file dpm/dpm.c at line 433\n-[profitbricks-build5-amd64:36366] *** An error occurred in MPI_Init_thread\n-[profitbricks-build5-amd64:36366] *** reported by process [1240268802,1]\n-[profitbricks-build5-amd64:36366] *** on a NULL communicator\n-[profitbricks-build5-amd64:36366] *** Unknown error\n-[profitbricks-build5-amd64:36366] *** MPI_ERRORS_ARE_FATAL (processes in this communicator will now abort,\n-[profitbricks-build5-amd64:36366] *** and potentially your MPI job)\n-[profitbricks-build5-amd64:36159] 3 more processes have sent help message help-mca-bml-r2.txt / unreachable proc\n-[profitbricks-build5-amd64:36159] Set MCA parameter "orte_base_help_aggregate" to 0 to see all help / error messages\n-[profitbricks-build5-amd64:36159] 1 more process has sent help message help-mpi-runtime.txt / mpi_init:startup:internal-failure\n-[profitbricks-build5-amd64:36159] 1 more process has sent help message help-mpi-errors.txt / mpi_errors_are_fatal\n-[profitbricks-build5-amd64:36159] 1 more process has sent help message help-mpi-errors.txt / mpi_errors_are_fatal unknown handle\n+[i-capture-the-hostname:46869] [[15263,1],0] ORTE_ERROR_LOG: Unreachable in file dpm/dpm.c at line 433\n+[i-capture-the-hostname:46869] *** An error occurred in MPI_Comm_spawn_multiple\n+[i-capture-the-hostname:46869] *** reported by process [1000275969,0]\n+[i-capture-the-hostname:46869] *** on communicator MPI_COMM_WORLD\n+[i-capture-the-hostname:46869] *** MPI_ERR_INTERN: internal error\n+[i-capture-the-hostname:46869] *** MPI_ERRORS_ARE_FATAL (processes in this communicator will now abort,\n+[i-capture-the-hostname:46869] *** and potentially your MPI job)\n+[i-capture-the-hostname:46798] 3 more processes have sent help message help-mca-bml-r2.txt / unreachable proc\n+[i-capture-the-hostname:46798] Set MCA parameter "orte_base_help_aggregate" to 0 to see all help / error messages\n \n \n spaconacc\n 1\n ./spawn\n fail\n --------------------------------------------------------------------------\n At least one pair of MPI processes are unable to reach each other for\n MPI communications. This means that no Open MPI device has indicated\n that it can be used to communicate between these processes. This is\n an error; Open MPI requires that all MPI processes be able to reach\n each other. This error can sometimes be the result of forgetting to\n specify the "self" BTL.\n \n- Process 1 ([[18951,1],0]) is on host: profitbricks-build5-amd64\n- Process 2 ([[18951,2],0]) is on host: profitbricks-build5-amd64\n+ Process 1 ([[15061,1],0]) is on host: i-capture-the-hostname\n+ Process 2 ([[15061,2],0]) is on host: i-capture-the-hostname\n BTLs attempted: self\n \n Your MPI job is now going to abort; sorry.\n --------------------------------------------------------------------------\n+[i-capture-the-hostname:47010] [[15061,1],0] ORTE_ERROR_LOG: Unreachable in file dpm/dpm.c at line 433\n MPI_Comm_spawn failed: MPI_ERR_INTERN: internal error\n-[profitbricks-build5-amd64:36667] [[18951,1],0] ORTE_ERROR_LOG: Unreachable in file dpm/dpm.c at line 433\n+[i-capture-the-hostname:47022] [[15061,2],0] ORTE_ERROR_LOG: Unreachable in file dpm/dpm.c at line 433\n --------------------------------------------------------------------------\n MPI_ABORT was invoked on rank 0 in communicator MPI_COMM_WORLD\n with errorcode 17.\n \n NOTE: invoking MPI_ABORT causes Open MPI to kill all MPI processes.\n You may or may not see output from other processes, depending on\n exactly when Open MPI kills them.\n --------------------------------------------------------------------------\n-[profitbricks-build5-amd64:36718] [[18951,2],0] ORTE_ERROR_LOG: Unreachable in file dpm/dpm.c at line 433\n --------------------------------------------------------------------------\n It looks like MPI_INIT failed for some reason; your parallel process is\n likely to abort. There are many reasons that a parallel process can\n fail during MPI_INIT; some of which are due to configuration or environment\n problems. This failure appears to be an internal failure; here's some\n additional information (which may only be relevant to an Open MPI\n developer):\n \n ompi_dpm_dyn_init() failed\n --> Returned "Unreachable" (-12) instead of "Success" (0)\n --------------------------------------------------------------------------\n-[profitbricks-build5-amd64:36718] *** An error occurred in MPI_Init\n-[profitbricks-build5-amd64:36718] *** reported by process [1241972738,0]\n-[profitbricks-build5-amd64:36718] *** on a NULL communicator\n-[profitbricks-build5-amd64:36718] *** Unknown error\n-[profitbricks-build5-amd64:36718] *** MPI_ERRORS_ARE_FATAL (processes in this communicator will now abort,\n-[profitbricks-build5-amd64:36718] *** and potentially your MPI job)\n-[profitbricks-build5-amd64:36565] 1 more process has sent help message help-mca-bml-r2.txt / unreachable proc\n-[profitbricks-build5-amd64:36565] Set MCA parameter "orte_base_help_aggregate" to 0 to see all help / error messages\n+[i-capture-the-hostname:47022] *** An error occurred in MPI_Init\n+[i-capture-the-hostname:47022] *** reported by process [987037698,0]\n+[i-capture-the-hostname:47022] *** on a NULL communicator\n+[i-capture-the-hostname:47022] *** Unknown error\n+[i-capture-the-hostname:47022] *** MPI_ERRORS_ARE_FATAL (processes in this communicator will now abort,\n+[i-capture-the-hostname:47022] *** and potentially your MPI job)\n+[i-capture-the-hostname:46980] 1 more process has sent help message help-mca-bml-r2.txt / unreachable proc\n+[i-capture-the-hostname:46980] Set MCA parameter "orte_base_help_aggregate" to 0 to see all help / error messages\n \n \n spaconacc2\n 1\n ./spawn\n fail\n --------------------------------------------------------------------------\n At least one pair of MPI processes are unable to reach each other for\n MPI communications. This means that no Open MPI device has indicated\n that it can be used to communicate between these processes. This is\n an error; Open MPI requires that all MPI processes be able to reach\n each other. This error can sometimes be the result of forgetting to\n specify the "self" BTL.\n \n- Process 1 ([[19251,1],0]) is on host: profitbricks-build5-amd64\n- Process 2 ([[19251,2],0]) is on host: profitbricks-build5-amd64\n+ Process 1 ([[14994,1],0]) is on host: i-capture-the-hostname\n+ Process 2 ([[14994,2],0]) is on host: i-capture-the-hostname\n BTLs attempted: self\n \n Your MPI job is now going to abort; sorry.\n --------------------------------------------------------------------------\n+[i-capture-the-hostname:47049] [[14994,1],0] ORTE_ERROR_LOG: Unreachable in file dpm/dpm.c at line 433\n+[i-capture-the-hostname:47049] *** An error occurred in MPI_Comm_spawn\n+[i-capture-the-hostname:47049] *** reported by process [982646785,0]\n+[i-capture-the-hostname:47049] *** on communicator MPI_COMM_SELF\n+[i-capture-the-hostname:47049] *** MPI_ERR_INTERN: internal error\n+[i-capture-the-hostname:47049] *** MPI_ERRORS_ARE_FATAL (processes in this communicator will now abort,\n+[i-capture-the-hostname:47049] *** and potentially your MPI job)\n --------------------------------------------------------------------------\n It looks like MPI_INIT failed for some reason; your parallel process is\n likely to abort. There are many reasons that a parallel process can\n fail during MPI_INIT; some of which are due to configuration or environment\n problems. This failure appears to be an internal failure; here's some\n additional information (which may only be relevant to an Open MPI\n developer):\n \n ompi_dpm_dyn_init() failed\n --> Returned "Unreachable" (-12) instead of "Success" (0)\n --------------------------------------------------------------------------\n-[profitbricks-build5-amd64:36984] [[19251,2],0] ORTE_ERROR_LOG: Unreachable in file dpm/dpm.c at line 433\n-[profitbricks-build5-amd64:36984] *** An error occurred in MPI_Init\n-[profitbricks-build5-amd64:36984] *** reported by process [1261633538,0]\n-[profitbricks-build5-amd64:36984] *** on a NULL communicator\n-[profitbricks-build5-amd64:36984] *** Unknown error\n-[profitbricks-build5-amd64:36984] *** MPI_ERRORS_ARE_FATAL (processes in this communicator will now abort,\n-[profitbricks-build5-amd64:36984] *** and potentially your MPI job)\n-[profitbricks-build5-amd64:36874] [[19251,1],0] ORTE_ERROR_LOG: Unreachable in file dpm/dpm.c at line 433\n-[profitbricks-build5-amd64:36874] *** An error occurred in MPI_Comm_spawn\n-[profitbricks-build5-amd64:36874] *** reported by process [1261633537,0]\n-[profitbricks-build5-amd64:36874] *** on communicator MPI_COMM_SELF\n-[profitbricks-build5-amd64:36874] *** MPI_ERR_INTERN: internal error\n-[profitbricks-build5-amd64:36874] *** MPI_ERRORS_ARE_FATAL (processes in this communicator will now abort,\n-[profitbricks-build5-amd64:36874] *** and potentially your MPI job)\n-[profitbricks-build5-amd64:36833] 1 more process has sent help message help-mca-bml-r2.txt / unreachable proc\n-[profitbricks-build5-amd64:36833] Set MCA parameter "orte_base_help_aggregate" to 0 to see all help / error messages\n+[i-capture-the-hostname:47063] [[14994,2],0] ORTE_ERROR_LOG: Unreachable in file dpm/dpm.c at line 433\n+[i-capture-the-hostname:47043] 1 more process has sent help message help-mca-bml-r2.txt / unreachable proc\n+[i-capture-the-hostname:47043] Set MCA parameter "orte_base_help_aggregate" to 0 to see all help / error messages\n \n \n selfconacc\n 2\n ./spawn\n pass\n \n@@ -1000088,140 +1000072,129 @@\n At least one pair of MPI processes are unable to reach each other for\n MPI communications. This means that no Open MPI device has indicated\n that it can be used to communicate between these processes. This is\n an error; Open MPI requires that all MPI processes be able to reach\n each other. This error can sometimes be the result of forgetting to\n specify the "self" BTL.\n \n- Process 1 ([[22167,2],1]) is on host: profitbricks-build5-amd64\n- Process 2 ([[22167,1],0]) is on host: profitbricks-build5-amd64\n+ Process 1 ([[13590,1],0]) is on host: i-capture-the-hostname\n+ Process 2 ([[13590,2],0]) is on host: i-capture-the-hostname\n BTLs attempted: vader self\n \n Your MPI job is now going to abort; sorry.\n --------------------------------------------------------------------------\n-[profitbricks-build5-amd64:37597] [[22167,2],1] ORTE_ERROR_LOG: Unreachable in file dpm/dpm.c at line 433\n+[i-capture-the-hostname:47194] [[13590,1],0] ORTE_ERROR_LOG: Unreachable in file dpm/dpm.c at line 433\n+[i-capture-the-hostname:47303] [[13590,2],0] ORTE_ERROR_LOG: Unreachable in file dpm/dpm.c at line 433\n+[i-capture-the-hostname:47306] [[13590,2],1] ORTE_ERROR_LOG: Unreachable in file dpm/dpm.c at line 433\n --------------------------------------------------------------------------\n It looks like MPI_INIT failed for some reason; your parallel process is\n likely to abort. There are many reasons that a parallel process can\n fail during MPI_INIT; some of which are due to configuration or environment\n problems. This failure appears to be an internal failure; here's some\n additional information (which may only be relevant to an Open MPI\n developer):\n \n ompi_dpm_dyn_init() failed\n --> Returned "Unreachable" (-12) instead of "Success" (0)\n --------------------------------------------------------------------------\n-[profitbricks-build5-amd64:37597] *** An error occurred in MPI_Init_thread\n-[profitbricks-build5-amd64:37597] *** reported by process [1452736514,1]\n-[profitbricks-build5-amd64:37597] *** on a NULL communicator\n-[profitbricks-build5-amd64:37597] *** Unknown error\n-[profitbricks-build5-amd64:37597] *** MPI_ERRORS_ARE_FATAL (processes in this communicator will now abort,\n-[profitbricks-build5-amd64:37597] *** and potentially your MPI job)\n-[profitbricks-build5-amd64:37596] [[22167,2],0] ORTE_ERROR_LOG: Unreachable in file dpm/dpm.c at line 433\n-[profitbricks-build5-amd64:37501] [[22167,1],0] ORTE_ERROR_LOG: Unreachable in file dpm/dpm.c at line 433\n-[profitbricks-build5-amd64:37501] *** An error occurred in MPI_Comm_spawn\n-[profitbricks-build5-amd64:37501] *** reported by process [1452736513,0]\n-[profitbricks-build5-amd64:37501] *** on communicator MPI_COMM_SELF\n-[profitbricks-build5-amd64:37501] *** MPI_ERR_INTERN: internal error\n-[profitbricks-build5-amd64:37501] *** MPI_ERRORS_ARE_FATAL (processes in this communicator will now abort,\n-[profitbricks-build5-amd64:37501] *** and potentially your MPI job)\n-[profitbricks-build5-amd64:37445] 2 more processes have sent help message help-mca-bml-r2.txt / unreachable proc\n-[profitbricks-build5-amd64:37445] Set MCA parameter "orte_base_help_aggregate" to 0 to see all help / error messages\n-[profitbricks-build5-amd64:37445] 1 more process has sent help message help-mpi-runtime.txt / mpi_init:startup:internal-failure\n-[profitbricks-build5-amd64:37445] 1 more process has sent help message help-mpi-errors.txt / mpi_errors_are_fatal unknown handle\n+[i-capture-the-hostname:47194] *** An error occurred in MPI_Comm_spawn\n+[i-capture-the-hostname:47194] *** reported by process [890634241,0]\n+[i-capture-the-hostname:47194] *** on communicator MPI_COMM_SELF\n+[i-capture-the-hostname:47194] *** MPI_ERR_INTERN: internal error\n+[i-capture-the-hostname:47194] *** MPI_ERRORS_ARE_FATAL (processes in this communicator will now abort,\n+[i-capture-the-hostname:47194] *** and potentially your MPI job)\n+[i-capture-the-hostname:47303] *** An error occurred in MPI_Init_thread\n+[i-capture-the-hostname:47303] *** reported by process [890634242,0]\n+[i-capture-the-hostname:47303] *** on a NULL communicator\n+[i-capture-the-hostname:47303] *** Unknown error\n+[i-capture-the-hostname:47303] *** MPI_ERRORS_ARE_FATAL (processes in this communicator will now abort,\n+[i-capture-the-hostname:47303] *** and potentially your MPI job)\n+[i-capture-the-hostname:47175] 2 more processes have sent help message help-mca-bml-r2.txt / unreachable proc\n+[i-capture-the-hostname:47175] Set MCA parameter "orte_base_help_aggregate" to 0 to see all help / error messages\n+[i-capture-the-hostname:47175] 1 more process has sent help message help-mpi-runtime.txt / mpi_init:startup:internal-failure\n+[i-capture-the-hostname:47175] 1 more process has sent help message help-mpi-errors.txt / mpi_errors_are_fatal unknown handle\n \n \n taskmaster\n 1\n ./spawn\n fail\n --------------------------------------------------------------------------\n At least one pair of MPI processes are unable to reach each other for\n MPI communications. This means that no Open MPI device has indicated\n that it can be used to communicate between these processes. This is\n an error; Open MPI requires that all MPI processes be able to reach\n each other. This error can sometimes be the result of forgetting to\n specify the "self" BTL.\n \n- Process 1 ([[22462,2],0]) is on host: profitbricks-build5-amd64\n- Process 2 ([[22462,1],0]) is on host: profitbricks-build5-amd64\n+ Process 1 ([[13351,1],0]) is on host: i-capture-the-hostname\n+ Process 2 ([[13351,2],0]) is on host: i-capture-the-hostname\n BTLs attempted: self\n \n Your MPI job is now going to abort; sorry.\n --------------------------------------------------------------------------\n-[profitbricks-build5-amd64:37968] [[22462,2],0] ORTE_ERROR_LOG: Unreachable in file dpm/dpm.c at line 433\n+[i-capture-the-hostname:47649] [[13351,2],0] ORTE_ERROR_LOG: Unreachable in file dpm/dpm.c at line 433\n+[i-capture-the-hostname:47531] [[13351,1],0] ORTE_ERROR_LOG: Unreachable in file dpm/dpm.c at line 433\n --------------------------------------------------------------------------\n It looks like MPI_INIT failed for some reason; your parallel process is\n likely to abort. There are many reasons that a parallel process can\n fail during MPI_INIT; some of which are due to configuration or environment\n problems. This failure appears to be an internal failure; here's some\n additional information (which may only be relevant to an Open MPI\n developer):\n \n ompi_dpm_dyn_init() failed\n --> Returned "Unreachable" (-12) instead of "Success" (0)\n --------------------------------------------------------------------------\n-[profitbricks-build5-amd64:37968] *** An error occurred in MPI_Init\n-[profitbricks-build5-amd64:37968] *** reported by process [1472069634,0]\n-[profitbricks-build5-amd64:37968] *** on a NULL communicator\n-[profitbricks-build5-amd64:37968] *** Unknown error\n-[profitbricks-build5-amd64:37968] *** MPI_ERRORS_ARE_FATAL (processes in this communicator will now abort,\n-[profitbricks-build5-amd64:37968] *** and potentially your MPI job)\n-[profitbricks-build5-amd64:37840] *** An error occurred in MPI_Comm_spawn\n-[profitbricks-build5-amd64:37840] *** reported by process [1472069633,0]\n-[profitbricks-build5-amd64:37840] *** on communicator MPI_COMM_WORLD\n-[profitbricks-build5-amd64:37840] *** MPI_ERR_INTERN: internal error\n-[profitbricks-build5-amd64:37840] *** MPI_ERRORS_ARE_FATAL (processes in this communicator will now abort,\n-[profitbricks-build5-amd64:37840] *** and potentially your MPI job)\n-[profitbricks-build5-amd64:37840] [[22462,1],0] ORTE_ERROR_LOG: Unreachable in file dpm/dpm.c at line 433\n-[profitbricks-build5-amd64:37740] 1 more process has sent help message help-mca-bml-r2.txt / unreachable proc\n-[profitbricks-build5-amd64:37740] Set MCA parameter "orte_base_help_aggregate" to 0 to see all help / error messages\n+[i-capture-the-hostname:47531] *** An error occurred in MPI_Comm_spawn\n+[i-capture-the-hostname:47531] *** reported by process [874971137,0]\n+[i-capture-the-hostname:47531] *** on communicator MPI_COMM_WORLD\n+[i-capture-the-hostname:47531] *** MPI_ERR_INTERN: internal error\n+[i-capture-the-hostname:47531] *** MPI_ERRORS_ARE_FATAL (processes in this communicator will now abort,\n+[i-capture-the-hostname:47531] *** and potentially your MPI job)\n+[i-capture-the-hostname:47649] *** An error occurred in MPI_Init\n+[i-capture-the-hostname:47649] *** reported by process [874971138,0]\n+[i-capture-the-hostname:47649] *** on a NULL communicator\n+[i-capture-the-hostname:47649] *** Unknown error\n+[i-capture-the-hostname:47649] *** MPI_ERRORS_ARE_FATAL (processes in this communicator will now abort,\n+[i-capture-the-hostname:47649] *** and potentially your MPI job)\n+[i-capture-the-hostname:47478] 1 more process has sent help message help-mca-bml-r2.txt / unreachable proc\n+[i-capture-the-hostname:47478] Set MCA parameter "orte_base_help_aggregate" to 0 to see all help / error messages\n \n \n taskmaster\n 2\n ./spawn\n fail\n --------------------------------------------------------------------------\n At least one pair of MPI processes are unable to reach each other for\n MPI communications. This means that no Open MPI device has indicated\n that it can be used to communicate between these processes. This is\n an error; Open MPI requires that all MPI processes be able to reach\n each other. This error can sometimes be the result of forgetting to\n specify the "self" BTL.\n \n- Process 1 ([[20979,1],1]) is on host: profitbricks-build5-amd64\n- Process 2 ([[20979,2],0]) is on host: profitbricks-build5-amd64\n+ Process 1 ([[12569,1],1]) is on host: i-capture-the-hostname\n+ Process 2 ([[12569,2],0]) is on host: i-capture-the-hostname\n BTLs attempted: vader self\n \n Your MPI job is now going to abort; sorry.\n --------------------------------------------------------------------------\n-[profitbricks-build5-amd64:38233] [[20979,1],1] ORTE_ERROR_LOG: Unreachable in file dpm/dpm.c at line 433\n-[profitbricks-build5-amd64:38233] *** An error occurred in MPI_Comm_spawn\n-[profitbricks-build5-amd64:38233] *** reported by process [1374879745,1]\n-[profitbricks-build5-amd64:38233] *** on communicator MPI_COMM_WORLD\n-[profitbricks-build5-amd64:38233] *** MPI_ERR_INTERN: internal error\n-[profitbricks-build5-amd64:38233] *** MPI_ERRORS_ARE_FATAL (processes in this communicator will now abort,\n-[profitbricks-build5-amd64:38233] *** and potentially your MPI job)\n-[profitbricks-build5-amd64:38232] [[20979,1],0] ORTE_ERROR_LOG: Unreachable in file dpm/dpm.c at line 433\n---------------------------------------------------------------------------\n-It looks like MPI_INIT failed for some reason; your parallel process is\n-likely to abort. There are many reasons that a parallel process can\n-fail during MPI_INIT; some of which are due to configuration or environment\n-problems. This failure appears to be an internal failure; here's some\n-additional information (which may only be relevant to an Open MPI\n-developer):\n-\n- ompi_dpm_dyn_init() failed\n- --> Returned "Unreachable" (-12) instead of "Success" (0)\n---------------------------------------------------------------------------\n-[profitbricks-build5-amd64:38240] [[20979,2],0] ORTE_ERROR_LOG: Unreachable in file dpm/dpm.c at line 433\n-[profitbricks-build5-amd64:38177] 2 more processes have sent help message help-mca-bml-r2.txt / unreachable proc\n-[profitbricks-build5-amd64:38177] Set MCA parameter "orte_base_help_aggregate" to 0 to see all help / error messages\n-[profitbricks-build5-amd64:38177] 1 more process has sent help message help-mpi-errors.txt / mpi_errors_are_fatal\n+[i-capture-the-hostname:48848] [[12569,1],0] ORTE_ERROR_LOG: Unreachable in file dpm/dpm.c at line 433\n+[i-capture-the-hostname:48848] *** An error occurred in MPI_Comm_spawn\n+[i-capture-the-hostname:48848] *** reported by process [823721985,0]\n+[i-capture-the-hostname:48848] *** on communicator MPI_COMM_WORLD\n+[i-capture-the-hostname:48848] *** MPI_ERR_INTERN: internal error\n+[i-capture-the-hostname:48848] *** MPI_ERRORS_ARE_FATAL (processes in this communicator will now abort,\n+[i-capture-the-hostname:48848] *** and potentially your MPI job)\n+[i-capture-the-hostname:48854] [[12569,1],1] ORTE_ERROR_LOG: Unreachable in file dpm/dpm.c at line 433\n+[i-capture-the-hostname:48200] 2 more processes have sent help message help-mca-bml-r2.txt / unreachable proc\n+[i-capture-the-hostname:48200] Set MCA parameter "orte_base_help_aggregate" to 0 to see all help / error messages\n+[i-capture-the-hostname:48200] 1 more process has sent help message help-mpi-errors.txt / mpi_errors_are_fatal\n+[warn] Epoll MOD(1) on fd 29 failed. Old events were 6; read change was 0 (none); write change was 2 (del); close change was 0 (none): Bad file descriptor\n \n \n join\n 2\n ./spawn\n fail\n gethostbyname failed\n@@ -1000243,148 +1000216,152 @@\n At least one pair of MPI processes are unable to reach each other for\n MPI communications. This means that no Open MPI device has indicated\n that it can be used to communicate between these processes. This is\n an error; Open MPI requires that all MPI processes be able to reach\n each other. This error can sometimes be the result of forgetting to\n specify the "self" BTL.\n \n- Process 1 ([[21189,1],1]) is on host: profitbricks-build5-amd64\n- Process 2 ([[21189,2],0]) is on host: profitbricks-build5-amd64\n+ Process 1 ([[19253,2],2]) is on host: i-capture-the-hostname\n+ Process 2 ([[19253,1],0]) is on host: i-capture-the-hostname\n BTLs attempted: vader self\n \n Your MPI job is now going to abort; sorry.\n --------------------------------------------------------------------------\n-[profitbricks-build5-amd64:38469] [[21189,1],1] ORTE_ERROR_LOG: Unreachable in file dpm/dpm.c at line 433\n-[profitbricks-build5-amd64:38615] [[21189,2],2] ORTE_ERROR_LOG: Unreachable in file dpm/dpm.c at line 433\n-[profitbricks-build5-amd64:38469] *** An error occurred in MPI_Comm_spawn\n-[profitbricks-build5-amd64:38469] *** reported by process [1388642305,1]\n-[profitbricks-build5-amd64:38469] *** on communicator MPI_COMM_WORLD\n-[profitbricks-build5-amd64:38469] *** MPI_ERR_INTERN: internal error\n-[profitbricks-build5-amd64:38469] *** MPI_ERRORS_ARE_FATAL (processes in this communicator will now abort,\n-[profitbricks-build5-amd64:38469] *** and potentially your MPI job)\n+[i-capture-the-hostname:51399] [[19253,2],0] ORTE_ERROR_LOG: Unreachable in file dpm/dpm.c at line 433\n --------------------------------------------------------------------------\n It looks like MPI_INIT failed for some reason; your parallel process is\n likely to abort. There are many reasons that a parallel process can\n fail during MPI_INIT; some of which are due to configuration or environment\n problems. This failure appears to be an internal failure; here's some\n additional information (which may only be relevant to an Open MPI\n developer):\n \n ompi_dpm_dyn_init() failed\n --> Returned "Unreachable" (-12) instead of "Success" (0)\n --------------------------------------------------------------------------\n-[profitbricks-build5-amd64:38468] [[21189,1],0] ORTE_ERROR_LOG: Unreachable in file dpm/dpm.c at line 433\n-[profitbricks-build5-amd64:38615] *** An error occurred in MPI_Init_thread\n-[profitbricks-build5-amd64:38615] *** reported by process [1388642306,2]\n-[profitbricks-build5-amd64:38615] *** on a NULL communicator\n-[profitbricks-build5-amd64:38615] *** Unknown error\n-[profitbricks-build5-amd64:38615] *** MPI_ERRORS_ARE_FATAL (processes in this communicator will now abort,\n-[profitbricks-build5-amd64:38615] *** and potentially your MPI job)\n-[profitbricks-build5-amd64:38423] 4 more processes have sent help message help-mca-bml-r2.txt / unreachable proc\n-[profitbricks-build5-amd64:38423] Set MCA parameter "orte_base_help_aggregate" to 0 to see all help / error messages\n-[profitbricks-build5-amd64:38423] 1 more process has sent help message help-mpi-errors.txt / mpi_errors_are_fatal\n+[i-capture-the-hostname:51043] [[19253,1],1] ORTE_ERROR_LOG: Unreachable in file dpm/dpm.c at line 433\n+[i-capture-the-hostname:51043] *** An error occurred in MPI_Comm_spawn\n+[i-capture-the-hostname:51043] *** reported by process [1261764609,1]\n+[i-capture-the-hostname:51043] *** on communicator MPI_COMM_WORLD\n+[i-capture-the-hostname:51043] *** MPI_ERR_INTERN: internal error\n+[i-capture-the-hostname:51043] *** MPI_ERRORS_ARE_FATAL (processes in this communicator will now abort,\n+[i-capture-the-hostname:51043] *** and potentially your MPI job)\n+[i-capture-the-hostname:51399] *** An error occurred in MPI_Init_thread\n+[i-capture-the-hostname:51399] *** reported by process [1261764610,0]\n+[i-capture-the-hostname:51399] *** on a NULL communicator\n+[i-capture-the-hostname:51399] *** Unknown error\n+[i-capture-the-hostname:51399] *** MPI_ERRORS_ARE_FATAL (processes in this communicator will now abort,\n+[i-capture-the-hostname:51399] *** and potentially your MPI job)\n+[i-capture-the-hostname:51036] [[19253,1],0] ORTE_ERROR_LOG: Unreachable in file dpm/dpm.c at line 433\n+[i-capture-the-hostname:51417] [[19253,2],2] ORTE_ERROR_LOG: Unreachable in file dpm/dpm.c at line 433\n+[i-capture-the-hostname:50788] 4 more processes have sent help message help-mca-bml-r2.txt / unreachable proc\n+[i-capture-the-hostname:50788] Set MCA parameter "orte_base_help_aggregate" to 0 to see all help / error messages\n+[i-capture-the-hostname:50788] 1 more process has sent help message help-mpi-runtime.txt / mpi_init:startup:internal-failure\n+[i-capture-the-hostname:50788] 1 more process has sent help message help-mpi-errors.txt / mpi_errors_are_fatal\n \n \n disconnect_reconnect2\n 2\n ./spawn\n fail\n --------------------------------------------------------------------------\n At least one pair of MPI processes are unable to reach each other for\n MPI communications. This means that no Open MPI device has indicated\n that it can be used to communicate between these processes. This is\n an error; Open MPI requires that all MPI processes be able to reach\n each other. This error can sometimes be the result of forgetting to\n specify the "self" BTL.\n \n- Process 1 ([[23641,1],1]) is on host: profitbricks-build5-amd64\n- Process 2 ([[23641,2],0]) is on host: profitbricks-build5-amd64\n+ Process 1 ([[18221,2],1]) is on host: i-capture-the-hostname\n+ Process 2 ([[18221,1],0]) is on host: i-capture-the-hostname\n BTLs attempted: vader self\n \n Your MPI job is now going to abort; sorry.\n --------------------------------------------------------------------------\n-[profitbricks-build5-amd64:39150] [[23641,1],1] ORTE_ERROR_LOG: Unreachable in file dpm/dpm.c at line 433\n-[profitbricks-build5-amd64:39150] *** An error occurred in MPI_Comm_spawn\n-[profitbricks-build5-amd64:39150] *** reported by process [1549336577,1]\n-[profitbricks-build5-amd64:39150] *** on communicator MPI_COMM_WORLD\n-[profitbricks-build5-amd64:39150] *** MPI_ERR_INTERN: internal error\n-[profitbricks-build5-amd64:39150] *** MPI_ERRORS_ARE_FATAL (processes in this communicator will now abort,\n-[profitbricks-build5-amd64:39150] *** and potentially your MPI job)\n+[i-capture-the-hostname:52518] [[18221,2],1] ORTE_ERROR_LOG: Unreachable in file dpm/dpm.c at line 433\n+[i-capture-the-hostname:52119] [[18221,1],1] ORTE_ERROR_LOG: Unreachable in file dpm/dpm.c at line 433\n --------------------------------------------------------------------------\n It looks like MPI_INIT failed for some reason; your parallel process is\n likely to abort. There are many reasons that a parallel process can\n fail during MPI_INIT; some of which are due to configuration or environment\n problems. This failure appears to be an internal failure; here's some\n additional information (which may only be relevant to an Open MPI\n developer):\n \n ompi_dpm_dyn_init() failed\n --> Returned "Unreachable" (-12) instead of "Success" (0)\n --------------------------------------------------------------------------\n-[profitbricks-build5-amd64:39148] [[23641,1],0] ORTE_ERROR_LOG: Unreachable in file dpm/dpm.c at line 433\n-[profitbricks-build5-amd64:39271] [[23641,2],2] ORTE_ERROR_LOG: Unreachable in file dpm/dpm.c at line 433\n-[profitbricks-build5-amd64:39270] [[23641,2],1] ORTE_ERROR_LOG: Unreachable in file dpm/dpm.c at line 433\n-[profitbricks-build5-amd64:39051] 4 more processes have sent help message help-mca-bml-r2.txt / unreachable proc\n-[profitbricks-build5-amd64:39051] Set MCA parameter "orte_base_help_aggregate" to 0 to see all help / error messages\n-[profitbricks-build5-amd64:39051] 1 more process has sent help message help-mpi-errors.txt / mpi_errors_are_fatal\n-[profitbricks-build5-amd64:39051] 1 more process has sent help message help-mpi-runtime.txt / mpi_init:startup:internal-failure\n+[i-capture-the-hostname:52119] *** An error occurred in MPI_Comm_spawn\n+[i-capture-the-hostname:52119] *** reported by process [1194131457,1]\n+[i-capture-the-hostname:52119] *** on communicator MPI_COMM_WORLD\n+[i-capture-the-hostname:52119] *** MPI_ERR_INTERN: internal error\n+[i-capture-the-hostname:52119] *** MPI_ERRORS_ARE_FATAL (processes in this communicator will now abort,\n+[i-capture-the-hostname:52119] *** and potentially your MPI job)\n+[i-capture-the-hostname:52518] *** An error occurred in MPI_Init\n+[i-capture-the-hostname:52518] *** reported by process [1194131458,1]\n+[i-capture-the-hostname:52518] *** on a NULL communicator\n+[i-capture-the-hostname:52518] *** Unknown error\n+[i-capture-the-hostname:52518] *** MPI_ERRORS_ARE_FATAL (processes in this communicator will now abort,\n+[i-capture-the-hostname:52518] *** and potentially your MPI job)\n+[i-capture-the-hostname:51836] 4 more processes have sent help message help-mca-bml-r2.txt / unreachable proc\n+[i-capture-the-hostname:51836] Set MCA parameter "orte_base_help_aggregate" to 0 to see all help / error messages\n \n \n disconnect_reconnect3\n 2\n ./spawn\n fail\n --------------------------------------------------------------------------\n At least one pair of MPI processes are unable to reach each other for\n MPI communications. This means that no Open MPI device has indicated\n that it can be used to communicate between these processes. This is\n an error; Open MPI requires that all MPI processes be able to reach\n each other. This error can sometimes be the result of forgetting to\n specify the "self" BTL.\n \n- Process 1 ([[24307,2],0]) is on host: profitbricks-build5-amd64\n- Process 2 ([[24307,1],0]) is on host: profitbricks-build5-amd64\n+ Process 1 ([[16910,2],0]) is on host: i-capture-the-hostname\n+ Process 2 ([[16910,1],0]) is on host: i-capture-the-hostname\n BTLs attempted: vader self\n \n Your MPI job is now going to abort; sorry.\n --------------------------------------------------------------------------\n-[profitbricks-build5-amd64:39665] [[24307,2],0] ORTE_ERROR_LOG: Unreachable in file dpm/dpm.c at line 433\n-[profitbricks-build5-amd64:39668] [[24307,2],1] ORTE_ERROR_LOG: Unreachable in file dpm/dpm.c at line 433\n-[profitbricks-build5-amd64:39672] [[24307,2],2] ORTE_ERROR_LOG: Unreachable in file dpm/dpm.c at line 433\n-[profitbricks-build5-amd64:39677] [[24307,2],3] ORTE_ERROR_LOG: Unreachable in file dpm/dpm.c at line 433\n-[profitbricks-build5-amd64:39562] [[24307,1],0] ORTE_ERROR_LOG: Unreachable in file dpm/dpm.c at line 433\n-[profitbricks-build5-amd64:39563] [[24307,1],1] ORTE_ERROR_LOG: Unreachable in file dpm/dpm.c at line 433\n-[profitbricks-build5-amd64:39562] *** An error occurred in MPI_Comm_spawn\n-[profitbricks-build5-amd64:39562] *** reported by process [1592983553,0]\n-[profitbricks-build5-amd64:39562] *** on communicator MPI_COMM_WORLD\n-[profitbricks-build5-amd64:39562] *** MPI_ERR_INTERN: internal error\n-[profitbricks-build5-amd64:39562] *** MPI_ERRORS_ARE_FATAL (processes in this communicator will now abort,\n-[profitbricks-build5-amd64:39562] *** and potentially your MPI job)\n+[i-capture-the-hostname:53900] [[16910,2],0] ORTE_ERROR_LOG: Unreachable in file dpm/dpm.c at line 433\n+[i-capture-the-hostname:53921] [[16910,2],2] ORTE_ERROR_LOG: Unreachable in file dpm/dpm.c at line 433\n --------------------------------------------------------------------------\n It looks like MPI_INIT failed for some reason; your parallel process is\n likely to abort. There are many reasons that a parallel process can\n fail during MPI_INIT; some of which are due to configuration or environment\n problems. This failure appears to be an internal failure; here's some\n additional information (which may only be relevant to an Open MPI\n developer):\n \n ompi_dpm_dyn_init() failed\n --> Returned "Unreachable" (-12) instead of "Success" (0)\n --------------------------------------------------------------------------\n-[profitbricks-build5-amd64:39672] *** An error occurred in MPI_Init_thread\n-[profitbricks-build5-amd64:39672] *** reported by process [1592983554,2]\n-[profitbricks-build5-amd64:39672] *** on a NULL communicator\n-[profitbricks-build5-amd64:39672] *** Unknown error\n-[profitbricks-build5-amd64:39672] *** MPI_ERRORS_ARE_FATAL (processes in this communicator will now abort,\n-[profitbricks-build5-amd64:39672] *** and potentially your MPI job)\n-[profitbricks-build5-amd64:39457] 5 more processes have sent help message help-mca-bml-r2.txt / unreachable proc\n-[profitbricks-build5-amd64:39457] Set MCA parameter "orte_base_help_aggregate" to 0 to see all help / error messages\n-[profitbricks-build5-amd64:39457] 1 more process has sent help message help-mpi-errors.txt / mpi_errors_are_fatal\n-[profitbricks-build5-amd64:39457] 3 more processes have sent help message help-mpi-runtime.txt / mpi_init:startup:internal-failure\n-[profitbricks-build5-amd64:39457] 3 more processes have sent help message help-mpi-errors.txt / mpi_errors_are_fatal unknown handle\n+[i-capture-the-hostname:53436] *** An error occurred in MPI_Comm_spawn\n+[i-capture-the-hostname:53436] *** reported by process [1108213761,1]\n+[i-capture-the-hostname:53436] *** on communicator MPI_COMM_WORLD\n+[i-capture-the-hostname:53436] *** MPI_ERR_INTERN: internal error\n+[i-capture-the-hostname:53436] *** MPI_ERRORS_ARE_FATAL (processes in this communicator will now abort,\n+[i-capture-the-hostname:53436] *** and potentially your MPI job)\n+[i-capture-the-hostname:53921] *** An error occurred in MPI_Init_thread\n+[i-capture-the-hostname:53921] *** reported by process [1108213762,2]\n+[i-capture-the-hostname:53921] *** on a NULL communicator\n+[i-capture-the-hostname:53921] *** Unknown error\n+[i-capture-the-hostname:53921] *** MPI_ERRORS_ARE_FATAL (processes in this communicator will now abort,\n+[i-capture-the-hostname:53921] *** and potentially your MPI job)\n+[i-capture-the-hostname:53428] [[16910,1],0] ORTE_ERROR_LOG: Unreachable in file dpm/dpm.c at line 433\n+[i-capture-the-hostname:53436] [[16910,1],1] ORTE_ERROR_LOG: Unreachable in file dpm/dpm.c at line 433\n+[i-capture-the-hostname:53930] [[16910,2],3] ORTE_ERROR_LOG: Unreachable in file dpm/dpm.c at line 433\n+[i-capture-the-hostname:53911] [[16910,2],1] ORTE_ERROR_LOG: Unreachable in file dpm/dpm.c at line 433\n+[i-capture-the-hostname:53087] 5 more processes have sent help message help-mca-bml-r2.txt / unreachable proc\n+[i-capture-the-hostname:53087] Set MCA parameter "orte_base_help_aggregate" to 0 to see all help / error messages\n+[i-capture-the-hostname:53087] 3 more processes have sent help message help-mpi-runtime.txt / mpi_init:startup:internal-failure\n+[i-capture-the-hostname:53087] 1 more process has sent help message help-mpi-errors.txt / mpi_errors_are_fatal\n+[i-capture-the-hostname:53087] 3 more processes have sent help message help-mpi-errors.txt / mpi_errors_are_fatal unknown handle\n \n \n multiple_ports\n 2\n ./spawn\n fail\n Three processes needed to run this test.\n@@ -1000407,174 +1000384,199 @@\n At least one pair of MPI processes are unable to reach each other for\n MPI communications. This means that no Open MPI device has indicated\n that it can be used to communicate between these processes. This is\n an error; Open MPI requires that all MPI processes be able to reach\n each other. This error can sometimes be the result of forgetting to\n specify the "self" BTL.\n \n- Process 1 ([[23265,2],2]) is on host: profitbricks-build5-amd64\n- Process 2 ([[23265,1],0]) is on host: profitbricks-build5-amd64\n+ Process 1 ([[21815,1],0]) is on host: i-capture-the-hostname\n+ Process 2 ([[21815,2],0]) is on host: i-capture-the-hostname\n BTLs attempted: vader self\n \n Your MPI job is now going to abort; sorry.\n --------------------------------------------------------------------------\n-[profitbricks-build5-amd64:40873] [[23265,2],2] ORTE_ERROR_LOG: Unreachable in file dpm/dpm.c at line 433\n+[i-capture-the-hostname:55553] [[21815,1],0] ORTE_ERROR_LOG: Unreachable in file dpm/dpm.c at line 433\n+[i-capture-the-hostname:55553] *** An error occurred in MPI_Comm_spawn\n+[i-capture-the-hostname:55553] *** reported by process [1429667841,0]\n+[i-capture-the-hostname:55553] *** on communicator MPI_COMM_WORLD\n+[i-capture-the-hostname:55553] *** MPI_ERR_INTERN: internal error\n+[i-capture-the-hostname:55553] *** MPI_ERRORS_ARE_FATAL (processes in this communicator will now abort,\n+[i-capture-the-hostname:55553] *** and potentially your MPI job)\n+[i-capture-the-hostname:55756] [[21815,2],0] ORTE_ERROR_LOG: Unreachable in file dpm/dpm.c at line 433\n+[i-capture-the-hostname:55762] [[21815,2],1] ORTE_ERROR_LOG: Unreachable in file dpm/dpm.c at line 433\n+[i-capture-the-hostname:55766] [[21815,2],2] ORTE_ERROR_LOG: Unreachable in file dpm/dpm.c at line 433\n+[i-capture-the-hostname:55560] [[21815,1],1] ORTE_ERROR_LOG: Unreachable in file dpm/dpm.c at line 433\n --------------------------------------------------------------------------\n It looks like MPI_INIT failed for some reason; your parallel process is\n likely to abort. There are many reasons that a parallel process can\n fail during MPI_INIT; some of which are due to configuration or environment\n problems. This failure appears to be an internal failure; here's some\n additional information (which may only be relevant to an Open MPI\n developer):\n \n ompi_dpm_dyn_init() failed\n --> Returned "Unreachable" (-12) instead of "Success" (0)\n --------------------------------------------------------------------------\n-[profitbricks-build5-amd64:40873] *** An error occurred in MPI_Init_thread\n-[profitbricks-build5-amd64:40873] *** reported by process [1524695042,2]\n-[profitbricks-build5-amd64:40873] *** on a NULL communicator\n-[profitbricks-build5-amd64:40873] *** Unknown error\n-[profitbricks-build5-amd64:40873] *** MPI_ERRORS_ARE_FATAL (processes in this communicator will now abort,\n-[profitbricks-build5-amd64:40873] *** and potentially your MPI job)\n-[profitbricks-build5-amd64:40499] 4 more processes have sent help message help-mca-bml-r2.txt / unreachable proc\n-[profitbricks-build5-amd64:40499] Set MCA parameter "orte_base_help_aggregate" to 0 to see all help / error messages\n+[i-capture-the-hostname:55766] *** An error occurred in MPI_Init_thread\n+[i-capture-the-hostname:55766] *** reported by process [1429667842,2]\n+[i-capture-the-hostname:55766] *** on a NULL communicator\n+[i-capture-the-hostname:55766] *** Unknown error\n+[i-capture-the-hostname:55766] *** MPI_ERRORS_ARE_FATAL (processes in this communicator will now abort,\n+[i-capture-the-hostname:55766] *** and potentially your MPI job)\n+[i-capture-the-hostname:55398] 4 more processes have sent help message help-mca-bml-r2.txt / unreachable proc\n+[i-capture-the-hostname:55398] Set MCA parameter "orte_base_help_aggregate" to 0 to see all help / error messages\n+[i-capture-the-hostname:55398] 1 more process has sent help message help-mpi-errors.txt / mpi_errors_are_fatal\n+[i-capture-the-hostname:55398] 2 more processes have sent help message help-mpi-runtime.txt / mpi_init:startup:internal-failure\n+[i-capture-the-hostname:55398] 2 more processes have sent help message help-mpi-errors.txt / mpi_errors_are_fatal unknown handle\n \n \n disconnect2\n 2\n ./spawn\n fail\n --------------------------------------------------------------------------\n At least one pair of MPI processes are unable to reach each other for\n MPI communications. This means that no Open MPI device has indicated\n that it can be used to communicate between these processes. This is\n an error; Open MPI requires that all MPI processes be able to reach\n each other. This error can sometimes be the result of forgetting to\n specify the "self" BTL.\n \n- Process 1 ([[25777,1],1]) is on host: profitbricks-build5-amd64\n- Process 2 ([[25777,2],0]) is on host: profitbricks-build5-amd64\n+ Process 1 ([[22521,2],0]) is on host: i-capture-the-hostname\n+ Process 2 ([[22521,1],0]) is on host: i-capture-the-hostname\n BTLs attempted: vader self\n \n Your MPI job is now going to abort; sorry.\n --------------------------------------------------------------------------\n-[profitbricks-build5-amd64:41184] [[25777,1],1] ORTE_ERROR_LOG: Unreachable in file dpm/dpm.c at line 433\n-[profitbricks-build5-amd64:41184] *** An error occurred in MPI_Comm_spawn\n-[profitbricks-build5-amd64:41184] *** reported by process [1689321473,1]\n-[profitbricks-build5-amd64:41184] *** on communicator MPI_COMM_WORLD\n-[profitbricks-build5-amd64:41184] *** MPI_ERR_INTERN: internal error\n-[profitbricks-build5-amd64:41184] *** MPI_ERRORS_ARE_FATAL (processes in this communicator will now abort,\n-[profitbricks-build5-amd64:41184] *** and potentially your MPI job)\n-[profitbricks-build5-amd64:41311] [[25777,2],2] ORTE_ERROR_LOG: Unreachable in file dpm/dpm.c at line 433\n --------------------------------------------------------------------------\n It looks like MPI_INIT failed for some reason; your parallel process is\n likely to abort. There are many reasons that a parallel process can\n fail during MPI_INIT; some of which are due to configuration or environment\n problems. This failure appears to be an internal failure; here's some\n additional information (which may only be relevant to an Open MPI\n developer):\n \n ompi_dpm_dyn_init() failed\n --> Returned "Unreachable" (-12) instead of "Success" (0)\n --------------------------------------------------------------------------\n-[profitbricks-build5-amd64:41301] [[25777,2],0] ORTE_ERROR_LOG: Unreachable in file dpm/dpm.c at line 433\n-[profitbricks-build5-amd64:41301] *** An error occurred in MPI_Init_thread\n-[profitbricks-build5-amd64:41301] *** reported by process [1689321474,0]\n-[profitbricks-build5-amd64:41301] *** on a NULL communicator\n-[profitbricks-build5-amd64:41301] *** Unknown error\n-[profitbricks-build5-amd64:41301] *** MPI_ERRORS_ARE_FATAL (processes in this communicator will now abort,\n-[profitbricks-build5-amd64:41301] *** and potentially your MPI job)\n-[profitbricks-build5-amd64:41182] [[25777,1],0] ORTE_ERROR_LOG: Unreachable in file dpm/dpm.c at line 433\n-[profitbricks-build5-amd64:41307] [[25777,2],1] ORTE_ERROR_LOG: Unreachable in file dpm/dpm.c at line 433\n-[profitbricks-build5-amd64:41059] 4 more processes have sent help message help-mca-bml-r2.txt / unreachable proc\n-[profitbricks-build5-amd64:41059] Set MCA parameter "orte_base_help_aggregate" to 0 to see all help / error messages\n-[profitbricks-build5-amd64:41059] 1 more process has sent help message help-mpi-errors.txt / mpi_errors_are_fatal\n-[profitbricks-build5-amd64:41059] 2 more processes have sent help message help-mpi-runtime.txt / mpi_init:startup:internal-failure\n-[profitbricks-build5-amd64:41059] 1 more process has sent help message help-mpi-errors.txt / mpi_errors_are_fatal unknown handle\n+[i-capture-the-hostname:56076] [[22521,1],0] ORTE_ERROR_LOG: Unreachable in file dpm/dpm.c at line 433\n+[i-capture-the-hostname:56085] [[22521,1],1] ORTE_ERROR_LOG: Unreachable in file dpm/dpm.c at line 433\n+[i-capture-the-hostname:56224] [[22521,2],0] ORTE_ERROR_LOG: Unreachable in file dpm/dpm.c at line 433\n+[i-capture-the-hostname:56233] [[22521,2],2] ORTE_ERROR_LOG: Unreachable in file dpm/dpm.c at line 433\n+[i-capture-the-hostname:56229] [[22521,2],1] ORTE_ERROR_LOG: Unreachable in file dpm/dpm.c at line 433\n+[i-capture-the-hostname:56085] *** An error occurred in MPI_Comm_spawn\n+[i-capture-the-hostname:56085] *** reported by process [1475936257,1]\n+[i-capture-the-hostname:56085] *** on communicator MPI_COMM_WORLD\n+[i-capture-the-hostname:56085] *** MPI_ERR_INTERN: internal error\n+[i-capture-the-hostname:56085] *** MPI_ERRORS_ARE_FATAL (processes in this communicator will now abort,\n+[i-capture-the-hostname:56085] *** and potentially your MPI job)\n+[i-capture-the-hostname:56224] *** An error occurred in MPI_Init_thread\n+[i-capture-the-hostname:56224] *** reported by process [1475936258,0]\n+[i-capture-the-hostname:56224] *** on a NULL communicator\n+[i-capture-the-hostname:56224] *** Unknown error\n+[i-capture-the-hostname:56224] *** MPI_ERRORS_ARE_FATAL (processes in this communicator will now abort,\n+[i-capture-the-hostname:56224] *** and potentially your MPI job)\n+[i-capture-the-hostname:55976] 4 more processes have sent help message help-mca-bml-r2.txt / unreachable proc\n+[i-capture-the-hostname:55976] Set MCA parameter "orte_base_help_aggregate" to 0 to see all help / error messages\n+[i-capture-the-hostname:55976] 2 more processes have sent help message help-mpi-runtime.txt / mpi_init:startup:internal-failure\n+[i-capture-the-hostname:55976] 1 more process has sent help message help-mpi-errors.txt / mpi_errors_are_fatal\n+[i-capture-the-hostname:55976] 2 more processes have sent help message help-mpi-errors.txt / mpi_errors_are_fatal unknown handle\n \n \n disconnect3\n 2\n ./spawn\n fail\n --------------------------------------------------------------------------\n At least one pair of MPI processes are unable to reach each other for\n MPI communications. This means that no Open MPI device has indicated\n that it can be used to communicate between these processes. This is\n an error; Open MPI requires that all MPI processes be able to reach\n each other. This error can sometimes be the result of forgetting to\n specify the "self" BTL.\n \n- Process 1 ([[26263,2],0]) is on host: profitbricks-build5-amd64\n- Process 2 ([[26263,1],0]) is on host: profitbricks-build5-amd64\n+ Process 1 ([[20933,1],0]) is on host: i-capture-the-hostname\n+ Process 2 ([[20933,2],0]) is on host: i-capture-the-hostname\n BTLs attempted: vader self\n \n Your MPI job is now going to abort; sorry.\n --------------------------------------------------------------------------\n-[profitbricks-build5-amd64:41806] [[26263,2],0] ORTE_ERROR_LOG: Unreachable in file dpm/dpm.c at line 433\n-[profitbricks-build5-amd64:41692] [[26263,1],1] ORTE_ERROR_LOG: Unreachable in file dpm/dpm.c at line 433\n-[profitbricks-build5-amd64:41810] [[26263,2],1] ORTE_ERROR_LOG: Unreachable in file dpm/dpm.c at line 433\n+[i-capture-the-hostname:56575] [[20933,1],0] ORTE_ERROR_LOG: Unreachable in file dpm/dpm.c at line 433\n+[i-capture-the-hostname:56575] *** An error occurred in MPI_Comm_spawn\n+[i-capture-the-hostname:56575] *** reported by process [1371865089,0]\n+[i-capture-the-hostname:56575] *** on communicator MPI_COMM_WORLD\n+[i-capture-the-hostname:56575] *** MPI_ERR_INTERN: internal error\n+[i-capture-the-hostname:56575] *** MPI_ERRORS_ARE_FATAL (processes in this communicator will now abort,\n+[i-capture-the-hostname:56575] *** and potentially your MPI job)\n+[i-capture-the-hostname:56578] [[20933,1],1] ORTE_ERROR_LOG: Unreachable in file dpm/dpm.c at line 433\n --------------------------------------------------------------------------\n It looks like MPI_INIT failed for some reason; your parallel process is\n likely to abort. There are many reasons that a parallel process can\n fail during MPI_INIT; some of which are due to configuration or environment\n problems. This failure appears to be an internal failure; here's some\n additional information (which may only be relevant to an Open MPI\n developer):\n \n ompi_dpm_dyn_init() failed\n --> Returned "Unreachable" (-12) instead of "Success" (0)\n --------------------------------------------------------------------------\n-[profitbricks-build5-amd64:41692] *** An error occurred in MPI_Comm_spawn\n-[profitbricks-build5-amd64:41692] *** reported by process [1721171969,1]\n-[profitbricks-build5-amd64:41692] *** on communicator MPI_COMM_WORLD\n-[profitbricks-build5-amd64:41692] *** MPI_ERR_INTERN: internal error\n-[profitbricks-build5-amd64:41692] *** MPI_ERRORS_ARE_FATAL (processes in this communicator will now abort,\n-[profitbricks-build5-amd64:41692] *** and potentially your MPI job)\n-[profitbricks-build5-amd64:41811] [[26263,2],2] ORTE_ERROR_LOG: Unreachable in file dpm/dpm.c at line 433\n-[profitbricks-build5-amd64:41810] *** An error occurred in MPI_Init_thread\n-[profitbricks-build5-amd64:41810] *** reported by process [1721171970,1]\n-[profitbricks-build5-amd64:41810] *** on a NULL communicator\n-[profitbricks-build5-amd64:41810] *** Unknown error\n-[profitbricks-build5-amd64:41810] *** MPI_ERRORS_ARE_FATAL (processes in this communicator will now abort,\n-[profitbricks-build5-amd64:41810] *** and potentially your MPI job)\n-[profitbricks-build5-amd64:41541] 4 more processes have sent help message help-mca-bml-r2.txt / unreachable proc\n-[profitbricks-build5-amd64:41541] Set MCA parameter "orte_base_help_aggregate" to 0 to see all help / error messages\n-[profitbricks-build5-amd64:41541] 2 more processes have sent help message help-mpi-runtime.txt / mpi_init:startup:internal-failure\n-[profitbricks-build5-amd64:41541] 2 more processes have sent help message help-mpi-errors.txt / mpi_errors_are_fatal unknown handle\n-[warn] Epoll MOD(1) on fd 31 failed. Old events were 6; read change was 0 (none); write change was 2 (del); close change was 0 (none): Bad file descriptor\n+[i-capture-the-hostname:56698] [[20933,2],2] ORTE_ERROR_LOG: Unreachable in file dpm/dpm.c at line 433\n+[i-capture-the-hostname:56695] [[20933,2],0] ORTE_ERROR_LOG: Unreachable in file dpm/dpm.c at line 433\n+[i-capture-the-hostname:56697] [[20933,2],1] ORTE_ERROR_LOG: Unreachable in file dpm/dpm.c at line 433\n+[i-capture-the-hostname:56468] 4 more processes have sent help message help-mca-bml-r2.txt / unreachable proc\n+[i-capture-the-hostname:56468] Set MCA parameter "orte_base_help_aggregate" to 0 to see all help / error messages\n+[i-capture-the-hostname:56468] 1 more process has sent help message help-mpi-errors.txt / mpi_errors_are_fatal\n+[i-capture-the-hostname:56468] 2 more processes have sent help message help-mpi-runtime.txt / mpi_init:startup:internal-failure\n \n \n concurrent_spawns\n 1\n ./spawn\n fail\n --------------------------------------------------------------------------\n At least one pair of MPI processes are unable to reach each other for\n MPI communications. This means that no Open MPI device has indicated\n that it can be used to communicate between these processes. This is\n an error; Open MPI requires that all MPI processes be able to reach\n each other. This error can sometimes be the result of forgetting to\n specify the "self" BTL.\n \n- Process 1 ([[24801,1],0]) is on host: profitbricks-build5-amd64\n- Process 2 ([[24801,2],0]) is on host: profitbricks-build5-amd64\n+ Process 1 ([[21366,2],0]) is on host: i-capture-the-hostname\n+ Process 2 ([[21366,1],0]) is on host: i-capture-the-hostname\n BTLs attempted: self\n \n Your MPI job is now going to abort; sorry.\n --------------------------------------------------------------------------\n-[profitbricks-build5-amd64:42156] [[24801,1],0] ORTE_ERROR_LOG: Unreachable in file dpm/dpm.c at line 433\n-[profitbricks-build5-amd64:42156] *** An error occurred in MPI_Comm_spawn\n-[profitbricks-build5-amd64:42156] *** reported by process [1625358337,0]\n-[profitbricks-build5-amd64:42156] *** on communicator MPI_COMM_WORLD\n-[profitbricks-build5-amd64:42156] *** MPI_ERR_INTERN: internal error\n-[profitbricks-build5-amd64:42156] *** MPI_ERRORS_ARE_FATAL (processes in this communicator will now abort,\n-[profitbricks-build5-amd64:42156] *** and potentially your MPI job)\n-[profitbricks-build5-amd64:42035] 1 more process has sent help message help-mca-bml-r2.txt / unreachable proc\n-[profitbricks-build5-amd64:42035] Set MCA parameter "orte_base_help_aggregate" to 0 to see all help / error messages\n+[i-capture-the-hostname:57123] [[21366,2],0] ORTE_ERROR_LOG: Unreachable in file dpm/dpm.c at line 433\n+--------------------------------------------------------------------------\n+It looks like MPI_INIT failed for some reason; your parallel process is\n+likely to abort. There are many reasons that a parallel process can\n+fail during MPI_INIT; some of which are due to configuration or environment\n+problems. This failure appears to be an internal failure; here's some\n+additional information (which may only be relevant to an Open MPI\n+developer):\n+\n+ ompi_dpm_dyn_init() failed\n+ --> Returned "Unreachable" (-12) instead of "Success" (0)\n+--------------------------------------------------------------------------\n+[i-capture-the-hostname:56968] [[21366,1],0] ORTE_ERROR_LOG: Unreachable in file dpm/dpm.c at line 433\n+[i-capture-the-hostname:56968] *** An error occurred in MPI_Comm_spawn\n+[i-capture-the-hostname:56968] *** reported by process [1400242177,0]\n+[i-capture-the-hostname:56968] *** on communicator MPI_COMM_WORLD\n+[i-capture-the-hostname:56968] *** MPI_ERR_INTERN: internal error\n+[i-capture-the-hostname:56968] *** MPI_ERRORS_ARE_FATAL (processes in this communicator will now abort,\n+[i-capture-the-hostname:56968] *** and potentially your MPI job)\n+[i-capture-the-hostname:57123] *** An error occurred in MPI_Init_thread\n+[i-capture-the-hostname:57123] *** reported by process [1400242178,0]\n+[i-capture-the-hostname:57123] *** on a NULL communicator\n+[i-capture-the-hostname:57123] *** Unknown error\n+[i-capture-the-hostname:57123] *** MPI_ERRORS_ARE_FATAL (processes in this communicator will now abort,\n+[i-capture-the-hostname:57123] *** and potentially your MPI job)\n+[i-capture-the-hostname:56871] 1 more process has sent help message help-mca-bml-r2.txt / unreachable proc\n+[i-capture-the-hostname:56871] Set MCA parameter "orte_base_help_aggregate" to 0 to see all help / error messages\n \n \n pgroup_connect_test\n 2\n ./spawn\n pass\n \n@@ -1000590,28 +1000592,28 @@\n ./spawn\n fail\n --------------------------------------------------------------------------\n mpiexec.openmpi was unable to launch the specified application as it could not access\n or execute an executable:\n \n Executable: ./spawn-nullargs\n-Node: profitbricks-build5-amd64\n+Node: i-capture-the-hostname\n \n while attempting to start process rank 0.\n --------------------------------------------------------------------------\n 5 total processes failed to start\n \n \n cartmap1\n 2\n ./topo\n fail\n- Found 2 errors\n+ rank outside of input communicator not UNDEFINED\n rank outside of input communicator not UNDEFINED\n-rank outside of input communicator not UNDEFINED\n+ Found 2 errors\n \n \n cartzero\n 2\n ./topo\n pass\n \n@@ -1000660,16 +1000662,16 @@\n pass\n \n \n graphmap1\n 2\n ./topo\n fail\n- Found 1 errors\n-Graph map with no local nodes did not return MPI_UNDEFINED\n+ Graph map with no local nodes did not return MPI_UNDEFINED\n+ Found 1 errors\n \n \n topotest\n 2\n ./topo\n pass\n \n@@ -1000692,65 +1000694,65 @@\n pass\n \n \n distgraph1\n 2\n ./topo\n fail\n- Error: Cannot partition 16 elements in 17 parts\n-[profitbricks-build5-amd64:52290] *** Process received signal ***\n-[profitbricks-build5-amd64:52290] Signal: Segmentation fault (11)\n-[profitbricks-build5-amd64:52290] Signal code: Address not mapped (1)\n-[profitbricks-build5-amd64:52290] Failing at address: (nil)\n-[profitbricks-build5-amd64:52290] [ 0] /lib/x86_64-linux-gnu/libpthread.so.0(+0x126b0)[0x7fa3a945a6b0]\n-[profitbricks-build5-amd64:52290] [ 1] /usr/lib/x86_64-linux-gnu/openmpi/lib/openmpi3/mca_topo_treematch.so(split_com_mat+0xab)[0x7fa3a5d6de8b]\n-[profitbricks-build5-amd64:52290] [ 2] /usr/lib/x86_64-linux-gnu/openmpi/lib/openmpi3/mca_topo_treematch.so(kpartition_build_level_topology+0xa3)[0x7fa3a5d6e423]\n-[profitbricks-build5-amd64:52290] [ 3] /usr/lib/x86_64-linux-gnu/openmpi/lib/openmpi3/mca_topo_treematch.so(kpartition_build_tree_from_topology+0x172)[0x7fa3a5d6e7c2]\n-[profitbricks-build5-amd64:52290] [ 4] /usr/lib/x86_64-linux-gnu/openmpi/lib/openmpi3/mca_topo_treematch.so(tm_build_tree_from_topology+0xbd)[0x7fa3a5d6b28d]\n-[profitbricks-build5-amd64:52290] [ 5] /usr/lib/x86_64-linux-gnu/openmpi/lib/openmpi3/mca_topo_treematch.so(mca_topo_treematch_dist_graph_create+0x2959)[0x7fa3a5d619e9]\n-[profitbricks-build5-amd64:52290] [ 6] /usr/lib/x86_64-linux-gnu/libmpi.so.40(MPI_Dist_graph_create+0x21d)[0x7fa3a94d93cd]\n-[profitbricks-build5-amd64:52290] [ 7] ./distgraph1(+0x39e7)[0x556bf805a9e7]\n-[profitbricks-build5-amd64:52290] [ 8] /lib/x86_64-linux-gnu/libc.so.6(__libc_start_main+0xeb)[0x7fa3a92a409b]\n-[profitbricks-build5-amd64:52290] [ 9] ./distgraph1(+0x497a)[0x556bf805b97a]\n-[profitbricks-build5-amd64:52290] *** End of error message ***\n+ Error: Cannot partition 15 elements in 2 parts\n+[i-capture-the-hostname:60427] *** Process received signal ***\n+[i-capture-the-hostname:60427] Signal: Segmentation fault (11)\n+[i-capture-the-hostname:60427] Signal code: Address not mapped (1)\n+[i-capture-the-hostname:60427] Failing at address: (nil)\n+[i-capture-the-hostname:60427] [ 0] /lib/x86_64-linux-gnu/libpthread.so.0(+0x126b0)[0x7fd3796d56b0]\n+[i-capture-the-hostname:60427] [ 1] /usr/lib/x86_64-linux-gnu/openmpi/lib/openmpi3/mca_topo_treematch.so(split_com_mat+0xab)[0x7fd375fe8e8b]\n+[i-capture-the-hostname:60427] [ 2] /usr/lib/x86_64-linux-gnu/openmpi/lib/openmpi3/mca_topo_treematch.so(kpartition_build_level_topology+0xa3)[0x7fd375fe9423]\n+[i-capture-the-hostname:60427] [ 3] /usr/lib/x86_64-linux-gnu/openmpi/lib/openmpi3/mca_topo_treematch.so(kpartition_build_tree_from_topology+0x172)[0x7fd375fe97c2]\n+[i-capture-the-hostname:60427] [ 4] /usr/lib/x86_64-linux-gnu/openmpi/lib/openmpi3/mca_topo_treematch.so(tm_build_tree_from_topology+0xbd)[0x7fd375fe628d]\n+[i-capture-the-hostname:60427] [ 5] /usr/lib/x86_64-linux-gnu/openmpi/lib/openmpi3/mca_topo_treematch.so(mca_topo_treematch_dist_graph_create+0x2959)[0x7fd375fdc9e9]\n+[i-capture-the-hostname:60427] [ 6] /usr/lib/x86_64-linux-gnu/libmpi.so.40(MPI_Dist_graph_create+0x21d)[0x7fd3797543cd]\n+[i-capture-the-hostname:60427] [ 7] ./distgraph1(+0x39e7)[0x55bccf6859e7]\n+[i-capture-the-hostname:60427] [ 8] /lib/x86_64-linux-gnu/libc.so.6(__libc_start_main+0xeb)[0x7fd37951f09b]\n+[i-capture-the-hostname:60427] [ 9] ./distgraph1(+0x497a)[0x55bccf68697a]\n+[i-capture-the-hostname:60427] *** End of error message ***\n --------------------------------------------------------------------------\n Primary job terminated normally, but 1 process returned\n a non-zero exit code. Per user-direction, the job has been aborted.\n --------------------------------------------------------------------------\n --------------------------------------------------------------------------\n-mpiexec.openmpi noticed that process rank 0 with PID 0 on node profitbricks-build5-amd64 exited on signal 11 (Segmentation fault).\n+mpiexec.openmpi noticed that process rank 0 with PID 0 on node i-capture-the-hostname exited on signal 11 (Segmentation fault).\n --------------------------------------------------------------------------\n \n \n dgraph_unwgt\n 2\n ./topo\n fail\n- Error: Cannot partition 16 elements in 17 parts\n-[profitbricks-build5-amd64:52743] *** Process received signal ***\n-[profitbricks-build5-amd64:52743] Signal: Segmentation fault (11)\n-[profitbricks-build5-amd64:52743] Signal code: Address not mapped (1)\n-[profitbricks-build5-amd64:52743] Failing at address: (nil)\n-[profitbricks-build5-amd64:52743] [ 0] /lib/x86_64-linux-gnu/libpthread.so.0(+0x126b0)[0x7f30493126b0]\n-[profitbricks-build5-amd64:52743] [ 1] /usr/lib/x86_64-linux-gnu/openmpi/lib/openmpi3/mca_topo_treematch.so(split_com_mat+0xab)[0x7f3045c25e8b]\n-[profitbricks-build5-amd64:52743] [ 2] /usr/lib/x86_64-linux-gnu/openmpi/lib/openmpi3/mca_topo_treematch.so(kpartition_build_level_topology+0xa3)[0x7f3045c26423]\n-[profitbricks-build5-amd64:52743] [ 3] /usr/lib/x86_64-linux-gnu/openmpi/lib/openmpi3/mca_topo_treematch.so(kpartition_build_tree_from_topology+0x172)[0x7f3045c267c2]\n-[profitbricks-build5-amd64:52743] [ 4] /usr/lib/x86_64-linux-gnu/openmpi/lib/openmpi3/mca_topo_treematch.so(tm_build_tree_from_topology+0xbd)[0x7f3045c2328d]\n-[profitbricks-build5-amd64:52743] [ 5] /usr/lib/x86_64-linux-gnu/openmpi/lib/openmpi3/mca_topo_treematch.so(mca_topo_treematch_dist_graph_create+0x2959)[0x7f3045c199e9]\n-[profitbricks-build5-amd64:52743] [ 6] /usr/lib/x86_64-linux-gnu/libmpi.so.40(MPI_Dist_graph_create+0x21d)[0x7f30493913cd]\n-[profitbricks-build5-amd64:52743] [ 7] ./dgraph_unwgt(+0x25e7)[0x556c34e935e7]\n-[profitbricks-build5-amd64:52743] [ 8] /lib/x86_64-linux-gnu/libc.so.6(__libc_start_main+0xeb)[0x7f304915c09b]\n-[profitbricks-build5-amd64:52743] [ 9] ./dgraph_unwgt(+0x278a)[0x556c34e9378a]\n-[profitbricks-build5-amd64:52743] *** End of error message ***\n+ Error: Cannot partition 15 elements in 2 parts\n+[i-capture-the-hostname:60703] *** Process received signal ***\n+[i-capture-the-hostname:60703] Signal: Segmentation fault (11)\n+[i-capture-the-hostname:60703] Signal code: Address not mapped (1)\n+[i-capture-the-hostname:60703] Failing at address: (nil)\n+[i-capture-the-hostname:60703] [ 0] /lib/x86_64-linux-gnu/libpthread.so.0(+0x126b0)[0x7f1d29a356b0]\n+[i-capture-the-hostname:60703] [ 1] /usr/lib/x86_64-linux-gnu/openmpi/lib/openmpi3/mca_topo_treematch.so(split_com_mat+0xab)[0x7f1d26348e8b]\n+[i-capture-the-hostname:60703] [ 2] /usr/lib/x86_64-linux-gnu/openmpi/lib/openmpi3/mca_topo_treematch.so(kpartition_build_level_topology+0xa3)[0x7f1d26349423]\n+[i-capture-the-hostname:60703] [ 3] /usr/lib/x86_64-linux-gnu/openmpi/lib/openmpi3/mca_topo_treematch.so(kpartition_build_tree_from_topology+0x172)[0x7f1d263497c2]\n+[i-capture-the-hostname:60703] [ 4] /usr/lib/x86_64-linux-gnu/openmpi/lib/openmpi3/mca_topo_treematch.so(tm_build_tree_from_topology+0xbd)[0x7f1d2634628d]\n+[i-capture-the-hostname:60703] [ 5] /usr/lib/x86_64-linux-gnu/openmpi/lib/openmpi3/mca_topo_treematch.so(mca_topo_treematch_dist_graph_create+0x2959)[0x7f1d2633c9e9]\n+[i-capture-the-hostname:60703] [ 6] /usr/lib/x86_64-linux-gnu/libmpi.so.40(MPI_Dist_graph_create+0x21d)[0x7f1d29ab43cd]\n+[i-capture-the-hostname:60703] [ 7] ./dgraph_unwgt(+0x25e7)[0x55b68d46f5e7]\n+[i-capture-the-hostname:60703] [ 8] /lib/x86_64-linux-gnu/libc.so.6(__libc_start_main+0xeb)[0x7f1d2987f09b]\n+[i-capture-the-hostname:60703] [ 9] ./dgraph_unwgt(+0x278a)[0x55b68d46f78a]\n+[i-capture-the-hostname:60703] *** End of error message ***\n --------------------------------------------------------------------------\n Primary job terminated normally, but 1 process returned\n a non-zero exit code. Per user-direction, the job has been aborted.\n --------------------------------------------------------------------------\n --------------------------------------------------------------------------\n-mpiexec.openmpi noticed that process rank 0 with PID 0 on node profitbricks-build5-amd64 exited on signal 11 (Segmentation fault).\n+mpiexec.openmpi noticed that process rank 0 with PID 0 on node i-capture-the-hostname exited on signal 11 (Segmentation fault).\n --------------------------------------------------------------------------\n \n \n neighb_coll\n 2\n ./topo\n pass\n@@ -1000780,19 +1000782,19 @@\n pass\n \n \n setviewcur\n 2\n ./io\n fail\n- mca_fbtl_posix_preadv: error in (p)readv:Invalid argument\n+ Found 2 errors\n+mca_fbtl_posix_preadv: error in (p)readv:Invalid argument\n 0: buf[0] = -1\n mca_fbtl_posix_preadv: error in (p)readv:Invalid argument\n-1: buf[0] = -1\n- Found 2 errors\n+1: buf[0] = -1\n \n \n i_noncontig\n 2\n ./io\n pass\n \n@@ -1000997,34 +1000999,34 @@\n typecntsf\n 1\n ./f77/datatype\n fail\n Program received signal SIGSEGV: Segmentation fault - invalid memory reference.\n \n Backtrace for this error:\n-#0 0x7fb55383d89d in ???\n-#1 0x7fb55383cae3 in ???\n-#2 0x7fb5534778df in ???\n-#3 0x7fb55357c63e in ???\n-#4 0x7fb552cf0415 in ???\n-#5 0x7fb553aeb05e in ???\n-#6 0x7fb553b1cb26 in ???\n-#7 0x7fb553bed6d3 in ???\n-#8 0x562440b7f53b in explore_\n+#0 0x7f5f53d4089d in ???\n+#1 0x7f5f53d3fae3 in ???\n+#2 0x7f5f5397a8df in ???\n+#3 0x7f5f53a7ce90 in ???\n+#4 0x7f5f531f3415 in ???\n+#5 0x7f5f53fee05e in ???\n+#6 0x7f5f5401fb26 in ???\n+#7 0x7f5f540f06d3 in ???\n+#8 0x55610e8de53b in explore_\n \tat ../../../f77/datatype/typecntsf.f:52\n-#9 0x562440b7f27c in MAIN__\n+#9 0x55610e8de27c in MAIN__\n \tat ../../../f77/datatype/typecntsf.f:24\n-#10 0x562440b7f27c in main\n+#10 0x55610e8de27c in main\n \tat ../../../f77/datatype/typecntsf.f:33\n --------------------------------------------------------------------------\n Primary job terminated normally, but 1 process returned\n a non-zero exit code. Per user-direction, the job has been aborted.\n --------------------------------------------------------------------------\n --------------------------------------------------------------------------\n-mpiexec.openmpi noticed that process rank 0 with PID 0 on node profitbricks-build5-amd64 exited on signal 11 (Segmentation fault).\n+mpiexec.openmpi noticed that process rank 0 with PID 0 on node i-capture-the-hostname exited on signal 11 (Segmentation fault).\n --------------------------------------------------------------------------\n \n \n typem2f\n 1\n ./f77/datatype\n pass\n@@ -1001100,14 +1001102,15 @@\n s2(MPI_ERROR) != MPI_ERR_OTHER at T4 Mrecv().\n s1(MPI_ERROR) != MPI_ERR_DIMS at T5 Mprobe().\n s2(MPI_ERROR) != MPI_ERR_OTHER at T5 Imrecv().\n s1(MPI_ERROR) != MPI_ERR_DIMS at T6 Improbe().\n s2(MPI_ERROR) != MPI_ERR_OTHER at T6 Mrecv().\n s1(MPI_ERROR) != MPI_ERR_DIMS at T7 Improbe().\n s2(MPI_ERROR) != MPI_ERR_OTHER at T7 Imrecv().\n+ Found 24 errors\n s1(MPI_ERROR) != MPI_ERR_DIMS at T0 Mprobe().\n s2(MPI_ERROR) != MPI_ERR_OTHER at T0 Mrecv().\n s1(MPI_ERROR) != MPI_ERR_DIMS at T1 Mprobe().\n s2(MPI_ERROR) != MPI_ERR_OTHER at T1 Imrecv().\n s1(MPI_ERROR) != MPI_ERR_DIMS at T2 Improbe().\n s2(MPI_ERROR) != MPI_ERR_OTHER at T2 Mrecv().\n s1(MPI_ERROR) != MPI_ERR_DIMS at T3 Improbe().\n@@ -1001115,16 +1001118,15 @@\n s1(MPI_ERROR) != MPI_ERR_DIMS at T4 Mprobe().\n s2(MPI_ERROR) != MPI_ERR_OTHER at T4 Mrecv().\n s1(MPI_ERROR) != MPI_ERR_DIMS at T5 Mprobe().\n s2(MPI_ERROR) != MPI_ERR_OTHER at T5 Imrecv().\n s1(MPI_ERROR) != MPI_ERR_DIMS at T6 Improbe().\n s2(MPI_ERROR) != MPI_ERR_OTHER at T6 Mrecv().\n s1(MPI_ERROR) != MPI_ERR_DIMS at T7 Improbe().\n- s2(MPI_ERROR) != MPI_ERR_OTHER at T7 Imrecv().\n- Found 24 errors\n+ s2(MPI_ERROR) != MPI_ERR_OTHER at T7 Imrecv().\n \n \n infotestf\n 1\n ./f77/info\n pass\n \n@@ -1001149,99 +1001151,100 @@\n At least one pair of MPI processes are unable to reach each other for\n MPI communications. This means that no Open MPI device has indicated\n that it can be used to communicate between these processes. This is\n an error; Open MPI requires that all MPI processes be able to reach\n each other. This error can sometimes be the result of forgetting to\n specify the "self" BTL.\n \n- Process 1 ([[55085,1],0]) is on host: profitbricks-build5-amd64\n- Process 2 ([[55085,2],0]) is on host: profitbricks-build5-amd64\n+ Process 1 ([[39012,1],0]) is on host: i-capture-the-hostname\n+ Process 2 ([[39012,2],0]) is on host: i-capture-the-hostname\n BTLs attempted: self\n \n Your MPI job is now going to abort; sorry.\n --------------------------------------------------------------------------\n-[profitbricks-build5-amd64:05143] [[55085,1],0] ORTE_ERROR_LOG: Unreachable in file dpm/dpm.c at line 433\n-[profitbricks-build5-amd64:05162] [[55085,2],0] ORTE_ERROR_LOG: Unreachable in file dpm/dpm.c at line 433\n-[profitbricks-build5-amd64:05163] [[55085,2],1] ORTE_ERROR_LOG: Unreachable in file dpm/dpm.c at line 433\n+[i-capture-the-hostname:05460] [[39012,1],0] ORTE_ERROR_LOG: Unreachable in file dpm/dpm.c at line 433\n+[i-capture-the-hostname:05501] [[39012,2],0] ORTE_ERROR_LOG: Unreachable in file dpm/dpm.c at line 433\n+[i-capture-the-hostname:05503] [[39012,2],1] ORTE_ERROR_LOG: Unreachable in file dpm/dpm.c at line 433\n --------------------------------------------------------------------------\n It looks like MPI_INIT failed for some reason; your parallel process is\n likely to abort. There are many reasons that a parallel process can\n fail during MPI_INIT; some of which are due to configuration or environment\n problems. This failure appears to be an internal failure; here's some\n additional information (which may only be relevant to an Open MPI\n developer):\n \n ompi_dpm_dyn_init() failed\n --> Returned "Unreachable" (-12) instead of "Success" (0)\n --------------------------------------------------------------------------\n-[profitbricks-build5-amd64:05143] *** An error occurred in MPI_Comm_spawn\n-[profitbricks-build5-amd64:05143] *** reported by process [3610050561,0]\n-[profitbricks-build5-amd64:05143] *** on communicator MPI_COMM_WORLD\n-[profitbricks-build5-amd64:05143] *** MPI_ERR_INTERN: internal error\n-[profitbricks-build5-amd64:05143] *** MPI_ERRORS_ARE_FATAL (processes in this communicator will now abort,\n-[profitbricks-build5-amd64:05143] *** and potentially your MPI job)\n-[profitbricks-build5-amd64:05162] *** An error occurred in MPI_Init\n-[profitbricks-build5-amd64:05162] *** reported by process [3610050562,0]\n-[profitbricks-build5-amd64:05162] *** on a NULL communicator\n-[profitbricks-build5-amd64:05162] *** Unknown error\n-[profitbricks-build5-amd64:05162] *** MPI_ERRORS_ARE_FATAL (processes in this communicator will now abort,\n-[profitbricks-build5-amd64:05162] *** and potentially your MPI job)\n-[profitbricks-build5-amd64:05119] 2 more processes have sent help message help-mca-bml-r2.txt / unreachable proc\n-[profitbricks-build5-amd64:05119] Set MCA parameter "orte_base_help_aggregate" to 0 to see all help / error messages\n-[profitbricks-build5-amd64:05119] 1 more process has sent help message help-mpi-runtime.txt / mpi_init:startup:internal-failure\n-[profitbricks-build5-amd64:05119] 1 more process has sent help message help-mpi-errors.txt / mpi_errors_are_fatal unknown handle\n+[i-capture-the-hostname:05460] *** An error occurred in MPI_Comm_spawn\n+[i-capture-the-hostname:05460] *** reported by process [2556690433,0]\n+[i-capture-the-hostname:05460] *** on communicator MPI_COMM_WORLD\n+[i-capture-the-hostname:05460] *** MPI_ERR_INTERN: internal error\n+[i-capture-the-hostname:05460] *** MPI_ERRORS_ARE_FATAL (processes in this communicator will now abort,\n+[i-capture-the-hostname:05460] *** and potentially your MPI job)\n+[i-capture-the-hostname:05501] *** An error occurred in MPI_Init\n+[i-capture-the-hostname:05501] *** reported by process [2556690434,0]\n+[i-capture-the-hostname:05501] *** on a NULL communicator\n+[i-capture-the-hostname:05501] *** Unknown error\n+[i-capture-the-hostname:05501] *** MPI_ERRORS_ARE_FATAL (processes in this communicator will now abort,\n+[i-capture-the-hostname:05501] *** and potentially your MPI job)\n+[i-capture-the-hostname:05429] 2 more processes have sent help message help-mca-bml-r2.txt / unreachable proc\n+[i-capture-the-hostname:05429] Set MCA parameter "orte_base_help_aggregate" to 0 to see all help / error messages\n+[i-capture-the-hostname:05429] 1 more process has sent help message help-mpi-runtime.txt / mpi_init:startup:internal-failure\n+[i-capture-the-hostname:05429] 1 more process has sent help message help-mpi-errors.txt / mpi_errors_are_fatal unknown handle\n \n \n spawnargvf\n 1\n ./f77/spawn\n fail\n --------------------------------------------------------------------------\n At least one pair of MPI processes are unable to reach each other for\n MPI communications. This means that no Open MPI device has indicated\n that it can be used to communicate between these processes. This is\n an error; Open MPI requires that all MPI processes be able to reach\n each other. This error can sometimes be the result of forgetting to\n specify the "self" BTL.\n \n- Process 1 ([[53436,2],1]) is on host: profitbricks-build5-amd64\n- Process 2 ([[53436,1],0]) is on host: profitbricks-build5-amd64\n- BTLs attempted: vader self\n+ Process 1 ([[39149,1],0]) is on host: i-capture-the-hostname\n+ Process 2 ([[39149,2],0]) is on host: i-capture-the-hostname\n+ BTLs attempted: self\n \n Your MPI job is now going to abort; sorry.\n --------------------------------------------------------------------------\n+[i-capture-the-hostname:05592] [[39149,1],0] ORTE_ERROR_LOG: Unreachable in file dpm/dpm.c at line 433\n+[i-capture-the-hostname:05592] *** An error occurred in MPI_Comm_spawn\n+[i-capture-the-hostname:05592] *** reported by process [2565668865,0]\n+[i-capture-the-hostname:05592] *** on communicator MPI_COMM_WORLD\n+[i-capture-the-hostname:05592] *** MPI_ERR_INTERN: internal error\n+[i-capture-the-hostname:05592] *** MPI_ERRORS_ARE_FATAL (processes in this communicator will now abort,\n+[i-capture-the-hostname:05592] *** and potentially your MPI job)\n+[i-capture-the-hostname:05635] [[39149,2],1] ORTE_ERROR_LOG: Unreachable in file dpm/dpm.c at line 433\n+[i-capture-the-hostname:05632] [[39149,2],0] ORTE_ERROR_LOG: Unreachable in file dpm/dpm.c at line 433\n --------------------------------------------------------------------------\n It looks like MPI_INIT failed for some reason; your parallel process is\n likely to abort. There are many reasons that a parallel process can\n fail during MPI_INIT; some of which are due to configuration or environment\n problems. This failure appears to be an internal failure; here's some\n additional information (which may only be relevant to an Open MPI\n developer):\n \n ompi_dpm_dyn_init() failed\n --> Returned "Unreachable" (-12) instead of "Success" (0)\n --------------------------------------------------------------------------\n-[profitbricks-build5-amd64:05418] [[53436,2],1] ORTE_ERROR_LOG: Unreachable in file dpm/dpm.c at line 433\n-[profitbricks-build5-amd64:05418] *** An error occurred in MPI_Init\n-[profitbricks-build5-amd64:05418] *** reported by process [3501981698,1]\n-[profitbricks-build5-amd64:05418] *** on a NULL communicator\n-[profitbricks-build5-amd64:05418] *** Unknown error\n-[profitbricks-build5-amd64:05418] *** MPI_ERRORS_ARE_FATAL (processes in this communicator will now abort,\n-[profitbricks-build5-amd64:05418] *** and potentially your MPI job)\n-[profitbricks-build5-amd64:05258] *** An error occurred in MPI_Comm_spawn\n-[profitbricks-build5-amd64:05258] *** reported by process [3501981697,0]\n-[profitbricks-build5-amd64:05258] *** on communicator MPI_COMM_WORLD\n-[profitbricks-build5-amd64:05258] *** MPI_ERR_INTERN: internal error\n-[profitbricks-build5-amd64:05258] *** MPI_ERRORS_ARE_FATAL (processes in this communicator will now abort,\n-[profitbricks-build5-amd64:05258] *** and potentially your MPI job)\n-[profitbricks-build5-amd64:05258] [[53436,1],0] ORTE_ERROR_LOG: Unreachable in file dpm/dpm.c at line 433\n-[profitbricks-build5-amd64:05414] [[53436,2],0] ORTE_ERROR_LOG: Unreachable in file dpm/dpm.c at line 433\n-[profitbricks-build5-amd64:05230] 2 more processes have sent help message help-mca-bml-r2.txt / unreachable proc\n-[profitbricks-build5-amd64:05230] Set MCA parameter "orte_base_help_aggregate" to 0 to see all help / error messages\n-[profitbricks-build5-amd64:05230] 1 more process has sent help message help-mpi-runtime.txt / mpi_init:startup:internal-failure\n+[i-capture-the-hostname:05632] *** An error occurred in MPI_Init\n+[i-capture-the-hostname:05632] *** reported by process [2565668866,0]\n+[i-capture-the-hostname:05632] *** on a NULL communicator\n+[i-capture-the-hostname:05632] *** Unknown error\n+[i-capture-the-hostname:05632] *** MPI_ERRORS_ARE_FATAL (processes in this communicator will now abort,\n+[i-capture-the-hostname:05632] *** and potentially your MPI job)\n+[i-capture-the-hostname:05564] 2 more processes have sent help message help-mca-bml-r2.txt / unreachable proc\n+[i-capture-the-hostname:05564] Set MCA parameter "orte_base_help_aggregate" to 0 to see all help / error messages\n+[i-capture-the-hostname:05564] 1 more process has sent help message help-mpi-runtime.txt / mpi_init:startup:internal-failure\n+[i-capture-the-hostname:05564] 1 more process has sent help message help-mpi-errors.txt / mpi_errors_are_fatal unknown handle\n \n \n connaccf\n 2\n ./f77/spawn\n pass\n \n@@ -1001254,50 +1001257,50 @@\n At least one pair of MPI processes are unable to reach each other for\n MPI communications. This means that no Open MPI device has indicated\n that it can be used to communicate between these processes. This is\n an error; Open MPI requires that all MPI processes be able to reach\n each other. This error can sometimes be the result of forgetting to\n specify the "self" BTL.\n \n- Process 1 ([[54051,2],0]) is on host: profitbricks-build5-amd64\n- Process 2 ([[54051,1],0]) is on host: profitbricks-build5-amd64\n+ Process 1 ([[39509,2],1]) is on host: i-capture-the-hostname\n+ Process 2 ([[39509,1],0]) is on host: i-capture-the-hostname\n BTLs attempted: vader self\n \n Your MPI job is now going to abort; sorry.\n --------------------------------------------------------------------------\n+[i-capture-the-hostname:05989] [[39509,2],1] ORTE_ERROR_LOG: Unreachable in file dpm/dpm.c at line 433\n+[i-capture-the-hostname:05924] [[39509,1],0] ORTE_ERROR_LOG: Unreachable in file dpm/dpm.c at line 433\n+[i-capture-the-hostname:05984] [[39509,2],0] ORTE_ERROR_LOG: Unreachable in file dpm/dpm.c at line 433\n --------------------------------------------------------------------------\n It looks like MPI_INIT failed for some reason; your parallel process is\n likely to abort. There are many reasons that a parallel process can\n fail during MPI_INIT; some of which are due to configuration or environment\n problems. This failure appears to be an internal failure; here's some\n additional information (which may only be relevant to an Open MPI\n developer):\n \n ompi_dpm_dyn_init() failed\n --> Returned "Unreachable" (-12) instead of "Success" (0)\n --------------------------------------------------------------------------\n-[profitbricks-build5-amd64:06382] [[54051,2],0] ORTE_ERROR_LOG: Unreachable in file dpm/dpm.c at line 433\n-[profitbricks-build5-amd64:06386] [[54051,2],1] ORTE_ERROR_LOG: Unreachable in file dpm/dpm.c at line 433\n-[profitbricks-build5-amd64:06382] *** An error occurred in MPI_Init\n-[profitbricks-build5-amd64:06382] *** reported by process [3542286338,0]\n-[profitbricks-build5-amd64:06382] *** on a NULL communicator\n-[profitbricks-build5-amd64:06382] *** Unknown error\n-[profitbricks-build5-amd64:06382] *** MPI_ERRORS_ARE_FATAL (processes in this communicator will now abort,\n-[profitbricks-build5-amd64:06382] *** and potentially your MPI job)\n-[profitbricks-build5-amd64:06192] [[54051,1],0] ORTE_ERROR_LOG: Unreachable in file dpm/dpm.c at line 433\n-[profitbricks-build5-amd64:06192] *** An error occurred in MPI_Comm_spawn_multiple\n-[profitbricks-build5-amd64:06192] *** reported by process [3542286337,0]\n-[profitbricks-build5-amd64:06192] *** on communicator MPI_COMM_WORLD\n-[profitbricks-build5-amd64:06192] *** MPI_ERR_INTERN: internal error\n-[profitbricks-build5-amd64:06192] *** MPI_ERRORS_ARE_FATAL (processes in this communicator will now abort,\n-[profitbricks-build5-amd64:06192] *** and potentially your MPI job)\n-[profitbricks-build5-amd64:06129] 2 more processes have sent help message help-mca-bml-r2.txt / unreachable proc\n-[profitbricks-build5-amd64:06129] Set MCA parameter "orte_base_help_aggregate" to 0 to see all help / error messages\n-[profitbricks-build5-amd64:06129] 1 more process has sent help message help-mpi-runtime.txt / mpi_init:startup:internal-failure\n-[profitbricks-build5-amd64:06129] 1 more process has sent help message help-mpi-errors.txt / mpi_errors_are_fatal unknown handle\n+[i-capture-the-hostname:05924] *** An error occurred in MPI_Comm_spawn_multiple\n+[i-capture-the-hostname:05924] *** reported by process [2589261825,0]\n+[i-capture-the-hostname:05924] *** on communicator MPI_COMM_WORLD\n+[i-capture-the-hostname:05924] *** MPI_ERR_INTERN: internal error\n+[i-capture-the-hostname:05924] *** MPI_ERRORS_ARE_FATAL (processes in this communicator will now abort,\n+[i-capture-the-hostname:05924] *** and potentially your MPI job)\n+[i-capture-the-hostname:05984] *** An error occurred in MPI_Init\n+[i-capture-the-hostname:05984] *** reported by process [2589261826,0]\n+[i-capture-the-hostname:05984] *** on a NULL communicator\n+[i-capture-the-hostname:05984] *** Unknown error\n+[i-capture-the-hostname:05984] *** MPI_ERRORS_ARE_FATAL (processes in this communicator will now abort,\n+[i-capture-the-hostname:05984] *** and potentially your MPI job)\n+[i-capture-the-hostname:05892] 2 more processes have sent help message help-mca-bml-r2.txt / unreachable proc\n+[i-capture-the-hostname:05892] Set MCA parameter "orte_base_help_aggregate" to 0 to see all help / error messages\n+[i-capture-the-hostname:05892] 1 more process has sent help message help-mpi-runtime.txt / mpi_init:startup:internal-failure\n+[i-capture-the-hostname:05892] 1 more process has sent help message help-mpi-errors.txt / mpi_errors_are_fatal unknown handle\n [warn] Epoll MOD(1) on fd 29 failed. Old events were 6; read change was 0 (none); write change was 2 (del); close change was 0 (none): Bad file descriptor\n \n \n spawnmult2f\n 2\n ./f77/spawn\n fail\n@@ -1001305,64 +1001308,73 @@\n At least one pair of MPI processes are unable to reach each other for\n MPI communications. This means that no Open MPI device has indicated\n that it can be used to communicate between these processes. This is\n an error; Open MPI requires that all MPI processes be able to reach\n each other. This error can sometimes be the result of forgetting to\n specify the "self" BTL.\n \n- Process 1 ([[56658,1],1]) is on host: profitbricks-build5-amd64\n- Process 2 ([[56658,2],0]) is on host: profitbricks-build5-amd64\n+ Process 1 ([[39594,1],1]) is on host: i-capture-the-hostname\n+ Process 2 ([[39594,2],0]) is on host: i-capture-the-hostname\n BTLs attempted: vader self\n \n Your MPI job is now going to abort; sorry.\n --------------------------------------------------------------------------\n-[profitbricks-build5-amd64:06724] [[56658,2],1] ORTE_ERROR_LOG: Unreachable in file dpm/dpm.c at line 433\n-[profitbricks-build5-amd64:06720] [[56658,2],0] ORTE_ERROR_LOG: Unreachable in file dpm/dpm.c at line 433\n --------------------------------------------------------------------------\n It looks like MPI_INIT failed for some reason; your parallel process is\n likely to abort. There are many reasons that a parallel process can\n fail during MPI_INIT; some of which are due to configuration or environment\n problems. This failure appears to be an internal failure; here's some\n additional information (which may only be relevant to an Open MPI\n developer):\n \n ompi_dpm_dyn_init() failed\n --> Returned "Unreachable" (-12) instead of "Success" (0)\n --------------------------------------------------------------------------\n-[profitbricks-build5-amd64:06724] *** An error occurred in MPI_Init\n-[profitbricks-build5-amd64:06724] *** reported by process [3713138690,1]\n-[profitbricks-build5-amd64:06724] *** on a NULL communicator\n-[profitbricks-build5-amd64:06724] *** Unknown error\n-[profitbricks-build5-amd64:06724] *** MPI_ERRORS_ARE_FATAL (processes in this communicator will now abort,\n-[profitbricks-build5-amd64:06724] *** and potentially your MPI job)\n-[profitbricks-build5-amd64:06613] *** An error occurred in MPI_Comm_spawn_multiple\n-[profitbricks-build5-amd64:06613] *** reported by process [3713138689,0]\n-[profitbricks-build5-amd64:06613] *** on communicator MPI_COMM_WORLD\n-[profitbricks-build5-amd64:06613] *** MPI_ERR_INTERN: internal error\n-[profitbricks-build5-amd64:06613] *** MPI_ERRORS_ARE_FATAL (processes in this communicator will now abort,\n-[profitbricks-build5-amd64:06613] *** and potentially your MPI job)\n-[profitbricks-build5-amd64:06615] [[56658,1],1] ORTE_ERROR_LOG: Unreachable in file dpm/dpm.c at line 433\n-[profitbricks-build5-amd64:06613] [[56658,1],0] ORTE_ERROR_LOG: Unreachable in file dpm/dpm.c at line 433\n-[profitbricks-build5-amd64:06528] 3 more processes have sent help message help-mca-bml-r2.txt / unreachable proc\n-[profitbricks-build5-amd64:06528] Set MCA parameter "orte_base_help_aggregate" to 0 to see all help / error messages\n-[profitbricks-build5-amd64:06528] 1 more process has sent help message help-mpi-runtime.txt / mpi_init:startup:internal-failure\n-[profitbricks-build5-amd64:06528] 1 more process has sent help message help-mpi-errors.txt / mpi_errors_are_fatal unknown handle\n-[profitbricks-build5-amd64:06528] 1 more process has sent help message help-mpi-errors.txt / mpi_errors_are_fatal\n+[i-capture-the-hostname:06230] [[39594,1],1] ORTE_ERROR_LOG: Unreachable in file dpm/dpm.c at line 433\n+[i-capture-the-hostname:06352] [[39594,2],0] ORTE_ERROR_LOG: Unreachable in file dpm/dpm.c at line 433\n+[i-capture-the-hostname:06224] [[39594,1],0] ORTE_ERROR_LOG: Unreachable in file dpm/dpm.c at line 433\n+[i-capture-the-hostname:06230] *** An error occurred in MPI_Comm_spawn_multiple\n+[i-capture-the-hostname:06230] *** reported by process [2594832385,1]\n+[i-capture-the-hostname:06230] *** on communicator MPI_COMM_WORLD\n+[i-capture-the-hostname:06230] *** MPI_ERR_INTERN: internal error\n+[i-capture-the-hostname:06230] *** MPI_ERRORS_ARE_FATAL (processes in this communicator will now abort,\n+[i-capture-the-hostname:06230] *** and potentially your MPI job)\n+[i-capture-the-hostname:06358] [[39594,2],1] ORTE_ERROR_LOG: Unreachable in file dpm/dpm.c at line 433\n+[i-capture-the-hostname:06352] *** An error occurred in MPI_Init\n+[i-capture-the-hostname:06352] *** reported by process [2594832386,0]\n+[i-capture-the-hostname:06352] *** on a NULL communicator\n+[i-capture-the-hostname:06352] *** Unknown error\n+[i-capture-the-hostname:06352] *** MPI_ERRORS_ARE_FATAL (processes in this communicator will now abort,\n+[i-capture-the-hostname:06352] *** and potentially your MPI job)\n+[i-capture-the-hostname:06139] 3 more processes have sent help message help-mca-bml-r2.txt / unreachable proc\n+[i-capture-the-hostname:06139] Set MCA parameter "orte_base_help_aggregate" to 0 to see all help / error messages\n+[i-capture-the-hostname:06139] 1 more process has sent help message help-mpi-runtime.txt / mpi_init:startup:internal-failure\n+[i-capture-the-hostname:06139] 1 more process has sent help message help-mpi-errors.txt / mpi_errors_are_fatal\n+[i-capture-the-hostname:06139] 1 more process has sent help message help-mpi-errors.txt / mpi_errors_are_fatal unknown handle\n \n \n iwriteatf\n 2\n ./f77/io\n pass\n \n \n iwritef\n 2\n ./f77/io\n- pass\n+ fail\n+ --------------------------------------------------------------------------\n+The user-provided time limit for job execution has been reached:\n+\n+ Timeout: 180 seconds\n+\n+The job will now be aborted. Please check your code and/or\n+adjust/remove the job execution time limit (as specified by --timeout\n+command line option or MPIEXEC_TIMEOUT environment variable).\n+--------------------------------------------------------------------------\n \n \n iwriteshf\n 2\n ./f77/io\n fail\n --------------------------------------------------------------------------\n@@ -1001411,27 +1001423,54 @@\n ./f77/io\n pass\n \n \n writeatallbef\n 2\n ./f77/io\n- pass\n+ fail\n+ --------------------------------------------------------------------------\n+The user-provided time limit for job execution has been reached:\n+\n+ Timeout: 180 seconds\n+\n+The job will now be aborted. Please check your code and/or\n+adjust/remove the job execution time limit (as specified by --timeout\n+command line option or MPIEXEC_TIMEOUT environment variable).\n+--------------------------------------------------------------------------\n \n \n writeallbef\n 2\n ./f77/io\n- pass\n+ fail\n+ --------------------------------------------------------------------------\n+The user-provided time limit for job execution has been reached:\n+\n+ Timeout: 180 seconds\n+\n+The job will now be aborted. Please check your code and/or\n+adjust/remove the job execution time limit (as specified by --timeout\n+command line option or MPIEXEC_TIMEOUT environment variable).\n+--------------------------------------------------------------------------\n \n \n writeordbef\n 2\n ./f77/io\n- pass\n+ fail\n+ --------------------------------------------------------------------------\n+The user-provided time limit for job execution has been reached:\n+\n+ Timeout: 180 seconds\n+\n+The job will now be aborted. Please check your code and/or\n+adjust/remove the job execution time limit (as specified by --timeout\n+command line option or MPIEXEC_TIMEOUT environment variable).\n+--------------------------------------------------------------------------\n \n \n fileerrf\n 1\n ./f77/io\n pass\n \n@@ -1001447,37 +1001486,46 @@\n ./f77/io\n pass\n \n \n atomicityf\n 2\n ./f77/io\n- pass\n+ fail\n+ --------------------------------------------------------------------------\n+MPI_ABORT was invoked on rank 1 in communicator MPI_COMM_WORLD\n+with errorcode 1.\n+\n+NOTE: invoking MPI_ABORT causes Open MPI to kill all MPI processes.\n+You may or may not see output from other processes, depending on\n+exactly when Open MPI kills them.\n+--------------------------------------------------------------------------\n+ (noncontig)Process 1 : readbuf( 2277 ) is 0 , should be 10\n \n \n miscfilef\n 2\n ./f77/io\n fail\n File pointer position = 275 , should be 20\n- File pointer position in bytes = 2680 , should be 1000\n File pointer position = 275 , should be 20\n File pointer position in bytes = 2680 , should be 1000\n+ File pointer position in bytes = 2680 , should be 1000\n Found 4 errors\n \n \n setviewcurf\n 2\n ./f77/io\n fail\n 0 : buf(1) = -1\n- Found 2 errors\n 1 : buf(1) = -1\n mca_fbtl_posix_preadv: error in (p)readv:Invalid argument\n-mca_fbtl_posix_preadv: error in (p)readv:Invalid argument\n+mca_fbtl_posix_preadv: error in (p)readv:Invalid argument\n+ Found 2 errors\n \n \n c2f2ciof\n 1\n ./f77/io\n pass\n \n@@ -1001610,89 +1001658,89 @@\n pass\n \n \n dgraph_wgtf\n 2\n ./f77/topo\n fail\n- Error: Cannot partition 16 elements in 17 parts\n+ Error: Cannot partition 15 elements in 2 parts\n \n Program received signal SIGSEGV: Segmentation fault - invalid memory reference.\n \n Backtrace for this error:\n-#0 0x7f40c401d89d in ???\n-#1 0x7f40c401cae3 in ???\n-#2 0x7f40c3c578df in ???\n-#3 0x7f40c0895e8b in ???\n-#4 0x7f40c0896422 in ???\n-#5 0x7f40c08967c1 in ???\n-#6 0x7f40c089328c in ???\n-#7 0x7f40c08899e8 in ???\n-#8 0x7f40c42d93cc in ???\n-#9 0x7f40c43c4bca in ???\n-#10 0x55ac5ff3aa1b in dgraph_unwgt\n+#0 0x7f30d2fd089d in ???\n+#1 0x7f30d2fcfae3 in ???\n+#2 0x7f30d2c0a8df in ???\n+#3 0x7f30d0148e8b in ???\n+#4 0x7f30d0149422 in ???\n+#5 0x7f30d01497c1 in ???\n+#6 0x7f30d014628c in ???\n+#7 0x7f30d013c9e8 in ???\n+#8 0x7f30d328c3cc in ???\n+#9 0x7f30d3377bca in ???\n+#10 0x561e752aea1b in dgraph_unwgt\n \tat ../../../f77/topo/dgraph_wgtf.f:157\n-#11 0x55ac5ff3a1fe in main\n+#11 0x561e752ae1fe in main\n \tat ../../../f77/topo/dgraph_wgtf.f:201\n --------------------------------------------------------------------------\n Primary job terminated normally, but 1 process returned\n a non-zero exit code. Per user-direction, the job has been aborted.\n --------------------------------------------------------------------------\n --------------------------------------------------------------------------\n-mpiexec.openmpi noticed that process rank 0 with PID 0 on node profitbricks-build5-amd64 exited on signal 11 (Segmentation fault).\n+mpiexec.openmpi noticed that process rank 0 with PID 0 on node i-capture-the-hostname exited on signal 11 (Segmentation fault).\n --------------------------------------------------------------------------\n \n \n dgraph_unwgtf\n 2\n ./f77/topo\n fail\n- Error: Cannot partition 16 elements in 17 parts\n+ Error: Cannot partition 15 elements in 2 parts\n \n Program received signal SIGSEGV: Segmentation fault - invalid memory reference.\n \n Backtrace for this error:\n-#0 0x7ffabbf0589d in ???\n-#1 0x7ffabbf04ae3 in ???\n-#2 0x7ffabbb3f8df in ???\n-#3 0x7ffab877de8b in ???\n-#4 0x7ffab877e422 in ???\n-#5 0x7ffab877e7c1 in ???\n-#6 0x7ffab877b28c in ???\n-#7 0x7ffab87719e8 in ???\n-#8 0x7ffabc1c13cc in ???\n-#9 0x7ffabc2acbca in ???\n-#10 0x55f07e31590d in dgraph_unwgt\n+#0 0x7fd401c2889d in ???\n+#1 0x7fd401c27ae3 in ???\n+#2 0x7fd4018628df in ???\n+#3 0x7fd3fa498e8b in ???\n+#4 0x7fd3fa499422 in ???\n+#5 0x7fd3fa4997c1 in ???\n+#6 0x7fd3fa49628c in ???\n+#7 0x7fd3fa48c9e8 in ???\n+#8 0x7fd401ee43cc in ???\n+#9 0x7fd401fcfbca in ???\n+#10 0x55bebef6a90d in dgraph_unwgt\n \tat ../../../f77/topo/dgraph_unwgtf.f:141\n-#11 0x55f07e3151fe in main\n+#11 0x55bebef6a1fe in main\n \tat ../../../f77/topo/dgraph_unwgtf.f:216\n --------------------------------------------------------------------------\n Primary job terminated normally, but 1 process returned\n a non-zero exit code. Per user-direction, the job has been aborted.\n --------------------------------------------------------------------------\n --------------------------------------------------------------------------\n-mpiexec.openmpi noticed that process rank 0 with PID 0 on node profitbricks-build5-amd64 exited on signal 11 (Segmentation fault).\n+mpiexec.openmpi noticed that process rank 0 with PID 0 on node i-capture-the-hostname exited on signal 11 (Segmentation fault).\n --------------------------------------------------------------------------\n \n \n profile1f\n 2\n ./f77/profile\n pass\n \n \n attrtx\n \n ./cxx/attr\n fail\n- Failed to build attrtx; make[2]: Entering directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/attr'\n+ Failed to build attrtx; make[2]: Entering directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/attr'\n (cd ../../cxx/util && make mtest.o)\n-make[3]: Entering directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/util'\n+make[3]: Entering directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/util'\n depbase=`echo mtest.o | sed 's|[^/]*$|.deps/&|;s|\\.o$||'`;\\\n-/usr/bin/mpicxx.openmpi -DHAVE_CONFIG_H -I. -I../../../cxx/util -I../../include -I../../include -I../../../include -Wdate-time -D_FORTIFY_SOURCE=2 -g -O2 -ffile-prefix-map=/build/1st/mpi-testsuite-3.2+dfsg=. -fstack-protector-strong -Wformat -Werror=format-security -MT mtest.o -MD -MP -MF $depbase.Tpo -c -o mtest.o ../../../cxx/util/mtest.cxx &&\\\n+/usr/bin/mpicxx.openmpi -DHAVE_CONFIG_H -I. -I../../../cxx/util -I../../include -I../../include -I../../../include -Wdate-time -D_FORTIFY_SOURCE=2 -g -O2 -ffile-prefix-map=/build/mpi-testsuite-3.2+dfsg/2nd=. -fstack-protector-strong -Wformat -Werror=format-security -MT mtest.o -MD -MP -MF $depbase.Tpo -c -o mtest.o ../../../cxx/util/mtest.cxx &&\\\n mv -f $depbase.Tpo $depbase.Po\n ../../../cxx/util/mtest.cxx: In function 'void* MTestTypeContigInitRecv(MTestDatatype*)':\n ../../../cxx/util/mtest.cxx:189:42: warning: 'int MPI_Type_extent(MPI_Datatype, MPI_Aint*)' is deprecated: MPI_Type_extent is superseded by MPI_Type_get_extent in MPI-2.0 [-Wdeprecated-declarations]\n MPI_Type_extent( mtype->datatype, &size );\n ^\n In file included from ../../../cxx/util/mtest.cxx:7:\n /usr/lib/x86_64-linux-gnu/openmpi/include/mpi.h:1758:20: note: declared here\n@@ -1001724,28 +1001772,28 @@\n ../../../cxx/util/mtest.cxx:848:50: error: 'NULL_COPY_FN' is not a member of 'MPI::Win'\n mem_keyval = MPI::Win::Create_keyval( MPI::Win::NULL_COPY_FN,\n ^~~~~~~~~~~~\n ../../../cxx/util/mtest.cxx:849:22: error: 'NULL_DELETE_FN' is not a member of 'MPI::Win'\n MPI::Win::NULL_DELETE_FN, 0 );\n ^~~~~~~~~~~~~~\n make[3]: *** [Makefile:416: mtest.o] Error 1\n-make[3]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/util'\n+make[3]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/util'\n make[2]: *** [Makefile:695: ../../cxx/util/mtest.o] Error 2\n-make[2]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/attr'\n+make[2]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/attr'\n \n \n attricx\n \n ./cxx/attr\n fail\n- Failed to build attricx; make[2]: Entering directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/attr'\n+ Failed to build attricx; make[2]: Entering directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/attr'\n (cd ../../cxx/util && make mtest.o)\n-make[3]: Entering directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/util'\n+make[3]: Entering directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/util'\n depbase=`echo mtest.o | sed 's|[^/]*$|.deps/&|;s|\\.o$||'`;\\\n-/usr/bin/mpicxx.openmpi -DHAVE_CONFIG_H -I. -I../../../cxx/util -I../../include -I../../include -I../../../include -Wdate-time -D_FORTIFY_SOURCE=2 -g -O2 -ffile-prefix-map=/build/1st/mpi-testsuite-3.2+dfsg=. -fstack-protector-strong -Wformat -Werror=format-security -MT mtest.o -MD -MP -MF $depbase.Tpo -c -o mtest.o ../../../cxx/util/mtest.cxx &&\\\n+/usr/bin/mpicxx.openmpi -DHAVE_CONFIG_H -I. -I../../../cxx/util -I../../include -I../../include -I../../../include -Wdate-time -D_FORTIFY_SOURCE=2 -g -O2 -ffile-prefix-map=/build/mpi-testsuite-3.2+dfsg/2nd=. -fstack-protector-strong -Wformat -Werror=format-security -MT mtest.o -MD -MP -MF $depbase.Tpo -c -o mtest.o ../../../cxx/util/mtest.cxx &&\\\n mv -f $depbase.Tpo $depbase.Po\n ../../../cxx/util/mtest.cxx: In function 'void* MTestTypeContigInitRecv(MTestDatatype*)':\n ../../../cxx/util/mtest.cxx:189:42: warning: 'int MPI_Type_extent(MPI_Datatype, MPI_Aint*)' is deprecated: MPI_Type_extent is superseded by MPI_Type_get_extent in MPI-2.0 [-Wdeprecated-declarations]\n MPI_Type_extent( mtype->datatype, &size );\n ^\n In file included from ../../../cxx/util/mtest.cxx:7:\n /usr/lib/x86_64-linux-gnu/openmpi/include/mpi.h:1758:20: note: declared here\n@@ -1001777,28 +1001825,28 @@\n ../../../cxx/util/mtest.cxx:848:50: error: 'NULL_COPY_FN' is not a member of 'MPI::Win'\n mem_keyval = MPI::Win::Create_keyval( MPI::Win::NULL_COPY_FN,\n ^~~~~~~~~~~~\n ../../../cxx/util/mtest.cxx:849:22: error: 'NULL_DELETE_FN' is not a member of 'MPI::Win'\n MPI::Win::NULL_DELETE_FN, 0 );\n ^~~~~~~~~~~~~~\n make[3]: *** [Makefile:416: mtest.o] Error 1\n-make[3]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/util'\n+make[3]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/util'\n make[2]: *** [Makefile:695: ../../cxx/util/mtest.o] Error 2\n-make[2]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/attr'\n+make[2]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/attr'\n \n \n baseattrcommx\n \n ./cxx/attr\n fail\n- Failed to build baseattrcommx; make[2]: Entering directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/attr'\n+ Failed to build baseattrcommx; make[2]: Entering directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/attr'\n (cd ../../cxx/util && make mtest.o)\n-make[3]: Entering directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/util'\n+make[3]: Entering directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/util'\n depbase=`echo mtest.o | sed 's|[^/]*$|.deps/&|;s|\\.o$||'`;\\\n-/usr/bin/mpicxx.openmpi -DHAVE_CONFIG_H -I. -I../../../cxx/util -I../../include -I../../include -I../../../include -Wdate-time -D_FORTIFY_SOURCE=2 -g -O2 -ffile-prefix-map=/build/1st/mpi-testsuite-3.2+dfsg=. -fstack-protector-strong -Wformat -Werror=format-security -MT mtest.o -MD -MP -MF $depbase.Tpo -c -o mtest.o ../../../cxx/util/mtest.cxx &&\\\n+/usr/bin/mpicxx.openmpi -DHAVE_CONFIG_H -I. -I../../../cxx/util -I../../include -I../../include -I../../../include -Wdate-time -D_FORTIFY_SOURCE=2 -g -O2 -ffile-prefix-map=/build/mpi-testsuite-3.2+dfsg/2nd=. -fstack-protector-strong -Wformat -Werror=format-security -MT mtest.o -MD -MP -MF $depbase.Tpo -c -o mtest.o ../../../cxx/util/mtest.cxx &&\\\n mv -f $depbase.Tpo $depbase.Po\n ../../../cxx/util/mtest.cxx: In function 'void* MTestTypeContigInitRecv(MTestDatatype*)':\n ../../../cxx/util/mtest.cxx:189:42: warning: 'int MPI_Type_extent(MPI_Datatype, MPI_Aint*)' is deprecated: MPI_Type_extent is superseded by MPI_Type_get_extent in MPI-2.0 [-Wdeprecated-declarations]\n MPI_Type_extent( mtype->datatype, &size );\n ^\n In file included from ../../../cxx/util/mtest.cxx:7:\n /usr/lib/x86_64-linux-gnu/openmpi/include/mpi.h:1758:20: note: declared here\n@@ -1001830,28 +1001878,28 @@\n ../../../cxx/util/mtest.cxx:848:50: error: 'NULL_COPY_FN' is not a member of 'MPI::Win'\n mem_keyval = MPI::Win::Create_keyval( MPI::Win::NULL_COPY_FN,\n ^~~~~~~~~~~~\n ../../../cxx/util/mtest.cxx:849:22: error: 'NULL_DELETE_FN' is not a member of 'MPI::Win'\n MPI::Win::NULL_DELETE_FN, 0 );\n ^~~~~~~~~~~~~~\n make[3]: *** [Makefile:416: mtest.o] Error 1\n-make[3]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/util'\n+make[3]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/util'\n make[2]: *** [Makefile:695: ../../cxx/util/mtest.o] Error 2\n-make[2]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/attr'\n+make[2]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/attr'\n \n \n fkeyvalcommx\n \n ./cxx/attr\n fail\n- Failed to build fkeyvalcommx; make[2]: Entering directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/attr'\n+ Failed to build fkeyvalcommx; make[2]: Entering directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/attr'\n (cd ../../cxx/util && make mtest.o)\n-make[3]: Entering directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/util'\n+make[3]: Entering directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/util'\n depbase=`echo mtest.o | sed 's|[^/]*$|.deps/&|;s|\\.o$||'`;\\\n-/usr/bin/mpicxx.openmpi -DHAVE_CONFIG_H -I. -I../../../cxx/util -I../../include -I../../include -I../../../include -Wdate-time -D_FORTIFY_SOURCE=2 -g -O2 -ffile-prefix-map=/build/1st/mpi-testsuite-3.2+dfsg=. -fstack-protector-strong -Wformat -Werror=format-security -MT mtest.o -MD -MP -MF $depbase.Tpo -c -o mtest.o ../../../cxx/util/mtest.cxx &&\\\n+/usr/bin/mpicxx.openmpi -DHAVE_CONFIG_H -I. -I../../../cxx/util -I../../include -I../../include -I../../../include -Wdate-time -D_FORTIFY_SOURCE=2 -g -O2 -ffile-prefix-map=/build/mpi-testsuite-3.2+dfsg/2nd=. -fstack-protector-strong -Wformat -Werror=format-security -MT mtest.o -MD -MP -MF $depbase.Tpo -c -o mtest.o ../../../cxx/util/mtest.cxx &&\\\n mv -f $depbase.Tpo $depbase.Po\n ../../../cxx/util/mtest.cxx: In function 'void* MTestTypeContigInitRecv(MTestDatatype*)':\n ../../../cxx/util/mtest.cxx:189:42: warning: 'int MPI_Type_extent(MPI_Datatype, MPI_Aint*)' is deprecated: MPI_Type_extent is superseded by MPI_Type_get_extent in MPI-2.0 [-Wdeprecated-declarations]\n MPI_Type_extent( mtype->datatype, &size );\n ^\n In file included from ../../../cxx/util/mtest.cxx:7:\n /usr/lib/x86_64-linux-gnu/openmpi/include/mpi.h:1758:20: note: declared here\n@@ -1001883,26 +1001931,26 @@\n ../../../cxx/util/mtest.cxx:848:50: error: 'NULL_COPY_FN' is not a member of 'MPI::Win'\n mem_keyval = MPI::Win::Create_keyval( MPI::Win::NULL_COPY_FN,\n ^~~~~~~~~~~~\n ../../../cxx/util/mtest.cxx:849:22: error: 'NULL_DELETE_FN' is not a member of 'MPI::Win'\n MPI::Win::NULL_DELETE_FN, 0 );\n ^~~~~~~~~~~~~~\n make[3]: *** [Makefile:416: mtest.o] Error 1\n-make[3]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/util'\n+make[3]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/util'\n make[2]: *** [Makefile:695: ../../cxx/util/mtest.o] Error 2\n-make[2]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/attr'\n+make[2]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/attr'\n \n \n fkeyvaltypex\n \n ./cxx/attr\n fail\n- Failed to build fkeyvaltypex; make[2]: Entering directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/attr'\n+ Failed to build fkeyvaltypex; make[2]: Entering directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/attr'\n depbase=`echo fkeyvaltypex.o | sed 's|[^/]*$|.deps/&|;s|\\.o$||'`;\\\n-/usr/bin/mpicxx.openmpi -DHAVE_CONFIG_H -I. -I../../../cxx/attr -I../../include -I../../include -I../../../include -Wdate-time -D_FORTIFY_SOURCE=2 -g -O2 -ffile-prefix-map=/build/1st/mpi-testsuite-3.2+dfsg=. -fstack-protector-strong -Wformat -Werror=format-security -MT fkeyvaltypex.o -MD -MP -MF $depbase.Tpo -c -o fkeyvaltypex.o ../../../cxx/attr/fkeyvaltypex.cxx &&\\\n+/usr/bin/mpicxx.openmpi -DHAVE_CONFIG_H -I. -I../../../cxx/attr -I../../include -I../../include -I../../../include -Wdate-time -D_FORTIFY_SOURCE=2 -g -O2 -ffile-prefix-map=/build/mpi-testsuite-3.2+dfsg/2nd=. -fstack-protector-strong -Wformat -Werror=format-security -MT fkeyvaltypex.o -MD -MP -MF $depbase.Tpo -c -o fkeyvaltypex.o ../../../cxx/attr/fkeyvaltypex.cxx &&\\\n mv -f $depbase.Tpo $depbase.Po\n ../../../cxx/attr/fkeyvaltypex.cxx: In function 'int main(int, char**)':\n ../../../cxx/attr/fkeyvaltypex.cxx:67:71: error: no matching function for call to 'Create_keyval(int (&)(const MPI::Datatype&, int, void*, void*, void*, bool&), int (&)(MPI::Datatype&, int, void*, void*), void*)'\n keyval = MPI::Datatype::Create_keyval( copy_fn, delete_fn, (void *)0 );\n ^\n In file included from /usr/lib/x86_64-linux-gnu/openmpi/include/openmpi/ompi/mpi/cxx/mpicxx.h:269,\n from /usr/lib/x86_64-linux-gnu/openmpi/include/mpi.h:2684,\n@@ -1001947,26 +1001995,26 @@\n ../../../cxx/attr/fkeyvaltypex.cxx:79:18: error: 'NULL_COPY_FN' is not a member of 'MPI::Datatype'\n MPI::Datatype::NULL_COPY_FN, MPI::Datatype::NULL_DELETE_FN,\n ^~~~~~~~~~~~\n ../../../cxx/attr/fkeyvaltypex.cxx:79:47: error: 'NULL_DELETE_FN' is not a member of 'MPI::Datatype'\n MPI::Datatype::NULL_COPY_FN, MPI::Datatype::NULL_DELETE_FN,\n ^~~~~~~~~~~~~~\n make[2]: *** [Makefile:475: fkeyvaltypex.o] Error 1\n-make[2]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/attr'\n+make[2]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/attr'\n \n \n bsend1cxx\n \n ./cxx/pt2pt\n fail\n- Failed to build bsend1cxx; make[2]: Entering directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/pt2pt'\n+ Failed to build bsend1cxx; make[2]: Entering directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/pt2pt'\n (cd ../../cxx/util && make mtest.o)\n-make[3]: Entering directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/util'\n+make[3]: Entering directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/util'\n depbase=`echo mtest.o | sed 's|[^/]*$|.deps/&|;s|\\.o$||'`;\\\n-/usr/bin/mpicxx.openmpi -DHAVE_CONFIG_H -I. -I../../../cxx/util -I../../include -I../../include -I../../../include -Wdate-time -D_FORTIFY_SOURCE=2 -g -O2 -ffile-prefix-map=/build/1st/mpi-testsuite-3.2+dfsg=. -fstack-protector-strong -Wformat -Werror=format-security -MT mtest.o -MD -MP -MF $depbase.Tpo -c -o mtest.o ../../../cxx/util/mtest.cxx &&\\\n+/usr/bin/mpicxx.openmpi -DHAVE_CONFIG_H -I. -I../../../cxx/util -I../../include -I../../include -I../../../include -Wdate-time -D_FORTIFY_SOURCE=2 -g -O2 -ffile-prefix-map=/build/mpi-testsuite-3.2+dfsg/2nd=. -fstack-protector-strong -Wformat -Werror=format-security -MT mtest.o -MD -MP -MF $depbase.Tpo -c -o mtest.o ../../../cxx/util/mtest.cxx &&\\\n mv -f $depbase.Tpo $depbase.Po\n ../../../cxx/util/mtest.cxx: In function 'void* MTestTypeContigInitRecv(MTestDatatype*)':\n ../../../cxx/util/mtest.cxx:189:42: warning: 'int MPI_Type_extent(MPI_Datatype, MPI_Aint*)' is deprecated: MPI_Type_extent is superseded by MPI_Type_get_extent in MPI-2.0 [-Wdeprecated-declarations]\n MPI_Type_extent( mtype->datatype, &size );\n ^\n In file included from ../../../cxx/util/mtest.cxx:7:\n /usr/lib/x86_64-linux-gnu/openmpi/include/mpi.h:1758:20: note: declared here\n@@ -1001998,28 +1002046,28 @@\n ../../../cxx/util/mtest.cxx:848:50: error: 'NULL_COPY_FN' is not a member of 'MPI::Win'\n mem_keyval = MPI::Win::Create_keyval( MPI::Win::NULL_COPY_FN,\n ^~~~~~~~~~~~\n ../../../cxx/util/mtest.cxx:849:22: error: 'NULL_DELETE_FN' is not a member of 'MPI::Win'\n MPI::Win::NULL_DELETE_FN, 0 );\n ^~~~~~~~~~~~~~\n make[3]: *** [Makefile:416: mtest.o] Error 1\n-make[3]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/util'\n+make[3]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/util'\n make[2]: *** [Makefile:650: ../../cxx/util/mtest.o] Error 2\n-make[2]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/pt2pt'\n+make[2]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/pt2pt'\n \n \n sendrecvx\n \n ./cxx/pt2pt\n fail\n- Failed to build sendrecvx; make[2]: Entering directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/pt2pt'\n+ Failed to build sendrecvx; make[2]: Entering directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/pt2pt'\n (cd ../../cxx/util && make mtest.o)\n-make[3]: Entering directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/util'\n+make[3]: Entering directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/util'\n depbase=`echo mtest.o | sed 's|[^/]*$|.deps/&|;s|\\.o$||'`;\\\n-/usr/bin/mpicxx.openmpi -DHAVE_CONFIG_H -I. -I../../../cxx/util -I../../include -I../../include -I../../../include -Wdate-time -D_FORTIFY_SOURCE=2 -g -O2 -ffile-prefix-map=/build/1st/mpi-testsuite-3.2+dfsg=. -fstack-protector-strong -Wformat -Werror=format-security -MT mtest.o -MD -MP -MF $depbase.Tpo -c -o mtest.o ../../../cxx/util/mtest.cxx &&\\\n+/usr/bin/mpicxx.openmpi -DHAVE_CONFIG_H -I. -I../../../cxx/util -I../../include -I../../include -I../../../include -Wdate-time -D_FORTIFY_SOURCE=2 -g -O2 -ffile-prefix-map=/build/mpi-testsuite-3.2+dfsg/2nd=. -fstack-protector-strong -Wformat -Werror=format-security -MT mtest.o -MD -MP -MF $depbase.Tpo -c -o mtest.o ../../../cxx/util/mtest.cxx &&\\\n mv -f $depbase.Tpo $depbase.Po\n ../../../cxx/util/mtest.cxx: In function 'void* MTestTypeContigInitRecv(MTestDatatype*)':\n ../../../cxx/util/mtest.cxx:189:42: warning: 'int MPI_Type_extent(MPI_Datatype, MPI_Aint*)' is deprecated: MPI_Type_extent is superseded by MPI_Type_get_extent in MPI-2.0 [-Wdeprecated-declarations]\n MPI_Type_extent( mtype->datatype, &size );\n ^\n In file included from ../../../cxx/util/mtest.cxx:7:\n /usr/lib/x86_64-linux-gnu/openmpi/include/mpi.h:1758:20: note: declared here\n@@ -1002051,28 +1002099,28 @@\n ../../../cxx/util/mtest.cxx:848:50: error: 'NULL_COPY_FN' is not a member of 'MPI::Win'\n mem_keyval = MPI::Win::Create_keyval( MPI::Win::NULL_COPY_FN,\n ^~~~~~~~~~~~\n ../../../cxx/util/mtest.cxx:849:22: error: 'NULL_DELETE_FN' is not a member of 'MPI::Win'\n MPI::Win::NULL_DELETE_FN, 0 );\n ^~~~~~~~~~~~~~\n make[3]: *** [Makefile:416: mtest.o] Error 1\n-make[3]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/util'\n+make[3]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/util'\n make[2]: *** [Makefile:650: ../../cxx/util/mtest.o] Error 2\n-make[2]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/pt2pt'\n+make[2]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/pt2pt'\n \n \n commname2\n \n ./cxx/comm\n fail\n- Failed to build commname2; make[2]: Entering directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/comm'\n+ Failed to build commname2; make[2]: Entering directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/comm'\n (cd ../../cxx/util && make mtest.o)\n-make[3]: Entering directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/util'\n+make[3]: Entering directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/util'\n depbase=`echo mtest.o | sed 's|[^/]*$|.deps/&|;s|\\.o$||'`;\\\n-/usr/bin/mpicxx.openmpi -DHAVE_CONFIG_H -I. -I../../../cxx/util -I../../include -I../../include -I../../../include -Wdate-time -D_FORTIFY_SOURCE=2 -g -O2 -ffile-prefix-map=/build/1st/mpi-testsuite-3.2+dfsg=. -fstack-protector-strong -Wformat -Werror=format-security -MT mtest.o -MD -MP -MF $depbase.Tpo -c -o mtest.o ../../../cxx/util/mtest.cxx &&\\\n+/usr/bin/mpicxx.openmpi -DHAVE_CONFIG_H -I. -I../../../cxx/util -I../../include -I../../include -I../../../include -Wdate-time -D_FORTIFY_SOURCE=2 -g -O2 -ffile-prefix-map=/build/mpi-testsuite-3.2+dfsg/2nd=. -fstack-protector-strong -Wformat -Werror=format-security -MT mtest.o -MD -MP -MF $depbase.Tpo -c -o mtest.o ../../../cxx/util/mtest.cxx &&\\\n mv -f $depbase.Tpo $depbase.Po\n ../../../cxx/util/mtest.cxx: In function 'void* MTestTypeContigInitRecv(MTestDatatype*)':\n ../../../cxx/util/mtest.cxx:189:42: warning: 'int MPI_Type_extent(MPI_Datatype, MPI_Aint*)' is deprecated: MPI_Type_extent is superseded by MPI_Type_get_extent in MPI-2.0 [-Wdeprecated-declarations]\n MPI_Type_extent( mtype->datatype, &size );\n ^\n In file included from ../../../cxx/util/mtest.cxx:7:\n /usr/lib/x86_64-linux-gnu/openmpi/include/mpi.h:1758:20: note: declared here\n@@ -1002104,28 +1002152,28 @@\n ../../../cxx/util/mtest.cxx:848:50: error: 'NULL_COPY_FN' is not a member of 'MPI::Win'\n mem_keyval = MPI::Win::Create_keyval( MPI::Win::NULL_COPY_FN,\n ^~~~~~~~~~~~\n ../../../cxx/util/mtest.cxx:849:22: error: 'NULL_DELETE_FN' is not a member of 'MPI::Win'\n MPI::Win::NULL_DELETE_FN, 0 );\n ^~~~~~~~~~~~~~\n make[3]: *** [Makefile:416: mtest.o] Error 1\n-make[3]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/util'\n+make[3]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/util'\n make[2]: *** [Makefile:640: ../../cxx/util/mtest.o] Error 2\n-make[2]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/comm'\n+make[2]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/comm'\n \n \n arcomplex\n \n ./cxx/coll\n fail\n- Failed to build arcomplex; make[2]: Entering directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/coll'\n+ Failed to build arcomplex; make[2]: Entering directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/coll'\n (cd ../../cxx/util && make mtest.o)\n-make[3]: Entering directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/util'\n+make[3]: Entering directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/util'\n depbase=`echo mtest.o | sed 's|[^/]*$|.deps/&|;s|\\.o$||'`;\\\n-/usr/bin/mpicxx.openmpi -DHAVE_CONFIG_H -I. -I../../../cxx/util -I../../include -I../../include -I../../../include -Wdate-time -D_FORTIFY_SOURCE=2 -g -O2 -ffile-prefix-map=/build/1st/mpi-testsuite-3.2+dfsg=. -fstack-protector-strong -Wformat -Werror=format-security -MT mtest.o -MD -MP -MF $depbase.Tpo -c -o mtest.o ../../../cxx/util/mtest.cxx &&\\\n+/usr/bin/mpicxx.openmpi -DHAVE_CONFIG_H -I. -I../../../cxx/util -I../../include -I../../include -I../../../include -Wdate-time -D_FORTIFY_SOURCE=2 -g -O2 -ffile-prefix-map=/build/mpi-testsuite-3.2+dfsg/2nd=. -fstack-protector-strong -Wformat -Werror=format-security -MT mtest.o -MD -MP -MF $depbase.Tpo -c -o mtest.o ../../../cxx/util/mtest.cxx &&\\\n mv -f $depbase.Tpo $depbase.Po\n ../../../cxx/util/mtest.cxx: In function 'void* MTestTypeContigInitRecv(MTestDatatype*)':\n ../../../cxx/util/mtest.cxx:189:42: warning: 'int MPI_Type_extent(MPI_Datatype, MPI_Aint*)' is deprecated: MPI_Type_extent is superseded by MPI_Type_get_extent in MPI-2.0 [-Wdeprecated-declarations]\n MPI_Type_extent( mtype->datatype, &size );\n ^\n In file included from ../../../cxx/util/mtest.cxx:7:\n /usr/lib/x86_64-linux-gnu/openmpi/include/mpi.h:1758:20: note: declared here\n@@ -1002157,28 +1002205,28 @@\n ../../../cxx/util/mtest.cxx:848:50: error: 'NULL_COPY_FN' is not a member of 'MPI::Win'\n mem_keyval = MPI::Win::Create_keyval( MPI::Win::NULL_COPY_FN,\n ^~~~~~~~~~~~\n ../../../cxx/util/mtest.cxx:849:22: error: 'NULL_DELETE_FN' is not a member of 'MPI::Win'\n MPI::Win::NULL_DELETE_FN, 0 );\n ^~~~~~~~~~~~~~\n make[3]: *** [Makefile:416: mtest.o] Error 1\n-make[3]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/util'\n+make[3]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/util'\n make[2]: *** [Makefile:865: ../../cxx/util/mtest.o] Error 2\n-make[2]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/coll'\n+make[2]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/coll'\n \n \n uallredx\n \n ./cxx/coll\n fail\n- Failed to build uallredx; make[2]: Entering directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/coll'\n+ Failed to build uallredx; make[2]: Entering directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/coll'\n (cd ../../cxx/util && make mtest.o)\n-make[3]: Entering directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/util'\n+make[3]: Entering directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/util'\n depbase=`echo mtest.o | sed 's|[^/]*$|.deps/&|;s|\\.o$||'`;\\\n-/usr/bin/mpicxx.openmpi -DHAVE_CONFIG_H -I. -I../../../cxx/util -I../../include -I../../include -I../../../include -Wdate-time -D_FORTIFY_SOURCE=2 -g -O2 -ffile-prefix-map=/build/1st/mpi-testsuite-3.2+dfsg=. -fstack-protector-strong -Wformat -Werror=format-security -MT mtest.o -MD -MP -MF $depbase.Tpo -c -o mtest.o ../../../cxx/util/mtest.cxx &&\\\n+/usr/bin/mpicxx.openmpi -DHAVE_CONFIG_H -I. -I../../../cxx/util -I../../include -I../../include -I../../../include -Wdate-time -D_FORTIFY_SOURCE=2 -g -O2 -ffile-prefix-map=/build/mpi-testsuite-3.2+dfsg/2nd=. -fstack-protector-strong -Wformat -Werror=format-security -MT mtest.o -MD -MP -MF $depbase.Tpo -c -o mtest.o ../../../cxx/util/mtest.cxx &&\\\n mv -f $depbase.Tpo $depbase.Po\n ../../../cxx/util/mtest.cxx: In function 'void* MTestTypeContigInitRecv(MTestDatatype*)':\n ../../../cxx/util/mtest.cxx:189:42: warning: 'int MPI_Type_extent(MPI_Datatype, MPI_Aint*)' is deprecated: MPI_Type_extent is superseded by MPI_Type_get_extent in MPI-2.0 [-Wdeprecated-declarations]\n MPI_Type_extent( mtype->datatype, &size );\n ^\n In file included from ../../../cxx/util/mtest.cxx:7:\n /usr/lib/x86_64-linux-gnu/openmpi/include/mpi.h:1758:20: note: declared here\n@@ -1002210,28 +1002258,28 @@\n ../../../cxx/util/mtest.cxx:848:50: error: 'NULL_COPY_FN' is not a member of 'MPI::Win'\n mem_keyval = MPI::Win::Create_keyval( MPI::Win::NULL_COPY_FN,\n ^~~~~~~~~~~~\n ../../../cxx/util/mtest.cxx:849:22: error: 'NULL_DELETE_FN' is not a member of 'MPI::Win'\n MPI::Win::NULL_DELETE_FN, 0 );\n ^~~~~~~~~~~~~~\n make[3]: *** [Makefile:416: mtest.o] Error 1\n-make[3]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/util'\n+make[3]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/util'\n make[2]: *** [Makefile:865: ../../cxx/util/mtest.o] Error 2\n-make[2]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/coll'\n+make[2]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/coll'\n \n \n uallreduce\n \n ./cxx/coll\n fail\n- Failed to build uallreduce; make[2]: Entering directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/coll'\n+ Failed to build uallreduce; make[2]: Entering directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/coll'\n (cd ../../cxx/util && make mtest.o)\n-make[3]: Entering directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/util'\n+make[3]: Entering directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/util'\n depbase=`echo mtest.o | sed 's|[^/]*$|.deps/&|;s|\\.o$||'`;\\\n-/usr/bin/mpicxx.openmpi -DHAVE_CONFIG_H -I. -I../../../cxx/util -I../../include -I../../include -I../../../include -Wdate-time -D_FORTIFY_SOURCE=2 -g -O2 -ffile-prefix-map=/build/1st/mpi-testsuite-3.2+dfsg=. -fstack-protector-strong -Wformat -Werror=format-security -MT mtest.o -MD -MP -MF $depbase.Tpo -c -o mtest.o ../../../cxx/util/mtest.cxx &&\\\n+/usr/bin/mpicxx.openmpi -DHAVE_CONFIG_H -I. -I../../../cxx/util -I../../include -I../../include -I../../../include -Wdate-time -D_FORTIFY_SOURCE=2 -g -O2 -ffile-prefix-map=/build/mpi-testsuite-3.2+dfsg/2nd=. -fstack-protector-strong -Wformat -Werror=format-security -MT mtest.o -MD -MP -MF $depbase.Tpo -c -o mtest.o ../../../cxx/util/mtest.cxx &&\\\n mv -f $depbase.Tpo $depbase.Po\n ../../../cxx/util/mtest.cxx: In function 'void* MTestTypeContigInitRecv(MTestDatatype*)':\n ../../../cxx/util/mtest.cxx:189:42: warning: 'int MPI_Type_extent(MPI_Datatype, MPI_Aint*)' is deprecated: MPI_Type_extent is superseded by MPI_Type_get_extent in MPI-2.0 [-Wdeprecated-declarations]\n MPI_Type_extent( mtype->datatype, &size );\n ^\n In file included from ../../../cxx/util/mtest.cxx:7:\n /usr/lib/x86_64-linux-gnu/openmpi/include/mpi.h:1758:20: note: declared here\n@@ -1002263,28 +1002311,28 @@\n ../../../cxx/util/mtest.cxx:848:50: error: 'NULL_COPY_FN' is not a member of 'MPI::Win'\n mem_keyval = MPI::Win::Create_keyval( MPI::Win::NULL_COPY_FN,\n ^~~~~~~~~~~~\n ../../../cxx/util/mtest.cxx:849:22: error: 'NULL_DELETE_FN' is not a member of 'MPI::Win'\n MPI::Win::NULL_DELETE_FN, 0 );\n ^~~~~~~~~~~~~~\n make[3]: *** [Makefile:416: mtest.o] Error 1\n-make[3]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/util'\n+make[3]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/util'\n make[2]: *** [Makefile:865: ../../cxx/util/mtest.o] Error 2\n-make[2]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/coll'\n+make[2]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/coll'\n \n \n ureduce\n \n ./cxx/coll\n fail\n- Failed to build ureduce; make[2]: Entering directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/coll'\n+ Failed to build ureduce; make[2]: Entering directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/coll'\n (cd ../../cxx/util && make mtest.o)\n-make[3]: Entering directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/util'\n+make[3]: Entering directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/util'\n depbase=`echo mtest.o | sed 's|[^/]*$|.deps/&|;s|\\.o$||'`;\\\n-/usr/bin/mpicxx.openmpi -DHAVE_CONFIG_H -I. -I../../../cxx/util -I../../include -I../../include -I../../../include -Wdate-time -D_FORTIFY_SOURCE=2 -g -O2 -ffile-prefix-map=/build/1st/mpi-testsuite-3.2+dfsg=. -fstack-protector-strong -Wformat -Werror=format-security -MT mtest.o -MD -MP -MF $depbase.Tpo -c -o mtest.o ../../../cxx/util/mtest.cxx &&\\\n+/usr/bin/mpicxx.openmpi -DHAVE_CONFIG_H -I. -I../../../cxx/util -I../../include -I../../include -I../../../include -Wdate-time -D_FORTIFY_SOURCE=2 -g -O2 -ffile-prefix-map=/build/mpi-testsuite-3.2+dfsg/2nd=. -fstack-protector-strong -Wformat -Werror=format-security -MT mtest.o -MD -MP -MF $depbase.Tpo -c -o mtest.o ../../../cxx/util/mtest.cxx &&\\\n mv -f $depbase.Tpo $depbase.Po\n ../../../cxx/util/mtest.cxx: In function 'void* MTestTypeContigInitRecv(MTestDatatype*)':\n ../../../cxx/util/mtest.cxx:189:42: warning: 'int MPI_Type_extent(MPI_Datatype, MPI_Aint*)' is deprecated: MPI_Type_extent is superseded by MPI_Type_get_extent in MPI-2.0 [-Wdeprecated-declarations]\n MPI_Type_extent( mtype->datatype, &size );\n ^\n In file included from ../../../cxx/util/mtest.cxx:7:\n /usr/lib/x86_64-linux-gnu/openmpi/include/mpi.h:1758:20: note: declared here\n@@ -1002316,28 +1002364,28 @@\n ../../../cxx/util/mtest.cxx:848:50: error: 'NULL_COPY_FN' is not a member of 'MPI::Win'\n mem_keyval = MPI::Win::Create_keyval( MPI::Win::NULL_COPY_FN,\n ^~~~~~~~~~~~\n ../../../cxx/util/mtest.cxx:849:22: error: 'NULL_DELETE_FN' is not a member of 'MPI::Win'\n MPI::Win::NULL_DELETE_FN, 0 );\n ^~~~~~~~~~~~~~\n make[3]: *** [Makefile:416: mtest.o] Error 1\n-make[3]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/util'\n+make[3]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/util'\n make[2]: *** [Makefile:865: ../../cxx/util/mtest.o] Error 2\n-make[2]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/coll'\n+make[2]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/coll'\n \n \n ureducelocal\n \n ./cxx/coll\n fail\n- Failed to build ureducelocal; make[2]: Entering directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/coll'\n+ Failed to build ureducelocal; make[2]: Entering directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/coll'\n (cd ../../cxx/util && make mtest.o)\n-make[3]: Entering directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/util'\n+make[3]: Entering directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/util'\n depbase=`echo mtest.o | sed 's|[^/]*$|.deps/&|;s|\\.o$||'`;\\\n-/usr/bin/mpicxx.openmpi -DHAVE_CONFIG_H -I. -I../../../cxx/util -I../../include -I../../include -I../../../include -Wdate-time -D_FORTIFY_SOURCE=2 -g -O2 -ffile-prefix-map=/build/1st/mpi-testsuite-3.2+dfsg=. -fstack-protector-strong -Wformat -Werror=format-security -MT mtest.o -MD -MP -MF $depbase.Tpo -c -o mtest.o ../../../cxx/util/mtest.cxx &&\\\n+/usr/bin/mpicxx.openmpi -DHAVE_CONFIG_H -I. -I../../../cxx/util -I../../include -I../../include -I../../../include -Wdate-time -D_FORTIFY_SOURCE=2 -g -O2 -ffile-prefix-map=/build/mpi-testsuite-3.2+dfsg/2nd=. -fstack-protector-strong -Wformat -Werror=format-security -MT mtest.o -MD -MP -MF $depbase.Tpo -c -o mtest.o ../../../cxx/util/mtest.cxx &&\\\n mv -f $depbase.Tpo $depbase.Po\n ../../../cxx/util/mtest.cxx: In function 'void* MTestTypeContigInitRecv(MTestDatatype*)':\n ../../../cxx/util/mtest.cxx:189:42: warning: 'int MPI_Type_extent(MPI_Datatype, MPI_Aint*)' is deprecated: MPI_Type_extent is superseded by MPI_Type_get_extent in MPI-2.0 [-Wdeprecated-declarations]\n MPI_Type_extent( mtype->datatype, &size );\n ^\n In file included from ../../../cxx/util/mtest.cxx:7:\n /usr/lib/x86_64-linux-gnu/openmpi/include/mpi.h:1758:20: note: declared here\n@@ -1002369,28 +1002417,28 @@\n ../../../cxx/util/mtest.cxx:848:50: error: 'NULL_COPY_FN' is not a member of 'MPI::Win'\n mem_keyval = MPI::Win::Create_keyval( MPI::Win::NULL_COPY_FN,\n ^~~~~~~~~~~~\n ../../../cxx/util/mtest.cxx:849:22: error: 'NULL_DELETE_FN' is not a member of 'MPI::Win'\n MPI::Win::NULL_DELETE_FN, 0 );\n ^~~~~~~~~~~~~~\n make[3]: *** [Makefile:416: mtest.o] Error 1\n-make[3]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/util'\n+make[3]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/util'\n make[2]: *** [Makefile:865: ../../cxx/util/mtest.o] Error 2\n-make[2]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/coll'\n+make[2]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/coll'\n \n \n uscan\n \n ./cxx/coll\n fail\n- Failed to build uscan; make[2]: Entering directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/coll'\n+ Failed to build uscan; make[2]: Entering directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/coll'\n (cd ../../cxx/util && make mtest.o)\n-make[3]: Entering directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/util'\n+make[3]: Entering directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/util'\n depbase=`echo mtest.o | sed 's|[^/]*$|.deps/&|;s|\\.o$||'`;\\\n-/usr/bin/mpicxx.openmpi -DHAVE_CONFIG_H -I. -I../../../cxx/util -I../../include -I../../include -I../../../include -Wdate-time -D_FORTIFY_SOURCE=2 -g -O2 -ffile-prefix-map=/build/1st/mpi-testsuite-3.2+dfsg=. -fstack-protector-strong -Wformat -Werror=format-security -MT mtest.o -MD -MP -MF $depbase.Tpo -c -o mtest.o ../../../cxx/util/mtest.cxx &&\\\n+/usr/bin/mpicxx.openmpi -DHAVE_CONFIG_H -I. -I../../../cxx/util -I../../include -I../../include -I../../../include -Wdate-time -D_FORTIFY_SOURCE=2 -g -O2 -ffile-prefix-map=/build/mpi-testsuite-3.2+dfsg/2nd=. -fstack-protector-strong -Wformat -Werror=format-security -MT mtest.o -MD -MP -MF $depbase.Tpo -c -o mtest.o ../../../cxx/util/mtest.cxx &&\\\n mv -f $depbase.Tpo $depbase.Po\n ../../../cxx/util/mtest.cxx: In function 'void* MTestTypeContigInitRecv(MTestDatatype*)':\n ../../../cxx/util/mtest.cxx:189:42: warning: 'int MPI_Type_extent(MPI_Datatype, MPI_Aint*)' is deprecated: MPI_Type_extent is superseded by MPI_Type_get_extent in MPI-2.0 [-Wdeprecated-declarations]\n MPI_Type_extent( mtype->datatype, &size );\n ^\n In file included from ../../../cxx/util/mtest.cxx:7:\n /usr/lib/x86_64-linux-gnu/openmpi/include/mpi.h:1758:20: note: declared here\n@@ -1002422,28 +1002470,28 @@\n ../../../cxx/util/mtest.cxx:848:50: error: 'NULL_COPY_FN' is not a member of 'MPI::Win'\n mem_keyval = MPI::Win::Create_keyval( MPI::Win::NULL_COPY_FN,\n ^~~~~~~~~~~~\n ../../../cxx/util/mtest.cxx:849:22: error: 'NULL_DELETE_FN' is not a member of 'MPI::Win'\n MPI::Win::NULL_DELETE_FN, 0 );\n ^~~~~~~~~~~~~~\n make[3]: *** [Makefile:416: mtest.o] Error 1\n-make[3]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/util'\n+make[3]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/util'\n make[2]: *** [Makefile:865: ../../cxx/util/mtest.o] Error 2\n-make[2]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/coll'\n+make[2]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/coll'\n \n \n uexscan\n \n ./cxx/coll\n fail\n- Failed to build uexscan; make[2]: Entering directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/coll'\n+ Failed to build uexscan; make[2]: Entering directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/coll'\n (cd ../../cxx/util && make mtest.o)\n-make[3]: Entering directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/util'\n+make[3]: Entering directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/util'\n depbase=`echo mtest.o | sed 's|[^/]*$|.deps/&|;s|\\.o$||'`;\\\n-/usr/bin/mpicxx.openmpi -DHAVE_CONFIG_H -I. -I../../../cxx/util -I../../include -I../../include -I../../../include -Wdate-time -D_FORTIFY_SOURCE=2 -g -O2 -ffile-prefix-map=/build/1st/mpi-testsuite-3.2+dfsg=. -fstack-protector-strong -Wformat -Werror=format-security -MT mtest.o -MD -MP -MF $depbase.Tpo -c -o mtest.o ../../../cxx/util/mtest.cxx &&\\\n+/usr/bin/mpicxx.openmpi -DHAVE_CONFIG_H -I. -I../../../cxx/util -I../../include -I../../include -I../../../include -Wdate-time -D_FORTIFY_SOURCE=2 -g -O2 -ffile-prefix-map=/build/mpi-testsuite-3.2+dfsg/2nd=. -fstack-protector-strong -Wformat -Werror=format-security -MT mtest.o -MD -MP -MF $depbase.Tpo -c -o mtest.o ../../../cxx/util/mtest.cxx &&\\\n mv -f $depbase.Tpo $depbase.Po\n ../../../cxx/util/mtest.cxx: In function 'void* MTestTypeContigInitRecv(MTestDatatype*)':\n ../../../cxx/util/mtest.cxx:189:42: warning: 'int MPI_Type_extent(MPI_Datatype, MPI_Aint*)' is deprecated: MPI_Type_extent is superseded by MPI_Type_get_extent in MPI-2.0 [-Wdeprecated-declarations]\n MPI_Type_extent( mtype->datatype, &size );\n ^\n In file included from ../../../cxx/util/mtest.cxx:7:\n /usr/lib/x86_64-linux-gnu/openmpi/include/mpi.h:1758:20: note: declared here\n@@ -1002475,28 +1002523,28 @@\n ../../../cxx/util/mtest.cxx:848:50: error: 'NULL_COPY_FN' is not a member of 'MPI::Win'\n mem_keyval = MPI::Win::Create_keyval( MPI::Win::NULL_COPY_FN,\n ^~~~~~~~~~~~\n ../../../cxx/util/mtest.cxx:849:22: error: 'NULL_DELETE_FN' is not a member of 'MPI::Win'\n MPI::Win::NULL_DELETE_FN, 0 );\n ^~~~~~~~~~~~~~\n make[3]: *** [Makefile:416: mtest.o] Error 1\n-make[3]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/util'\n+make[3]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/util'\n make[2]: *** [Makefile:865: ../../cxx/util/mtest.o] Error 2\n-make[2]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/coll'\n+make[2]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/coll'\n \n \n alltoallw2x\n \n ./cxx/coll\n fail\n- Failed to build alltoallw2x; make[2]: Entering directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/coll'\n+ Failed to build alltoallw2x; make[2]: Entering directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/coll'\n (cd ../../cxx/util && make mtest.o)\n-make[3]: Entering directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/util'\n+make[3]: Entering directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/util'\n depbase=`echo mtest.o | sed 's|[^/]*$|.deps/&|;s|\\.o$||'`;\\\n-/usr/bin/mpicxx.openmpi -DHAVE_CONFIG_H -I. -I../../../cxx/util -I../../include -I../../include -I../../../include -Wdate-time -D_FORTIFY_SOURCE=2 -g -O2 -ffile-prefix-map=/build/1st/mpi-testsuite-3.2+dfsg=. -fstack-protector-strong -Wformat -Werror=format-security -MT mtest.o -MD -MP -MF $depbase.Tpo -c -o mtest.o ../../../cxx/util/mtest.cxx &&\\\n+/usr/bin/mpicxx.openmpi -DHAVE_CONFIG_H -I. -I../../../cxx/util -I../../include -I../../include -I../../../include -Wdate-time -D_FORTIFY_SOURCE=2 -g -O2 -ffile-prefix-map=/build/mpi-testsuite-3.2+dfsg/2nd=. -fstack-protector-strong -Wformat -Werror=format-security -MT mtest.o -MD -MP -MF $depbase.Tpo -c -o mtest.o ../../../cxx/util/mtest.cxx &&\\\n mv -f $depbase.Tpo $depbase.Po\n ../../../cxx/util/mtest.cxx: In function 'void* MTestTypeContigInitRecv(MTestDatatype*)':\n ../../../cxx/util/mtest.cxx:189:42: warning: 'int MPI_Type_extent(MPI_Datatype, MPI_Aint*)' is deprecated: MPI_Type_extent is superseded by MPI_Type_get_extent in MPI-2.0 [-Wdeprecated-declarations]\n MPI_Type_extent( mtype->datatype, &size );\n ^\n In file included from ../../../cxx/util/mtest.cxx:7:\n /usr/lib/x86_64-linux-gnu/openmpi/include/mpi.h:1758:20: note: declared here\n@@ -1002528,28 +1002576,28 @@\n ../../../cxx/util/mtest.cxx:848:50: error: 'NULL_COPY_FN' is not a member of 'MPI::Win'\n mem_keyval = MPI::Win::Create_keyval( MPI::Win::NULL_COPY_FN,\n ^~~~~~~~~~~~\n ../../../cxx/util/mtest.cxx:849:22: error: 'NULL_DELETE_FN' is not a member of 'MPI::Win'\n MPI::Win::NULL_DELETE_FN, 0 );\n ^~~~~~~~~~~~~~\n make[3]: *** [Makefile:416: mtest.o] Error 1\n-make[3]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/util'\n+make[3]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/util'\n make[2]: *** [Makefile:865: ../../cxx/util/mtest.o] Error 2\n-make[2]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/coll'\n+make[2]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/coll'\n \n \n icbcastx\n \n ./cxx/coll\n fail\n- Failed to build icbcastx; make[2]: Entering directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/coll'\n+ Failed to build icbcastx; make[2]: Entering directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/coll'\n (cd ../../cxx/util && make mtest.o)\n-make[3]: Entering directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/util'\n+make[3]: Entering directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/util'\n depbase=`echo mtest.o | sed 's|[^/]*$|.deps/&|;s|\\.o$||'`;\\\n-/usr/bin/mpicxx.openmpi -DHAVE_CONFIG_H -I. -I../../../cxx/util -I../../include -I../../include -I../../../include -Wdate-time -D_FORTIFY_SOURCE=2 -g -O2 -ffile-prefix-map=/build/1st/mpi-testsuite-3.2+dfsg=. -fstack-protector-strong -Wformat -Werror=format-security -MT mtest.o -MD -MP -MF $depbase.Tpo -c -o mtest.o ../../../cxx/util/mtest.cxx &&\\\n+/usr/bin/mpicxx.openmpi -DHAVE_CONFIG_H -I. -I../../../cxx/util -I../../include -I../../include -I../../../include -Wdate-time -D_FORTIFY_SOURCE=2 -g -O2 -ffile-prefix-map=/build/mpi-testsuite-3.2+dfsg/2nd=. -fstack-protector-strong -Wformat -Werror=format-security -MT mtest.o -MD -MP -MF $depbase.Tpo -c -o mtest.o ../../../cxx/util/mtest.cxx &&\\\n mv -f $depbase.Tpo $depbase.Po\n ../../../cxx/util/mtest.cxx: In function 'void* MTestTypeContigInitRecv(MTestDatatype*)':\n ../../../cxx/util/mtest.cxx:189:42: warning: 'int MPI_Type_extent(MPI_Datatype, MPI_Aint*)' is deprecated: MPI_Type_extent is superseded by MPI_Type_get_extent in MPI-2.0 [-Wdeprecated-declarations]\n MPI_Type_extent( mtype->datatype, &size );\n ^\n In file included from ../../../cxx/util/mtest.cxx:7:\n /usr/lib/x86_64-linux-gnu/openmpi/include/mpi.h:1758:20: note: declared here\n@@ -1002581,31 +1002629,31 @@\n ../../../cxx/util/mtest.cxx:848:50: error: 'NULL_COPY_FN' is not a member of 'MPI::Win'\n mem_keyval = MPI::Win::Create_keyval( MPI::Win::NULL_COPY_FN,\n ^~~~~~~~~~~~\n ../../../cxx/util/mtest.cxx:849:22: error: 'NULL_DELETE_FN' is not a member of 'MPI::Win'\n MPI::Win::NULL_DELETE_FN, 0 );\n ^~~~~~~~~~~~~~\n make[3]: *** [Makefile:416: mtest.o] Error 1\n-make[3]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/util'\n+make[3]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/util'\n make[2]: *** [Makefile:865: ../../cxx/util/mtest.o] Error 2\n-make[2]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/coll'\n+make[2]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/coll'\n \n \n icbcastx\n \n ./cxx/coll\n fail\n- Failed to build icbcastx; make[2]: Entering directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/coll'\n+ Failed to build icbcastx; make[2]: Entering directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/coll'\n depbase=`echo icbcastx.o | sed 's|[^/]*$|.deps/&|;s|\\.o$||'`;\\\n-/usr/bin/mpicxx.openmpi -DHAVE_CONFIG_H -I. -I../../../cxx/coll -I../../include -I../../include -I../../../include -Wdate-time -D_FORTIFY_SOURCE=2 -g -O2 -ffile-prefix-map=/build/1st/mpi-testsuite-3.2+dfsg=. -fstack-protector-strong -Wformat -Werror=format-security -MT icbcastx.o -MD -MP -MF $depbase.Tpo -c -o icbcastx.o ../../../cxx/coll/icbcastx.cxx &&\\\n+/usr/bin/mpicxx.openmpi -DHAVE_CONFIG_H -I. -I../../../cxx/coll -I../../include -I../../include -I../../../include -Wdate-time -D_FORTIFY_SOURCE=2 -g -O2 -ffile-prefix-map=/build/mpi-testsuite-3.2+dfsg/2nd=. -fstack-protector-strong -Wformat -Werror=format-security -MT icbcastx.o -MD -MP -MF $depbase.Tpo -c -o icbcastx.o ../../../cxx/coll/icbcastx.cxx &&\\\n mv -f $depbase.Tpo $depbase.Po\n (cd ../../cxx/util && make mtest.o)\n-make[3]: Entering directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/util'\n+make[3]: Entering directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/util'\n depbase=`echo mtest.o | sed 's|[^/]*$|.deps/&|;s|\\.o$||'`;\\\n-/usr/bin/mpicxx.openmpi -DHAVE_CONFIG_H -I. -I../../../cxx/util -I../../include -I../../include -I../../../include -Wdate-time -D_FORTIFY_SOURCE=2 -g -O2 -ffile-prefix-map=/build/1st/mpi-testsuite-3.2+dfsg=. -fstack-protector-strong -Wformat -Werror=format-security -MT mtest.o -MD -MP -MF $depbase.Tpo -c -o mtest.o ../../../cxx/util/mtest.cxx &&\\\n+/usr/bin/mpicxx.openmpi -DHAVE_CONFIG_H -I. -I../../../cxx/util -I../../include -I../../include -I../../../include -Wdate-time -D_FORTIFY_SOURCE=2 -g -O2 -ffile-prefix-map=/build/mpi-testsuite-3.2+dfsg/2nd=. -fstack-protector-strong -Wformat -Werror=format-security -MT mtest.o -MD -MP -MF $depbase.Tpo -c -o mtest.o ../../../cxx/util/mtest.cxx &&\\\n mv -f $depbase.Tpo $depbase.Po\n ../../../cxx/util/mtest.cxx: In function 'void* MTestTypeContigInitRecv(MTestDatatype*)':\n ../../../cxx/util/mtest.cxx:189:42: warning: 'int MPI_Type_extent(MPI_Datatype, MPI_Aint*)' is deprecated: MPI_Type_extent is superseded by MPI_Type_get_extent in MPI-2.0 [-Wdeprecated-declarations]\n MPI_Type_extent( mtype->datatype, &size );\n ^\n In file included from ../../../cxx/util/mtest.cxx:7:\n /usr/lib/x86_64-linux-gnu/openmpi/include/mpi.h:1758:20: note: declared here\n@@ -1002637,28 +1002685,28 @@\n ../../../cxx/util/mtest.cxx:848:50: error: 'NULL_COPY_FN' is not a member of 'MPI::Win'\n mem_keyval = MPI::Win::Create_keyval( MPI::Win::NULL_COPY_FN,\n ^~~~~~~~~~~~\n ../../../cxx/util/mtest.cxx:849:22: error: 'NULL_DELETE_FN' is not a member of 'MPI::Win'\n MPI::Win::NULL_DELETE_FN, 0 );\n ^~~~~~~~~~~~~~\n make[3]: *** [Makefile:416: mtest.o] Error 1\n-make[3]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/util'\n+make[3]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/util'\n make[2]: *** [Makefile:865: ../../cxx/util/mtest.o] Error 2\n-make[2]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/coll'\n+make[2]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/coll'\n \n \n icallreducex\n \n ./cxx/coll\n fail\n- Failed to build icallreducex; make[2]: Entering directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/coll'\n+ Failed to build icallreducex; make[2]: Entering directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/coll'\n (cd ../../cxx/util && make mtest.o)\n-make[3]: Entering directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/util'\n+make[3]: Entering directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/util'\n depbase=`echo mtest.o | sed 's|[^/]*$|.deps/&|;s|\\.o$||'`;\\\n-/usr/bin/mpicxx.openmpi -DHAVE_CONFIG_H -I. -I../../../cxx/util -I../../include -I../../include -I../../../include -Wdate-time -D_FORTIFY_SOURCE=2 -g -O2 -ffile-prefix-map=/build/1st/mpi-testsuite-3.2+dfsg=. -fstack-protector-strong -Wformat -Werror=format-security -MT mtest.o -MD -MP -MF $depbase.Tpo -c -o mtest.o ../../../cxx/util/mtest.cxx &&\\\n+/usr/bin/mpicxx.openmpi -DHAVE_CONFIG_H -I. -I../../../cxx/util -I../../include -I../../include -I../../../include -Wdate-time -D_FORTIFY_SOURCE=2 -g -O2 -ffile-prefix-map=/build/mpi-testsuite-3.2+dfsg/2nd=. -fstack-protector-strong -Wformat -Werror=format-security -MT mtest.o -MD -MP -MF $depbase.Tpo -c -o mtest.o ../../../cxx/util/mtest.cxx &&\\\n mv -f $depbase.Tpo $depbase.Po\n ../../../cxx/util/mtest.cxx: In function 'void* MTestTypeContigInitRecv(MTestDatatype*)':\n ../../../cxx/util/mtest.cxx:189:42: warning: 'int MPI_Type_extent(MPI_Datatype, MPI_Aint*)' is deprecated: MPI_Type_extent is superseded by MPI_Type_get_extent in MPI-2.0 [-Wdeprecated-declarations]\n MPI_Type_extent( mtype->datatype, &size );\n ^\n In file included from ../../../cxx/util/mtest.cxx:7:\n /usr/lib/x86_64-linux-gnu/openmpi/include/mpi.h:1758:20: note: declared here\n@@ -1002690,28 +1002738,28 @@\n ../../../cxx/util/mtest.cxx:848:50: error: 'NULL_COPY_FN' is not a member of 'MPI::Win'\n mem_keyval = MPI::Win::Create_keyval( MPI::Win::NULL_COPY_FN,\n ^~~~~~~~~~~~\n ../../../cxx/util/mtest.cxx:849:22: error: 'NULL_DELETE_FN' is not a member of 'MPI::Win'\n MPI::Win::NULL_DELETE_FN, 0 );\n ^~~~~~~~~~~~~~\n make[3]: *** [Makefile:416: mtest.o] Error 1\n-make[3]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/util'\n+make[3]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/util'\n make[2]: *** [Makefile:865: ../../cxx/util/mtest.o] Error 2\n-make[2]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/coll'\n+make[2]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/coll'\n \n \n icreducex\n \n ./cxx/coll\n fail\n- Failed to build icreducex; make[2]: Entering directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/coll'\n+ Failed to build icreducex; make[2]: Entering directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/coll'\n (cd ../../cxx/util && make mtest.o)\n-make[3]: Entering directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/util'\n+make[3]: Entering directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/util'\n depbase=`echo mtest.o | sed 's|[^/]*$|.deps/&|;s|\\.o$||'`;\\\n-/usr/bin/mpicxx.openmpi -DHAVE_CONFIG_H -I. -I../../../cxx/util -I../../include -I../../include -I../../../include -Wdate-time -D_FORTIFY_SOURCE=2 -g -O2 -ffile-prefix-map=/build/1st/mpi-testsuite-3.2+dfsg=. -fstack-protector-strong -Wformat -Werror=format-security -MT mtest.o -MD -MP -MF $depbase.Tpo -c -o mtest.o ../../../cxx/util/mtest.cxx &&\\\n+/usr/bin/mpicxx.openmpi -DHAVE_CONFIG_H -I. -I../../../cxx/util -I../../include -I../../include -I../../../include -Wdate-time -D_FORTIFY_SOURCE=2 -g -O2 -ffile-prefix-map=/build/mpi-testsuite-3.2+dfsg/2nd=. -fstack-protector-strong -Wformat -Werror=format-security -MT mtest.o -MD -MP -MF $depbase.Tpo -c -o mtest.o ../../../cxx/util/mtest.cxx &&\\\n mv -f $depbase.Tpo $depbase.Po\n ../../../cxx/util/mtest.cxx: In function 'void* MTestTypeContigInitRecv(MTestDatatype*)':\n ../../../cxx/util/mtest.cxx:189:42: warning: 'int MPI_Type_extent(MPI_Datatype, MPI_Aint*)' is deprecated: MPI_Type_extent is superseded by MPI_Type_get_extent in MPI-2.0 [-Wdeprecated-declarations]\n MPI_Type_extent( mtype->datatype, &size );\n ^\n In file included from ../../../cxx/util/mtest.cxx:7:\n /usr/lib/x86_64-linux-gnu/openmpi/include/mpi.h:1758:20: note: declared here\n@@ -1002743,28 +1002791,28 @@\n ../../../cxx/util/mtest.cxx:848:50: error: 'NULL_COPY_FN' is not a member of 'MPI::Win'\n mem_keyval = MPI::Win::Create_keyval( MPI::Win::NULL_COPY_FN,\n ^~~~~~~~~~~~\n ../../../cxx/util/mtest.cxx:849:22: error: 'NULL_DELETE_FN' is not a member of 'MPI::Win'\n MPI::Win::NULL_DELETE_FN, 0 );\n ^~~~~~~~~~~~~~\n make[3]: *** [Makefile:416: mtest.o] Error 1\n-make[3]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/util'\n+make[3]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/util'\n make[2]: *** [Makefile:865: ../../cxx/util/mtest.o] Error 2\n-make[2]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/coll'\n+make[2]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/coll'\n \n \n icscatterx\n \n ./cxx/coll\n fail\n- Failed to build icscatterx; make[2]: Entering directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/coll'\n+ Failed to build icscatterx; make[2]: Entering directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/coll'\n (cd ../../cxx/util && make mtest.o)\n-make[3]: Entering directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/util'\n+make[3]: Entering directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/util'\n depbase=`echo mtest.o | sed 's|[^/]*$|.deps/&|;s|\\.o$||'`;\\\n-/usr/bin/mpicxx.openmpi -DHAVE_CONFIG_H -I. -I../../../cxx/util -I../../include -I../../include -I../../../include -Wdate-time -D_FORTIFY_SOURCE=2 -g -O2 -ffile-prefix-map=/build/1st/mpi-testsuite-3.2+dfsg=. -fstack-protector-strong -Wformat -Werror=format-security -MT mtest.o -MD -MP -MF $depbase.Tpo -c -o mtest.o ../../../cxx/util/mtest.cxx &&\\\n+/usr/bin/mpicxx.openmpi -DHAVE_CONFIG_H -I. -I../../../cxx/util -I../../include -I../../include -I../../../include -Wdate-time -D_FORTIFY_SOURCE=2 -g -O2 -ffile-prefix-map=/build/mpi-testsuite-3.2+dfsg/2nd=. -fstack-protector-strong -Wformat -Werror=format-security -MT mtest.o -MD -MP -MF $depbase.Tpo -c -o mtest.o ../../../cxx/util/mtest.cxx &&\\\n mv -f $depbase.Tpo $depbase.Po\n ../../../cxx/util/mtest.cxx: In function 'void* MTestTypeContigInitRecv(MTestDatatype*)':\n ../../../cxx/util/mtest.cxx:189:42: warning: 'int MPI_Type_extent(MPI_Datatype, MPI_Aint*)' is deprecated: MPI_Type_extent is superseded by MPI_Type_get_extent in MPI-2.0 [-Wdeprecated-declarations]\n MPI_Type_extent( mtype->datatype, &size );\n ^\n In file included from ../../../cxx/util/mtest.cxx:7:\n /usr/lib/x86_64-linux-gnu/openmpi/include/mpi.h:1758:20: note: declared here\n@@ -1002796,28 +1002844,28 @@\n ../../../cxx/util/mtest.cxx:848:50: error: 'NULL_COPY_FN' is not a member of 'MPI::Win'\n mem_keyval = MPI::Win::Create_keyval( MPI::Win::NULL_COPY_FN,\n ^~~~~~~~~~~~\n ../../../cxx/util/mtest.cxx:849:22: error: 'NULL_DELETE_FN' is not a member of 'MPI::Win'\n MPI::Win::NULL_DELETE_FN, 0 );\n ^~~~~~~~~~~~~~\n make[3]: *** [Makefile:416: mtest.o] Error 1\n-make[3]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/util'\n+make[3]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/util'\n make[2]: *** [Makefile:865: ../../cxx/util/mtest.o] Error 2\n-make[2]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/coll'\n+make[2]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/coll'\n \n \n icgatherx\n \n ./cxx/coll\n fail\n- Failed to build icgatherx; make[2]: Entering directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/coll'\n+ Failed to build icgatherx; make[2]: Entering directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/coll'\n (cd ../../cxx/util && make mtest.o)\n-make[3]: Entering directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/util'\n+make[3]: Entering directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/util'\n depbase=`echo mtest.o | sed 's|[^/]*$|.deps/&|;s|\\.o$||'`;\\\n-/usr/bin/mpicxx.openmpi -DHAVE_CONFIG_H -I. -I../../../cxx/util -I../../include -I../../include -I../../../include -Wdate-time -D_FORTIFY_SOURCE=2 -g -O2 -ffile-prefix-map=/build/1st/mpi-testsuite-3.2+dfsg=. -fstack-protector-strong -Wformat -Werror=format-security -MT mtest.o -MD -MP -MF $depbase.Tpo -c -o mtest.o ../../../cxx/util/mtest.cxx &&\\\n+/usr/bin/mpicxx.openmpi -DHAVE_CONFIG_H -I. -I../../../cxx/util -I../../include -I../../include -I../../../include -Wdate-time -D_FORTIFY_SOURCE=2 -g -O2 -ffile-prefix-map=/build/mpi-testsuite-3.2+dfsg/2nd=. -fstack-protector-strong -Wformat -Werror=format-security -MT mtest.o -MD -MP -MF $depbase.Tpo -c -o mtest.o ../../../cxx/util/mtest.cxx &&\\\n mv -f $depbase.Tpo $depbase.Po\n ../../../cxx/util/mtest.cxx: In function 'void* MTestTypeContigInitRecv(MTestDatatype*)':\n ../../../cxx/util/mtest.cxx:189:42: warning: 'int MPI_Type_extent(MPI_Datatype, MPI_Aint*)' is deprecated: MPI_Type_extent is superseded by MPI_Type_get_extent in MPI-2.0 [-Wdeprecated-declarations]\n MPI_Type_extent( mtype->datatype, &size );\n ^\n In file included from ../../../cxx/util/mtest.cxx:7:\n /usr/lib/x86_64-linux-gnu/openmpi/include/mpi.h:1758:20: note: declared here\n@@ -1002849,28 +1002897,28 @@\n ../../../cxx/util/mtest.cxx:848:50: error: 'NULL_COPY_FN' is not a member of 'MPI::Win'\n mem_keyval = MPI::Win::Create_keyval( MPI::Win::NULL_COPY_FN,\n ^~~~~~~~~~~~\n ../../../cxx/util/mtest.cxx:849:22: error: 'NULL_DELETE_FN' is not a member of 'MPI::Win'\n MPI::Win::NULL_DELETE_FN, 0 );\n ^~~~~~~~~~~~~~\n make[3]: *** [Makefile:416: mtest.o] Error 1\n-make[3]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/util'\n+make[3]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/util'\n make[2]: *** [Makefile:865: ../../cxx/util/mtest.o] Error 2\n-make[2]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/coll'\n+make[2]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/coll'\n \n \n icallgatherx\n \n ./cxx/coll\n fail\n- Failed to build icallgatherx; make[2]: Entering directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/coll'\n+ Failed to build icallgatherx; make[2]: Entering directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/coll'\n (cd ../../cxx/util && make mtest.o)\n-make[3]: Entering directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/util'\n+make[3]: Entering directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/util'\n depbase=`echo mtest.o | sed 's|[^/]*$|.deps/&|;s|\\.o$||'`;\\\n-/usr/bin/mpicxx.openmpi -DHAVE_CONFIG_H -I. -I../../../cxx/util -I../../include -I../../include -I../../../include -Wdate-time -D_FORTIFY_SOURCE=2 -g -O2 -ffile-prefix-map=/build/1st/mpi-testsuite-3.2+dfsg=. -fstack-protector-strong -Wformat -Werror=format-security -MT mtest.o -MD -MP -MF $depbase.Tpo -c -o mtest.o ../../../cxx/util/mtest.cxx &&\\\n+/usr/bin/mpicxx.openmpi -DHAVE_CONFIG_H -I. -I../../../cxx/util -I../../include -I../../include -I../../../include -Wdate-time -D_FORTIFY_SOURCE=2 -g -O2 -ffile-prefix-map=/build/mpi-testsuite-3.2+dfsg/2nd=. -fstack-protector-strong -Wformat -Werror=format-security -MT mtest.o -MD -MP -MF $depbase.Tpo -c -o mtest.o ../../../cxx/util/mtest.cxx &&\\\n mv -f $depbase.Tpo $depbase.Po\n ../../../cxx/util/mtest.cxx: In function 'void* MTestTypeContigInitRecv(MTestDatatype*)':\n ../../../cxx/util/mtest.cxx:189:42: warning: 'int MPI_Type_extent(MPI_Datatype, MPI_Aint*)' is deprecated: MPI_Type_extent is superseded by MPI_Type_get_extent in MPI-2.0 [-Wdeprecated-declarations]\n MPI_Type_extent( mtype->datatype, &size );\n ^\n In file included from ../../../cxx/util/mtest.cxx:7:\n /usr/lib/x86_64-linux-gnu/openmpi/include/mpi.h:1758:20: note: declared here\n@@ -1002902,28 +1002950,28 @@\n ../../../cxx/util/mtest.cxx:848:50: error: 'NULL_COPY_FN' is not a member of 'MPI::Win'\n mem_keyval = MPI::Win::Create_keyval( MPI::Win::NULL_COPY_FN,\n ^~~~~~~~~~~~\n ../../../cxx/util/mtest.cxx:849:22: error: 'NULL_DELETE_FN' is not a member of 'MPI::Win'\n MPI::Win::NULL_DELETE_FN, 0 );\n ^~~~~~~~~~~~~~\n make[3]: *** [Makefile:416: mtest.o] Error 1\n-make[3]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/util'\n+make[3]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/util'\n make[2]: *** [Makefile:865: ../../cxx/util/mtest.o] Error 2\n-make[2]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/coll'\n+make[2]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/coll'\n \n \n icbarrierx\n \n ./cxx/coll\n fail\n- Failed to build icbarrierx; make[2]: Entering directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/coll'\n+ Failed to build icbarrierx; make[2]: Entering directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/coll'\n (cd ../../cxx/util && make mtest.o)\n-make[3]: Entering directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/util'\n+make[3]: Entering directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/util'\n depbase=`echo mtest.o | sed 's|[^/]*$|.deps/&|;s|\\.o$||'`;\\\n-/usr/bin/mpicxx.openmpi -DHAVE_CONFIG_H -I. -I../../../cxx/util -I../../include -I../../include -I../../../include -Wdate-time -D_FORTIFY_SOURCE=2 -g -O2 -ffile-prefix-map=/build/1st/mpi-testsuite-3.2+dfsg=. -fstack-protector-strong -Wformat -Werror=format-security -MT mtest.o -MD -MP -MF $depbase.Tpo -c -o mtest.o ../../../cxx/util/mtest.cxx &&\\\n+/usr/bin/mpicxx.openmpi -DHAVE_CONFIG_H -I. -I../../../cxx/util -I../../include -I../../include -I../../../include -Wdate-time -D_FORTIFY_SOURCE=2 -g -O2 -ffile-prefix-map=/build/mpi-testsuite-3.2+dfsg/2nd=. -fstack-protector-strong -Wformat -Werror=format-security -MT mtest.o -MD -MP -MF $depbase.Tpo -c -o mtest.o ../../../cxx/util/mtest.cxx &&\\\n mv -f $depbase.Tpo $depbase.Po\n ../../../cxx/util/mtest.cxx: In function 'void* MTestTypeContigInitRecv(MTestDatatype*)':\n ../../../cxx/util/mtest.cxx:189:42: warning: 'int MPI_Type_extent(MPI_Datatype, MPI_Aint*)' is deprecated: MPI_Type_extent is superseded by MPI_Type_get_extent in MPI-2.0 [-Wdeprecated-declarations]\n MPI_Type_extent( mtype->datatype, &size );\n ^\n In file included from ../../../cxx/util/mtest.cxx:7:\n /usr/lib/x86_64-linux-gnu/openmpi/include/mpi.h:1758:20: note: declared here\n@@ -1002955,28 +1003003,28 @@\n ../../../cxx/util/mtest.cxx:848:50: error: 'NULL_COPY_FN' is not a member of 'MPI::Win'\n mem_keyval = MPI::Win::Create_keyval( MPI::Win::NULL_COPY_FN,\n ^~~~~~~~~~~~\n ../../../cxx/util/mtest.cxx:849:22: error: 'NULL_DELETE_FN' is not a member of 'MPI::Win'\n MPI::Win::NULL_DELETE_FN, 0 );\n ^~~~~~~~~~~~~~\n make[3]: *** [Makefile:416: mtest.o] Error 1\n-make[3]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/util'\n+make[3]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/util'\n make[2]: *** [Makefile:865: ../../cxx/util/mtest.o] Error 2\n-make[2]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/coll'\n+make[2]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/coll'\n \n \n icallgathervx\n \n ./cxx/coll\n fail\n- Failed to build icallgathervx; make[2]: Entering directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/coll'\n+ Failed to build icallgathervx; make[2]: Entering directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/coll'\n (cd ../../cxx/util && make mtest.o)\n-make[3]: Entering directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/util'\n+make[3]: Entering directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/util'\n depbase=`echo mtest.o | sed 's|[^/]*$|.deps/&|;s|\\.o$||'`;\\\n-/usr/bin/mpicxx.openmpi -DHAVE_CONFIG_H -I. -I../../../cxx/util -I../../include -I../../include -I../../../include -Wdate-time -D_FORTIFY_SOURCE=2 -g -O2 -ffile-prefix-map=/build/1st/mpi-testsuite-3.2+dfsg=. -fstack-protector-strong -Wformat -Werror=format-security -MT mtest.o -MD -MP -MF $depbase.Tpo -c -o mtest.o ../../../cxx/util/mtest.cxx &&\\\n+/usr/bin/mpicxx.openmpi -DHAVE_CONFIG_H -I. -I../../../cxx/util -I../../include -I../../include -I../../../include -Wdate-time -D_FORTIFY_SOURCE=2 -g -O2 -ffile-prefix-map=/build/mpi-testsuite-3.2+dfsg/2nd=. -fstack-protector-strong -Wformat -Werror=format-security -MT mtest.o -MD -MP -MF $depbase.Tpo -c -o mtest.o ../../../cxx/util/mtest.cxx &&\\\n mv -f $depbase.Tpo $depbase.Po\n ../../../cxx/util/mtest.cxx: In function 'void* MTestTypeContigInitRecv(MTestDatatype*)':\n ../../../cxx/util/mtest.cxx:189:42: warning: 'int MPI_Type_extent(MPI_Datatype, MPI_Aint*)' is deprecated: MPI_Type_extent is superseded by MPI_Type_get_extent in MPI-2.0 [-Wdeprecated-declarations]\n MPI_Type_extent( mtype->datatype, &size );\n ^\n In file included from ../../../cxx/util/mtest.cxx:7:\n /usr/lib/x86_64-linux-gnu/openmpi/include/mpi.h:1758:20: note: declared here\n@@ -1003008,28 +1003056,28 @@\n ../../../cxx/util/mtest.cxx:848:50: error: 'NULL_COPY_FN' is not a member of 'MPI::Win'\n mem_keyval = MPI::Win::Create_keyval( MPI::Win::NULL_COPY_FN,\n ^~~~~~~~~~~~\n ../../../cxx/util/mtest.cxx:849:22: error: 'NULL_DELETE_FN' is not a member of 'MPI::Win'\n MPI::Win::NULL_DELETE_FN, 0 );\n ^~~~~~~~~~~~~~\n make[3]: *** [Makefile:416: mtest.o] Error 1\n-make[3]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/util'\n+make[3]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/util'\n make[2]: *** [Makefile:865: ../../cxx/util/mtest.o] Error 2\n-make[2]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/coll'\n+make[2]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/coll'\n \n \n icgathervx\n \n ./cxx/coll\n fail\n- Failed to build icgathervx; make[2]: Entering directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/coll'\n+ Failed to build icgathervx; make[2]: Entering directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/coll'\n (cd ../../cxx/util && make mtest.o)\n-make[3]: Entering directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/util'\n+make[3]: Entering directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/util'\n depbase=`echo mtest.o | sed 's|[^/]*$|.deps/&|;s|\\.o$||'`;\\\n-/usr/bin/mpicxx.openmpi -DHAVE_CONFIG_H -I. -I../../../cxx/util -I../../include -I../../include -I../../../include -Wdate-time -D_FORTIFY_SOURCE=2 -g -O2 -ffile-prefix-map=/build/1st/mpi-testsuite-3.2+dfsg=. -fstack-protector-strong -Wformat -Werror=format-security -MT mtest.o -MD -MP -MF $depbase.Tpo -c -o mtest.o ../../../cxx/util/mtest.cxx &&\\\n+/usr/bin/mpicxx.openmpi -DHAVE_CONFIG_H -I. -I../../../cxx/util -I../../include -I../../include -I../../../include -Wdate-time -D_FORTIFY_SOURCE=2 -g -O2 -ffile-prefix-map=/build/mpi-testsuite-3.2+dfsg/2nd=. -fstack-protector-strong -Wformat -Werror=format-security -MT mtest.o -MD -MP -MF $depbase.Tpo -c -o mtest.o ../../../cxx/util/mtest.cxx &&\\\n mv -f $depbase.Tpo $depbase.Po\n ../../../cxx/util/mtest.cxx: In function 'void* MTestTypeContigInitRecv(MTestDatatype*)':\n ../../../cxx/util/mtest.cxx:189:42: warning: 'int MPI_Type_extent(MPI_Datatype, MPI_Aint*)' is deprecated: MPI_Type_extent is superseded by MPI_Type_get_extent in MPI-2.0 [-Wdeprecated-declarations]\n MPI_Type_extent( mtype->datatype, &size );\n ^\n In file included from ../../../cxx/util/mtest.cxx:7:\n /usr/lib/x86_64-linux-gnu/openmpi/include/mpi.h:1758:20: note: declared here\n@@ -1003061,28 +1003109,28 @@\n ../../../cxx/util/mtest.cxx:848:50: error: 'NULL_COPY_FN' is not a member of 'MPI::Win'\n mem_keyval = MPI::Win::Create_keyval( MPI::Win::NULL_COPY_FN,\n ^~~~~~~~~~~~\n ../../../cxx/util/mtest.cxx:849:22: error: 'NULL_DELETE_FN' is not a member of 'MPI::Win'\n MPI::Win::NULL_DELETE_FN, 0 );\n ^~~~~~~~~~~~~~\n make[3]: *** [Makefile:416: mtest.o] Error 1\n-make[3]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/util'\n+make[3]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/util'\n make[2]: *** [Makefile:865: ../../cxx/util/mtest.o] Error 2\n-make[2]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/coll'\n+make[2]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/coll'\n \n \n icscattervx\n \n ./cxx/coll\n fail\n- Failed to build icscattervx; make[2]: Entering directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/coll'\n+ Failed to build icscattervx; make[2]: Entering directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/coll'\n (cd ../../cxx/util && make mtest.o)\n-make[3]: Entering directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/util'\n+make[3]: Entering directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/util'\n depbase=`echo mtest.o | sed 's|[^/]*$|.deps/&|;s|\\.o$||'`;\\\n-/usr/bin/mpicxx.openmpi -DHAVE_CONFIG_H -I. -I../../../cxx/util -I../../include -I../../include -I../../../include -Wdate-time -D_FORTIFY_SOURCE=2 -g -O2 -ffile-prefix-map=/build/1st/mpi-testsuite-3.2+dfsg=. -fstack-protector-strong -Wformat -Werror=format-security -MT mtest.o -MD -MP -MF $depbase.Tpo -c -o mtest.o ../../../cxx/util/mtest.cxx &&\\\n+/usr/bin/mpicxx.openmpi -DHAVE_CONFIG_H -I. -I../../../cxx/util -I../../include -I../../include -I../../../include -Wdate-time -D_FORTIFY_SOURCE=2 -g -O2 -ffile-prefix-map=/build/mpi-testsuite-3.2+dfsg/2nd=. -fstack-protector-strong -Wformat -Werror=format-security -MT mtest.o -MD -MP -MF $depbase.Tpo -c -o mtest.o ../../../cxx/util/mtest.cxx &&\\\n mv -f $depbase.Tpo $depbase.Po\n ../../../cxx/util/mtest.cxx: In function 'void* MTestTypeContigInitRecv(MTestDatatype*)':\n ../../../cxx/util/mtest.cxx:189:42: warning: 'int MPI_Type_extent(MPI_Datatype, MPI_Aint*)' is deprecated: MPI_Type_extent is superseded by MPI_Type_get_extent in MPI-2.0 [-Wdeprecated-declarations]\n MPI_Type_extent( mtype->datatype, &size );\n ^\n In file included from ../../../cxx/util/mtest.cxx:7:\n /usr/lib/x86_64-linux-gnu/openmpi/include/mpi.h:1758:20: note: declared here\n@@ -1003114,28 +1003162,28 @@\n ../../../cxx/util/mtest.cxx:848:50: error: 'NULL_COPY_FN' is not a member of 'MPI::Win'\n mem_keyval = MPI::Win::Create_keyval( MPI::Win::NULL_COPY_FN,\n ^~~~~~~~~~~~\n ../../../cxx/util/mtest.cxx:849:22: error: 'NULL_DELETE_FN' is not a member of 'MPI::Win'\n MPI::Win::NULL_DELETE_FN, 0 );\n ^~~~~~~~~~~~~~\n make[3]: *** [Makefile:416: mtest.o] Error 1\n-make[3]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/util'\n+make[3]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/util'\n make[2]: *** [Makefile:865: ../../cxx/util/mtest.o] Error 2\n-make[2]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/coll'\n+make[2]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/coll'\n \n \n icalltoallx\n \n ./cxx/coll\n fail\n- Failed to build icalltoallx; make[2]: Entering directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/coll'\n+ Failed to build icalltoallx; make[2]: Entering directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/coll'\n (cd ../../cxx/util && make mtest.o)\n-make[3]: Entering directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/util'\n+make[3]: Entering directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/util'\n depbase=`echo mtest.o | sed 's|[^/]*$|.deps/&|;s|\\.o$||'`;\\\n-/usr/bin/mpicxx.openmpi -DHAVE_CONFIG_H -I. -I../../../cxx/util -I../../include -I../../include -I../../../include -Wdate-time -D_FORTIFY_SOURCE=2 -g -O2 -ffile-prefix-map=/build/1st/mpi-testsuite-3.2+dfsg=. -fstack-protector-strong -Wformat -Werror=format-security -MT mtest.o -MD -MP -MF $depbase.Tpo -c -o mtest.o ../../../cxx/util/mtest.cxx &&\\\n+/usr/bin/mpicxx.openmpi -DHAVE_CONFIG_H -I. -I../../../cxx/util -I../../include -I../../include -I../../../include -Wdate-time -D_FORTIFY_SOURCE=2 -g -O2 -ffile-prefix-map=/build/mpi-testsuite-3.2+dfsg/2nd=. -fstack-protector-strong -Wformat -Werror=format-security -MT mtest.o -MD -MP -MF $depbase.Tpo -c -o mtest.o ../../../cxx/util/mtest.cxx &&\\\n mv -f $depbase.Tpo $depbase.Po\n ../../../cxx/util/mtest.cxx: In function 'void* MTestTypeContigInitRecv(MTestDatatype*)':\n ../../../cxx/util/mtest.cxx:189:42: warning: 'int MPI_Type_extent(MPI_Datatype, MPI_Aint*)' is deprecated: MPI_Type_extent is superseded by MPI_Type_get_extent in MPI-2.0 [-Wdeprecated-declarations]\n MPI_Type_extent( mtype->datatype, &size );\n ^\n In file included from ../../../cxx/util/mtest.cxx:7:\n /usr/lib/x86_64-linux-gnu/openmpi/include/mpi.h:1758:20: note: declared here\n@@ -1003167,28 +1003215,28 @@\n ../../../cxx/util/mtest.cxx:848:50: error: 'NULL_COPY_FN' is not a member of 'MPI::Win'\n mem_keyval = MPI::Win::Create_keyval( MPI::Win::NULL_COPY_FN,\n ^~~~~~~~~~~~\n ../../../cxx/util/mtest.cxx:849:22: error: 'NULL_DELETE_FN' is not a member of 'MPI::Win'\n MPI::Win::NULL_DELETE_FN, 0 );\n ^~~~~~~~~~~~~~\n make[3]: *** [Makefile:416: mtest.o] Error 1\n-make[3]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/util'\n+make[3]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/util'\n make[2]: *** [Makefile:865: ../../cxx/util/mtest.o] Error 2\n-make[2]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/coll'\n+make[2]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/coll'\n \n \n reduceboolx\n \n ./cxx/coll\n fail\n- Failed to build reduceboolx; make[2]: Entering directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/coll'\n+ Failed to build reduceboolx; make[2]: Entering directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/coll'\n (cd ../../cxx/util && make mtest.o)\n-make[3]: Entering directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/util'\n+make[3]: Entering directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/util'\n depbase=`echo mtest.o | sed 's|[^/]*$|.deps/&|;s|\\.o$||'`;\\\n-/usr/bin/mpicxx.openmpi -DHAVE_CONFIG_H -I. -I../../../cxx/util -I../../include -I../../include -I../../../include -Wdate-time -D_FORTIFY_SOURCE=2 -g -O2 -ffile-prefix-map=/build/1st/mpi-testsuite-3.2+dfsg=. -fstack-protector-strong -Wformat -Werror=format-security -MT mtest.o -MD -MP -MF $depbase.Tpo -c -o mtest.o ../../../cxx/util/mtest.cxx &&\\\n+/usr/bin/mpicxx.openmpi -DHAVE_CONFIG_H -I. -I../../../cxx/util -I../../include -I../../include -I../../../include -Wdate-time -D_FORTIFY_SOURCE=2 -g -O2 -ffile-prefix-map=/build/mpi-testsuite-3.2+dfsg/2nd=. -fstack-protector-strong -Wformat -Werror=format-security -MT mtest.o -MD -MP -MF $depbase.Tpo -c -o mtest.o ../../../cxx/util/mtest.cxx &&\\\n mv -f $depbase.Tpo $depbase.Po\n ../../../cxx/util/mtest.cxx: In function 'void* MTestTypeContigInitRecv(MTestDatatype*)':\n ../../../cxx/util/mtest.cxx:189:42: warning: 'int MPI_Type_extent(MPI_Datatype, MPI_Aint*)' is deprecated: MPI_Type_extent is superseded by MPI_Type_get_extent in MPI-2.0 [-Wdeprecated-declarations]\n MPI_Type_extent( mtype->datatype, &size );\n ^\n In file included from ../../../cxx/util/mtest.cxx:7:\n /usr/lib/x86_64-linux-gnu/openmpi/include/mpi.h:1758:20: note: declared here\n@@ -1003220,49 +1003268,49 @@\n ../../../cxx/util/mtest.cxx:848:50: error: 'NULL_COPY_FN' is not a member of 'MPI::Win'\n mem_keyval = MPI::Win::Create_keyval( MPI::Win::NULL_COPY_FN,\n ^~~~~~~~~~~~\n ../../../cxx/util/mtest.cxx:849:22: error: 'NULL_DELETE_FN' is not a member of 'MPI::Win'\n MPI::Win::NULL_DELETE_FN, 0 );\n ^~~~~~~~~~~~~~\n make[3]: *** [Makefile:416: mtest.o] Error 1\n-make[3]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/util'\n+make[3]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/util'\n make[2]: *** [Makefile:865: ../../cxx/util/mtest.o] Error 2\n-make[2]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/coll'\n+make[2]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/coll'\n \n \n redscatblk\n \n ./cxx/coll\n fail\n- Failed to build redscatblk; make[2]: Entering directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/coll'\n+ Failed to build redscatblk; make[2]: Entering directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/coll'\n depbase=`echo redscatblk.o | sed 's|[^/]*$|.deps/&|;s|\\.o$||'`;\\\n-/usr/bin/mpicxx.openmpi -DHAVE_CONFIG_H -I. -I../../../cxx/coll -I../../include -I../../include -I../../../include -Wdate-time -D_FORTIFY_SOURCE=2 -g -O2 -ffile-prefix-map=/build/1st/mpi-testsuite-3.2+dfsg=. -fstack-protector-strong -Wformat -Werror=format-security -MT redscatblk.o -MD -MP -MF $depbase.Tpo -c -o redscatblk.o ../../../cxx/coll/redscatblk.cxx &&\\\n+/usr/bin/mpicxx.openmpi -DHAVE_CONFIG_H -I. -I../../../cxx/coll -I../../include -I../../include -I../../../include -Wdate-time -D_FORTIFY_SOURCE=2 -g -O2 -ffile-prefix-map=/build/mpi-testsuite-3.2+dfsg/2nd=. -fstack-protector-strong -Wformat -Werror=format-security -MT redscatblk.o -MD -MP -MF $depbase.Tpo -c -o redscatblk.o ../../../cxx/coll/redscatblk.cxx &&\\\n mv -f $depbase.Tpo $depbase.Po\n ../../../cxx/coll/redscatblk.cxx: In function 'int main(int, char**)':\n ../../../cxx/coll/redscatblk.cxx:42:21: error: 'class MPI::Intracomm' has no member named 'Reduce_scatter_block'; did you mean 'Reduce_scatter'?\n MPI::COMM_WORLD.Reduce_scatter_block(sendbuf, recvbuf, mycount, MPI::INT, MPI::SUM);\n ^~~~~~~~~~~~~~~~~~~~\n Reduce_scatter\n ../../../cxx/coll/redscatblk.cxx:54:21: error: 'class MPI::Intracomm' has no member named 'Reduce_scatter_block'; did you mean 'Reduce_scatter'?\n MPI::COMM_WORLD.Reduce_scatter_block(MPI_IN_PLACE, sendbuf, mycount, MPI::INT, MPI::SUM);\n ^~~~~~~~~~~~~~~~~~~~\n Reduce_scatter\n make[2]: *** [Makefile:645: redscatblk.o] Error 1\n-make[2]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/coll'\n+make[2]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/coll'\n \n \n commcallx\n \n ./cxx/errhan\n fail\n- Failed to build commcallx; make[2]: Entering directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/errhan'\n+ Failed to build commcallx; make[2]: Entering directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/errhan'\n (cd ../../cxx/util && make mtest.o)\n-make[3]: Entering directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/util'\n+make[3]: Entering directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/util'\n depbase=`echo mtest.o | sed 's|[^/]*$|.deps/&|;s|\\.o$||'`;\\\n-/usr/bin/mpicxx.openmpi -DHAVE_CONFIG_H -I. -I../../../cxx/util -I../../include -I../../include -I../../../include -Wdate-time -D_FORTIFY_SOURCE=2 -g -O2 -ffile-prefix-map=/build/1st/mpi-testsuite-3.2+dfsg=. -fstack-protector-strong -Wformat -Werror=format-security -MT mtest.o -MD -MP -MF $depbase.Tpo -c -o mtest.o ../../../cxx/util/mtest.cxx &&\\\n+/usr/bin/mpicxx.openmpi -DHAVE_CONFIG_H -I. -I../../../cxx/util -I../../include -I../../include -I../../../include -Wdate-time -D_FORTIFY_SOURCE=2 -g -O2 -ffile-prefix-map=/build/mpi-testsuite-3.2+dfsg/2nd=. -fstack-protector-strong -Wformat -Werror=format-security -MT mtest.o -MD -MP -MF $depbase.Tpo -c -o mtest.o ../../../cxx/util/mtest.cxx &&\\\n mv -f $depbase.Tpo $depbase.Po\n ../../../cxx/util/mtest.cxx: In function 'void* MTestTypeContigInitRecv(MTestDatatype*)':\n ../../../cxx/util/mtest.cxx:189:42: warning: 'int MPI_Type_extent(MPI_Datatype, MPI_Aint*)' is deprecated: MPI_Type_extent is superseded by MPI_Type_get_extent in MPI-2.0 [-Wdeprecated-declarations]\n MPI_Type_extent( mtype->datatype, &size );\n ^\n In file included from ../../../cxx/util/mtest.cxx:7:\n /usr/lib/x86_64-linux-gnu/openmpi/include/mpi.h:1758:20: note: declared here\n@@ -1003294,28 +1003342,28 @@\n ../../../cxx/util/mtest.cxx:848:50: error: 'NULL_COPY_FN' is not a member of 'MPI::Win'\n mem_keyval = MPI::Win::Create_keyval( MPI::Win::NULL_COPY_FN,\n ^~~~~~~~~~~~\n ../../../cxx/util/mtest.cxx:849:22: error: 'NULL_DELETE_FN' is not a member of 'MPI::Win'\n MPI::Win::NULL_DELETE_FN, 0 );\n ^~~~~~~~~~~~~~\n make[3]: *** [Makefile:416: mtest.o] Error 1\n-make[3]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/util'\n+make[3]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/util'\n make[2]: *** [Makefile:640: ../../cxx/util/mtest.o] Error 2\n-make[2]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/errhan'\n+make[2]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/errhan'\n \n \n baseenv\n \n ./cxx/init\n fail\n- Failed to build baseenv; make[2]: Entering directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/init'\n+ Failed to build baseenv; make[2]: Entering directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/init'\n (cd ../../cxx/util && make mtest.o)\n-make[3]: Entering directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/util'\n+make[3]: Entering directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/util'\n depbase=`echo mtest.o | sed 's|[^/]*$|.deps/&|;s|\\.o$||'`;\\\n-/usr/bin/mpicxx.openmpi -DHAVE_CONFIG_H -I. -I../../../cxx/util -I../../include -I../../include -I../../../include -Wdate-time -D_FORTIFY_SOURCE=2 -g -O2 -ffile-prefix-map=/build/1st/mpi-testsuite-3.2+dfsg=. -fstack-protector-strong -Wformat -Werror=format-security -MT mtest.o -MD -MP -MF $depbase.Tpo -c -o mtest.o ../../../cxx/util/mtest.cxx &&\\\n+/usr/bin/mpicxx.openmpi -DHAVE_CONFIG_H -I. -I../../../cxx/util -I../../include -I../../include -I../../../include -Wdate-time -D_FORTIFY_SOURCE=2 -g -O2 -ffile-prefix-map=/build/mpi-testsuite-3.2+dfsg/2nd=. -fstack-protector-strong -Wformat -Werror=format-security -MT mtest.o -MD -MP -MF $depbase.Tpo -c -o mtest.o ../../../cxx/util/mtest.cxx &&\\\n mv -f $depbase.Tpo $depbase.Po\n ../../../cxx/util/mtest.cxx: In function 'void* MTestTypeContigInitRecv(MTestDatatype*)':\n ../../../cxx/util/mtest.cxx:189:42: warning: 'int MPI_Type_extent(MPI_Datatype, MPI_Aint*)' is deprecated: MPI_Type_extent is superseded by MPI_Type_get_extent in MPI-2.0 [-Wdeprecated-declarations]\n MPI_Type_extent( mtype->datatype, &size );\n ^\n In file included from ../../../cxx/util/mtest.cxx:7:\n /usr/lib/x86_64-linux-gnu/openmpi/include/mpi.h:1758:20: note: declared here\n@@ -1003347,28 +1003395,28 @@\n ../../../cxx/util/mtest.cxx:848:50: error: 'NULL_COPY_FN' is not a member of 'MPI::Win'\n mem_keyval = MPI::Win::Create_keyval( MPI::Win::NULL_COPY_FN,\n ^~~~~~~~~~~~\n ../../../cxx/util/mtest.cxx:849:22: error: 'NULL_DELETE_FN' is not a member of 'MPI::Win'\n MPI::Win::NULL_DELETE_FN, 0 );\n ^~~~~~~~~~~~~~\n make[3]: *** [Makefile:416: mtest.o] Error 1\n-make[3]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/util'\n+make[3]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/util'\n make[2]: *** [Makefile:663: ../../cxx/util/mtest.o] Error 2\n-make[2]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/init'\n+make[2]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/init'\n \n \n initstatx\n \n ./cxx/init\n fail\n- Failed to build initstatx; make[2]: Entering directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/init'\n+ Failed to build initstatx; make[2]: Entering directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/init'\n (cd ../../cxx/util && make mtest.o)\n-make[3]: Entering directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/util'\n+make[3]: Entering directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/util'\n depbase=`echo mtest.o | sed 's|[^/]*$|.deps/&|;s|\\.o$||'`;\\\n-/usr/bin/mpicxx.openmpi -DHAVE_CONFIG_H -I. -I../../../cxx/util -I../../include -I../../include -I../../../include -Wdate-time -D_FORTIFY_SOURCE=2 -g -O2 -ffile-prefix-map=/build/1st/mpi-testsuite-3.2+dfsg=. -fstack-protector-strong -Wformat -Werror=format-security -MT mtest.o -MD -MP -MF $depbase.Tpo -c -o mtest.o ../../../cxx/util/mtest.cxx &&\\\n+/usr/bin/mpicxx.openmpi -DHAVE_CONFIG_H -I. -I../../../cxx/util -I../../include -I../../include -I../../../include -Wdate-time -D_FORTIFY_SOURCE=2 -g -O2 -ffile-prefix-map=/build/mpi-testsuite-3.2+dfsg/2nd=. -fstack-protector-strong -Wformat -Werror=format-security -MT mtest.o -MD -MP -MF $depbase.Tpo -c -o mtest.o ../../../cxx/util/mtest.cxx &&\\\n mv -f $depbase.Tpo $depbase.Po\n ../../../cxx/util/mtest.cxx: In function 'void* MTestTypeContigInitRecv(MTestDatatype*)':\n ../../../cxx/util/mtest.cxx:189:42: warning: 'int MPI_Type_extent(MPI_Datatype, MPI_Aint*)' is deprecated: MPI_Type_extent is superseded by MPI_Type_get_extent in MPI-2.0 [-Wdeprecated-declarations]\n MPI_Type_extent( mtype->datatype, &size );\n ^\n In file included from ../../../cxx/util/mtest.cxx:7:\n /usr/lib/x86_64-linux-gnu/openmpi/include/mpi.h:1758:20: note: declared here\n@@ -1003400,28 +1003448,28 @@\n ../../../cxx/util/mtest.cxx:848:50: error: 'NULL_COPY_FN' is not a member of 'MPI::Win'\n mem_keyval = MPI::Win::Create_keyval( MPI::Win::NULL_COPY_FN,\n ^~~~~~~~~~~~\n ../../../cxx/util/mtest.cxx:849:22: error: 'NULL_DELETE_FN' is not a member of 'MPI::Win'\n MPI::Win::NULL_DELETE_FN, 0 );\n ^~~~~~~~~~~~~~\n make[3]: *** [Makefile:416: mtest.o] Error 1\n-make[3]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/util'\n+make[3]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/util'\n make[2]: *** [Makefile:663: ../../cxx/util/mtest.o] Error 2\n-make[2]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/init'\n+make[2]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/init'\n \n \n initstat2x\n \n ./cxx/init\n fail\n- Failed to build initstat2x; make[2]: Entering directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/init'\n+ Failed to build initstat2x; make[2]: Entering directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/init'\n (cd ../../cxx/util && make mtest.o)\n-make[3]: Entering directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/util'\n+make[3]: Entering directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/util'\n depbase=`echo mtest.o | sed 's|[^/]*$|.deps/&|;s|\\.o$||'`;\\\n-/usr/bin/mpicxx.openmpi -DHAVE_CONFIG_H -I. -I../../../cxx/util -I../../include -I../../include -I../../../include -Wdate-time -D_FORTIFY_SOURCE=2 -g -O2 -ffile-prefix-map=/build/1st/mpi-testsuite-3.2+dfsg=. -fstack-protector-strong -Wformat -Werror=format-security -MT mtest.o -MD -MP -MF $depbase.Tpo -c -o mtest.o ../../../cxx/util/mtest.cxx &&\\\n+/usr/bin/mpicxx.openmpi -DHAVE_CONFIG_H -I. -I../../../cxx/util -I../../include -I../../include -I../../../include -Wdate-time -D_FORTIFY_SOURCE=2 -g -O2 -ffile-prefix-map=/build/mpi-testsuite-3.2+dfsg/2nd=. -fstack-protector-strong -Wformat -Werror=format-security -MT mtest.o -MD -MP -MF $depbase.Tpo -c -o mtest.o ../../../cxx/util/mtest.cxx &&\\\n mv -f $depbase.Tpo $depbase.Po\n ../../../cxx/util/mtest.cxx: In function 'void* MTestTypeContigInitRecv(MTestDatatype*)':\n ../../../cxx/util/mtest.cxx:189:42: warning: 'int MPI_Type_extent(MPI_Datatype, MPI_Aint*)' is deprecated: MPI_Type_extent is superseded by MPI_Type_get_extent in MPI-2.0 [-Wdeprecated-declarations]\n MPI_Type_extent( mtype->datatype, &size );\n ^\n In file included from ../../../cxx/util/mtest.cxx:7:\n /usr/lib/x86_64-linux-gnu/openmpi/include/mpi.h:1758:20: note: declared here\n@@ -1003453,28 +1003501,28 @@\n ../../../cxx/util/mtest.cxx:848:50: error: 'NULL_COPY_FN' is not a member of 'MPI::Win'\n mem_keyval = MPI::Win::Create_keyval( MPI::Win::NULL_COPY_FN,\n ^~~~~~~~~~~~\n ../../../cxx/util/mtest.cxx:849:22: error: 'NULL_DELETE_FN' is not a member of 'MPI::Win'\n MPI::Win::NULL_DELETE_FN, 0 );\n ^~~~~~~~~~~~~~\n make[3]: *** [Makefile:416: mtest.o] Error 1\n-make[3]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/util'\n+make[3]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/util'\n make[2]: *** [Makefile:663: ../../cxx/util/mtest.o] Error 2\n-make[2]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/init'\n+make[2]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/init'\n \n \n infodupx\n \n ./cxx/info\n fail\n- Failed to build infodupx; make[2]: Entering directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/info'\n+ Failed to build infodupx; make[2]: Entering directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/info'\n (cd ../../cxx/util && make mtest.o)\n-make[3]: Entering directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/util'\n+make[3]: Entering directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/util'\n depbase=`echo mtest.o | sed 's|[^/]*$|.deps/&|;s|\\.o$||'`;\\\n-/usr/bin/mpicxx.openmpi -DHAVE_CONFIG_H -I. -I../../../cxx/util -I../../include -I../../include -I../../../include -Wdate-time -D_FORTIFY_SOURCE=2 -g -O2 -ffile-prefix-map=/build/1st/mpi-testsuite-3.2+dfsg=. -fstack-protector-strong -Wformat -Werror=format-security -MT mtest.o -MD -MP -MF $depbase.Tpo -c -o mtest.o ../../../cxx/util/mtest.cxx &&\\\n+/usr/bin/mpicxx.openmpi -DHAVE_CONFIG_H -I. -I../../../cxx/util -I../../include -I../../include -I../../../include -Wdate-time -D_FORTIFY_SOURCE=2 -g -O2 -ffile-prefix-map=/build/mpi-testsuite-3.2+dfsg/2nd=. -fstack-protector-strong -Wformat -Werror=format-security -MT mtest.o -MD -MP -MF $depbase.Tpo -c -o mtest.o ../../../cxx/util/mtest.cxx &&\\\n mv -f $depbase.Tpo $depbase.Po\n ../../../cxx/util/mtest.cxx: In function 'void* MTestTypeContigInitRecv(MTestDatatype*)':\n ../../../cxx/util/mtest.cxx:189:42: warning: 'int MPI_Type_extent(MPI_Datatype, MPI_Aint*)' is deprecated: MPI_Type_extent is superseded by MPI_Type_get_extent in MPI-2.0 [-Wdeprecated-declarations]\n MPI_Type_extent( mtype->datatype, &size );\n ^\n In file included from ../../../cxx/util/mtest.cxx:7:\n /usr/lib/x86_64-linux-gnu/openmpi/include/mpi.h:1758:20: note: declared here\n@@ -1003506,28 +1003554,28 @@\n ../../../cxx/util/mtest.cxx:848:50: error: 'NULL_COPY_FN' is not a member of 'MPI::Win'\n mem_keyval = MPI::Win::Create_keyval( MPI::Win::NULL_COPY_FN,\n ^~~~~~~~~~~~\n ../../../cxx/util/mtest.cxx:849:22: error: 'NULL_DELETE_FN' is not a member of 'MPI::Win'\n MPI::Win::NULL_DELETE_FN, 0 );\n ^~~~~~~~~~~~~~\n make[3]: *** [Makefile:416: mtest.o] Error 1\n-make[3]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/util'\n+make[3]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/util'\n make[2]: *** [Makefile:673: ../../cxx/util/mtest.o] Error 2\n-make[2]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/info'\n+make[2]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/info'\n \n \n infodelx\n \n ./cxx/info\n fail\n- Failed to build infodelx; make[2]: Entering directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/info'\n+ Failed to build infodelx; make[2]: Entering directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/info'\n (cd ../../cxx/util && make mtest.o)\n-make[3]: Entering directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/util'\n+make[3]: Entering directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/util'\n depbase=`echo mtest.o | sed 's|[^/]*$|.deps/&|;s|\\.o$||'`;\\\n-/usr/bin/mpicxx.openmpi -DHAVE_CONFIG_H -I. -I../../../cxx/util -I../../include -I../../include -I../../../include -Wdate-time -D_FORTIFY_SOURCE=2 -g -O2 -ffile-prefix-map=/build/1st/mpi-testsuite-3.2+dfsg=. -fstack-protector-strong -Wformat -Werror=format-security -MT mtest.o -MD -MP -MF $depbase.Tpo -c -o mtest.o ../../../cxx/util/mtest.cxx &&\\\n+/usr/bin/mpicxx.openmpi -DHAVE_CONFIG_H -I. -I../../../cxx/util -I../../include -I../../include -I../../../include -Wdate-time -D_FORTIFY_SOURCE=2 -g -O2 -ffile-prefix-map=/build/mpi-testsuite-3.2+dfsg/2nd=. -fstack-protector-strong -Wformat -Werror=format-security -MT mtest.o -MD -MP -MF $depbase.Tpo -c -o mtest.o ../../../cxx/util/mtest.cxx &&\\\n mv -f $depbase.Tpo $depbase.Po\n ../../../cxx/util/mtest.cxx: In function 'void* MTestTypeContigInitRecv(MTestDatatype*)':\n ../../../cxx/util/mtest.cxx:189:42: warning: 'int MPI_Type_extent(MPI_Datatype, MPI_Aint*)' is deprecated: MPI_Type_extent is superseded by MPI_Type_get_extent in MPI-2.0 [-Wdeprecated-declarations]\n MPI_Type_extent( mtype->datatype, &size );\n ^\n In file included from ../../../cxx/util/mtest.cxx:7:\n /usr/lib/x86_64-linux-gnu/openmpi/include/mpi.h:1758:20: note: declared here\n@@ -1003559,28 +1003607,28 @@\n ../../../cxx/util/mtest.cxx:848:50: error: 'NULL_COPY_FN' is not a member of 'MPI::Win'\n mem_keyval = MPI::Win::Create_keyval( MPI::Win::NULL_COPY_FN,\n ^~~~~~~~~~~~\n ../../../cxx/util/mtest.cxx:849:22: error: 'NULL_DELETE_FN' is not a member of 'MPI::Win'\n MPI::Win::NULL_DELETE_FN, 0 );\n ^~~~~~~~~~~~~~\n make[3]: *** [Makefile:416: mtest.o] Error 1\n-make[3]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/util'\n+make[3]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/util'\n make[2]: *** [Makefile:673: ../../cxx/util/mtest.o] Error 2\n-make[2]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/info'\n+make[2]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/info'\n \n \n infovallenx\n \n ./cxx/info\n fail\n- Failed to build infovallenx; make[2]: Entering directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/info'\n+ Failed to build infovallenx; make[2]: Entering directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/info'\n (cd ../../cxx/util && make mtest.o)\n-make[3]: Entering directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/util'\n+make[3]: Entering directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/util'\n depbase=`echo mtest.o | sed 's|[^/]*$|.deps/&|;s|\\.o$||'`;\\\n-/usr/bin/mpicxx.openmpi -DHAVE_CONFIG_H -I. -I../../../cxx/util -I../../include -I../../include -I../../../include -Wdate-time -D_FORTIFY_SOURCE=2 -g -O2 -ffile-prefix-map=/build/1st/mpi-testsuite-3.2+dfsg=. -fstack-protector-strong -Wformat -Werror=format-security -MT mtest.o -MD -MP -MF $depbase.Tpo -c -o mtest.o ../../../cxx/util/mtest.cxx &&\\\n+/usr/bin/mpicxx.openmpi -DHAVE_CONFIG_H -I. -I../../../cxx/util -I../../include -I../../include -I../../../include -Wdate-time -D_FORTIFY_SOURCE=2 -g -O2 -ffile-prefix-map=/build/mpi-testsuite-3.2+dfsg/2nd=. -fstack-protector-strong -Wformat -Werror=format-security -MT mtest.o -MD -MP -MF $depbase.Tpo -c -o mtest.o ../../../cxx/util/mtest.cxx &&\\\n mv -f $depbase.Tpo $depbase.Po\n ../../../cxx/util/mtest.cxx: In function 'void* MTestTypeContigInitRecv(MTestDatatype*)':\n ../../../cxx/util/mtest.cxx:189:42: warning: 'int MPI_Type_extent(MPI_Datatype, MPI_Aint*)' is deprecated: MPI_Type_extent is superseded by MPI_Type_get_extent in MPI-2.0 [-Wdeprecated-declarations]\n MPI_Type_extent( mtype->datatype, &size );\n ^\n In file included from ../../../cxx/util/mtest.cxx:7:\n /usr/lib/x86_64-linux-gnu/openmpi/include/mpi.h:1758:20: note: declared here\n@@ -1003612,28 +1003660,28 @@\n ../../../cxx/util/mtest.cxx:848:50: error: 'NULL_COPY_FN' is not a member of 'MPI::Win'\n mem_keyval = MPI::Win::Create_keyval( MPI::Win::NULL_COPY_FN,\n ^~~~~~~~~~~~\n ../../../cxx/util/mtest.cxx:849:22: error: 'NULL_DELETE_FN' is not a member of 'MPI::Win'\n MPI::Win::NULL_DELETE_FN, 0 );\n ^~~~~~~~~~~~~~\n make[3]: *** [Makefile:416: mtest.o] Error 1\n-make[3]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/util'\n+make[3]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/util'\n make[2]: *** [Makefile:673: ../../cxx/util/mtest.o] Error 2\n-make[2]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/info'\n+make[2]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/info'\n \n \n infoorderx\n \n ./cxx/info\n fail\n- Failed to build infoorderx; make[2]: Entering directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/info'\n+ Failed to build infoorderx; make[2]: Entering directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/info'\n (cd ../../cxx/util && make mtest.o)\n-make[3]: Entering directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/util'\n+make[3]: Entering directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/util'\n depbase=`echo mtest.o | sed 's|[^/]*$|.deps/&|;s|\\.o$||'`;\\\n-/usr/bin/mpicxx.openmpi -DHAVE_CONFIG_H -I. -I../../../cxx/util -I../../include -I../../include -I../../../include -Wdate-time -D_FORTIFY_SOURCE=2 -g -O2 -ffile-prefix-map=/build/1st/mpi-testsuite-3.2+dfsg=. -fstack-protector-strong -Wformat -Werror=format-security -MT mtest.o -MD -MP -MF $depbase.Tpo -c -o mtest.o ../../../cxx/util/mtest.cxx &&\\\n+/usr/bin/mpicxx.openmpi -DHAVE_CONFIG_H -I. -I../../../cxx/util -I../../include -I../../include -I../../../include -Wdate-time -D_FORTIFY_SOURCE=2 -g -O2 -ffile-prefix-map=/build/mpi-testsuite-3.2+dfsg/2nd=. -fstack-protector-strong -Wformat -Werror=format-security -MT mtest.o -MD -MP -MF $depbase.Tpo -c -o mtest.o ../../../cxx/util/mtest.cxx &&\\\n mv -f $depbase.Tpo $depbase.Po\n ../../../cxx/util/mtest.cxx: In function 'void* MTestTypeContigInitRecv(MTestDatatype*)':\n ../../../cxx/util/mtest.cxx:189:42: warning: 'int MPI_Type_extent(MPI_Datatype, MPI_Aint*)' is deprecated: MPI_Type_extent is superseded by MPI_Type_get_extent in MPI-2.0 [-Wdeprecated-declarations]\n MPI_Type_extent( mtype->datatype, &size );\n ^\n In file included from ../../../cxx/util/mtest.cxx:7:\n /usr/lib/x86_64-linux-gnu/openmpi/include/mpi.h:1758:20: note: declared here\n@@ -1003665,28 +1003713,28 @@\n ../../../cxx/util/mtest.cxx:848:50: error: 'NULL_COPY_FN' is not a member of 'MPI::Win'\n mem_keyval = MPI::Win::Create_keyval( MPI::Win::NULL_COPY_FN,\n ^~~~~~~~~~~~\n ../../../cxx/util/mtest.cxx:849:22: error: 'NULL_DELETE_FN' is not a member of 'MPI::Win'\n MPI::Win::NULL_DELETE_FN, 0 );\n ^~~~~~~~~~~~~~\n make[3]: *** [Makefile:416: mtest.o] Error 1\n-make[3]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/util'\n+make[3]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/util'\n make[2]: *** [Makefile:673: ../../cxx/util/mtest.o] Error 2\n-make[2]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/info'\n+make[2]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/info'\n \n \n packsizex\n \n ./cxx/datatype\n fail\n- Failed to build packsizex; make[2]: Entering directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/datatype'\n+ Failed to build packsizex; make[2]: Entering directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/datatype'\n (cd ../../cxx/util && make mtest.o)\n-make[3]: Entering directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/util'\n+make[3]: Entering directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/util'\n depbase=`echo mtest.o | sed 's|[^/]*$|.deps/&|;s|\\.o$||'`;\\\n-/usr/bin/mpicxx.openmpi -DHAVE_CONFIG_H -I. -I../../../cxx/util -I../../include -I../../include -I../../../include -Wdate-time -D_FORTIFY_SOURCE=2 -g -O2 -ffile-prefix-map=/build/1st/mpi-testsuite-3.2+dfsg=. -fstack-protector-strong -Wformat -Werror=format-security -MT mtest.o -MD -MP -MF $depbase.Tpo -c -o mtest.o ../../../cxx/util/mtest.cxx &&\\\n+/usr/bin/mpicxx.openmpi -DHAVE_CONFIG_H -I. -I../../../cxx/util -I../../include -I../../include -I../../../include -Wdate-time -D_FORTIFY_SOURCE=2 -g -O2 -ffile-prefix-map=/build/mpi-testsuite-3.2+dfsg/2nd=. -fstack-protector-strong -Wformat -Werror=format-security -MT mtest.o -MD -MP -MF $depbase.Tpo -c -o mtest.o ../../../cxx/util/mtest.cxx &&\\\n mv -f $depbase.Tpo $depbase.Po\n ../../../cxx/util/mtest.cxx: In function 'void* MTestTypeContigInitRecv(MTestDatatype*)':\n ../../../cxx/util/mtest.cxx:189:42: warning: 'int MPI_Type_extent(MPI_Datatype, MPI_Aint*)' is deprecated: MPI_Type_extent is superseded by MPI_Type_get_extent in MPI-2.0 [-Wdeprecated-declarations]\n MPI_Type_extent( mtype->datatype, &size );\n ^\n In file included from ../../../cxx/util/mtest.cxx:7:\n /usr/lib/x86_64-linux-gnu/openmpi/include/mpi.h:1758:20: note: declared here\n@@ -1003718,28 +1003766,28 @@\n ../../../cxx/util/mtest.cxx:848:50: error: 'NULL_COPY_FN' is not a member of 'MPI::Win'\n mem_keyval = MPI::Win::Create_keyval( MPI::Win::NULL_COPY_FN,\n ^~~~~~~~~~~~\n ../../../cxx/util/mtest.cxx:849:22: error: 'NULL_DELETE_FN' is not a member of 'MPI::Win'\n MPI::Win::NULL_DELETE_FN, 0 );\n ^~~~~~~~~~~~~~\n make[3]: *** [Makefile:416: mtest.o] Error 1\n-make[3]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/util'\n+make[3]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/util'\n make[2]: *** [Makefile:673: ../../cxx/util/mtest.o] Error 2\n-make[2]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/datatype'\n+make[2]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/datatype'\n \n \n typecntsx\n \n ./cxx/datatype\n fail\n- Failed to build typecntsx; make[2]: Entering directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/datatype'\n+ Failed to build typecntsx; make[2]: Entering directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/datatype'\n (cd ../../cxx/util && make mtest.o)\n-make[3]: Entering directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/util'\n+make[3]: Entering directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/util'\n depbase=`echo mtest.o | sed 's|[^/]*$|.deps/&|;s|\\.o$||'`;\\\n-/usr/bin/mpicxx.openmpi -DHAVE_CONFIG_H -I. -I../../../cxx/util -I../../include -I../../include -I../../../include -Wdate-time -D_FORTIFY_SOURCE=2 -g -O2 -ffile-prefix-map=/build/1st/mpi-testsuite-3.2+dfsg=. -fstack-protector-strong -Wformat -Werror=format-security -MT mtest.o -MD -MP -MF $depbase.Tpo -c -o mtest.o ../../../cxx/util/mtest.cxx &&\\\n+/usr/bin/mpicxx.openmpi -DHAVE_CONFIG_H -I. -I../../../cxx/util -I../../include -I../../include -I../../../include -Wdate-time -D_FORTIFY_SOURCE=2 -g -O2 -ffile-prefix-map=/build/mpi-testsuite-3.2+dfsg/2nd=. -fstack-protector-strong -Wformat -Werror=format-security -MT mtest.o -MD -MP -MF $depbase.Tpo -c -o mtest.o ../../../cxx/util/mtest.cxx &&\\\n mv -f $depbase.Tpo $depbase.Po\n ../../../cxx/util/mtest.cxx: In function 'void* MTestTypeContigInitRecv(MTestDatatype*)':\n ../../../cxx/util/mtest.cxx:189:42: warning: 'int MPI_Type_extent(MPI_Datatype, MPI_Aint*)' is deprecated: MPI_Type_extent is superseded by MPI_Type_get_extent in MPI-2.0 [-Wdeprecated-declarations]\n MPI_Type_extent( mtype->datatype, &size );\n ^\n In file included from ../../../cxx/util/mtest.cxx:7:\n /usr/lib/x86_64-linux-gnu/openmpi/include/mpi.h:1758:20: note: declared here\n@@ -1003771,26 +1003819,26 @@\n ../../../cxx/util/mtest.cxx:848:50: error: 'NULL_COPY_FN' is not a member of 'MPI::Win'\n mem_keyval = MPI::Win::Create_keyval( MPI::Win::NULL_COPY_FN,\n ^~~~~~~~~~~~\n ../../../cxx/util/mtest.cxx:849:22: error: 'NULL_DELETE_FN' is not a member of 'MPI::Win'\n MPI::Win::NULL_DELETE_FN, 0 );\n ^~~~~~~~~~~~~~\n make[3]: *** [Makefile:416: mtest.o] Error 1\n-make[3]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/util'\n+make[3]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/util'\n make[2]: *** [Makefile:673: ../../cxx/util/mtest.o] Error 2\n-make[2]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/datatype'\n+make[2]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/datatype'\n \n \n typenamex\n \n ./cxx/datatype\n fail\n- Failed to build typenamex; make[2]: Entering directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/datatype'\n+ Failed to build typenamex; make[2]: Entering directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/datatype'\n depbase=`echo typenamex.o | sed 's|[^/]*$|.deps/&|;s|\\.o$||'`;\\\n-/usr/bin/mpicxx.openmpi -DHAVE_CONFIG_H -I. -I../../../cxx/datatype -I../../include -I../../include -I../../../include -Wdate-time -D_FORTIFY_SOURCE=2 -g -O2 -ffile-prefix-map=/build/1st/mpi-testsuite-3.2+dfsg=. -fstack-protector-strong -Wformat -Werror=format-security -MT typenamex.o -MD -MP -MF $depbase.Tpo -c -o typenamex.o ../../../cxx/datatype/typenamex.cxx &&\\\n+/usr/bin/mpicxx.openmpi -DHAVE_CONFIG_H -I. -I../../../cxx/datatype -I../../include -I../../include -I../../../include -Wdate-time -D_FORTIFY_SOURCE=2 -g -O2 -ffile-prefix-map=/build/mpi-testsuite-3.2+dfsg/2nd=. -fstack-protector-strong -Wformat -Werror=format-security -MT typenamex.o -MD -MP -MF $depbase.Tpo -c -o typenamex.o ../../../cxx/datatype/typenamex.cxx &&\\\n mv -f $depbase.Tpo $depbase.Po\n ../../../cxx/datatype/typenamex.cxx: In function 'int main(int, char**)':\n ../../../cxx/datatype/typenamex.cxx:78:30: error: passing 'const MPI::Datatype' as 'this' argument discards qualifiers [-fpermissive]\n MPI::INT.Set_name( "int" );\n ^\n In file included from /usr/lib/x86_64-linux-gnu/openmpi/include/openmpi/ompi/mpi/cxx/mpicxx.h:269,\n from /usr/lib/x86_64-linux-gnu/openmpi/include/mpi.h:2684,\n@@ -1003838,24 +1003886,24 @@\n { MPI::INTEGER16, "MPI_INTEGER16" },\n ^~~~~~~~~\n ../../../cxx/datatype/typenamex.cxx:158:9: note: suggested alternative: 'INTEGER1'\n { MPI::INTEGER16, "MPI_INTEGER16" },\n ^~~~~~~~~\n INTEGER1\n make[2]: *** [Makefile:453: typenamex.o] Error 1\n-make[2]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/datatype'\n+make[2]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/datatype'\n \n \n typemiscx\n \n ./cxx/datatype\n fail\n- Failed to build typemiscx; make[2]: Entering directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/datatype'\n+ Failed to build typemiscx; make[2]: Entering directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/datatype'\n depbase=`echo typemiscx.o | sed 's|[^/]*$|.deps/&|;s|\\.o$||'`;\\\n-/usr/bin/mpicxx.openmpi -DHAVE_CONFIG_H -I. -I../../../cxx/datatype -I../../include -I../../include -I../../../include -Wdate-time -D_FORTIFY_SOURCE=2 -g -O2 -ffile-prefix-map=/build/1st/mpi-testsuite-3.2+dfsg=. -fstack-protector-strong -Wformat -Werror=format-security -MT typemiscx.o -MD -MP -MF $depbase.Tpo -c -o typemiscx.o ../../../cxx/datatype/typemiscx.cxx &&\\\n+/usr/bin/mpicxx.openmpi -DHAVE_CONFIG_H -I. -I../../../cxx/datatype -I../../include -I../../include -I../../../include -Wdate-time -D_FORTIFY_SOURCE=2 -g -O2 -ffile-prefix-map=/build/mpi-testsuite-3.2+dfsg/2nd=. -fstack-protector-strong -Wformat -Werror=format-security -MT typemiscx.o -MD -MP -MF $depbase.Tpo -c -o typemiscx.o ../../../cxx/datatype/typemiscx.cxx &&\\\n mv -f $depbase.Tpo $depbase.Po\n ../../../cxx/datatype/typemiscx.cxx: In function 'int main(int, char**)':\n ../../../cxx/datatype/typemiscx.cxx:49:14: error: 'REAL16' is not a member of 'MPI'\n if (MPI::REAL16 != MPI::DATATYPE_NULL) {\n ^~~~~~\n ../../../cxx/datatype/typemiscx.cxx:49:14: note: suggested alternative: 'REAL8'\n if (MPI::REAL16 != MPI::DATATYPE_NULL) {\n@@ -1003921,26 +1003969,26 @@\n MPI::INTEGER8.Get_envelope( num_ints, num_adds, num_types, combiner );\n ^~~~~~~~\n ../../../cxx/datatype/typemiscx.cxx:99:7: note: suggested alternative: 'INTEGER4'\n MPI::INTEGER8.Get_envelope( num_ints, num_adds, num_types, combiner );\n ^~~~~~~~\n INTEGER4\n make[2]: *** [Makefile:453: typemiscx.o] Error 1\n-make[2]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/datatype'\n+make[2]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/datatype'\n \n \n iwriteatx\n \n ./cxx/io\n fail\n- Failed to build iwriteatx; make[2]: Entering directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/io'\n+ Failed to build iwriteatx; make[2]: Entering directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/io'\n (cd ../../cxx/util && make mtest.o)\n-make[3]: Entering directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/util'\n+make[3]: Entering directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/util'\n depbase=`echo mtest.o | sed 's|[^/]*$|.deps/&|;s|\\.o$||'`;\\\n-/usr/bin/mpicxx.openmpi -DHAVE_CONFIG_H -I. -I../../../cxx/util -I../../include -I../../include -I../../../include -Wdate-time -D_FORTIFY_SOURCE=2 -g -O2 -ffile-prefix-map=/build/1st/mpi-testsuite-3.2+dfsg=. -fstack-protector-strong -Wformat -Werror=format-security -MT mtest.o -MD -MP -MF $depbase.Tpo -c -o mtest.o ../../../cxx/util/mtest.cxx &&\\\n+/usr/bin/mpicxx.openmpi -DHAVE_CONFIG_H -I. -I../../../cxx/util -I../../include -I../../include -I../../../include -Wdate-time -D_FORTIFY_SOURCE=2 -g -O2 -ffile-prefix-map=/build/mpi-testsuite-3.2+dfsg/2nd=. -fstack-protector-strong -Wformat -Werror=format-security -MT mtest.o -MD -MP -MF $depbase.Tpo -c -o mtest.o ../../../cxx/util/mtest.cxx &&\\\n mv -f $depbase.Tpo $depbase.Po\n ../../../cxx/util/mtest.cxx: In function 'void* MTestTypeContigInitRecv(MTestDatatype*)':\n ../../../cxx/util/mtest.cxx:189:42: warning: 'int MPI_Type_extent(MPI_Datatype, MPI_Aint*)' is deprecated: MPI_Type_extent is superseded by MPI_Type_get_extent in MPI-2.0 [-Wdeprecated-declarations]\n MPI_Type_extent( mtype->datatype, &size );\n ^\n In file included from ../../../cxx/util/mtest.cxx:7:\n /usr/lib/x86_64-linux-gnu/openmpi/include/mpi.h:1758:20: note: declared here\n@@ -1003972,28 +1004020,28 @@\n ../../../cxx/util/mtest.cxx:848:50: error: 'NULL_COPY_FN' is not a member of 'MPI::Win'\n mem_keyval = MPI::Win::Create_keyval( MPI::Win::NULL_COPY_FN,\n ^~~~~~~~~~~~\n ../../../cxx/util/mtest.cxx:849:22: error: 'NULL_DELETE_FN' is not a member of 'MPI::Win'\n MPI::Win::NULL_DELETE_FN, 0 );\n ^~~~~~~~~~~~~~\n make[3]: *** [Makefile:416: mtest.o] Error 1\n-make[3]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/util'\n+make[3]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/util'\n make[2]: *** [Makefile:989: ../../cxx/util/mtest.o] Error 2\n-make[2]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/io'\n+make[2]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/io'\n \n \n iwritex\n \n ./cxx/io\n fail\n- Failed to build iwritex; make[2]: Entering directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/io'\n+ Failed to build iwritex; make[2]: Entering directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/io'\n (cd ../../cxx/util && make mtest.o)\n-make[3]: Entering directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/util'\n+make[3]: Entering directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/util'\n depbase=`echo mtest.o | sed 's|[^/]*$|.deps/&|;s|\\.o$||'`;\\\n-/usr/bin/mpicxx.openmpi -DHAVE_CONFIG_H -I. -I../../../cxx/util -I../../include -I../../include -I../../../include -Wdate-time -D_FORTIFY_SOURCE=2 -g -O2 -ffile-prefix-map=/build/1st/mpi-testsuite-3.2+dfsg=. -fstack-protector-strong -Wformat -Werror=format-security -MT mtest.o -MD -MP -MF $depbase.Tpo -c -o mtest.o ../../../cxx/util/mtest.cxx &&\\\n+/usr/bin/mpicxx.openmpi -DHAVE_CONFIG_H -I. -I../../../cxx/util -I../../include -I../../include -I../../../include -Wdate-time -D_FORTIFY_SOURCE=2 -g -O2 -ffile-prefix-map=/build/mpi-testsuite-3.2+dfsg/2nd=. -fstack-protector-strong -Wformat -Werror=format-security -MT mtest.o -MD -MP -MF $depbase.Tpo -c -o mtest.o ../../../cxx/util/mtest.cxx &&\\\n mv -f $depbase.Tpo $depbase.Po\n ../../../cxx/util/mtest.cxx: In function 'void* MTestTypeContigInitRecv(MTestDatatype*)':\n ../../../cxx/util/mtest.cxx:189:42: warning: 'int MPI_Type_extent(MPI_Datatype, MPI_Aint*)' is deprecated: MPI_Type_extent is superseded by MPI_Type_get_extent in MPI-2.0 [-Wdeprecated-declarations]\n MPI_Type_extent( mtype->datatype, &size );\n ^\n In file included from ../../../cxx/util/mtest.cxx:7:\n /usr/lib/x86_64-linux-gnu/openmpi/include/mpi.h:1758:20: note: declared here\n@@ -1004025,28 +1004073,28 @@\n ../../../cxx/util/mtest.cxx:848:50: error: 'NULL_COPY_FN' is not a member of 'MPI::Win'\n mem_keyval = MPI::Win::Create_keyval( MPI::Win::NULL_COPY_FN,\n ^~~~~~~~~~~~\n ../../../cxx/util/mtest.cxx:849:22: error: 'NULL_DELETE_FN' is not a member of 'MPI::Win'\n MPI::Win::NULL_DELETE_FN, 0 );\n ^~~~~~~~~~~~~~\n make[3]: *** [Makefile:416: mtest.o] Error 1\n-make[3]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/util'\n+make[3]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/util'\n make[2]: *** [Makefile:989: ../../cxx/util/mtest.o] Error 2\n-make[2]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/io'\n+make[2]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/io'\n \n \n iwriteshx\n \n ./cxx/io\n fail\n- Failed to build iwriteshx; make[2]: Entering directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/io'\n+ Failed to build iwriteshx; make[2]: Entering directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/io'\n (cd ../../cxx/util && make mtest.o)\n-make[3]: Entering directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/util'\n+make[3]: Entering directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/util'\n depbase=`echo mtest.o | sed 's|[^/]*$|.deps/&|;s|\\.o$||'`;\\\n-/usr/bin/mpicxx.openmpi -DHAVE_CONFIG_H -I. -I../../../cxx/util -I../../include -I../../include -I../../../include -Wdate-time -D_FORTIFY_SOURCE=2 -g -O2 -ffile-prefix-map=/build/1st/mpi-testsuite-3.2+dfsg=. -fstack-protector-strong -Wformat -Werror=format-security -MT mtest.o -MD -MP -MF $depbase.Tpo -c -o mtest.o ../../../cxx/util/mtest.cxx &&\\\n+/usr/bin/mpicxx.openmpi -DHAVE_CONFIG_H -I. -I../../../cxx/util -I../../include -I../../include -I../../../include -Wdate-time -D_FORTIFY_SOURCE=2 -g -O2 -ffile-prefix-map=/build/mpi-testsuite-3.2+dfsg/2nd=. -fstack-protector-strong -Wformat -Werror=format-security -MT mtest.o -MD -MP -MF $depbase.Tpo -c -o mtest.o ../../../cxx/util/mtest.cxx &&\\\n mv -f $depbase.Tpo $depbase.Po\n ../../../cxx/util/mtest.cxx: In function 'void* MTestTypeContigInitRecv(MTestDatatype*)':\n ../../../cxx/util/mtest.cxx:189:42: warning: 'int MPI_Type_extent(MPI_Datatype, MPI_Aint*)' is deprecated: MPI_Type_extent is superseded by MPI_Type_get_extent in MPI-2.0 [-Wdeprecated-declarations]\n MPI_Type_extent( mtype->datatype, &size );\n ^\n In file included from ../../../cxx/util/mtest.cxx:7:\n /usr/lib/x86_64-linux-gnu/openmpi/include/mpi.h:1758:20: note: declared here\n@@ -1004078,28 +1004126,28 @@\n ../../../cxx/util/mtest.cxx:848:50: error: 'NULL_COPY_FN' is not a member of 'MPI::Win'\n mem_keyval = MPI::Win::Create_keyval( MPI::Win::NULL_COPY_FN,\n ^~~~~~~~~~~~\n ../../../cxx/util/mtest.cxx:849:22: error: 'NULL_DELETE_FN' is not a member of 'MPI::Win'\n MPI::Win::NULL_DELETE_FN, 0 );\n ^~~~~~~~~~~~~~\n make[3]: *** [Makefile:416: mtest.o] Error 1\n-make[3]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/util'\n+make[3]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/util'\n make[2]: *** [Makefile:989: ../../cxx/util/mtest.o] Error 2\n-make[2]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/io'\n+make[2]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/io'\n \n \n writex\n \n ./cxx/io\n fail\n- Failed to build writex; make[2]: Entering directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/io'\n+ Failed to build writex; make[2]: Entering directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/io'\n (cd ../../cxx/util && make mtest.o)\n-make[3]: Entering directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/util'\n+make[3]: Entering directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/util'\n depbase=`echo mtest.o | sed 's|[^/]*$|.deps/&|;s|\\.o$||'`;\\\n-/usr/bin/mpicxx.openmpi -DHAVE_CONFIG_H -I. -I../../../cxx/util -I../../include -I../../include -I../../../include -Wdate-time -D_FORTIFY_SOURCE=2 -g -O2 -ffile-prefix-map=/build/1st/mpi-testsuite-3.2+dfsg=. -fstack-protector-strong -Wformat -Werror=format-security -MT mtest.o -MD -MP -MF $depbase.Tpo -c -o mtest.o ../../../cxx/util/mtest.cxx &&\\\n+/usr/bin/mpicxx.openmpi -DHAVE_CONFIG_H -I. -I../../../cxx/util -I../../include -I../../include -I../../../include -Wdate-time -D_FORTIFY_SOURCE=2 -g -O2 -ffile-prefix-map=/build/mpi-testsuite-3.2+dfsg/2nd=. -fstack-protector-strong -Wformat -Werror=format-security -MT mtest.o -MD -MP -MF $depbase.Tpo -c -o mtest.o ../../../cxx/util/mtest.cxx &&\\\n mv -f $depbase.Tpo $depbase.Po\n ../../../cxx/util/mtest.cxx: In function 'void* MTestTypeContigInitRecv(MTestDatatype*)':\n ../../../cxx/util/mtest.cxx:189:42: warning: 'int MPI_Type_extent(MPI_Datatype, MPI_Aint*)' is deprecated: MPI_Type_extent is superseded by MPI_Type_get_extent in MPI-2.0 [-Wdeprecated-declarations]\n MPI_Type_extent( mtype->datatype, &size );\n ^\n In file included from ../../../cxx/util/mtest.cxx:7:\n /usr/lib/x86_64-linux-gnu/openmpi/include/mpi.h:1758:20: note: declared here\n@@ -1004131,28 +1004179,28 @@\n ../../../cxx/util/mtest.cxx:848:50: error: 'NULL_COPY_FN' is not a member of 'MPI::Win'\n mem_keyval = MPI::Win::Create_keyval( MPI::Win::NULL_COPY_FN,\n ^~~~~~~~~~~~\n ../../../cxx/util/mtest.cxx:849:22: error: 'NULL_DELETE_FN' is not a member of 'MPI::Win'\n MPI::Win::NULL_DELETE_FN, 0 );\n ^~~~~~~~~~~~~~\n make[3]: *** [Makefile:416: mtest.o] Error 1\n-make[3]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/util'\n+make[3]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/util'\n make[2]: *** [Makefile:989: ../../cxx/util/mtest.o] Error 2\n-make[2]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/io'\n+make[2]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/io'\n \n \n writeatx\n \n ./cxx/io\n fail\n- Failed to build writeatx; make[2]: Entering directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/io'\n+ Failed to build writeatx; make[2]: Entering directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/io'\n (cd ../../cxx/util && make mtest.o)\n-make[3]: Entering directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/util'\n+make[3]: Entering directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/util'\n depbase=`echo mtest.o | sed 's|[^/]*$|.deps/&|;s|\\.o$||'`;\\\n-/usr/bin/mpicxx.openmpi -DHAVE_CONFIG_H -I. -I../../../cxx/util -I../../include -I../../include -I../../../include -Wdate-time -D_FORTIFY_SOURCE=2 -g -O2 -ffile-prefix-map=/build/1st/mpi-testsuite-3.2+dfsg=. -fstack-protector-strong -Wformat -Werror=format-security -MT mtest.o -MD -MP -MF $depbase.Tpo -c -o mtest.o ../../../cxx/util/mtest.cxx &&\\\n+/usr/bin/mpicxx.openmpi -DHAVE_CONFIG_H -I. -I../../../cxx/util -I../../include -I../../include -I../../../include -Wdate-time -D_FORTIFY_SOURCE=2 -g -O2 -ffile-prefix-map=/build/mpi-testsuite-3.2+dfsg/2nd=. -fstack-protector-strong -Wformat -Werror=format-security -MT mtest.o -MD -MP -MF $depbase.Tpo -c -o mtest.o ../../../cxx/util/mtest.cxx &&\\\n mv -f $depbase.Tpo $depbase.Po\n ../../../cxx/util/mtest.cxx: In function 'void* MTestTypeContigInitRecv(MTestDatatype*)':\n ../../../cxx/util/mtest.cxx:189:42: warning: 'int MPI_Type_extent(MPI_Datatype, MPI_Aint*)' is deprecated: MPI_Type_extent is superseded by MPI_Type_get_extent in MPI-2.0 [-Wdeprecated-declarations]\n MPI_Type_extent( mtype->datatype, &size );\n ^\n In file included from ../../../cxx/util/mtest.cxx:7:\n /usr/lib/x86_64-linux-gnu/openmpi/include/mpi.h:1758:20: note: declared here\n@@ -1004184,28 +1004232,28 @@\n ../../../cxx/util/mtest.cxx:848:50: error: 'NULL_COPY_FN' is not a member of 'MPI::Win'\n mem_keyval = MPI::Win::Create_keyval( MPI::Win::NULL_COPY_FN,\n ^~~~~~~~~~~~\n ../../../cxx/util/mtest.cxx:849:22: error: 'NULL_DELETE_FN' is not a member of 'MPI::Win'\n MPI::Win::NULL_DELETE_FN, 0 );\n ^~~~~~~~~~~~~~\n make[3]: *** [Makefile:416: mtest.o] Error 1\n-make[3]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/util'\n+make[3]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/util'\n make[2]: *** [Makefile:989: ../../cxx/util/mtest.o] Error 2\n-make[2]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/io'\n+make[2]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/io'\n \n \n writeallx\n \n ./cxx/io\n fail\n- Failed to build writeallx; make[2]: Entering directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/io'\n+ Failed to build writeallx; make[2]: Entering directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/io'\n (cd ../../cxx/util && make mtest.o)\n-make[3]: Entering directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/util'\n+make[3]: Entering directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/util'\n depbase=`echo mtest.o | sed 's|[^/]*$|.deps/&|;s|\\.o$||'`;\\\n-/usr/bin/mpicxx.openmpi -DHAVE_CONFIG_H -I. -I../../../cxx/util -I../../include -I../../include -I../../../include -Wdate-time -D_FORTIFY_SOURCE=2 -g -O2 -ffile-prefix-map=/build/1st/mpi-testsuite-3.2+dfsg=. -fstack-protector-strong -Wformat -Werror=format-security -MT mtest.o -MD -MP -MF $depbase.Tpo -c -o mtest.o ../../../cxx/util/mtest.cxx &&\\\n+/usr/bin/mpicxx.openmpi -DHAVE_CONFIG_H -I. -I../../../cxx/util -I../../include -I../../include -I../../../include -Wdate-time -D_FORTIFY_SOURCE=2 -g -O2 -ffile-prefix-map=/build/mpi-testsuite-3.2+dfsg/2nd=. -fstack-protector-strong -Wformat -Werror=format-security -MT mtest.o -MD -MP -MF $depbase.Tpo -c -o mtest.o ../../../cxx/util/mtest.cxx &&\\\n mv -f $depbase.Tpo $depbase.Po\n ../../../cxx/util/mtest.cxx: In function 'void* MTestTypeContigInitRecv(MTestDatatype*)':\n ../../../cxx/util/mtest.cxx:189:42: warning: 'int MPI_Type_extent(MPI_Datatype, MPI_Aint*)' is deprecated: MPI_Type_extent is superseded by MPI_Type_get_extent in MPI-2.0 [-Wdeprecated-declarations]\n MPI_Type_extent( mtype->datatype, &size );\n ^\n In file included from ../../../cxx/util/mtest.cxx:7:\n /usr/lib/x86_64-linux-gnu/openmpi/include/mpi.h:1758:20: note: declared here\n@@ -1004237,28 +1004285,28 @@\n ../../../cxx/util/mtest.cxx:848:50: error: 'NULL_COPY_FN' is not a member of 'MPI::Win'\n mem_keyval = MPI::Win::Create_keyval( MPI::Win::NULL_COPY_FN,\n ^~~~~~~~~~~~\n ../../../cxx/util/mtest.cxx:849:22: error: 'NULL_DELETE_FN' is not a member of 'MPI::Win'\n MPI::Win::NULL_DELETE_FN, 0 );\n ^~~~~~~~~~~~~~\n make[3]: *** [Makefile:416: mtest.o] Error 1\n-make[3]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/util'\n+make[3]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/util'\n make[2]: *** [Makefile:989: ../../cxx/util/mtest.o] Error 2\n-make[2]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/io'\n+make[2]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/io'\n \n \n writeshx\n \n ./cxx/io\n fail\n- Failed to build writeshx; make[2]: Entering directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/io'\n+ Failed to build writeshx; make[2]: Entering directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/io'\n (cd ../../cxx/util && make mtest.o)\n-make[3]: Entering directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/util'\n+make[3]: Entering directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/util'\n depbase=`echo mtest.o | sed 's|[^/]*$|.deps/&|;s|\\.o$||'`;\\\n-/usr/bin/mpicxx.openmpi -DHAVE_CONFIG_H -I. -I../../../cxx/util -I../../include -I../../include -I../../../include -Wdate-time -D_FORTIFY_SOURCE=2 -g -O2 -ffile-prefix-map=/build/1st/mpi-testsuite-3.2+dfsg=. -fstack-protector-strong -Wformat -Werror=format-security -MT mtest.o -MD -MP -MF $depbase.Tpo -c -o mtest.o ../../../cxx/util/mtest.cxx &&\\\n+/usr/bin/mpicxx.openmpi -DHAVE_CONFIG_H -I. -I../../../cxx/util -I../../include -I../../include -I../../../include -Wdate-time -D_FORTIFY_SOURCE=2 -g -O2 -ffile-prefix-map=/build/mpi-testsuite-3.2+dfsg/2nd=. -fstack-protector-strong -Wformat -Werror=format-security -MT mtest.o -MD -MP -MF $depbase.Tpo -c -o mtest.o ../../../cxx/util/mtest.cxx &&\\\n mv -f $depbase.Tpo $depbase.Po\n ../../../cxx/util/mtest.cxx: In function 'void* MTestTypeContigInitRecv(MTestDatatype*)':\n ../../../cxx/util/mtest.cxx:189:42: warning: 'int MPI_Type_extent(MPI_Datatype, MPI_Aint*)' is deprecated: MPI_Type_extent is superseded by MPI_Type_get_extent in MPI-2.0 [-Wdeprecated-declarations]\n MPI_Type_extent( mtype->datatype, &size );\n ^\n In file included from ../../../cxx/util/mtest.cxx:7:\n /usr/lib/x86_64-linux-gnu/openmpi/include/mpi.h:1758:20: note: declared here\n@@ -1004290,28 +1004338,28 @@\n ../../../cxx/util/mtest.cxx:848:50: error: 'NULL_COPY_FN' is not a member of 'MPI::Win'\n mem_keyval = MPI::Win::Create_keyval( MPI::Win::NULL_COPY_FN,\n ^~~~~~~~~~~~\n ../../../cxx/util/mtest.cxx:849:22: error: 'NULL_DELETE_FN' is not a member of 'MPI::Win'\n MPI::Win::NULL_DELETE_FN, 0 );\n ^~~~~~~~~~~~~~\n make[3]: *** [Makefile:416: mtest.o] Error 1\n-make[3]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/util'\n+make[3]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/util'\n make[2]: *** [Makefile:989: ../../cxx/util/mtest.o] Error 2\n-make[2]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/io'\n+make[2]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/io'\n \n \n writeordx\n \n ./cxx/io\n fail\n- Failed to build writeordx; make[2]: Entering directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/io'\n+ Failed to build writeordx; make[2]: Entering directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/io'\n (cd ../../cxx/util && make mtest.o)\n-make[3]: Entering directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/util'\n+make[3]: Entering directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/util'\n depbase=`echo mtest.o | sed 's|[^/]*$|.deps/&|;s|\\.o$||'`;\\\n-/usr/bin/mpicxx.openmpi -DHAVE_CONFIG_H -I. -I../../../cxx/util -I../../include -I../../include -I../../../include -Wdate-time -D_FORTIFY_SOURCE=2 -g -O2 -ffile-prefix-map=/build/1st/mpi-testsuite-3.2+dfsg=. -fstack-protector-strong -Wformat -Werror=format-security -MT mtest.o -MD -MP -MF $depbase.Tpo -c -o mtest.o ../../../cxx/util/mtest.cxx &&\\\n+/usr/bin/mpicxx.openmpi -DHAVE_CONFIG_H -I. -I../../../cxx/util -I../../include -I../../include -I../../../include -Wdate-time -D_FORTIFY_SOURCE=2 -g -O2 -ffile-prefix-map=/build/mpi-testsuite-3.2+dfsg/2nd=. -fstack-protector-strong -Wformat -Werror=format-security -MT mtest.o -MD -MP -MF $depbase.Tpo -c -o mtest.o ../../../cxx/util/mtest.cxx &&\\\n mv -f $depbase.Tpo $depbase.Po\n ../../../cxx/util/mtest.cxx: In function 'void* MTestTypeContigInitRecv(MTestDatatype*)':\n ../../../cxx/util/mtest.cxx:189:42: warning: 'int MPI_Type_extent(MPI_Datatype, MPI_Aint*)' is deprecated: MPI_Type_extent is superseded by MPI_Type_get_extent in MPI-2.0 [-Wdeprecated-declarations]\n MPI_Type_extent( mtype->datatype, &size );\n ^\n In file included from ../../../cxx/util/mtest.cxx:7:\n /usr/lib/x86_64-linux-gnu/openmpi/include/mpi.h:1758:20: note: declared here\n@@ -1004343,28 +1004391,28 @@\n ../../../cxx/util/mtest.cxx:848:50: error: 'NULL_COPY_FN' is not a member of 'MPI::Win'\n mem_keyval = MPI::Win::Create_keyval( MPI::Win::NULL_COPY_FN,\n ^~~~~~~~~~~~\n ../../../cxx/util/mtest.cxx:849:22: error: 'NULL_DELETE_FN' is not a member of 'MPI::Win'\n MPI::Win::NULL_DELETE_FN, 0 );\n ^~~~~~~~~~~~~~\n make[3]: *** [Makefile:416: mtest.o] Error 1\n-make[3]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/util'\n+make[3]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/util'\n make[2]: *** [Makefile:989: ../../cxx/util/mtest.o] Error 2\n-make[2]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/io'\n+make[2]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/io'\n \n \n writeatallx\n \n ./cxx/io\n fail\n- Failed to build writeatallx; make[2]: Entering directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/io'\n+ Failed to build writeatallx; make[2]: Entering directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/io'\n (cd ../../cxx/util && make mtest.o)\n-make[3]: Entering directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/util'\n+make[3]: Entering directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/util'\n depbase=`echo mtest.o | sed 's|[^/]*$|.deps/&|;s|\\.o$||'`;\\\n-/usr/bin/mpicxx.openmpi -DHAVE_CONFIG_H -I. -I../../../cxx/util -I../../include -I../../include -I../../../include -Wdate-time -D_FORTIFY_SOURCE=2 -g -O2 -ffile-prefix-map=/build/1st/mpi-testsuite-3.2+dfsg=. -fstack-protector-strong -Wformat -Werror=format-security -MT mtest.o -MD -MP -MF $depbase.Tpo -c -o mtest.o ../../../cxx/util/mtest.cxx &&\\\n+/usr/bin/mpicxx.openmpi -DHAVE_CONFIG_H -I. -I../../../cxx/util -I../../include -I../../include -I../../../include -Wdate-time -D_FORTIFY_SOURCE=2 -g -O2 -ffile-prefix-map=/build/mpi-testsuite-3.2+dfsg/2nd=. -fstack-protector-strong -Wformat -Werror=format-security -MT mtest.o -MD -MP -MF $depbase.Tpo -c -o mtest.o ../../../cxx/util/mtest.cxx &&\\\n mv -f $depbase.Tpo $depbase.Po\n ../../../cxx/util/mtest.cxx: In function 'void* MTestTypeContigInitRecv(MTestDatatype*)':\n ../../../cxx/util/mtest.cxx:189:42: warning: 'int MPI_Type_extent(MPI_Datatype, MPI_Aint*)' is deprecated: MPI_Type_extent is superseded by MPI_Type_get_extent in MPI-2.0 [-Wdeprecated-declarations]\n MPI_Type_extent( mtype->datatype, &size );\n ^\n In file included from ../../../cxx/util/mtest.cxx:7:\n /usr/lib/x86_64-linux-gnu/openmpi/include/mpi.h:1758:20: note: declared here\n@@ -1004396,28 +1004444,28 @@\n ../../../cxx/util/mtest.cxx:848:50: error: 'NULL_COPY_FN' is not a member of 'MPI::Win'\n mem_keyval = MPI::Win::Create_keyval( MPI::Win::NULL_COPY_FN,\n ^~~~~~~~~~~~\n ../../../cxx/util/mtest.cxx:849:22: error: 'NULL_DELETE_FN' is not a member of 'MPI::Win'\n MPI::Win::NULL_DELETE_FN, 0 );\n ^~~~~~~~~~~~~~\n make[3]: *** [Makefile:416: mtest.o] Error 1\n-make[3]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/util'\n+make[3]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/util'\n make[2]: *** [Makefile:989: ../../cxx/util/mtest.o] Error 2\n-make[2]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/io'\n+make[2]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/io'\n \n \n writeatallbex\n \n ./cxx/io\n fail\n- Failed to build writeatallbex; make[2]: Entering directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/io'\n+ Failed to build writeatallbex; make[2]: Entering directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/io'\n (cd ../../cxx/util && make mtest.o)\n-make[3]: Entering directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/util'\n+make[3]: Entering directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/util'\n depbase=`echo mtest.o | sed 's|[^/]*$|.deps/&|;s|\\.o$||'`;\\\n-/usr/bin/mpicxx.openmpi -DHAVE_CONFIG_H -I. -I../../../cxx/util -I../../include -I../../include -I../../../include -Wdate-time -D_FORTIFY_SOURCE=2 -g -O2 -ffile-prefix-map=/build/1st/mpi-testsuite-3.2+dfsg=. -fstack-protector-strong -Wformat -Werror=format-security -MT mtest.o -MD -MP -MF $depbase.Tpo -c -o mtest.o ../../../cxx/util/mtest.cxx &&\\\n+/usr/bin/mpicxx.openmpi -DHAVE_CONFIG_H -I. -I../../../cxx/util -I../../include -I../../include -I../../../include -Wdate-time -D_FORTIFY_SOURCE=2 -g -O2 -ffile-prefix-map=/build/mpi-testsuite-3.2+dfsg/2nd=. -fstack-protector-strong -Wformat -Werror=format-security -MT mtest.o -MD -MP -MF $depbase.Tpo -c -o mtest.o ../../../cxx/util/mtest.cxx &&\\\n mv -f $depbase.Tpo $depbase.Po\n ../../../cxx/util/mtest.cxx: In function 'void* MTestTypeContigInitRecv(MTestDatatype*)':\n ../../../cxx/util/mtest.cxx:189:42: warning: 'int MPI_Type_extent(MPI_Datatype, MPI_Aint*)' is deprecated: MPI_Type_extent is superseded by MPI_Type_get_extent in MPI-2.0 [-Wdeprecated-declarations]\n MPI_Type_extent( mtype->datatype, &size );\n ^\n In file included from ../../../cxx/util/mtest.cxx:7:\n /usr/lib/x86_64-linux-gnu/openmpi/include/mpi.h:1758:20: note: declared here\n@@ -1004449,28 +1004497,28 @@\n ../../../cxx/util/mtest.cxx:848:50: error: 'NULL_COPY_FN' is not a member of 'MPI::Win'\n mem_keyval = MPI::Win::Create_keyval( MPI::Win::NULL_COPY_FN,\n ^~~~~~~~~~~~\n ../../../cxx/util/mtest.cxx:849:22: error: 'NULL_DELETE_FN' is not a member of 'MPI::Win'\n MPI::Win::NULL_DELETE_FN, 0 );\n ^~~~~~~~~~~~~~\n make[3]: *** [Makefile:416: mtest.o] Error 1\n-make[3]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/util'\n+make[3]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/util'\n make[2]: *** [Makefile:989: ../../cxx/util/mtest.o] Error 2\n-make[2]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/io'\n+make[2]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/io'\n \n \n writeallbex\n \n ./cxx/io\n fail\n- Failed to build writeallbex; make[2]: Entering directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/io'\n+ Failed to build writeallbex; make[2]: Entering directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/io'\n (cd ../../cxx/util && make mtest.o)\n-make[3]: Entering directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/util'\n+make[3]: Entering directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/util'\n depbase=`echo mtest.o | sed 's|[^/]*$|.deps/&|;s|\\.o$||'`;\\\n-/usr/bin/mpicxx.openmpi -DHAVE_CONFIG_H -I. -I../../../cxx/util -I../../include -I../../include -I../../../include -Wdate-time -D_FORTIFY_SOURCE=2 -g -O2 -ffile-prefix-map=/build/1st/mpi-testsuite-3.2+dfsg=. -fstack-protector-strong -Wformat -Werror=format-security -MT mtest.o -MD -MP -MF $depbase.Tpo -c -o mtest.o ../../../cxx/util/mtest.cxx &&\\\n+/usr/bin/mpicxx.openmpi -DHAVE_CONFIG_H -I. -I../../../cxx/util -I../../include -I../../include -I../../../include -Wdate-time -D_FORTIFY_SOURCE=2 -g -O2 -ffile-prefix-map=/build/mpi-testsuite-3.2+dfsg/2nd=. -fstack-protector-strong -Wformat -Werror=format-security -MT mtest.o -MD -MP -MF $depbase.Tpo -c -o mtest.o ../../../cxx/util/mtest.cxx &&\\\n mv -f $depbase.Tpo $depbase.Po\n ../../../cxx/util/mtest.cxx: In function 'void* MTestTypeContigInitRecv(MTestDatatype*)':\n ../../../cxx/util/mtest.cxx:189:42: warning: 'int MPI_Type_extent(MPI_Datatype, MPI_Aint*)' is deprecated: MPI_Type_extent is superseded by MPI_Type_get_extent in MPI-2.0 [-Wdeprecated-declarations]\n MPI_Type_extent( mtype->datatype, &size );\n ^\n In file included from ../../../cxx/util/mtest.cxx:7:\n /usr/lib/x86_64-linux-gnu/openmpi/include/mpi.h:1758:20: note: declared here\n@@ -1004502,28 +1004550,28 @@\n ../../../cxx/util/mtest.cxx:848:50: error: 'NULL_COPY_FN' is not a member of 'MPI::Win'\n mem_keyval = MPI::Win::Create_keyval( MPI::Win::NULL_COPY_FN,\n ^~~~~~~~~~~~\n ../../../cxx/util/mtest.cxx:849:22: error: 'NULL_DELETE_FN' is not a member of 'MPI::Win'\n MPI::Win::NULL_DELETE_FN, 0 );\n ^~~~~~~~~~~~~~\n make[3]: *** [Makefile:416: mtest.o] Error 1\n-make[3]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/util'\n+make[3]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/util'\n make[2]: *** [Makefile:989: ../../cxx/util/mtest.o] Error 2\n-make[2]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/io'\n+make[2]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/io'\n \n \n writeordbex\n \n ./cxx/io\n fail\n- Failed to build writeordbex; make[2]: Entering directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/io'\n+ Failed to build writeordbex; make[2]: Entering directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/io'\n (cd ../../cxx/util && make mtest.o)\n-make[3]: Entering directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/util'\n+make[3]: Entering directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/util'\n depbase=`echo mtest.o | sed 's|[^/]*$|.deps/&|;s|\\.o$||'`;\\\n-/usr/bin/mpicxx.openmpi -DHAVE_CONFIG_H -I. -I../../../cxx/util -I../../include -I../../include -I../../../include -Wdate-time -D_FORTIFY_SOURCE=2 -g -O2 -ffile-prefix-map=/build/1st/mpi-testsuite-3.2+dfsg=. -fstack-protector-strong -Wformat -Werror=format-security -MT mtest.o -MD -MP -MF $depbase.Tpo -c -o mtest.o ../../../cxx/util/mtest.cxx &&\\\n+/usr/bin/mpicxx.openmpi -DHAVE_CONFIG_H -I. -I../../../cxx/util -I../../include -I../../include -I../../../include -Wdate-time -D_FORTIFY_SOURCE=2 -g -O2 -ffile-prefix-map=/build/mpi-testsuite-3.2+dfsg/2nd=. -fstack-protector-strong -Wformat -Werror=format-security -MT mtest.o -MD -MP -MF $depbase.Tpo -c -o mtest.o ../../../cxx/util/mtest.cxx &&\\\n mv -f $depbase.Tpo $depbase.Po\n ../../../cxx/util/mtest.cxx: In function 'void* MTestTypeContigInitRecv(MTestDatatype*)':\n ../../../cxx/util/mtest.cxx:189:42: warning: 'int MPI_Type_extent(MPI_Datatype, MPI_Aint*)' is deprecated: MPI_Type_extent is superseded by MPI_Type_get_extent in MPI-2.0 [-Wdeprecated-declarations]\n MPI_Type_extent( mtype->datatype, &size );\n ^\n In file included from ../../../cxx/util/mtest.cxx:7:\n /usr/lib/x86_64-linux-gnu/openmpi/include/mpi.h:1758:20: note: declared here\n@@ -1004555,28 +1004603,28 @@\n ../../../cxx/util/mtest.cxx:848:50: error: 'NULL_COPY_FN' is not a member of 'MPI::Win'\n mem_keyval = MPI::Win::Create_keyval( MPI::Win::NULL_COPY_FN,\n ^~~~~~~~~~~~\n ../../../cxx/util/mtest.cxx:849:22: error: 'NULL_DELETE_FN' is not a member of 'MPI::Win'\n MPI::Win::NULL_DELETE_FN, 0 );\n ^~~~~~~~~~~~~~\n make[3]: *** [Makefile:416: mtest.o] Error 1\n-make[3]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/util'\n+make[3]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/util'\n make[2]: *** [Makefile:989: ../../cxx/util/mtest.o] Error 2\n-make[2]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/io'\n+make[2]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/io'\n \n \n iwriteatnosx\n \n ./cxx/io\n fail\n- Failed to build iwriteatnosx; make[2]: Entering directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/io'\n+ Failed to build iwriteatnosx; make[2]: Entering directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/io'\n (cd ../../cxx/util && make mtest.o)\n-make[3]: Entering directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/util'\n+make[3]: Entering directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/util'\n depbase=`echo mtest.o | sed 's|[^/]*$|.deps/&|;s|\\.o$||'`;\\\n-/usr/bin/mpicxx.openmpi -DHAVE_CONFIG_H -I. -I../../../cxx/util -I../../include -I../../include -I../../../include -Wdate-time -D_FORTIFY_SOURCE=2 -g -O2 -ffile-prefix-map=/build/1st/mpi-testsuite-3.2+dfsg=. -fstack-protector-strong -Wformat -Werror=format-security -MT mtest.o -MD -MP -MF $depbase.Tpo -c -o mtest.o ../../../cxx/util/mtest.cxx &&\\\n+/usr/bin/mpicxx.openmpi -DHAVE_CONFIG_H -I. -I../../../cxx/util -I../../include -I../../include -I../../../include -Wdate-time -D_FORTIFY_SOURCE=2 -g -O2 -ffile-prefix-map=/build/mpi-testsuite-3.2+dfsg/2nd=. -fstack-protector-strong -Wformat -Werror=format-security -MT mtest.o -MD -MP -MF $depbase.Tpo -c -o mtest.o ../../../cxx/util/mtest.cxx &&\\\n mv -f $depbase.Tpo $depbase.Po\n ../../../cxx/util/mtest.cxx: In function 'void* MTestTypeContigInitRecv(MTestDatatype*)':\n ../../../cxx/util/mtest.cxx:189:42: warning: 'int MPI_Type_extent(MPI_Datatype, MPI_Aint*)' is deprecated: MPI_Type_extent is superseded by MPI_Type_get_extent in MPI-2.0 [-Wdeprecated-declarations]\n MPI_Type_extent( mtype->datatype, &size );\n ^\n In file included from ../../../cxx/util/mtest.cxx:7:\n /usr/lib/x86_64-linux-gnu/openmpi/include/mpi.h:1758:20: note: declared here\n@@ -1004608,28 +1004656,28 @@\n ../../../cxx/util/mtest.cxx:848:50: error: 'NULL_COPY_FN' is not a member of 'MPI::Win'\n mem_keyval = MPI::Win::Create_keyval( MPI::Win::NULL_COPY_FN,\n ^~~~~~~~~~~~\n ../../../cxx/util/mtest.cxx:849:22: error: 'NULL_DELETE_FN' is not a member of 'MPI::Win'\n MPI::Win::NULL_DELETE_FN, 0 );\n ^~~~~~~~~~~~~~\n make[3]: *** [Makefile:416: mtest.o] Error 1\n-make[3]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/util'\n+make[3]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/util'\n make[2]: *** [Makefile:989: ../../cxx/util/mtest.o] Error 2\n-make[2]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/io'\n+make[2]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/io'\n \n \n iwritenosx\n \n ./cxx/io\n fail\n- Failed to build iwritenosx; make[2]: Entering directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/io'\n+ Failed to build iwritenosx; make[2]: Entering directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/io'\n (cd ../../cxx/util && make mtest.o)\n-make[3]: Entering directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/util'\n+make[3]: Entering directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/util'\n depbase=`echo mtest.o | sed 's|[^/]*$|.deps/&|;s|\\.o$||'`;\\\n-/usr/bin/mpicxx.openmpi -DHAVE_CONFIG_H -I. -I../../../cxx/util -I../../include -I../../include -I../../../include -Wdate-time -D_FORTIFY_SOURCE=2 -g -O2 -ffile-prefix-map=/build/1st/mpi-testsuite-3.2+dfsg=. -fstack-protector-strong -Wformat -Werror=format-security -MT mtest.o -MD -MP -MF $depbase.Tpo -c -o mtest.o ../../../cxx/util/mtest.cxx &&\\\n+/usr/bin/mpicxx.openmpi -DHAVE_CONFIG_H -I. -I../../../cxx/util -I../../include -I../../include -I../../../include -Wdate-time -D_FORTIFY_SOURCE=2 -g -O2 -ffile-prefix-map=/build/mpi-testsuite-3.2+dfsg/2nd=. -fstack-protector-strong -Wformat -Werror=format-security -MT mtest.o -MD -MP -MF $depbase.Tpo -c -o mtest.o ../../../cxx/util/mtest.cxx &&\\\n mv -f $depbase.Tpo $depbase.Po\n ../../../cxx/util/mtest.cxx: In function 'void* MTestTypeContigInitRecv(MTestDatatype*)':\n ../../../cxx/util/mtest.cxx:189:42: warning: 'int MPI_Type_extent(MPI_Datatype, MPI_Aint*)' is deprecated: MPI_Type_extent is superseded by MPI_Type_get_extent in MPI-2.0 [-Wdeprecated-declarations]\n MPI_Type_extent( mtype->datatype, &size );\n ^\n In file included from ../../../cxx/util/mtest.cxx:7:\n /usr/lib/x86_64-linux-gnu/openmpi/include/mpi.h:1758:20: note: declared here\n@@ -1004661,28 +1004709,28 @@\n ../../../cxx/util/mtest.cxx:848:50: error: 'NULL_COPY_FN' is not a member of 'MPI::Win'\n mem_keyval = MPI::Win::Create_keyval( MPI::Win::NULL_COPY_FN,\n ^~~~~~~~~~~~\n ../../../cxx/util/mtest.cxx:849:22: error: 'NULL_DELETE_FN' is not a member of 'MPI::Win'\n MPI::Win::NULL_DELETE_FN, 0 );\n ^~~~~~~~~~~~~~\n make[3]: *** [Makefile:416: mtest.o] Error 1\n-make[3]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/util'\n+make[3]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/util'\n make[2]: *** [Makefile:989: ../../cxx/util/mtest.o] Error 2\n-make[2]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/io'\n+make[2]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/io'\n \n \n iwriteshnosx\n \n ./cxx/io\n fail\n- Failed to build iwriteshnosx; make[2]: Entering directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/io'\n+ Failed to build iwriteshnosx; make[2]: Entering directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/io'\n (cd ../../cxx/util && make mtest.o)\n-make[3]: Entering directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/util'\n+make[3]: Entering directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/util'\n depbase=`echo mtest.o | sed 's|[^/]*$|.deps/&|;s|\\.o$||'`;\\\n-/usr/bin/mpicxx.openmpi -DHAVE_CONFIG_H -I. -I../../../cxx/util -I../../include -I../../include -I../../../include -Wdate-time -D_FORTIFY_SOURCE=2 -g -O2 -ffile-prefix-map=/build/1st/mpi-testsuite-3.2+dfsg=. -fstack-protector-strong -Wformat -Werror=format-security -MT mtest.o -MD -MP -MF $depbase.Tpo -c -o mtest.o ../../../cxx/util/mtest.cxx &&\\\n+/usr/bin/mpicxx.openmpi -DHAVE_CONFIG_H -I. -I../../../cxx/util -I../../include -I../../include -I../../../include -Wdate-time -D_FORTIFY_SOURCE=2 -g -O2 -ffile-prefix-map=/build/mpi-testsuite-3.2+dfsg/2nd=. -fstack-protector-strong -Wformat -Werror=format-security -MT mtest.o -MD -MP -MF $depbase.Tpo -c -o mtest.o ../../../cxx/util/mtest.cxx &&\\\n mv -f $depbase.Tpo $depbase.Po\n ../../../cxx/util/mtest.cxx: In function 'void* MTestTypeContigInitRecv(MTestDatatype*)':\n ../../../cxx/util/mtest.cxx:189:42: warning: 'int MPI_Type_extent(MPI_Datatype, MPI_Aint*)' is deprecated: MPI_Type_extent is superseded by MPI_Type_get_extent in MPI-2.0 [-Wdeprecated-declarations]\n MPI_Type_extent( mtype->datatype, &size );\n ^\n In file included from ../../../cxx/util/mtest.cxx:7:\n /usr/lib/x86_64-linux-gnu/openmpi/include/mpi.h:1758:20: note: declared here\n@@ -1004714,28 +1004762,28 @@\n ../../../cxx/util/mtest.cxx:848:50: error: 'NULL_COPY_FN' is not a member of 'MPI::Win'\n mem_keyval = MPI::Win::Create_keyval( MPI::Win::NULL_COPY_FN,\n ^~~~~~~~~~~~\n ../../../cxx/util/mtest.cxx:849:22: error: 'NULL_DELETE_FN' is not a member of 'MPI::Win'\n MPI::Win::NULL_DELETE_FN, 0 );\n ^~~~~~~~~~~~~~\n make[3]: *** [Makefile:416: mtest.o] Error 1\n-make[3]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/util'\n+make[3]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/util'\n make[2]: *** [Makefile:989: ../../cxx/util/mtest.o] Error 2\n-make[2]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/io'\n+make[2]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/io'\n \n \n writenosx\n \n ./cxx/io\n fail\n- Failed to build writenosx; make[2]: Entering directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/io'\n+ Failed to build writenosx; make[2]: Entering directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/io'\n (cd ../../cxx/util && make mtest.o)\n-make[3]: Entering directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/util'\n+make[3]: Entering directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/util'\n depbase=`echo mtest.o | sed 's|[^/]*$|.deps/&|;s|\\.o$||'`;\\\n-/usr/bin/mpicxx.openmpi -DHAVE_CONFIG_H -I. -I../../../cxx/util -I../../include -I../../include -I../../../include -Wdate-time -D_FORTIFY_SOURCE=2 -g -O2 -ffile-prefix-map=/build/1st/mpi-testsuite-3.2+dfsg=. -fstack-protector-strong -Wformat -Werror=format-security -MT mtest.o -MD -MP -MF $depbase.Tpo -c -o mtest.o ../../../cxx/util/mtest.cxx &&\\\n+/usr/bin/mpicxx.openmpi -DHAVE_CONFIG_H -I. -I../../../cxx/util -I../../include -I../../include -I../../../include -Wdate-time -D_FORTIFY_SOURCE=2 -g -O2 -ffile-prefix-map=/build/mpi-testsuite-3.2+dfsg/2nd=. -fstack-protector-strong -Wformat -Werror=format-security -MT mtest.o -MD -MP -MF $depbase.Tpo -c -o mtest.o ../../../cxx/util/mtest.cxx &&\\\n mv -f $depbase.Tpo $depbase.Po\n ../../../cxx/util/mtest.cxx: In function 'void* MTestTypeContigInitRecv(MTestDatatype*)':\n ../../../cxx/util/mtest.cxx:189:42: warning: 'int MPI_Type_extent(MPI_Datatype, MPI_Aint*)' is deprecated: MPI_Type_extent is superseded by MPI_Type_get_extent in MPI-2.0 [-Wdeprecated-declarations]\n MPI_Type_extent( mtype->datatype, &size );\n ^\n In file included from ../../../cxx/util/mtest.cxx:7:\n /usr/lib/x86_64-linux-gnu/openmpi/include/mpi.h:1758:20: note: declared here\n@@ -1004767,28 +1004815,28 @@\n ../../../cxx/util/mtest.cxx:848:50: error: 'NULL_COPY_FN' is not a member of 'MPI::Win'\n mem_keyval = MPI::Win::Create_keyval( MPI::Win::NULL_COPY_FN,\n ^~~~~~~~~~~~\n ../../../cxx/util/mtest.cxx:849:22: error: 'NULL_DELETE_FN' is not a member of 'MPI::Win'\n MPI::Win::NULL_DELETE_FN, 0 );\n ^~~~~~~~~~~~~~\n make[3]: *** [Makefile:416: mtest.o] Error 1\n-make[3]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/util'\n+make[3]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/util'\n make[2]: *** [Makefile:989: ../../cxx/util/mtest.o] Error 2\n-make[2]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/io'\n+make[2]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/io'\n \n \n writeatnosx\n \n ./cxx/io\n fail\n- Failed to build writeatnosx; make[2]: Entering directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/io'\n+ Failed to build writeatnosx; make[2]: Entering directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/io'\n (cd ../../cxx/util && make mtest.o)\n-make[3]: Entering directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/util'\n+make[3]: Entering directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/util'\n depbase=`echo mtest.o | sed 's|[^/]*$|.deps/&|;s|\\.o$||'`;\\\n-/usr/bin/mpicxx.openmpi -DHAVE_CONFIG_H -I. -I../../../cxx/util -I../../include -I../../include -I../../../include -Wdate-time -D_FORTIFY_SOURCE=2 -g -O2 -ffile-prefix-map=/build/1st/mpi-testsuite-3.2+dfsg=. -fstack-protector-strong -Wformat -Werror=format-security -MT mtest.o -MD -MP -MF $depbase.Tpo -c -o mtest.o ../../../cxx/util/mtest.cxx &&\\\n+/usr/bin/mpicxx.openmpi -DHAVE_CONFIG_H -I. -I../../../cxx/util -I../../include -I../../include -I../../../include -Wdate-time -D_FORTIFY_SOURCE=2 -g -O2 -ffile-prefix-map=/build/mpi-testsuite-3.2+dfsg/2nd=. -fstack-protector-strong -Wformat -Werror=format-security -MT mtest.o -MD -MP -MF $depbase.Tpo -c -o mtest.o ../../../cxx/util/mtest.cxx &&\\\n mv -f $depbase.Tpo $depbase.Po\n ../../../cxx/util/mtest.cxx: In function 'void* MTestTypeContigInitRecv(MTestDatatype*)':\n ../../../cxx/util/mtest.cxx:189:42: warning: 'int MPI_Type_extent(MPI_Datatype, MPI_Aint*)' is deprecated: MPI_Type_extent is superseded by MPI_Type_get_extent in MPI-2.0 [-Wdeprecated-declarations]\n MPI_Type_extent( mtype->datatype, &size );\n ^\n In file included from ../../../cxx/util/mtest.cxx:7:\n /usr/lib/x86_64-linux-gnu/openmpi/include/mpi.h:1758:20: note: declared here\n@@ -1004820,28 +1004868,28 @@\n ../../../cxx/util/mtest.cxx:848:50: error: 'NULL_COPY_FN' is not a member of 'MPI::Win'\n mem_keyval = MPI::Win::Create_keyval( MPI::Win::NULL_COPY_FN,\n ^~~~~~~~~~~~\n ../../../cxx/util/mtest.cxx:849:22: error: 'NULL_DELETE_FN' is not a member of 'MPI::Win'\n MPI::Win::NULL_DELETE_FN, 0 );\n ^~~~~~~~~~~~~~\n make[3]: *** [Makefile:416: mtest.o] Error 1\n-make[3]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/util'\n+make[3]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/util'\n make[2]: *** [Makefile:989: ../../cxx/util/mtest.o] Error 2\n-make[2]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/io'\n+make[2]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/io'\n \n \n writeallnosx\n \n ./cxx/io\n fail\n- Failed to build writeallnosx; make[2]: Entering directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/io'\n+ Failed to build writeallnosx; make[2]: Entering directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/io'\n (cd ../../cxx/util && make mtest.o)\n-make[3]: Entering directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/util'\n+make[3]: Entering directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/util'\n depbase=`echo mtest.o | sed 's|[^/]*$|.deps/&|;s|\\.o$||'`;\\\n-/usr/bin/mpicxx.openmpi -DHAVE_CONFIG_H -I. -I../../../cxx/util -I../../include -I../../include -I../../../include -Wdate-time -D_FORTIFY_SOURCE=2 -g -O2 -ffile-prefix-map=/build/1st/mpi-testsuite-3.2+dfsg=. -fstack-protector-strong -Wformat -Werror=format-security -MT mtest.o -MD -MP -MF $depbase.Tpo -c -o mtest.o ../../../cxx/util/mtest.cxx &&\\\n+/usr/bin/mpicxx.openmpi -DHAVE_CONFIG_H -I. -I../../../cxx/util -I../../include -I../../include -I../../../include -Wdate-time -D_FORTIFY_SOURCE=2 -g -O2 -ffile-prefix-map=/build/mpi-testsuite-3.2+dfsg/2nd=. -fstack-protector-strong -Wformat -Werror=format-security -MT mtest.o -MD -MP -MF $depbase.Tpo -c -o mtest.o ../../../cxx/util/mtest.cxx &&\\\n mv -f $depbase.Tpo $depbase.Po\n ../../../cxx/util/mtest.cxx: In function 'void* MTestTypeContigInitRecv(MTestDatatype*)':\n ../../../cxx/util/mtest.cxx:189:42: warning: 'int MPI_Type_extent(MPI_Datatype, MPI_Aint*)' is deprecated: MPI_Type_extent is superseded by MPI_Type_get_extent in MPI-2.0 [-Wdeprecated-declarations]\n MPI_Type_extent( mtype->datatype, &size );\n ^\n In file included from ../../../cxx/util/mtest.cxx:7:\n /usr/lib/x86_64-linux-gnu/openmpi/include/mpi.h:1758:20: note: declared here\n@@ -1004873,28 +1004921,28 @@\n ../../../cxx/util/mtest.cxx:848:50: error: 'NULL_COPY_FN' is not a member of 'MPI::Win'\n mem_keyval = MPI::Win::Create_keyval( MPI::Win::NULL_COPY_FN,\n ^~~~~~~~~~~~\n ../../../cxx/util/mtest.cxx:849:22: error: 'NULL_DELETE_FN' is not a member of 'MPI::Win'\n MPI::Win::NULL_DELETE_FN, 0 );\n ^~~~~~~~~~~~~~\n make[3]: *** [Makefile:416: mtest.o] Error 1\n-make[3]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/util'\n+make[3]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/util'\n make[2]: *** [Makefile:989: ../../cxx/util/mtest.o] Error 2\n-make[2]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/io'\n+make[2]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/io'\n \n \n writeshnosx\n \n ./cxx/io\n fail\n- Failed to build writeshnosx; make[2]: Entering directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/io'\n+ Failed to build writeshnosx; make[2]: Entering directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/io'\n (cd ../../cxx/util && make mtest.o)\n-make[3]: Entering directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/util'\n+make[3]: Entering directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/util'\n depbase=`echo mtest.o | sed 's|[^/]*$|.deps/&|;s|\\.o$||'`;\\\n-/usr/bin/mpicxx.openmpi -DHAVE_CONFIG_H -I. -I../../../cxx/util -I../../include -I../../include -I../../../include -Wdate-time -D_FORTIFY_SOURCE=2 -g -O2 -ffile-prefix-map=/build/1st/mpi-testsuite-3.2+dfsg=. -fstack-protector-strong -Wformat -Werror=format-security -MT mtest.o -MD -MP -MF $depbase.Tpo -c -o mtest.o ../../../cxx/util/mtest.cxx &&\\\n+/usr/bin/mpicxx.openmpi -DHAVE_CONFIG_H -I. -I../../../cxx/util -I../../include -I../../include -I../../../include -Wdate-time -D_FORTIFY_SOURCE=2 -g -O2 -ffile-prefix-map=/build/mpi-testsuite-3.2+dfsg/2nd=. -fstack-protector-strong -Wformat -Werror=format-security -MT mtest.o -MD -MP -MF $depbase.Tpo -c -o mtest.o ../../../cxx/util/mtest.cxx &&\\\n mv -f $depbase.Tpo $depbase.Po\n ../../../cxx/util/mtest.cxx: In function 'void* MTestTypeContigInitRecv(MTestDatatype*)':\n ../../../cxx/util/mtest.cxx:189:42: warning: 'int MPI_Type_extent(MPI_Datatype, MPI_Aint*)' is deprecated: MPI_Type_extent is superseded by MPI_Type_get_extent in MPI-2.0 [-Wdeprecated-declarations]\n MPI_Type_extent( mtype->datatype, &size );\n ^\n In file included from ../../../cxx/util/mtest.cxx:7:\n /usr/lib/x86_64-linux-gnu/openmpi/include/mpi.h:1758:20: note: declared here\n@@ -1004926,28 +1004974,28 @@\n ../../../cxx/util/mtest.cxx:848:50: error: 'NULL_COPY_FN' is not a member of 'MPI::Win'\n mem_keyval = MPI::Win::Create_keyval( MPI::Win::NULL_COPY_FN,\n ^~~~~~~~~~~~\n ../../../cxx/util/mtest.cxx:849:22: error: 'NULL_DELETE_FN' is not a member of 'MPI::Win'\n MPI::Win::NULL_DELETE_FN, 0 );\n ^~~~~~~~~~~~~~\n make[3]: *** [Makefile:416: mtest.o] Error 1\n-make[3]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/util'\n+make[3]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/util'\n make[2]: *** [Makefile:989: ../../cxx/util/mtest.o] Error 2\n-make[2]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/io'\n+make[2]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/io'\n \n \n writeordnosx\n \n ./cxx/io\n fail\n- Failed to build writeordnosx; make[2]: Entering directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/io'\n+ Failed to build writeordnosx; make[2]: Entering directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/io'\n (cd ../../cxx/util && make mtest.o)\n-make[3]: Entering directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/util'\n+make[3]: Entering directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/util'\n depbase=`echo mtest.o | sed 's|[^/]*$|.deps/&|;s|\\.o$||'`;\\\n-/usr/bin/mpicxx.openmpi -DHAVE_CONFIG_H -I. -I../../../cxx/util -I../../include -I../../include -I../../../include -Wdate-time -D_FORTIFY_SOURCE=2 -g -O2 -ffile-prefix-map=/build/1st/mpi-testsuite-3.2+dfsg=. -fstack-protector-strong -Wformat -Werror=format-security -MT mtest.o -MD -MP -MF $depbase.Tpo -c -o mtest.o ../../../cxx/util/mtest.cxx &&\\\n+/usr/bin/mpicxx.openmpi -DHAVE_CONFIG_H -I. -I../../../cxx/util -I../../include -I../../include -I../../../include -Wdate-time -D_FORTIFY_SOURCE=2 -g -O2 -ffile-prefix-map=/build/mpi-testsuite-3.2+dfsg/2nd=. -fstack-protector-strong -Wformat -Werror=format-security -MT mtest.o -MD -MP -MF $depbase.Tpo -c -o mtest.o ../../../cxx/util/mtest.cxx &&\\\n mv -f $depbase.Tpo $depbase.Po\n ../../../cxx/util/mtest.cxx: In function 'void* MTestTypeContigInitRecv(MTestDatatype*)':\n ../../../cxx/util/mtest.cxx:189:42: warning: 'int MPI_Type_extent(MPI_Datatype, MPI_Aint*)' is deprecated: MPI_Type_extent is superseded by MPI_Type_get_extent in MPI-2.0 [-Wdeprecated-declarations]\n MPI_Type_extent( mtype->datatype, &size );\n ^\n In file included from ../../../cxx/util/mtest.cxx:7:\n /usr/lib/x86_64-linux-gnu/openmpi/include/mpi.h:1758:20: note: declared here\n@@ -1004979,28 +1005027,28 @@\n ../../../cxx/util/mtest.cxx:848:50: error: 'NULL_COPY_FN' is not a member of 'MPI::Win'\n mem_keyval = MPI::Win::Create_keyval( MPI::Win::NULL_COPY_FN,\n ^~~~~~~~~~~~\n ../../../cxx/util/mtest.cxx:849:22: error: 'NULL_DELETE_FN' is not a member of 'MPI::Win'\n MPI::Win::NULL_DELETE_FN, 0 );\n ^~~~~~~~~~~~~~\n make[3]: *** [Makefile:416: mtest.o] Error 1\n-make[3]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/util'\n+make[3]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/util'\n make[2]: *** [Makefile:989: ../../cxx/util/mtest.o] Error 2\n-make[2]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/io'\n+make[2]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/io'\n \n \n writeatallnosx\n \n ./cxx/io\n fail\n- Failed to build writeatallnosx; make[2]: Entering directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/io'\n+ Failed to build writeatallnosx; make[2]: Entering directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/io'\n (cd ../../cxx/util && make mtest.o)\n-make[3]: Entering directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/util'\n+make[3]: Entering directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/util'\n depbase=`echo mtest.o | sed 's|[^/]*$|.deps/&|;s|\\.o$||'`;\\\n-/usr/bin/mpicxx.openmpi -DHAVE_CONFIG_H -I. -I../../../cxx/util -I../../include -I../../include -I../../../include -Wdate-time -D_FORTIFY_SOURCE=2 -g -O2 -ffile-prefix-map=/build/1st/mpi-testsuite-3.2+dfsg=. -fstack-protector-strong -Wformat -Werror=format-security -MT mtest.o -MD -MP -MF $depbase.Tpo -c -o mtest.o ../../../cxx/util/mtest.cxx &&\\\n+/usr/bin/mpicxx.openmpi -DHAVE_CONFIG_H -I. -I../../../cxx/util -I../../include -I../../include -I../../../include -Wdate-time -D_FORTIFY_SOURCE=2 -g -O2 -ffile-prefix-map=/build/mpi-testsuite-3.2+dfsg/2nd=. -fstack-protector-strong -Wformat -Werror=format-security -MT mtest.o -MD -MP -MF $depbase.Tpo -c -o mtest.o ../../../cxx/util/mtest.cxx &&\\\n mv -f $depbase.Tpo $depbase.Po\n ../../../cxx/util/mtest.cxx: In function 'void* MTestTypeContigInitRecv(MTestDatatype*)':\n ../../../cxx/util/mtest.cxx:189:42: warning: 'int MPI_Type_extent(MPI_Datatype, MPI_Aint*)' is deprecated: MPI_Type_extent is superseded by MPI_Type_get_extent in MPI-2.0 [-Wdeprecated-declarations]\n MPI_Type_extent( mtype->datatype, &size );\n ^\n In file included from ../../../cxx/util/mtest.cxx:7:\n /usr/lib/x86_64-linux-gnu/openmpi/include/mpi.h:1758:20: note: declared here\n@@ -1005032,28 +1005080,28 @@\n ../../../cxx/util/mtest.cxx:848:50: error: 'NULL_COPY_FN' is not a member of 'MPI::Win'\n mem_keyval = MPI::Win::Create_keyval( MPI::Win::NULL_COPY_FN,\n ^~~~~~~~~~~~\n ../../../cxx/util/mtest.cxx:849:22: error: 'NULL_DELETE_FN' is not a member of 'MPI::Win'\n MPI::Win::NULL_DELETE_FN, 0 );\n ^~~~~~~~~~~~~~\n make[3]: *** [Makefile:416: mtest.o] Error 1\n-make[3]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/util'\n+make[3]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/util'\n make[2]: *** [Makefile:989: ../../cxx/util/mtest.o] Error 2\n-make[2]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/io'\n+make[2]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/io'\n \n \n writeatallbenosx\n \n ./cxx/io\n fail\n- Failed to build writeatallbenosx; make[2]: Entering directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/io'\n+ Failed to build writeatallbenosx; make[2]: Entering directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/io'\n (cd ../../cxx/util && make mtest.o)\n-make[3]: Entering directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/util'\n+make[3]: Entering directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/util'\n depbase=`echo mtest.o | sed 's|[^/]*$|.deps/&|;s|\\.o$||'`;\\\n-/usr/bin/mpicxx.openmpi -DHAVE_CONFIG_H -I. -I../../../cxx/util -I../../include -I../../include -I../../../include -Wdate-time -D_FORTIFY_SOURCE=2 -g -O2 -ffile-prefix-map=/build/1st/mpi-testsuite-3.2+dfsg=. -fstack-protector-strong -Wformat -Werror=format-security -MT mtest.o -MD -MP -MF $depbase.Tpo -c -o mtest.o ../../../cxx/util/mtest.cxx &&\\\n+/usr/bin/mpicxx.openmpi -DHAVE_CONFIG_H -I. -I../../../cxx/util -I../../include -I../../include -I../../../include -Wdate-time -D_FORTIFY_SOURCE=2 -g -O2 -ffile-prefix-map=/build/mpi-testsuite-3.2+dfsg/2nd=. -fstack-protector-strong -Wformat -Werror=format-security -MT mtest.o -MD -MP -MF $depbase.Tpo -c -o mtest.o ../../../cxx/util/mtest.cxx &&\\\n mv -f $depbase.Tpo $depbase.Po\n ../../../cxx/util/mtest.cxx: In function 'void* MTestTypeContigInitRecv(MTestDatatype*)':\n ../../../cxx/util/mtest.cxx:189:42: warning: 'int MPI_Type_extent(MPI_Datatype, MPI_Aint*)' is deprecated: MPI_Type_extent is superseded by MPI_Type_get_extent in MPI-2.0 [-Wdeprecated-declarations]\n MPI_Type_extent( mtype->datatype, &size );\n ^\n In file included from ../../../cxx/util/mtest.cxx:7:\n /usr/lib/x86_64-linux-gnu/openmpi/include/mpi.h:1758:20: note: declared here\n@@ -1005085,28 +1005133,28 @@\n ../../../cxx/util/mtest.cxx:848:50: error: 'NULL_COPY_FN' is not a member of 'MPI::Win'\n mem_keyval = MPI::Win::Create_keyval( MPI::Win::NULL_COPY_FN,\n ^~~~~~~~~~~~\n ../../../cxx/util/mtest.cxx:849:22: error: 'NULL_DELETE_FN' is not a member of 'MPI::Win'\n MPI::Win::NULL_DELETE_FN, 0 );\n ^~~~~~~~~~~~~~\n make[3]: *** [Makefile:416: mtest.o] Error 1\n-make[3]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/util'\n+make[3]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/util'\n make[2]: *** [Makefile:989: ../../cxx/util/mtest.o] Error 2\n-make[2]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/io'\n+make[2]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/io'\n \n \n writeallbenosx\n \n ./cxx/io\n fail\n- Failed to build writeallbenosx; make[2]: Entering directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/io'\n+ Failed to build writeallbenosx; make[2]: Entering directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/io'\n (cd ../../cxx/util && make mtest.o)\n-make[3]: Entering directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/util'\n+make[3]: Entering directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/util'\n depbase=`echo mtest.o | sed 's|[^/]*$|.deps/&|;s|\\.o$||'`;\\\n-/usr/bin/mpicxx.openmpi -DHAVE_CONFIG_H -I. -I../../../cxx/util -I../../include -I../../include -I../../../include -Wdate-time -D_FORTIFY_SOURCE=2 -g -O2 -ffile-prefix-map=/build/1st/mpi-testsuite-3.2+dfsg=. -fstack-protector-strong -Wformat -Werror=format-security -MT mtest.o -MD -MP -MF $depbase.Tpo -c -o mtest.o ../../../cxx/util/mtest.cxx &&\\\n+/usr/bin/mpicxx.openmpi -DHAVE_CONFIG_H -I. -I../../../cxx/util -I../../include -I../../include -I../../../include -Wdate-time -D_FORTIFY_SOURCE=2 -g -O2 -ffile-prefix-map=/build/mpi-testsuite-3.2+dfsg/2nd=. -fstack-protector-strong -Wformat -Werror=format-security -MT mtest.o -MD -MP -MF $depbase.Tpo -c -o mtest.o ../../../cxx/util/mtest.cxx &&\\\n mv -f $depbase.Tpo $depbase.Po\n ../../../cxx/util/mtest.cxx: In function 'void* MTestTypeContigInitRecv(MTestDatatype*)':\n ../../../cxx/util/mtest.cxx:189:42: warning: 'int MPI_Type_extent(MPI_Datatype, MPI_Aint*)' is deprecated: MPI_Type_extent is superseded by MPI_Type_get_extent in MPI-2.0 [-Wdeprecated-declarations]\n MPI_Type_extent( mtype->datatype, &size );\n ^\n In file included from ../../../cxx/util/mtest.cxx:7:\n /usr/lib/x86_64-linux-gnu/openmpi/include/mpi.h:1758:20: note: declared here\n@@ -1005138,28 +1005186,28 @@\n ../../../cxx/util/mtest.cxx:848:50: error: 'NULL_COPY_FN' is not a member of 'MPI::Win'\n mem_keyval = MPI::Win::Create_keyval( MPI::Win::NULL_COPY_FN,\n ^~~~~~~~~~~~\n ../../../cxx/util/mtest.cxx:849:22: error: 'NULL_DELETE_FN' is not a member of 'MPI::Win'\n MPI::Win::NULL_DELETE_FN, 0 );\n ^~~~~~~~~~~~~~\n make[3]: *** [Makefile:416: mtest.o] Error 1\n-make[3]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/util'\n+make[3]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/util'\n make[2]: *** [Makefile:989: ../../cxx/util/mtest.o] Error 2\n-make[2]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/io'\n+make[2]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/io'\n \n \n writeordbenosx\n \n ./cxx/io\n fail\n- Failed to build writeordbenosx; make[2]: Entering directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/io'\n+ Failed to build writeordbenosx; make[2]: Entering directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/io'\n (cd ../../cxx/util && make mtest.o)\n-make[3]: Entering directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/util'\n+make[3]: Entering directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/util'\n depbase=`echo mtest.o | sed 's|[^/]*$|.deps/&|;s|\\.o$||'`;\\\n-/usr/bin/mpicxx.openmpi -DHAVE_CONFIG_H -I. -I../../../cxx/util -I../../include -I../../include -I../../../include -Wdate-time -D_FORTIFY_SOURCE=2 -g -O2 -ffile-prefix-map=/build/1st/mpi-testsuite-3.2+dfsg=. -fstack-protector-strong -Wformat -Werror=format-security -MT mtest.o -MD -MP -MF $depbase.Tpo -c -o mtest.o ../../../cxx/util/mtest.cxx &&\\\n+/usr/bin/mpicxx.openmpi -DHAVE_CONFIG_H -I. -I../../../cxx/util -I../../include -I../../include -I../../../include -Wdate-time -D_FORTIFY_SOURCE=2 -g -O2 -ffile-prefix-map=/build/mpi-testsuite-3.2+dfsg/2nd=. -fstack-protector-strong -Wformat -Werror=format-security -MT mtest.o -MD -MP -MF $depbase.Tpo -c -o mtest.o ../../../cxx/util/mtest.cxx &&\\\n mv -f $depbase.Tpo $depbase.Po\n ../../../cxx/util/mtest.cxx: In function 'void* MTestTypeContigInitRecv(MTestDatatype*)':\n ../../../cxx/util/mtest.cxx:189:42: warning: 'int MPI_Type_extent(MPI_Datatype, MPI_Aint*)' is deprecated: MPI_Type_extent is superseded by MPI_Type_get_extent in MPI-2.0 [-Wdeprecated-declarations]\n MPI_Type_extent( mtype->datatype, &size );\n ^\n In file included from ../../../cxx/util/mtest.cxx:7:\n /usr/lib/x86_64-linux-gnu/openmpi/include/mpi.h:1758:20: note: declared here\n@@ -1005191,28 +1005239,28 @@\n ../../../cxx/util/mtest.cxx:848:50: error: 'NULL_COPY_FN' is not a member of 'MPI::Win'\n mem_keyval = MPI::Win::Create_keyval( MPI::Win::NULL_COPY_FN,\n ^~~~~~~~~~~~\n ../../../cxx/util/mtest.cxx:849:22: error: 'NULL_DELETE_FN' is not a member of 'MPI::Win'\n MPI::Win::NULL_DELETE_FN, 0 );\n ^~~~~~~~~~~~~~\n make[3]: *** [Makefile:416: mtest.o] Error 1\n-make[3]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/util'\n+make[3]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/util'\n make[2]: *** [Makefile:989: ../../cxx/util/mtest.o] Error 2\n-make[2]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/io'\n+make[2]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/io'\n \n \n fileerrx\n \n ./cxx/io\n fail\n- Failed to build fileerrx; make[2]: Entering directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/io'\n+ Failed to build fileerrx; make[2]: Entering directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/io'\n (cd ../../cxx/util && make mtest.o)\n-make[3]: Entering directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/util'\n+make[3]: Entering directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/util'\n depbase=`echo mtest.o | sed 's|[^/]*$|.deps/&|;s|\\.o$||'`;\\\n-/usr/bin/mpicxx.openmpi -DHAVE_CONFIG_H -I. -I../../../cxx/util -I../../include -I../../include -I../../../include -Wdate-time -D_FORTIFY_SOURCE=2 -g -O2 -ffile-prefix-map=/build/1st/mpi-testsuite-3.2+dfsg=. -fstack-protector-strong -Wformat -Werror=format-security -MT mtest.o -MD -MP -MF $depbase.Tpo -c -o mtest.o ../../../cxx/util/mtest.cxx &&\\\n+/usr/bin/mpicxx.openmpi -DHAVE_CONFIG_H -I. -I../../../cxx/util -I../../include -I../../include -I../../../include -Wdate-time -D_FORTIFY_SOURCE=2 -g -O2 -ffile-prefix-map=/build/mpi-testsuite-3.2+dfsg/2nd=. -fstack-protector-strong -Wformat -Werror=format-security -MT mtest.o -MD -MP -MF $depbase.Tpo -c -o mtest.o ../../../cxx/util/mtest.cxx &&\\\n mv -f $depbase.Tpo $depbase.Po\n ../../../cxx/util/mtest.cxx: In function 'void* MTestTypeContigInitRecv(MTestDatatype*)':\n ../../../cxx/util/mtest.cxx:189:42: warning: 'int MPI_Type_extent(MPI_Datatype, MPI_Aint*)' is deprecated: MPI_Type_extent is superseded by MPI_Type_get_extent in MPI-2.0 [-Wdeprecated-declarations]\n MPI_Type_extent( mtype->datatype, &size );\n ^\n In file included from ../../../cxx/util/mtest.cxx:7:\n /usr/lib/x86_64-linux-gnu/openmpi/include/mpi.h:1758:20: note: declared here\n@@ -1005244,28 +1005292,28 @@\n ../../../cxx/util/mtest.cxx:848:50: error: 'NULL_COPY_FN' is not a member of 'MPI::Win'\n mem_keyval = MPI::Win::Create_keyval( MPI::Win::NULL_COPY_FN,\n ^~~~~~~~~~~~\n ../../../cxx/util/mtest.cxx:849:22: error: 'NULL_DELETE_FN' is not a member of 'MPI::Win'\n MPI::Win::NULL_DELETE_FN, 0 );\n ^~~~~~~~~~~~~~\n make[3]: *** [Makefile:416: mtest.o] Error 1\n-make[3]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/util'\n+make[3]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/util'\n make[2]: *** [Makefile:989: ../../cxx/util/mtest.o] Error 2\n-make[2]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/io'\n+make[2]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/io'\n \n \n fileinfox\n \n ./cxx/io\n fail\n- Failed to build fileinfox; make[2]: Entering directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/io'\n+ Failed to build fileinfox; make[2]: Entering directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/io'\n (cd ../../cxx/util && make mtest.o)\n-make[3]: Entering directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/util'\n+make[3]: Entering directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/util'\n depbase=`echo mtest.o | sed 's|[^/]*$|.deps/&|;s|\\.o$||'`;\\\n-/usr/bin/mpicxx.openmpi -DHAVE_CONFIG_H -I. -I../../../cxx/util -I../../include -I../../include -I../../../include -Wdate-time -D_FORTIFY_SOURCE=2 -g -O2 -ffile-prefix-map=/build/1st/mpi-testsuite-3.2+dfsg=. -fstack-protector-strong -Wformat -Werror=format-security -MT mtest.o -MD -MP -MF $depbase.Tpo -c -o mtest.o ../../../cxx/util/mtest.cxx &&\\\n+/usr/bin/mpicxx.openmpi -DHAVE_CONFIG_H -I. -I../../../cxx/util -I../../include -I../../include -I../../../include -Wdate-time -D_FORTIFY_SOURCE=2 -g -O2 -ffile-prefix-map=/build/mpi-testsuite-3.2+dfsg/2nd=. -fstack-protector-strong -Wformat -Werror=format-security -MT mtest.o -MD -MP -MF $depbase.Tpo -c -o mtest.o ../../../cxx/util/mtest.cxx &&\\\n mv -f $depbase.Tpo $depbase.Po\n ../../../cxx/util/mtest.cxx: In function 'void* MTestTypeContigInitRecv(MTestDatatype*)':\n ../../../cxx/util/mtest.cxx:189:42: warning: 'int MPI_Type_extent(MPI_Datatype, MPI_Aint*)' is deprecated: MPI_Type_extent is superseded by MPI_Type_get_extent in MPI-2.0 [-Wdeprecated-declarations]\n MPI_Type_extent( mtype->datatype, &size );\n ^\n In file included from ../../../cxx/util/mtest.cxx:7:\n /usr/lib/x86_64-linux-gnu/openmpi/include/mpi.h:1758:20: note: declared here\n@@ -1005297,28 +1005345,28 @@\n ../../../cxx/util/mtest.cxx:848:50: error: 'NULL_COPY_FN' is not a member of 'MPI::Win'\n mem_keyval = MPI::Win::Create_keyval( MPI::Win::NULL_COPY_FN,\n ^~~~~~~~~~~~\n ../../../cxx/util/mtest.cxx:849:22: error: 'NULL_DELETE_FN' is not a member of 'MPI::Win'\n MPI::Win::NULL_DELETE_FN, 0 );\n ^~~~~~~~~~~~~~\n make[3]: *** [Makefile:416: mtest.o] Error 1\n-make[3]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/util'\n+make[3]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/util'\n make[2]: *** [Makefile:989: ../../cxx/util/mtest.o] Error 2\n-make[2]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/io'\n+make[2]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/io'\n \n \n filemiscx\n \n ./cxx/io\n fail\n- Failed to build filemiscx; make[2]: Entering directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/io'\n+ Failed to build filemiscx; make[2]: Entering directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/io'\n (cd ../../cxx/util && make mtest.o)\n-make[3]: Entering directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/util'\n+make[3]: Entering directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/util'\n depbase=`echo mtest.o | sed 's|[^/]*$|.deps/&|;s|\\.o$||'`;\\\n-/usr/bin/mpicxx.openmpi -DHAVE_CONFIG_H -I. -I../../../cxx/util -I../../include -I../../include -I../../../include -Wdate-time -D_FORTIFY_SOURCE=2 -g -O2 -ffile-prefix-map=/build/1st/mpi-testsuite-3.2+dfsg=. -fstack-protector-strong -Wformat -Werror=format-security -MT mtest.o -MD -MP -MF $depbase.Tpo -c -o mtest.o ../../../cxx/util/mtest.cxx &&\\\n+/usr/bin/mpicxx.openmpi -DHAVE_CONFIG_H -I. -I../../../cxx/util -I../../include -I../../include -I../../../include -Wdate-time -D_FORTIFY_SOURCE=2 -g -O2 -ffile-prefix-map=/build/mpi-testsuite-3.2+dfsg/2nd=. -fstack-protector-strong -Wformat -Werror=format-security -MT mtest.o -MD -MP -MF $depbase.Tpo -c -o mtest.o ../../../cxx/util/mtest.cxx &&\\\n mv -f $depbase.Tpo $depbase.Po\n ../../../cxx/util/mtest.cxx: In function 'void* MTestTypeContigInitRecv(MTestDatatype*)':\n ../../../cxx/util/mtest.cxx:189:42: warning: 'int MPI_Type_extent(MPI_Datatype, MPI_Aint*)' is deprecated: MPI_Type_extent is superseded by MPI_Type_get_extent in MPI-2.0 [-Wdeprecated-declarations]\n MPI_Type_extent( mtype->datatype, &size );\n ^\n In file included from ../../../cxx/util/mtest.cxx:7:\n /usr/lib/x86_64-linux-gnu/openmpi/include/mpi.h:1758:20: note: declared here\n@@ -1005350,28 +1005398,28 @@\n ../../../cxx/util/mtest.cxx:848:50: error: 'NULL_COPY_FN' is not a member of 'MPI::Win'\n mem_keyval = MPI::Win::Create_keyval( MPI::Win::NULL_COPY_FN,\n ^~~~~~~~~~~~\n ../../../cxx/util/mtest.cxx:849:22: error: 'NULL_DELETE_FN' is not a member of 'MPI::Win'\n MPI::Win::NULL_DELETE_FN, 0 );\n ^~~~~~~~~~~~~~\n make[3]: *** [Makefile:416: mtest.o] Error 1\n-make[3]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/util'\n+make[3]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/util'\n make[2]: *** [Makefile:989: ../../cxx/util/mtest.o] Error 2\n-make[2]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/io'\n+make[2]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/io'\n \n \n shpositionx\n \n ./cxx/io\n fail\n- Failed to build shpositionx; make[2]: Entering directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/io'\n+ Failed to build shpositionx; make[2]: Entering directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/io'\n (cd ../../cxx/util && make mtest.o)\n-make[3]: Entering directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/util'\n+make[3]: Entering directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/util'\n depbase=`echo mtest.o | sed 's|[^/]*$|.deps/&|;s|\\.o$||'`;\\\n-/usr/bin/mpicxx.openmpi -DHAVE_CONFIG_H -I. -I../../../cxx/util -I../../include -I../../include -I../../../include -Wdate-time -D_FORTIFY_SOURCE=2 -g -O2 -ffile-prefix-map=/build/1st/mpi-testsuite-3.2+dfsg=. -fstack-protector-strong -Wformat -Werror=format-security -MT mtest.o -MD -MP -MF $depbase.Tpo -c -o mtest.o ../../../cxx/util/mtest.cxx &&\\\n+/usr/bin/mpicxx.openmpi -DHAVE_CONFIG_H -I. -I../../../cxx/util -I../../include -I../../include -I../../../include -Wdate-time -D_FORTIFY_SOURCE=2 -g -O2 -ffile-prefix-map=/build/mpi-testsuite-3.2+dfsg/2nd=. -fstack-protector-strong -Wformat -Werror=format-security -MT mtest.o -MD -MP -MF $depbase.Tpo -c -o mtest.o ../../../cxx/util/mtest.cxx &&\\\n mv -f $depbase.Tpo $depbase.Po\n ../../../cxx/util/mtest.cxx: In function 'void* MTestTypeContigInitRecv(MTestDatatype*)':\n ../../../cxx/util/mtest.cxx:189:42: warning: 'int MPI_Type_extent(MPI_Datatype, MPI_Aint*)' is deprecated: MPI_Type_extent is superseded by MPI_Type_get_extent in MPI-2.0 [-Wdeprecated-declarations]\n MPI_Type_extent( mtype->datatype, &size );\n ^\n In file included from ../../../cxx/util/mtest.cxx:7:\n /usr/lib/x86_64-linux-gnu/openmpi/include/mpi.h:1758:20: note: declared here\n@@ -1005403,28 +1005451,28 @@\n ../../../cxx/util/mtest.cxx:848:50: error: 'NULL_COPY_FN' is not a member of 'MPI::Win'\n mem_keyval = MPI::Win::Create_keyval( MPI::Win::NULL_COPY_FN,\n ^~~~~~~~~~~~\n ../../../cxx/util/mtest.cxx:849:22: error: 'NULL_DELETE_FN' is not a member of 'MPI::Win'\n MPI::Win::NULL_DELETE_FN, 0 );\n ^~~~~~~~~~~~~~\n make[3]: *** [Makefile:416: mtest.o] Error 1\n-make[3]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/util'\n+make[3]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/util'\n make[2]: *** [Makefile:989: ../../cxx/util/mtest.o] Error 2\n-make[2]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/io'\n+make[2]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/io'\n \n \n seekavail\n \n ./cxx/io\n fail\n- Failed to build seekavail; make[2]: Entering directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/io'\n+ Failed to build seekavail; make[2]: Entering directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/io'\n (cd ../../cxx/util && make mtest.o)\n-make[3]: Entering directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/util'\n+make[3]: Entering directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/util'\n depbase=`echo mtest.o | sed 's|[^/]*$|.deps/&|;s|\\.o$||'`;\\\n-/usr/bin/mpicxx.openmpi -DHAVE_CONFIG_H -I. -I../../../cxx/util -I../../include -I../../include -I../../../include -Wdate-time -D_FORTIFY_SOURCE=2 -g -O2 -ffile-prefix-map=/build/1st/mpi-testsuite-3.2+dfsg=. -fstack-protector-strong -Wformat -Werror=format-security -MT mtest.o -MD -MP -MF $depbase.Tpo -c -o mtest.o ../../../cxx/util/mtest.cxx &&\\\n+/usr/bin/mpicxx.openmpi -DHAVE_CONFIG_H -I. -I../../../cxx/util -I../../include -I../../include -I../../../include -Wdate-time -D_FORTIFY_SOURCE=2 -g -O2 -ffile-prefix-map=/build/mpi-testsuite-3.2+dfsg/2nd=. -fstack-protector-strong -Wformat -Werror=format-security -MT mtest.o -MD -MP -MF $depbase.Tpo -c -o mtest.o ../../../cxx/util/mtest.cxx &&\\\n mv -f $depbase.Tpo $depbase.Po\n ../../../cxx/util/mtest.cxx: In function 'void* MTestTypeContigInitRecv(MTestDatatype*)':\n ../../../cxx/util/mtest.cxx:189:42: warning: 'int MPI_Type_extent(MPI_Datatype, MPI_Aint*)' is deprecated: MPI_Type_extent is superseded by MPI_Type_get_extent in MPI-2.0 [-Wdeprecated-declarations]\n MPI_Type_extent( mtype->datatype, &size );\n ^\n In file included from ../../../cxx/util/mtest.cxx:7:\n /usr/lib/x86_64-linux-gnu/openmpi/include/mpi.h:1758:20: note: declared here\n@@ -1005456,28 +1005504,28 @@\n ../../../cxx/util/mtest.cxx:848:50: error: 'NULL_COPY_FN' is not a member of 'MPI::Win'\n mem_keyval = MPI::Win::Create_keyval( MPI::Win::NULL_COPY_FN,\n ^~~~~~~~~~~~\n ../../../cxx/util/mtest.cxx:849:22: error: 'NULL_DELETE_FN' is not a member of 'MPI::Win'\n MPI::Win::NULL_DELETE_FN, 0 );\n ^~~~~~~~~~~~~~\n make[3]: *** [Makefile:416: mtest.o] Error 1\n-make[3]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/util'\n+make[3]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/util'\n make[2]: *** [Makefile:989: ../../cxx/util/mtest.o] Error 2\n-make[2]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/io'\n+make[2]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/io'\n \n \n namepubx\n \n ./cxx/spawn\n fail\n- Failed to build namepubx; make[2]: Entering directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/spawn'\n+ Failed to build namepubx; make[2]: Entering directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/spawn'\n (cd ../../cxx/util && make mtest.o)\n-make[3]: Entering directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/util'\n+make[3]: Entering directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/util'\n depbase=`echo mtest.o | sed 's|[^/]*$|.deps/&|;s|\\.o$||'`;\\\n-/usr/bin/mpicxx.openmpi -DHAVE_CONFIG_H -I. -I../../../cxx/util -I../../include -I../../include -I../../../include -Wdate-time -D_FORTIFY_SOURCE=2 -g -O2 -ffile-prefix-map=/build/1st/mpi-testsuite-3.2+dfsg=. -fstack-protector-strong -Wformat -Werror=format-security -MT mtest.o -MD -MP -MF $depbase.Tpo -c -o mtest.o ../../../cxx/util/mtest.cxx &&\\\n+/usr/bin/mpicxx.openmpi -DHAVE_CONFIG_H -I. -I../../../cxx/util -I../../include -I../../include -I../../../include -Wdate-time -D_FORTIFY_SOURCE=2 -g -O2 -ffile-prefix-map=/build/mpi-testsuite-3.2+dfsg/2nd=. -fstack-protector-strong -Wformat -Werror=format-security -MT mtest.o -MD -MP -MF $depbase.Tpo -c -o mtest.o ../../../cxx/util/mtest.cxx &&\\\n mv -f $depbase.Tpo $depbase.Po\n ../../../cxx/util/mtest.cxx: In function 'void* MTestTypeContigInitRecv(MTestDatatype*)':\n ../../../cxx/util/mtest.cxx:189:42: warning: 'int MPI_Type_extent(MPI_Datatype, MPI_Aint*)' is deprecated: MPI_Type_extent is superseded by MPI_Type_get_extent in MPI-2.0 [-Wdeprecated-declarations]\n MPI_Type_extent( mtype->datatype, &size );\n ^\n In file included from ../../../cxx/util/mtest.cxx:7:\n /usr/lib/x86_64-linux-gnu/openmpi/include/mpi.h:1758:20: note: declared here\n@@ -1005509,28 +1005557,28 @@\n ../../../cxx/util/mtest.cxx:848:50: error: 'NULL_COPY_FN' is not a member of 'MPI::Win'\n mem_keyval = MPI::Win::Create_keyval( MPI::Win::NULL_COPY_FN,\n ^~~~~~~~~~~~\n ../../../cxx/util/mtest.cxx:849:22: error: 'NULL_DELETE_FN' is not a member of 'MPI::Win'\n MPI::Win::NULL_DELETE_FN, 0 );\n ^~~~~~~~~~~~~~\n make[3]: *** [Makefile:416: mtest.o] Error 1\n-make[3]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/util'\n+make[3]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/util'\n make[2]: *** [Makefile:673: ../../cxx/util/mtest.o] Error 2\n-make[2]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/spawn'\n+make[2]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/spawn'\n \n \n spawnintrax\n \n ./cxx/spawn\n fail\n- Failed to build spawnintrax; make[2]: Entering directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/spawn'\n+ Failed to build spawnintrax; make[2]: Entering directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/spawn'\n (cd ../../cxx/util && make mtest.o)\n-make[3]: Entering directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/util'\n+make[3]: Entering directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/util'\n depbase=`echo mtest.o | sed 's|[^/]*$|.deps/&|;s|\\.o$||'`;\\\n-/usr/bin/mpicxx.openmpi -DHAVE_CONFIG_H -I. -I../../../cxx/util -I../../include -I../../include -I../../../include -Wdate-time -D_FORTIFY_SOURCE=2 -g -O2 -ffile-prefix-map=/build/1st/mpi-testsuite-3.2+dfsg=. -fstack-protector-strong -Wformat -Werror=format-security -MT mtest.o -MD -MP -MF $depbase.Tpo -c -o mtest.o ../../../cxx/util/mtest.cxx &&\\\n+/usr/bin/mpicxx.openmpi -DHAVE_CONFIG_H -I. -I../../../cxx/util -I../../include -I../../include -I../../../include -Wdate-time -D_FORTIFY_SOURCE=2 -g -O2 -ffile-prefix-map=/build/mpi-testsuite-3.2+dfsg/2nd=. -fstack-protector-strong -Wformat -Werror=format-security -MT mtest.o -MD -MP -MF $depbase.Tpo -c -o mtest.o ../../../cxx/util/mtest.cxx &&\\\n mv -f $depbase.Tpo $depbase.Po\n ../../../cxx/util/mtest.cxx: In function 'void* MTestTypeContigInitRecv(MTestDatatype*)':\n ../../../cxx/util/mtest.cxx:189:42: warning: 'int MPI_Type_extent(MPI_Datatype, MPI_Aint*)' is deprecated: MPI_Type_extent is superseded by MPI_Type_get_extent in MPI-2.0 [-Wdeprecated-declarations]\n MPI_Type_extent( mtype->datatype, &size );\n ^\n In file included from ../../../cxx/util/mtest.cxx:7:\n /usr/lib/x86_64-linux-gnu/openmpi/include/mpi.h:1758:20: note: declared here\n@@ -1005562,31 +1005610,31 @@\n ../../../cxx/util/mtest.cxx:848:50: error: 'NULL_COPY_FN' is not a member of 'MPI::Win'\n mem_keyval = MPI::Win::Create_keyval( MPI::Win::NULL_COPY_FN,\n ^~~~~~~~~~~~\n ../../../cxx/util/mtest.cxx:849:22: error: 'NULL_DELETE_FN' is not a member of 'MPI::Win'\n MPI::Win::NULL_DELETE_FN, 0 );\n ^~~~~~~~~~~~~~\n make[3]: *** [Makefile:416: mtest.o] Error 1\n-make[3]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/util'\n+make[3]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/util'\n make[2]: *** [Makefile:673: ../../cxx/util/mtest.o] Error 2\n-make[2]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/spawn'\n+make[2]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/spawn'\n \n \n spawnintrax\n \n ./cxx/spawn\n fail\n- Failed to build spawnintrax; make[2]: Entering directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/spawn'\n+ Failed to build spawnintrax; make[2]: Entering directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/spawn'\n depbase=`echo spawnintrax.o | sed 's|[^/]*$|.deps/&|;s|\\.o$||'`;\\\n-/usr/bin/mpicxx.openmpi -DHAVE_CONFIG_H -I. -I../../../cxx/spawn -I../../include -I../../include -I../../../include -Wdate-time -D_FORTIFY_SOURCE=2 -g -O2 -ffile-prefix-map=/build/1st/mpi-testsuite-3.2+dfsg=. -fstack-protector-strong -Wformat -Werror=format-security -MT spawnintrax.o -MD -MP -MF $depbase.Tpo -c -o spawnintrax.o ../../../cxx/spawn/spawnintrax.cxx &&\\\n+/usr/bin/mpicxx.openmpi -DHAVE_CONFIG_H -I. -I../../../cxx/spawn -I../../include -I../../include -I../../../include -Wdate-time -D_FORTIFY_SOURCE=2 -g -O2 -ffile-prefix-map=/build/mpi-testsuite-3.2+dfsg/2nd=. -fstack-protector-strong -Wformat -Werror=format-security -MT spawnintrax.o -MD -MP -MF $depbase.Tpo -c -o spawnintrax.o ../../../cxx/spawn/spawnintrax.cxx &&\\\n mv -f $depbase.Tpo $depbase.Po\n (cd ../../cxx/util && make mtest.o)\n-make[3]: Entering directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/util'\n+make[3]: Entering directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/util'\n depbase=`echo mtest.o | sed 's|[^/]*$|.deps/&|;s|\\.o$||'`;\\\n-/usr/bin/mpicxx.openmpi -DHAVE_CONFIG_H -I. -I../../../cxx/util -I../../include -I../../include -I../../../include -Wdate-time -D_FORTIFY_SOURCE=2 -g -O2 -ffile-prefix-map=/build/1st/mpi-testsuite-3.2+dfsg=. -fstack-protector-strong -Wformat -Werror=format-security -MT mtest.o -MD -MP -MF $depbase.Tpo -c -o mtest.o ../../../cxx/util/mtest.cxx &&\\\n+/usr/bin/mpicxx.openmpi -DHAVE_CONFIG_H -I. -I../../../cxx/util -I../../include -I../../include -I../../../include -Wdate-time -D_FORTIFY_SOURCE=2 -g -O2 -ffile-prefix-map=/build/mpi-testsuite-3.2+dfsg/2nd=. -fstack-protector-strong -Wformat -Werror=format-security -MT mtest.o -MD -MP -MF $depbase.Tpo -c -o mtest.o ../../../cxx/util/mtest.cxx &&\\\n mv -f $depbase.Tpo $depbase.Po\n ../../../cxx/util/mtest.cxx: In function 'void* MTestTypeContigInitRecv(MTestDatatype*)':\n ../../../cxx/util/mtest.cxx:189:42: warning: 'int MPI_Type_extent(MPI_Datatype, MPI_Aint*)' is deprecated: MPI_Type_extent is superseded by MPI_Type_get_extent in MPI-2.0 [-Wdeprecated-declarations]\n MPI_Type_extent( mtype->datatype, &size );\n ^\n In file included from ../../../cxx/util/mtest.cxx:7:\n /usr/lib/x86_64-linux-gnu/openmpi/include/mpi.h:1758:20: note: declared here\n@@ -1005618,28 +1005666,28 @@\n ../../../cxx/util/mtest.cxx:848:50: error: 'NULL_COPY_FN' is not a member of 'MPI::Win'\n mem_keyval = MPI::Win::Create_keyval( MPI::Win::NULL_COPY_FN,\n ^~~~~~~~~~~~\n ../../../cxx/util/mtest.cxx:849:22: error: 'NULL_DELETE_FN' is not a member of 'MPI::Win'\n MPI::Win::NULL_DELETE_FN, 0 );\n ^~~~~~~~~~~~~~\n make[3]: *** [Makefile:416: mtest.o] Error 1\n-make[3]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/util'\n+make[3]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/util'\n make[2]: *** [Makefile:673: ../../cxx/util/mtest.o] Error 2\n-make[2]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/spawn'\n+make[2]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/spawn'\n \n \n spawnargvx\n \n ./cxx/spawn\n fail\n- Failed to build spawnargvx; make[2]: Entering directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/spawn'\n+ Failed to build spawnargvx; make[2]: Entering directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/spawn'\n (cd ../../cxx/util && make mtest.o)\n-make[3]: Entering directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/util'\n+make[3]: Entering directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/util'\n depbase=`echo mtest.o | sed 's|[^/]*$|.deps/&|;s|\\.o$||'`;\\\n-/usr/bin/mpicxx.openmpi -DHAVE_CONFIG_H -I. -I../../../cxx/util -I../../include -I../../include -I../../../include -Wdate-time -D_FORTIFY_SOURCE=2 -g -O2 -ffile-prefix-map=/build/1st/mpi-testsuite-3.2+dfsg=. -fstack-protector-strong -Wformat -Werror=format-security -MT mtest.o -MD -MP -MF $depbase.Tpo -c -o mtest.o ../../../cxx/util/mtest.cxx &&\\\n+/usr/bin/mpicxx.openmpi -DHAVE_CONFIG_H -I. -I../../../cxx/util -I../../include -I../../include -I../../../include -Wdate-time -D_FORTIFY_SOURCE=2 -g -O2 -ffile-prefix-map=/build/mpi-testsuite-3.2+dfsg/2nd=. -fstack-protector-strong -Wformat -Werror=format-security -MT mtest.o -MD -MP -MF $depbase.Tpo -c -o mtest.o ../../../cxx/util/mtest.cxx &&\\\n mv -f $depbase.Tpo $depbase.Po\n ../../../cxx/util/mtest.cxx: In function 'void* MTestTypeContigInitRecv(MTestDatatype*)':\n ../../../cxx/util/mtest.cxx:189:42: warning: 'int MPI_Type_extent(MPI_Datatype, MPI_Aint*)' is deprecated: MPI_Type_extent is superseded by MPI_Type_get_extent in MPI-2.0 [-Wdeprecated-declarations]\n MPI_Type_extent( mtype->datatype, &size );\n ^\n In file included from ../../../cxx/util/mtest.cxx:7:\n /usr/lib/x86_64-linux-gnu/openmpi/include/mpi.h:1758:20: note: declared here\n@@ -1005671,28 +1005719,28 @@\n ../../../cxx/util/mtest.cxx:848:50: error: 'NULL_COPY_FN' is not a member of 'MPI::Win'\n mem_keyval = MPI::Win::Create_keyval( MPI::Win::NULL_COPY_FN,\n ^~~~~~~~~~~~\n ../../../cxx/util/mtest.cxx:849:22: error: 'NULL_DELETE_FN' is not a member of 'MPI::Win'\n MPI::Win::NULL_DELETE_FN, 0 );\n ^~~~~~~~~~~~~~\n make[3]: *** [Makefile:416: mtest.o] Error 1\n-make[3]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/util'\n+make[3]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/util'\n make[2]: *** [Makefile:673: ../../cxx/util/mtest.o] Error 2\n-make[2]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/spawn'\n+make[2]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/spawn'\n \n \n selfconaccx\n \n ./cxx/spawn\n fail\n- Failed to build selfconaccx; make[2]: Entering directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/spawn'\n+ Failed to build selfconaccx; make[2]: Entering directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/spawn'\n (cd ../../cxx/util && make mtest.o)\n-make[3]: Entering directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/util'\n+make[3]: Entering directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/util'\n depbase=`echo mtest.o | sed 's|[^/]*$|.deps/&|;s|\\.o$||'`;\\\n-/usr/bin/mpicxx.openmpi -DHAVE_CONFIG_H -I. -I../../../cxx/util -I../../include -I../../include -I../../../include -Wdate-time -D_FORTIFY_SOURCE=2 -g -O2 -ffile-prefix-map=/build/1st/mpi-testsuite-3.2+dfsg=. -fstack-protector-strong -Wformat -Werror=format-security -MT mtest.o -MD -MP -MF $depbase.Tpo -c -o mtest.o ../../../cxx/util/mtest.cxx &&\\\n+/usr/bin/mpicxx.openmpi -DHAVE_CONFIG_H -I. -I../../../cxx/util -I../../include -I../../include -I../../../include -Wdate-time -D_FORTIFY_SOURCE=2 -g -O2 -ffile-prefix-map=/build/mpi-testsuite-3.2+dfsg/2nd=. -fstack-protector-strong -Wformat -Werror=format-security -MT mtest.o -MD -MP -MF $depbase.Tpo -c -o mtest.o ../../../cxx/util/mtest.cxx &&\\\n mv -f $depbase.Tpo $depbase.Po\n ../../../cxx/util/mtest.cxx: In function 'void* MTestTypeContigInitRecv(MTestDatatype*)':\n ../../../cxx/util/mtest.cxx:189:42: warning: 'int MPI_Type_extent(MPI_Datatype, MPI_Aint*)' is deprecated: MPI_Type_extent is superseded by MPI_Type_get_extent in MPI-2.0 [-Wdeprecated-declarations]\n MPI_Type_extent( mtype->datatype, &size );\n ^\n In file included from ../../../cxx/util/mtest.cxx:7:\n /usr/lib/x86_64-linux-gnu/openmpi/include/mpi.h:1758:20: note: declared here\n@@ -1005724,28 +1005772,28 @@\n ../../../cxx/util/mtest.cxx:848:50: error: 'NULL_COPY_FN' is not a member of 'MPI::Win'\n mem_keyval = MPI::Win::Create_keyval( MPI::Win::NULL_COPY_FN,\n ^~~~~~~~~~~~\n ../../../cxx/util/mtest.cxx:849:22: error: 'NULL_DELETE_FN' is not a member of 'MPI::Win'\n MPI::Win::NULL_DELETE_FN, 0 );\n ^~~~~~~~~~~~~~\n make[3]: *** [Makefile:416: mtest.o] Error 1\n-make[3]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/util'\n+make[3]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/util'\n make[2]: *** [Makefile:673: ../../cxx/util/mtest.o] Error 2\n-make[2]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/spawn'\n+make[2]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/spawn'\n \n \n winnamex\n \n ./cxx/rma\n fail\n- Failed to build winnamex; make[2]: Entering directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/rma'\n+ Failed to build winnamex; make[2]: Entering directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/rma'\n (cd ../../cxx/util && make mtest.o)\n-make[3]: Entering directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/util'\n+make[3]: Entering directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/util'\n depbase=`echo mtest.o | sed 's|[^/]*$|.deps/&|;s|\\.o$||'`;\\\n-/usr/bin/mpicxx.openmpi -DHAVE_CONFIG_H -I. -I../../../cxx/util -I../../include -I../../include -I../../../include -Wdate-time -D_FORTIFY_SOURCE=2 -g -O2 -ffile-prefix-map=/build/1st/mpi-testsuite-3.2+dfsg=. -fstack-protector-strong -Wformat -Werror=format-security -MT mtest.o -MD -MP -MF $depbase.Tpo -c -o mtest.o ../../../cxx/util/mtest.cxx &&\\\n+/usr/bin/mpicxx.openmpi -DHAVE_CONFIG_H -I. -I../../../cxx/util -I../../include -I../../include -I../../../include -Wdate-time -D_FORTIFY_SOURCE=2 -g -O2 -ffile-prefix-map=/build/mpi-testsuite-3.2+dfsg/2nd=. -fstack-protector-strong -Wformat -Werror=format-security -MT mtest.o -MD -MP -MF $depbase.Tpo -c -o mtest.o ../../../cxx/util/mtest.cxx &&\\\n mv -f $depbase.Tpo $depbase.Po\n ../../../cxx/util/mtest.cxx: In function 'void* MTestTypeContigInitRecv(MTestDatatype*)':\n ../../../cxx/util/mtest.cxx:189:42: warning: 'int MPI_Type_extent(MPI_Datatype, MPI_Aint*)' is deprecated: MPI_Type_extent is superseded by MPI_Type_get_extent in MPI-2.0 [-Wdeprecated-declarations]\n MPI_Type_extent( mtype->datatype, &size );\n ^\n In file included from ../../../cxx/util/mtest.cxx:7:\n /usr/lib/x86_64-linux-gnu/openmpi/include/mpi.h:1758:20: note: declared here\n@@ -1005777,28 +1005825,28 @@\n ../../../cxx/util/mtest.cxx:848:50: error: 'NULL_COPY_FN' is not a member of 'MPI::Win'\n mem_keyval = MPI::Win::Create_keyval( MPI::Win::NULL_COPY_FN,\n ^~~~~~~~~~~~\n ../../../cxx/util/mtest.cxx:849:22: error: 'NULL_DELETE_FN' is not a member of 'MPI::Win'\n MPI::Win::NULL_DELETE_FN, 0 );\n ^~~~~~~~~~~~~~\n make[3]: *** [Makefile:416: mtest.o] Error 1\n-make[3]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/util'\n+make[3]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/util'\n make[2]: *** [Makefile:700: ../../cxx/util/mtest.o] Error 2\n-make[2]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/rma'\n+make[2]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/rma'\n \n \n wincallx\n \n ./cxx/rma\n fail\n- Failed to build wincallx; make[2]: Entering directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/rma'\n+ Failed to build wincallx; make[2]: Entering directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/rma'\n (cd ../../cxx/util && make mtest.o)\n-make[3]: Entering directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/util'\n+make[3]: Entering directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/util'\n depbase=`echo mtest.o | sed 's|[^/]*$|.deps/&|;s|\\.o$||'`;\\\n-/usr/bin/mpicxx.openmpi -DHAVE_CONFIG_H -I. -I../../../cxx/util -I../../include -I../../include -I../../../include -Wdate-time -D_FORTIFY_SOURCE=2 -g -O2 -ffile-prefix-map=/build/1st/mpi-testsuite-3.2+dfsg=. -fstack-protector-strong -Wformat -Werror=format-security -MT mtest.o -MD -MP -MF $depbase.Tpo -c -o mtest.o ../../../cxx/util/mtest.cxx &&\\\n+/usr/bin/mpicxx.openmpi -DHAVE_CONFIG_H -I. -I../../../cxx/util -I../../include -I../../include -I../../../include -Wdate-time -D_FORTIFY_SOURCE=2 -g -O2 -ffile-prefix-map=/build/mpi-testsuite-3.2+dfsg/2nd=. -fstack-protector-strong -Wformat -Werror=format-security -MT mtest.o -MD -MP -MF $depbase.Tpo -c -o mtest.o ../../../cxx/util/mtest.cxx &&\\\n mv -f $depbase.Tpo $depbase.Po\n ../../../cxx/util/mtest.cxx: In function 'void* MTestTypeContigInitRecv(MTestDatatype*)':\n ../../../cxx/util/mtest.cxx:189:42: warning: 'int MPI_Type_extent(MPI_Datatype, MPI_Aint*)' is deprecated: MPI_Type_extent is superseded by MPI_Type_get_extent in MPI-2.0 [-Wdeprecated-declarations]\n MPI_Type_extent( mtype->datatype, &size );\n ^\n In file included from ../../../cxx/util/mtest.cxx:7:\n /usr/lib/x86_64-linux-gnu/openmpi/include/mpi.h:1758:20: note: declared here\n@@ -1005830,28 +1005878,28 @@\n ../../../cxx/util/mtest.cxx:848:50: error: 'NULL_COPY_FN' is not a member of 'MPI::Win'\n mem_keyval = MPI::Win::Create_keyval( MPI::Win::NULL_COPY_FN,\n ^~~~~~~~~~~~\n ../../../cxx/util/mtest.cxx:849:22: error: 'NULL_DELETE_FN' is not a member of 'MPI::Win'\n MPI::Win::NULL_DELETE_FN, 0 );\n ^~~~~~~~~~~~~~\n make[3]: *** [Makefile:416: mtest.o] Error 1\n-make[3]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/util'\n+make[3]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/util'\n make[2]: *** [Makefile:700: ../../cxx/util/mtest.o] Error 2\n-make[2]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/rma'\n+make[2]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/rma'\n \n \n getgroupx\n \n ./cxx/rma\n fail\n- Failed to build getgroupx; make[2]: Entering directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/rma'\n+ Failed to build getgroupx; make[2]: Entering directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/rma'\n (cd ../../cxx/util && make mtest.o)\n-make[3]: Entering directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/util'\n+make[3]: Entering directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/util'\n depbase=`echo mtest.o | sed 's|[^/]*$|.deps/&|;s|\\.o$||'`;\\\n-/usr/bin/mpicxx.openmpi -DHAVE_CONFIG_H -I. -I../../../cxx/util -I../../include -I../../include -I../../../include -Wdate-time -D_FORTIFY_SOURCE=2 -g -O2 -ffile-prefix-map=/build/1st/mpi-testsuite-3.2+dfsg=. -fstack-protector-strong -Wformat -Werror=format-security -MT mtest.o -MD -MP -MF $depbase.Tpo -c -o mtest.o ../../../cxx/util/mtest.cxx &&\\\n+/usr/bin/mpicxx.openmpi -DHAVE_CONFIG_H -I. -I../../../cxx/util -I../../include -I../../include -I../../../include -Wdate-time -D_FORTIFY_SOURCE=2 -g -O2 -ffile-prefix-map=/build/mpi-testsuite-3.2+dfsg/2nd=. -fstack-protector-strong -Wformat -Werror=format-security -MT mtest.o -MD -MP -MF $depbase.Tpo -c -o mtest.o ../../../cxx/util/mtest.cxx &&\\\n mv -f $depbase.Tpo $depbase.Po\n ../../../cxx/util/mtest.cxx: In function 'void* MTestTypeContigInitRecv(MTestDatatype*)':\n ../../../cxx/util/mtest.cxx:189:42: warning: 'int MPI_Type_extent(MPI_Datatype, MPI_Aint*)' is deprecated: MPI_Type_extent is superseded by MPI_Type_get_extent in MPI-2.0 [-Wdeprecated-declarations]\n MPI_Type_extent( mtype->datatype, &size );\n ^\n In file included from ../../../cxx/util/mtest.cxx:7:\n /usr/lib/x86_64-linux-gnu/openmpi/include/mpi.h:1758:20: note: declared here\n@@ -1005883,28 +1005931,28 @@\n ../../../cxx/util/mtest.cxx:848:50: error: 'NULL_COPY_FN' is not a member of 'MPI::Win'\n mem_keyval = MPI::Win::Create_keyval( MPI::Win::NULL_COPY_FN,\n ^~~~~~~~~~~~\n ../../../cxx/util/mtest.cxx:849:22: error: 'NULL_DELETE_FN' is not a member of 'MPI::Win'\n MPI::Win::NULL_DELETE_FN, 0 );\n ^~~~~~~~~~~~~~\n make[3]: *** [Makefile:416: mtest.o] Error 1\n-make[3]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/util'\n+make[3]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/util'\n make[2]: *** [Makefile:700: ../../cxx/util/mtest.o] Error 2\n-make[2]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/rma'\n+make[2]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/rma'\n \n \n winfencex\n \n ./cxx/rma\n fail\n- Failed to build winfencex; make[2]: Entering directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/rma'\n+ Failed to build winfencex; make[2]: Entering directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/rma'\n (cd ../../cxx/util && make mtest.o)\n-make[3]: Entering directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/util'\n+make[3]: Entering directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/util'\n depbase=`echo mtest.o | sed 's|[^/]*$|.deps/&|;s|\\.o$||'`;\\\n-/usr/bin/mpicxx.openmpi -DHAVE_CONFIG_H -I. -I../../../cxx/util -I../../include -I../../include -I../../../include -Wdate-time -D_FORTIFY_SOURCE=2 -g -O2 -ffile-prefix-map=/build/1st/mpi-testsuite-3.2+dfsg=. -fstack-protector-strong -Wformat -Werror=format-security -MT mtest.o -MD -MP -MF $depbase.Tpo -c -o mtest.o ../../../cxx/util/mtest.cxx &&\\\n+/usr/bin/mpicxx.openmpi -DHAVE_CONFIG_H -I. -I../../../cxx/util -I../../include -I../../include -I../../../include -Wdate-time -D_FORTIFY_SOURCE=2 -g -O2 -ffile-prefix-map=/build/mpi-testsuite-3.2+dfsg/2nd=. -fstack-protector-strong -Wformat -Werror=format-security -MT mtest.o -MD -MP -MF $depbase.Tpo -c -o mtest.o ../../../cxx/util/mtest.cxx &&\\\n mv -f $depbase.Tpo $depbase.Po\n ../../../cxx/util/mtest.cxx: In function 'void* MTestTypeContigInitRecv(MTestDatatype*)':\n ../../../cxx/util/mtest.cxx:189:42: warning: 'int MPI_Type_extent(MPI_Datatype, MPI_Aint*)' is deprecated: MPI_Type_extent is superseded by MPI_Type_get_extent in MPI-2.0 [-Wdeprecated-declarations]\n MPI_Type_extent( mtype->datatype, &size );\n ^\n In file included from ../../../cxx/util/mtest.cxx:7:\n /usr/lib/x86_64-linux-gnu/openmpi/include/mpi.h:1758:20: note: declared here\n@@ -1005936,28 +1005984,28 @@\n ../../../cxx/util/mtest.cxx:848:50: error: 'NULL_COPY_FN' is not a member of 'MPI::Win'\n mem_keyval = MPI::Win::Create_keyval( MPI::Win::NULL_COPY_FN,\n ^~~~~~~~~~~~\n ../../../cxx/util/mtest.cxx:849:22: error: 'NULL_DELETE_FN' is not a member of 'MPI::Win'\n MPI::Win::NULL_DELETE_FN, 0 );\n ^~~~~~~~~~~~~~\n make[3]: *** [Makefile:416: mtest.o] Error 1\n-make[3]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/util'\n+make[3]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/util'\n make[2]: *** [Makefile:700: ../../cxx/util/mtest.o] Error 2\n-make[2]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/rma'\n+make[2]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/rma'\n \n \n winscale1x\n \n ./cxx/rma\n fail\n- Failed to build winscale1x; make[2]: Entering directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/rma'\n+ Failed to build winscale1x; make[2]: Entering directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/rma'\n (cd ../../cxx/util && make mtest.o)\n-make[3]: Entering directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/util'\n+make[3]: Entering directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/util'\n depbase=`echo mtest.o | sed 's|[^/]*$|.deps/&|;s|\\.o$||'`;\\\n-/usr/bin/mpicxx.openmpi -DHAVE_CONFIG_H -I. -I../../../cxx/util -I../../include -I../../include -I../../../include -Wdate-time -D_FORTIFY_SOURCE=2 -g -O2 -ffile-prefix-map=/build/1st/mpi-testsuite-3.2+dfsg=. -fstack-protector-strong -Wformat -Werror=format-security -MT mtest.o -MD -MP -MF $depbase.Tpo -c -o mtest.o ../../../cxx/util/mtest.cxx &&\\\n+/usr/bin/mpicxx.openmpi -DHAVE_CONFIG_H -I. -I../../../cxx/util -I../../include -I../../include -I../../../include -Wdate-time -D_FORTIFY_SOURCE=2 -g -O2 -ffile-prefix-map=/build/mpi-testsuite-3.2+dfsg/2nd=. -fstack-protector-strong -Wformat -Werror=format-security -MT mtest.o -MD -MP -MF $depbase.Tpo -c -o mtest.o ../../../cxx/util/mtest.cxx &&\\\n mv -f $depbase.Tpo $depbase.Po\n ../../../cxx/util/mtest.cxx: In function 'void* MTestTypeContigInitRecv(MTestDatatype*)':\n ../../../cxx/util/mtest.cxx:189:42: warning: 'int MPI_Type_extent(MPI_Datatype, MPI_Aint*)' is deprecated: MPI_Type_extent is superseded by MPI_Type_get_extent in MPI-2.0 [-Wdeprecated-declarations]\n MPI_Type_extent( mtype->datatype, &size );\n ^\n In file included from ../../../cxx/util/mtest.cxx:7:\n /usr/lib/x86_64-linux-gnu/openmpi/include/mpi.h:1758:20: note: declared here\n@@ -1005989,28 +1006037,28 @@\n ../../../cxx/util/mtest.cxx:848:50: error: 'NULL_COPY_FN' is not a member of 'MPI::Win'\n mem_keyval = MPI::Win::Create_keyval( MPI::Win::NULL_COPY_FN,\n ^~~~~~~~~~~~\n ../../../cxx/util/mtest.cxx:849:22: error: 'NULL_DELETE_FN' is not a member of 'MPI::Win'\n MPI::Win::NULL_DELETE_FN, 0 );\n ^~~~~~~~~~~~~~\n make[3]: *** [Makefile:416: mtest.o] Error 1\n-make[3]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/util'\n+make[3]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/util'\n make[2]: *** [Makefile:700: ../../cxx/util/mtest.o] Error 2\n-make[2]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/rma'\n+make[2]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/rma'\n \n \n winscale2x\n \n ./cxx/rma\n fail\n- Failed to build winscale2x; make[2]: Entering directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/rma'\n+ Failed to build winscale2x; make[2]: Entering directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/rma'\n (cd ../../cxx/util && make mtest.o)\n-make[3]: Entering directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/util'\n+make[3]: Entering directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/util'\n depbase=`echo mtest.o | sed 's|[^/]*$|.deps/&|;s|\\.o$||'`;\\\n-/usr/bin/mpicxx.openmpi -DHAVE_CONFIG_H -I. -I../../../cxx/util -I../../include -I../../include -I../../../include -Wdate-time -D_FORTIFY_SOURCE=2 -g -O2 -ffile-prefix-map=/build/1st/mpi-testsuite-3.2+dfsg=. -fstack-protector-strong -Wformat -Werror=format-security -MT mtest.o -MD -MP -MF $depbase.Tpo -c -o mtest.o ../../../cxx/util/mtest.cxx &&\\\n+/usr/bin/mpicxx.openmpi -DHAVE_CONFIG_H -I. -I../../../cxx/util -I../../include -I../../include -I../../../include -Wdate-time -D_FORTIFY_SOURCE=2 -g -O2 -ffile-prefix-map=/build/mpi-testsuite-3.2+dfsg/2nd=. -fstack-protector-strong -Wformat -Werror=format-security -MT mtest.o -MD -MP -MF $depbase.Tpo -c -o mtest.o ../../../cxx/util/mtest.cxx &&\\\n mv -f $depbase.Tpo $depbase.Po\n ../../../cxx/util/mtest.cxx: In function 'void* MTestTypeContigInitRecv(MTestDatatype*)':\n ../../../cxx/util/mtest.cxx:189:42: warning: 'int MPI_Type_extent(MPI_Datatype, MPI_Aint*)' is deprecated: MPI_Type_extent is superseded by MPI_Type_get_extent in MPI-2.0 [-Wdeprecated-declarations]\n MPI_Type_extent( mtype->datatype, &size );\n ^\n In file included from ../../../cxx/util/mtest.cxx:7:\n /usr/lib/x86_64-linux-gnu/openmpi/include/mpi.h:1758:20: note: declared here\n@@ -1006042,36 +1006090,36 @@\n ../../../cxx/util/mtest.cxx:848:50: error: 'NULL_COPY_FN' is not a member of 'MPI::Win'\n mem_keyval = MPI::Win::Create_keyval( MPI::Win::NULL_COPY_FN,\n ^~~~~~~~~~~~\n ../../../cxx/util/mtest.cxx:849:22: error: 'NULL_DELETE_FN' is not a member of 'MPI::Win'\n MPI::Win::NULL_DELETE_FN, 0 );\n ^~~~~~~~~~~~~~\n make[3]: *** [Makefile:416: mtest.o] Error 1\n-make[3]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/util'\n+make[3]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/util'\n make[2]: *** [Makefile:700: ../../cxx/util/mtest.o] Error 2\n-make[2]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/rma'\n+make[2]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/rma'\n \n \n fkeyvalwinx\n \n ./cxx/rma\n fail\n- Failed to build fkeyvalwinx; make[2]: Entering directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/rma'\n+ Failed to build fkeyvalwinx; make[2]: Entering directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/rma'\n depbase=`echo fkeyvalwinx.o | sed 's|[^/]*$|.deps/&|;s|\\.o$||'`;\\\n-/usr/bin/mpicxx.openmpi -DHAVE_CONFIG_H -I. -I../../../cxx/rma -I../../include -I../../include -I../../../include -Wdate-time -D_FORTIFY_SOURCE=2 -g -O2 -ffile-prefix-map=/build/1st/mpi-testsuite-3.2+dfsg=. -fstack-protector-strong -Wformat -Werror=format-security -MT fkeyvalwinx.o -MD -MP -MF $depbase.Tpo -c -o fkeyvalwinx.o ../../../cxx/rma/fkeyvalwinx.cxx &&\\\n+/usr/bin/mpicxx.openmpi -DHAVE_CONFIG_H -I. -I../../../cxx/rma -I../../include -I../../include -I../../../include -Wdate-time -D_FORTIFY_SOURCE=2 -g -O2 -ffile-prefix-map=/build/mpi-testsuite-3.2+dfsg/2nd=. -fstack-protector-strong -Wformat -Werror=format-security -MT fkeyvalwinx.o -MD -MP -MF $depbase.Tpo -c -o fkeyvalwinx.o ../../../cxx/rma/fkeyvalwinx.cxx &&\\\n mv -f $depbase.Tpo $depbase.Po\n ../../../cxx/rma/fkeyvalwinx.cxx: In function 'int main(int, char**)':\n ../../../cxx/rma/fkeyvalwinx.cxx:73:13: error: 'NULL_COPY_FN' is not a member of 'MPI::Win'\n MPI::Win::NULL_COPY_FN, MPI::Win::NULL_DELETE_FN, (void *)0 );\n ^~~~~~~~~~~~\n ../../../cxx/rma/fkeyvalwinx.cxx:73:37: error: 'NULL_DELETE_FN' is not a member of 'MPI::Win'\n MPI::Win::NULL_COPY_FN, MPI::Win::NULL_DELETE_FN, (void *)0 );\n ^~~~~~~~~~~~~~\n make[2]: *** [Makefile:480: fkeyvalwinx.o] Error 1\n-make[2]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/rma'\n+make[2]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/rma'\n \n \n attrmpi1f90\n 1\n ./f90/attr\n pass\n \n@@ -1006166,28 +1006214,28 @@\n pass\n \n \n alltoallvf90\n \n ./f90/coll\n fail\n- Failed to build alltoallvf90; make[2]: Entering directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/f90/coll'\n-/usr/bin/mpif90.openmpi -g -O2 -ffile-prefix-map=/build/1st/mpi-testsuite-3.2+dfsg=. -fstack-protector-strong -fcray-pointer -c -o alltoallvf90.o ../../../f90/coll/alltoallvf90.f90\n+ Failed to build alltoallvf90; make[2]: Entering directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/f90/coll'\n+/usr/bin/mpif90.openmpi -g -O2 -ffile-prefix-map=/build/mpi-testsuite-3.2+dfsg/2nd=. -fstack-protector-strong -fcray-pointer -c -o alltoallvf90.o ../../../f90/coll/alltoallvf90.f90\n ../../../f90/coll/alltoallvf90.f90:50:64:\n \n & rbuf, rcounts, rdispls, rtypes, comm, ierr )\n 1\n Error: There is no specific subroutine for the generic 'mpi_alltoallv' at (1)\n ../../../f90/coll/alltoallvf90.f90:103:59:\n \n & rbuf, rcounts, rdispls, rtypes, comm, ierr )\n 1\n Error: There is no specific subroutine for the generic 'mpi_alltoallv' at (1)\n make[2]: *** [Makefile:552: alltoallvf90.o] Error 1\n-make[2]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/f90/coll'\n+make[2]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/f90/coll'\n \n \n inplacef90\n 2\n ./f90/coll\n pass\n \n@@ -1006204,75 +1006252,75 @@\n pass\n \n \n split_typef90\n \n ./f90/coll\n fail\n- Failed to build split_typef90; make[2]: Entering directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/f90/coll'\n-/usr/bin/mpif90.openmpi -g -O2 -ffile-prefix-map=/build/1st/mpi-testsuite-3.2+dfsg=. -fstack-protector-strong -fcray-pointer -c -o split_typef90.o ../../../f90/coll/split_typef90.f90\n+ Failed to build split_typef90; make[2]: Entering directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/f90/coll'\n+/usr/bin/mpif90.openmpi -g -O2 -ffile-prefix-map=/build/mpi-testsuite-3.2+dfsg/2nd=. -fstack-protector-strong -fcray-pointer -c -o split_typef90.o ../../../f90/coll/split_typef90.f90\n ../../../f90/coll/split_typef90.f90:39:59:\n \n & rbuf, rcounts, rdispls, rtypes, newcomm, ierr )\n 1\n Error: There is no specific subroutine for the generic 'mpi_alltoallv' at (1)\n make[2]: *** [Makefile:552: split_typef90.o] Error 1\n-make[2]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/f90/coll'\n+make[2]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/f90/coll'\n \n \n nonblockingf90\n 2\n ./f90/coll\n pass\n \n \n vw_inplacef90\n \n ./f90/coll\n fail\n- Failed to build vw_inplacef90; make[2]: Entering directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/f90/coll'\n-/usr/bin/mpif90.openmpi -g -O2 -ffile-prefix-map=/build/1st/mpi-testsuite-3.2+dfsg=. -fstack-protector-strong -fcray-pointer -c -o vw_inplacef90.o ../../../f90/coll/vw_inplacef90.f90\n+ Failed to build vw_inplacef90; make[2]: Entering directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/f90/coll'\n+/usr/bin/mpif90.openmpi -g -O2 -ffile-prefix-map=/build/mpi-testsuite-3.2+dfsg/2nd=. -fstack-protector-strong -fcray-pointer -c -o vw_inplacef90.o ../../../f90/coll/vw_inplacef90.f90\n ../../../f90/coll/vw_inplacef90.f90:63:40:\n \n & comm, ierr )\n 1\n Error: There is no specific subroutine for the generic 'mpi_alltoallv' at (1)\n ../../../f90/coll/vw_inplacef90.f90:92:40:\n \n & comm, ierr )\n 1\n Error: There is no specific subroutine for the generic 'mpi_alltoallw' at (1)\n make[2]: *** [Makefile:552: vw_inplacef90.o] Error 1\n-make[2]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/f90/coll'\n+make[2]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/f90/coll'\n \n \n red_scat_blockf90\n 2\n ./f90/coll\n pass\n \n \n nonblocking_inpf90\n \n ./f90/coll\n fail\n- Failed to build nonblocking_inpf90; make[2]: Entering directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/f90/coll'\n-/usr/bin/mpif90.openmpi -g -O2 -ffile-prefix-map=/build/1st/mpi-testsuite-3.2+dfsg=. -fstack-protector-strong -fcray-pointer -c -o nonblocking_inpf90.o ../../../f90/coll/nonblocking_inpf90.f90\n+ Failed to build nonblocking_inpf90; make[2]: Entering directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/f90/coll'\n+/usr/bin/mpif90.openmpi -g -O2 -ffile-prefix-map=/build/mpi-testsuite-3.2+dfsg/2nd=. -fstack-protector-strong -fcray-pointer -c -o nonblocking_inpf90.o ../../../f90/coll/nonblocking_inpf90.f90\n ../../../f90/coll/nonblocking_inpf90.f90:58:47:\n \n & comm, req, ierr )\n 1\n Error: There is no specific subroutine for the generic 'mpi_ialltoallv' at (1)\n ../../../f90/coll/nonblocking_inpf90.f90:87:47:\n \n & comm, req, ierr )\n 1\n Error: There is no specific subroutine for the generic 'mpi_ialltoallw' at (1)\n make[2]: *** [Makefile:552: nonblocking_inpf90.o] Error 1\n-make[2]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/f90/coll'\n+make[2]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/f90/coll'\n \n \n commnamef90\n 2\n ./f90/comm\n pass\n \n@@ -1006324,15 +1006372,24 @@\n ./f90/init\n pass\n \n \n iwriteatf90\n 2\n ./f90/io\n- pass\n+ fail\n+ --------------------------------------------------------------------------\n+The user-provided time limit for job execution has been reached:\n+\n+ Timeout: 180 seconds\n+\n+The job will now be aborted. Please check your code and/or\n+adjust/remove the job execution time limit (as specified by --timeout\n+command line option or MPIEXEC_TIMEOUT environment variable).\n+--------------------------------------------------------------------------\n \n \n iwritef90\n 2\n ./f90/io\n pass\n \n@@ -1006387,21 +1006444,39 @@\n ./f90/io\n pass\n \n \n writeatallbef90\n 2\n ./f90/io\n- pass\n+ fail\n+ --------------------------------------------------------------------------\n+The user-provided time limit for job execution has been reached:\n+\n+ Timeout: 180 seconds\n+\n+The job will now be aborted. Please check your code and/or\n+adjust/remove the job execution time limit (as specified by --timeout\n+command line option or MPIEXEC_TIMEOUT environment variable).\n+--------------------------------------------------------------------------\n \n \n writeallbef90\n 2\n ./f90/io\n- pass\n+ fail\n+ --------------------------------------------------------------------------\n+The user-provided time limit for job execution has been reached:\n+\n+ Timeout: 180 seconds\n+\n+The job will now be aborted. Please check your code and/or\n+adjust/remove the job execution time limit (as specified by --timeout\n+command line option or MPIEXEC_TIMEOUT environment variable).\n+--------------------------------------------------------------------------\n \n \n writeordbef90\n 2\n ./f90/io\n fail\n --------------------------------------------------------------------------\n@@ -1006432,46 +1006507,52 @@\n ./f90/io\n pass\n \n \n atomicityf90\n 2\n ./f90/io\n- fail\n- (noncontig)Process 1 : readbuf( 1939 ) is 0 , should be 10\n---------------------------------------------------------------------------\n-MPI_ABORT was invoked on rank 1 in communicator MPI_COMM_WORLD\n-with errorcode 1.\n-\n-NOTE: invoking MPI_ABORT causes Open MPI to kill all MPI processes.\n-You may or may not see output from other processes, depending on\n-exactly when Open MPI kills them.\n---------------------------------------------------------------------------\n+ pass\n \n \n miscfilef90\n 2\n ./f90/io\n fail\n- File pointer position = 275 , should be 20\n- File pointer position = 275 , should be 20\n- File pointer position in bytes = 2680 , should be 1000\n- File pointer position in bytes = 2680 , should be 1000\n- Found 4 errors\n+ [i-capture-the-hostname:00913] PMIX ERROR: PMIX TEMPORARILY UNAVAILABLE in file ../../../../../../src/mca/ptl/tcp/ptl_tcp.c at line 790\n+[i-capture-the-hostname:00913] OPAL ERROR: Unreachable in file ext2x_client.c at line 109\n+*** An error occurred in MPI_Init\n+*** on a NULL communicator\n+*** MPI_ERRORS_ARE_FATAL (processes in this communicator will now abort,\n+*** and potentially your MPI job)\n+[i-capture-the-hostname:00913] Local abort before MPI_INIT completed completed successfully, but am not able to aggregate error messages, and not able to guarantee that all other processes were killed!\n+--------------------------------------------------------------------------\n+Primary job terminated normally, but 1 process returned\n+a non-zero exit code. Per user-direction, the job has been aborted.\n+--------------------------------------------------------------------------\n+[i-capture-the-hostname:00880] PMIX ERROR: UNREACHABLE in file ../../../../../../src/mca/ptl/tcp/ptl_tcp_component.c at line 1423\n+[i-capture-the-hostname:00880] PMIX ERROR: UNREACHABLE in file ../../../../../../src/mca/ptl/tcp/ptl_tcp_component.c at line 1423\n+--------------------------------------------------------------------------\n+mpiexec.openmpi detected that one or more processes exited with non-zero status, thus causing\n+the job to be terminated. The first process to do so was:\n+\n+ Process name: [[36385,1],1]\n+ Exit code: 1\n+--------------------------------------------------------------------------\n \n \n setviewcurf90\n 2\n ./f90/io\n fail\n 0 : buf(1) = -1\n+ Found 2 errors\n 1 : buf(1) = -1\n mca_fbtl_posix_preadv: error in (p)readv:Invalid argument\n-mca_fbtl_posix_preadv: error in (p)readv:Invalid argument\n- Found 2 errors\n+mca_fbtl_posix_preadv: error in (p)readv:Invalid argument\n \n \n c2f2ciof90\n 1\n ./f90/io\n pass\n \n@@ -1006500,16 +1006581,16 @@\n pass\n \n \n allpairf90\n \n ./f90/pt2pt\n fail\n- Failed to build allpairf90; make[2]: Entering directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/f90/pt2pt'\n-/usr/bin/mpif90.openmpi -g -O2 -ffile-prefix-map=/build/1st/mpi-testsuite-3.2+dfsg=. -fstack-protector-strong -fcray-pointer -c -o allpairf90.o ../../../f90/pt2pt/allpairf90.f90\n+ Failed to build allpairf90; make[2]: Entering directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/f90/pt2pt'\n+/usr/bin/mpif90.openmpi -g -O2 -ffile-prefix-map=/build/mpi-testsuite-3.2+dfsg/2nd=. -fstack-protector-strong -fcray-pointer -c -o allpairf90.o ../../../f90/pt2pt/allpairf90.f90\n ../../../f90/pt2pt/allpairf90.f90:392:64:\n \n call MPI_Waitany(2, requests, index, statuses, ierr)\n 1\n Error: There is no specific subroutine for the generic 'mpi_waitany' at (1)\n ../../../f90/pt2pt/allpairf90.f90:500:50:\n \n@@ -1006518,46 +1006599,46 @@\n Error: There is no specific subroutine for the generic 'mpi_testany' at (1)\n ../../../f90/pt2pt/allpairf90.f90:744:50:\n \n & statuses(1,1), ierr)\n 1\n Error: There is no specific subroutine for the generic 'mpi_testany' at (1)\n make[2]: *** [Makefile:462: allpairf90.o] Error 1\n-make[2]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/f90/pt2pt'\n+make[2]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/f90/pt2pt'\n \n \n mprobef90\n 2\n ./f90/pt2pt\n fail\n s1(MPI_ERROR) != MPI_ERR_DIMS at T4 Mprobe().\n- s2(MPI_ERROR) != MPI_ERR_OTHER at T4 Mrecv().\n- s1(MPI_ERROR) != MPI_ERR_DIMS at T5 Mprobe().\n- s2(MPI_ERROR) != MPI_ERR_OTHER at T5 Imrecv().\n- s1(MPI_ERROR) != MPI_ERR_DIMS at T6 Improbe().\n- s2(MPI_ERROR) != MPI_ERR_OTHER at T6 Mrecv().\n- s1(MPI_ERROR) != MPI_ERR_DIMS at T7 Improbe().\n- s2(MPI_ERROR) != MPI_ERR_OTHER at T7 Imrecv().\n- Found 24 errors\n s1(MPI_ERROR) != MPI_ERR_DIMS at T0 Mprobe().\n s2(MPI_ERROR) != MPI_ERR_OTHER at T0 Mrecv().\n s1(MPI_ERROR) != MPI_ERR_DIMS at T1 Mprobe().\n+ s2(MPI_ERROR) != MPI_ERR_OTHER at T4 Mrecv().\n s2(MPI_ERROR) != MPI_ERR_OTHER at T1 Imrecv().\n s1(MPI_ERROR) != MPI_ERR_DIMS at T2 Improbe().\n s2(MPI_ERROR) != MPI_ERR_OTHER at T2 Mrecv().\n s1(MPI_ERROR) != MPI_ERR_DIMS at T3 Improbe().\n s2(MPI_ERROR) != MPI_ERR_OTHER at T3 Imrecv().\n s1(MPI_ERROR) != MPI_ERR_DIMS at T4 Mprobe().\n s2(MPI_ERROR) != MPI_ERR_OTHER at T4 Mrecv().\n s1(MPI_ERROR) != MPI_ERR_DIMS at T5 Mprobe().\n s2(MPI_ERROR) != MPI_ERR_OTHER at T5 Imrecv().\n s1(MPI_ERROR) != MPI_ERR_DIMS at T6 Improbe().\n s2(MPI_ERROR) != MPI_ERR_OTHER at T6 Mrecv().\n s1(MPI_ERROR) != MPI_ERR_DIMS at T7 Improbe().\n- s2(MPI_ERROR) != MPI_ERR_OTHER at T7 Imrecv().\n+ s2(MPI_ERROR) != MPI_ERR_OTHER at T7 Imrecv().\n+ s1(MPI_ERROR) != MPI_ERR_DIMS at T5 Mprobe().\n+ s2(MPI_ERROR) != MPI_ERR_OTHER at T5 Imrecv().\n+ s1(MPI_ERROR) != MPI_ERR_DIMS at T6 Improbe().\n+ s2(MPI_ERROR) != MPI_ERR_OTHER at T6 Mrecv().\n+ s1(MPI_ERROR) != MPI_ERR_DIMS at T7 Improbe().\n+ s2(MPI_ERROR) != MPI_ERR_OTHER at T7 Imrecv().\n+ Found 24 errors\n \n \n typenamef90\n 1\n ./f90/datatype\n pass\n \n@@ -1006580,34 +1006661,34 @@\n typecntsf90\n 1\n ./f90/datatype\n fail\n Program received signal SIGSEGV: Segmentation fault - invalid memory reference.\n \n Backtrace for this error:\n-#0 0x7fc031c6589d in ???\n-#1 0x7fc031c64ae3 in ???\n-#2 0x7fc03189f8df in ???\n-#3 0x7fc0319a463e in ???\n-#4 0x7fc031118415 in ???\n-#5 0x7fc031f1305e in ???\n-#6 0x7fc031f44b26 in ???\n-#7 0x7fc0320156d3 in ???\n-#8 0x56423c6e3535 in explore_\n+#0 0x7f33f0eb889d in ???\n+#1 0x7f33f0eb7ae3 in ???\n+#2 0x7f33f0af28df in ???\n+#3 0x7f33f0bf4e90 in ???\n+#4 0x7f33f036b415 in ???\n+#5 0x7f33f116605e in ???\n+#6 0x7f33f1197b26 in ???\n+#7 0x7f33f12686d3 in ???\n+#8 0x55868c871535 in explore_\n \tat ../../../f90/datatype/typecntsf90.f90:52\n-#9 0x56423c6e327a in MAIN__\n+#9 0x55868c87127a in MAIN__\n \tat ../../../f90/datatype/typecntsf90.f90:24\n-#10 0x56423c6e327a in main\n+#10 0x55868c87127a in main\n \tat ../../../f90/datatype/typecntsf90.f90:8\n --------------------------------------------------------------------------\n Primary job terminated normally, but 1 process returned\n a non-zero exit code. Per user-direction, the job has been aborted.\n --------------------------------------------------------------------------\n --------------------------------------------------------------------------\n-mpiexec.openmpi noticed that process rank 0 with PID 0 on node profitbricks-build5-amd64 exited on signal 11 (Segmentation fault).\n+mpiexec.openmpi noticed that process rank 0 with PID 0 on node i-capture-the-hostname exited on signal 11 (Segmentation fault).\n --------------------------------------------------------------------------\n \n \n typem2f90\n 1\n ./f90/datatype\n pass\n@@ -1006733,55 +1006814,55 @@\n pass\n \n \n createf90types\n 1\n ./f90/f90types\n fail\n- [profitbricks-build5-amd64:21633] *** Process received signal ***\n-[profitbricks-build5-amd64:21633] Signal: Segmentation fault (11)\n-[profitbricks-build5-amd64:21633] Signal code: Address not mapped (1)\n-[profitbricks-build5-amd64:21633] Failing at address: 0x11\n-[profitbricks-build5-amd64:21633] [ 0] /lib/x86_64-linux-gnu/libpthread.so.0(+0x126b0)[0x7f5b7a86a6b0]\n-[profitbricks-build5-amd64:21633] [ 1] /usr/lib/x86_64-linux-gnu/libmpi.so.40(MPI_Type_get_contents+0x139)[0x7f5b7a90cb19]\n-[profitbricks-build5-amd64:21633] [ 2] ./createf90types(+0x38c1)[0x5571534ee8c1]\n-[profitbricks-build5-amd64:21633] [ 3] ./createf90types(+0x35f9)[0x5571534ee5f9]\n-[profitbricks-build5-amd64:21633] [ 4] /lib/x86_64-linux-gnu/libc.so.6(__libc_start_main+0xeb)[0x7f5b7a6b409b]\n-[profitbricks-build5-amd64:21633] [ 5] ./createf90types(+0x373a)[0x5571534ee73a]\n-[profitbricks-build5-amd64:21633] *** End of error message ***\n+ [i-capture-the-hostname:05855] *** Process received signal ***\n+[i-capture-the-hostname:05855] Signal: Segmentation fault (11)\n+[i-capture-the-hostname:05855] Signal code: Address not mapped (1)\n+[i-capture-the-hostname:05855] Failing at address: 0x11\n+[i-capture-the-hostname:05855] [ 0] /lib/x86_64-linux-gnu/libpthread.so.0(+0x126b0)[0x7fd86ab956b0]\n+[i-capture-the-hostname:05855] [ 1] /usr/lib/x86_64-linux-gnu/libmpi.so.40(MPI_Type_get_contents+0x139)[0x7fd86ac37b19]\n+[i-capture-the-hostname:05855] [ 2] ./createf90types(+0x38c1)[0x5653f4a9b8c1]\n+[i-capture-the-hostname:05855] [ 3] ./createf90types(+0x35f9)[0x5653f4a9b5f9]\n+[i-capture-the-hostname:05855] [ 4] /lib/x86_64-linux-gnu/libc.so.6(__libc_start_main+0xeb)[0x7fd86a9df09b]\n+[i-capture-the-hostname:05855] [ 5] ./createf90types(+0x373a)[0x5653f4a9b73a]\n+[i-capture-the-hostname:05855] *** End of error message ***\n --------------------------------------------------------------------------\n Primary job terminated normally, but 1 process returned\n a non-zero exit code. Per user-direction, the job has been aborted.\n --------------------------------------------------------------------------\n --------------------------------------------------------------------------\n-mpiexec.openmpi noticed that process rank 0 with PID 0 on node profitbricks-build5-amd64 exited on signal 11 (Segmentation fault).\n+mpiexec.openmpi noticed that process rank 0 with PID 0 on node i-capture-the-hostname exited on signal 11 (Segmentation fault).\n --------------------------------------------------------------------------\n \n \n createf90types\n 1\n ./f90/f90types\n fail\n- [profitbricks-build5-amd64:22139] *** Process received signal ***\n-[profitbricks-build5-amd64:22139] Signal: Segmentation fault (11)\n-[profitbricks-build5-amd64:22139] Signal code: Address not mapped (1)\n-[profitbricks-build5-amd64:22139] Failing at address: 0x11\n-[profitbricks-build5-amd64:22139] [ 0] /lib/x86_64-linux-gnu/libpthread.so.0(+0x126b0)[0x7f185f1aa6b0]\n-[profitbricks-build5-amd64:22139] [ 1] /usr/lib/x86_64-linux-gnu/libmpi.so.40(MPI_Type_get_contents+0x139)[0x7f185f24cb19]\n-[profitbricks-build5-amd64:22139] [ 2] ./createf90types(+0x38c1)[0x563f76edf8c1]\n-[profitbricks-build5-amd64:22139] [ 3] ./createf90types(+0x35f9)[0x563f76edf5f9]\n-[profitbricks-build5-amd64:22139] [ 4] /lib/x86_64-linux-gnu/libc.so.6(__libc_start_main+0xeb)[0x7f185eff409b]\n-[profitbricks-build5-amd64:22139] [ 5] ./createf90types(+0x373a)[0x563f76edf73a]\n-[profitbricks-build5-amd64:22139] *** End of error message ***\n+ [i-capture-the-hostname:05920] *** Process received signal ***\n+[i-capture-the-hostname:05920] Signal: Segmentation fault (11)\n+[i-capture-the-hostname:05920] Signal code: Address not mapped (1)\n+[i-capture-the-hostname:05920] Failing at address: 0x11\n+[i-capture-the-hostname:05920] [ 0] /lib/x86_64-linux-gnu/libpthread.so.0(+0x126b0)[0x7fa2ded4d6b0]\n+[i-capture-the-hostname:05920] [ 1] /usr/lib/x86_64-linux-gnu/libmpi.so.40(MPI_Type_get_contents+0x139)[0x7fa2dedefb19]\n+[i-capture-the-hostname:05920] [ 2] ./createf90types(+0x38c1)[0x561d264c18c1]\n+[i-capture-the-hostname:05920] [ 3] ./createf90types(+0x35f9)[0x561d264c15f9]\n+[i-capture-the-hostname:05920] [ 4] /lib/x86_64-linux-gnu/libc.so.6(__libc_start_main+0xeb)[0x7fa2deb9709b]\n+[i-capture-the-hostname:05920] [ 5] ./createf90types(+0x373a)[0x561d264c173a]\n+[i-capture-the-hostname:05920] *** End of error message ***\n --------------------------------------------------------------------------\n Primary job terminated normally, but 1 process returned\n a non-zero exit code. Per user-direction, the job has been aborted.\n --------------------------------------------------------------------------\n --------------------------------------------------------------------------\n-mpiexec.openmpi noticed that process rank 0 with PID 0 on node profitbricks-build5-amd64 exited on signal 11 (Segmentation fault).\n+mpiexec.openmpi noticed that process rank 0 with PID 0 on node i-capture-the-hostname exited on signal 11 (Segmentation fault).\n --------------------------------------------------------------------------\n \n \n winscale1f90\n 2\n ./f90/rma\n pass\n@@ -1006869,102 +1006950,93 @@\n At least one pair of MPI processes are unable to reach each other for\n MPI communications. This means that no Open MPI device has indicated\n that it can be used to communicate between these processes. This is\n an error; Open MPI requires that all MPI processes be able to reach\n each other. This error can sometimes be the result of forgetting to\n specify the "self" BTL.\n \n- Process 1 ([[39792,2],0]) is on host: profitbricks-build5-amd64\n- Process 2 ([[39792,1],0]) is on host: profitbricks-build5-amd64\n+ Process 1 ([[37874,2],1]) is on host: i-capture-the-hostname\n+ Process 2 ([[37874,1],0]) is on host: i-capture-the-hostname\n BTLs attempted: vader self\n \n Your MPI job is now going to abort; sorry.\n --------------------------------------------------------------------------\n-[profitbricks-build5-amd64:24549] [[39792,1],0] ORTE_ERROR_LOG: Unreachable in file dpm/dpm.c at line 433\n-[profitbricks-build5-amd64:24613] [[39792,2],0] ORTE_ERROR_LOG: Unreachable in file dpm/dpm.c at line 433\n-[profitbricks-build5-amd64:24615] [[39792,2],1] ORTE_ERROR_LOG: Unreachable in file dpm/dpm.c at line 433\n --------------------------------------------------------------------------\n It looks like MPI_INIT failed for some reason; your parallel process is\n likely to abort. There are many reasons that a parallel process can\n fail during MPI_INIT; some of which are due to configuration or environment\n problems. This failure appears to be an internal failure; here's some\n additional information (which may only be relevant to an Open MPI\n developer):\n \n ompi_dpm_dyn_init() failed\n --> Returned "Unreachable" (-12) instead of "Success" (0)\n --------------------------------------------------------------------------\n-[profitbricks-build5-amd64:24549] *** An error occurred in MPI_Comm_spawn\n-[profitbricks-build5-amd64:24549] *** reported by process [2607808513,0]\n-[profitbricks-build5-amd64:24549] *** on communicator MPI_COMM_WORLD\n-[profitbricks-build5-amd64:24549] *** MPI_ERR_INTERN: internal error\n-[profitbricks-build5-amd64:24549] *** MPI_ERRORS_ARE_FATAL (processes in this communicator will now abort,\n-[profitbricks-build5-amd64:24549] *** and potentially your MPI job)\n-[profitbricks-build5-amd64:24613] *** An error occurred in MPI_Init\n-[profitbricks-build5-amd64:24613] *** reported by process [2607808514,0]\n-[profitbricks-build5-amd64:24613] *** on a NULL communicator\n-[profitbricks-build5-amd64:24613] *** Unknown error\n-[profitbricks-build5-amd64:24613] *** MPI_ERRORS_ARE_FATAL (processes in this communicator will now abort,\n-[profitbricks-build5-amd64:24613] *** and potentially your MPI job)\n-[profitbricks-build5-amd64:24482] 2 more processes have sent help message help-mca-bml-r2.txt / unreachable proc\n-[profitbricks-build5-amd64:24482] Set MCA parameter "orte_base_help_aggregate" to 0 to see all help / error messages\n-[profitbricks-build5-amd64:24482] 1 more process has sent help message help-mpi-runtime.txt / mpi_init:startup:internal-failure\n-[profitbricks-build5-amd64:24482] 1 more process has sent help message help-mpi-errors.txt / mpi_errors_are_fatal unknown handle\n-[warn] Epoll MOD(1) on fd 27 failed. Old events were 6; read change was 0 (none); write change was 2 (del); close change was 0 (none): Bad file descriptor\n+[i-capture-the-hostname:07987] [[37874,2],0] ORTE_ERROR_LOG: Unreachable in file dpm/dpm.c at line 433\n+[i-capture-the-hostname:07932] [[37874,1],0] ORTE_ERROR_LOG: Unreachable in file dpm/dpm.c at line 433\n+[i-capture-the-hostname:07988] [[37874,2],1] ORTE_ERROR_LOG: Unreachable in file dpm/dpm.c at line 433\n+[i-capture-the-hostname:07932] *** An error occurred in MPI_Comm_spawn\n+[i-capture-the-hostname:07932] *** reported by process [2482110465,0]\n+[i-capture-the-hostname:07932] *** on communicator MPI_COMM_WORLD\n+[i-capture-the-hostname:07932] *** MPI_ERR_INTERN: internal error\n+[i-capture-the-hostname:07932] *** MPI_ERRORS_ARE_FATAL (processes in this communicator will now abort,\n+[i-capture-the-hostname:07932] *** and potentially your MPI job)\n+[i-capture-the-hostname:07988] *** An error occurred in MPI_Init\n+[i-capture-the-hostname:07988] *** reported by process [2482110466,1]\n+[i-capture-the-hostname:07988] *** on a NULL communicator\n+[i-capture-the-hostname:07988] *** Unknown error\n+[i-capture-the-hostname:07988] *** MPI_ERRORS_ARE_FATAL (processes in this communicator will now abort,\n+[i-capture-the-hostname:07988] *** and potentially your MPI job)\n+[i-capture-the-hostname:07843] 2 more processes have sent help message help-mca-bml-r2.txt / unreachable proc\n+[i-capture-the-hostname:07843] Set MCA parameter "orte_base_help_aggregate" to 0 to see all help / error messages\n+[i-capture-the-hostname:07843] 1 more process has sent help message help-mpi-runtime.txt / mpi_init:startup:internal-failure\n+[i-capture-the-hostname:07843] 1 more process has sent help message help-mpi-errors.txt / mpi_errors_are_fatal unknown handle\n \n \n spawnargvf90\n 1\n ./f90/spawn\n fail\n --------------------------------------------------------------------------\n At least one pair of MPI processes are unable to reach each other for\n MPI communications. This means that no Open MPI device has indicated\n that it can be used to communicate between these processes. This is\n an error; Open MPI requires that all MPI processes be able to reach\n each other. This error can sometimes be the result of forgetting to\n specify the "self" BTL.\n \n- Process 1 ([[42070,2],0]) is on host: profitbricks-build5-amd64\n- Process 2 ([[42070,1],0]) is on host: profitbricks-build5-amd64\n- BTLs attempted: vader self\n+ Process 1 ([[44378,1],0]) is on host: i-capture-the-hostname\n+ Process 2 ([[44378,2],0]) is on host: i-capture-the-hostname\n+ BTLs attempted: self\n \n Your MPI job is now going to abort; sorry.\n --------------------------------------------------------------------------\n+[i-capture-the-hostname:08249] [[44378,1],0] ORTE_ERROR_LOG: Unreachable in file dpm/dpm.c at line 433\n+[i-capture-the-hostname:08249] *** An error occurred in MPI_Comm_spawn\n+[i-capture-the-hostname:08249] *** reported by process [2908356609,0]\n+[i-capture-the-hostname:08249] *** on communicator MPI_COMM_WORLD\n+[i-capture-the-hostname:08249] *** MPI_ERR_INTERN: internal error\n+[i-capture-the-hostname:08249] *** MPI_ERRORS_ARE_FATAL (processes in this communicator will now abort,\n+[i-capture-the-hostname:08249] *** and potentially your MPI job)\n --------------------------------------------------------------------------\n It looks like MPI_INIT failed for some reason; your parallel process is\n likely to abort. There are many reasons that a parallel process can\n fail during MPI_INIT; some of which are due to configuration or environment\n problems. This failure appears to be an internal failure; here's some\n additional information (which may only be relevant to an Open MPI\n developer):\n \n ompi_dpm_dyn_init() failed\n --> Returned "Unreachable" (-12) instead of "Success" (0)\n --------------------------------------------------------------------------\n-[profitbricks-build5-amd64:24799] [[42070,2],0] ORTE_ERROR_LOG: Unreachable in file dpm/dpm.c at line 433\n-[profitbricks-build5-amd64:24799] *** An error occurred in MPI_Init\n-[profitbricks-build5-amd64:24799] *** reported by process [2757099522,0]\n-[profitbricks-build5-amd64:24799] *** on a NULL communicator\n-[profitbricks-build5-amd64:24799] *** Unknown error\n-[profitbricks-build5-amd64:24799] *** MPI_ERRORS_ARE_FATAL (processes in this communicator will now abort,\n-[profitbricks-build5-amd64:24799] *** and potentially your MPI job)\n-[profitbricks-build5-amd64:24752] [[42070,1],0] ORTE_ERROR_LOG: Unreachable in file dpm/dpm.c at line 433\n-[profitbricks-build5-amd64:24800] [[42070,2],1] ORTE_ERROR_LOG: Unreachable in file dpm/dpm.c at line 433\n-[profitbricks-build5-amd64:24752] *** An error occurred in MPI_Comm_spawn\n-[profitbricks-build5-amd64:24752] *** reported by process [2757099521,0]\n-[profitbricks-build5-amd64:24752] *** on communicator MPI_COMM_WORLD\n-[profitbricks-build5-amd64:24752] *** MPI_ERR_INTERN: internal error\n-[profitbricks-build5-amd64:24752] *** MPI_ERRORS_ARE_FATAL (processes in this communicator will now abort,\n-[profitbricks-build5-amd64:24752] *** and potentially your MPI job)\n-[profitbricks-build5-amd64:24708] 2 more processes have sent help message help-mca-bml-r2.txt / unreachable proc\n-[profitbricks-build5-amd64:24708] Set MCA parameter "orte_base_help_aggregate" to 0 to see all help / error messages\n-[profitbricks-build5-amd64:24708] 1 more process has sent help message help-mpi-runtime.txt / mpi_init:startup:internal-failure\n-[profitbricks-build5-amd64:24708] 1 more process has sent help message help-mpi-errors.txt / mpi_errors_are_fatal unknown handle\n-[warn] Epoll MOD(1) on fd 29 failed. Old events were 6; read change was 0 (none); write change was 2 (del); close change was 0 (none): Bad file descriptor\n+[i-capture-the-hostname:08273] [[44378,2],0] ORTE_ERROR_LOG: Unreachable in file dpm/dpm.c at line 433\n+[i-capture-the-hostname:08274] [[44378,2],1] ORTE_ERROR_LOG: Unreachable in file dpm/dpm.c at line 433\n+[i-capture-the-hostname:08203] 2 more processes have sent help message help-mca-bml-r2.txt / unreachable proc\n+[i-capture-the-hostname:08203] Set MCA parameter "orte_base_help_aggregate" to 0 to see all help / error messages\n+[i-capture-the-hostname:08203] 1 more process has sent help message help-mpi-runtime.txt / mpi_init:startup:internal-failure\n \n \n connaccf90\n 2\n ./f90/spawn\n pass\n \n@@ -1006977,202 +1007049,197 @@\n At least one pair of MPI processes are unable to reach each other for\n MPI communications. This means that no Open MPI device has indicated\n that it can be used to communicate between these processes. This is\n an error; Open MPI requires that all MPI processes be able to reach\n each other. This error can sometimes be the result of forgetting to\n specify the "self" BTL.\n \n- Process 1 ([[42300,1],0]) is on host: profitbricks-build5-amd64\n- Process 2 ([[42300,2],0]) is on host: profitbricks-build5-amd64\n+ Process 1 ([[44498,1],0]) is on host: i-capture-the-hostname\n+ Process 2 ([[44498,2],0]) is on host: i-capture-the-hostname\n BTLs attempted: self\n \n Your MPI job is now going to abort; sorry.\n --------------------------------------------------------------------------\n-[profitbricks-build5-amd64:25125] [[42300,1],0] ORTE_ERROR_LOG: Unreachable in file dpm/dpm.c at line 433\n-[profitbricks-build5-amd64:25125] *** An error occurred in MPI_Comm_spawn_multiple\n-[profitbricks-build5-amd64:25125] *** reported by process [2772172801,0]\n-[profitbricks-build5-amd64:25125] *** on communicator MPI_COMM_WORLD\n-[profitbricks-build5-amd64:25125] *** MPI_ERR_INTERN: internal error\n-[profitbricks-build5-amd64:25125] *** MPI_ERRORS_ARE_FATAL (processes in this communicator will now abort,\n-[profitbricks-build5-amd64:25125] *** and potentially your MPI job)\n+[i-capture-the-hostname:08354] [[44498,1],0] ORTE_ERROR_LOG: Unreachable in file dpm/dpm.c at line 433\n+[i-capture-the-hostname:08354] *** An error occurred in MPI_Comm_spawn_multiple\n+[i-capture-the-hostname:08354] *** reported by process [2916220929,0]\n+[i-capture-the-hostname:08354] *** on communicator MPI_COMM_WORLD\n+[i-capture-the-hostname:08354] *** MPI_ERR_INTERN: internal error\n+[i-capture-the-hostname:08354] *** MPI_ERRORS_ARE_FATAL (processes in this communicator will now abort,\n+[i-capture-the-hostname:08354] *** and potentially your MPI job)\n --------------------------------------------------------------------------\n It looks like MPI_INIT failed for some reason; your parallel process is\n likely to abort. There are many reasons that a parallel process can\n fail during MPI_INIT; some of which are due to configuration or environment\n problems. This failure appears to be an internal failure; here's some\n additional information (which may only be relevant to an Open MPI\n developer):\n \n ompi_dpm_dyn_init() failed\n --> Returned "Unreachable" (-12) instead of "Success" (0)\n --------------------------------------------------------------------------\n-[profitbricks-build5-amd64:25217] [[42300,2],1] ORTE_ERROR_LOG: Unreachable in file dpm/dpm.c at line 433\n-[profitbricks-build5-amd64:25214] [[42300,2],0] ORTE_ERROR_LOG: Unreachable in file dpm/dpm.c at line 433\n-[profitbricks-build5-amd64:25217] *** An error occurred in MPI_Init\n-[profitbricks-build5-amd64:25217] *** reported by process [2772172802,1]\n-[profitbricks-build5-amd64:25217] *** on a NULL communicator\n-[profitbricks-build5-amd64:25217] *** Unknown error\n-[profitbricks-build5-amd64:25217] *** MPI_ERRORS_ARE_FATAL (processes in this communicator will now abort,\n-[profitbricks-build5-amd64:25217] *** and potentially your MPI job)\n-[profitbricks-build5-amd64:25070] 2 more processes have sent help message help-mca-bml-r2.txt / unreachable proc\n-[profitbricks-build5-amd64:25070] Set MCA parameter "orte_base_help_aggregate" to 0 to see all help / error messages\n-[profitbricks-build5-amd64:25070] 1 more process has sent help message help-mpi-runtime.txt / mpi_init:startup:internal-failure\n-[profitbricks-build5-amd64:25070] 1 more process has sent help message help-mpi-errors.txt / mpi_errors_are_fatal unknown handle\n+[i-capture-the-hostname:08387] [[44498,2],1] ORTE_ERROR_LOG: Unreachable in file dpm/dpm.c at line 433\n+[i-capture-the-hostname:08386] [[44498,2],0] ORTE_ERROR_LOG: Unreachable in file dpm/dpm.c at line 433\n+[i-capture-the-hostname:08323] 2 more processes have sent help message help-mca-bml-r2.txt / unreachable proc\n+[i-capture-the-hostname:08323] Set MCA parameter "orte_base_help_aggregate" to 0 to see all help / error messages\n+[i-capture-the-hostname:08323] 1 more process has sent help message help-mpi-runtime.txt / mpi_init:startup:internal-failure\n \n \n spawnmult2f90\n 2\n ./f90/spawn\n fail\n --------------------------------------------------------------------------\n At least one pair of MPI processes are unable to reach each other for\n MPI communications. This means that no Open MPI device has indicated\n that it can be used to communicate between these processes. This is\n an error; Open MPI requires that all MPI processes be able to reach\n each other. This error can sometimes be the result of forgetting to\n specify the "self" BTL.\n \n- Process 1 ([[43006,1],0]) is on host: profitbricks-build5-amd64\n- Process 2 ([[43006,2],0]) is on host: profitbricks-build5-amd64\n+ Process 1 ([[44099,1],1]) is on host: i-capture-the-hostname\n+ Process 2 ([[44099,2],0]) is on host: i-capture-the-hostname\n BTLs attempted: vader self\n \n Your MPI job is now going to abort; sorry.\n --------------------------------------------------------------------------\n-[profitbricks-build5-amd64:25470] [[43006,1],0] ORTE_ERROR_LOG: Unreachable in file dpm/dpm.c at line 433\n-[profitbricks-build5-amd64:25470] *** An error occurred in MPI_Comm_spawn_multiple\n-[profitbricks-build5-amd64:25470] *** reported by process [2818441217,0]\n-[profitbricks-build5-amd64:25470] *** on communicator MPI_COMM_WORLD\n-[profitbricks-build5-amd64:25470] *** MPI_ERR_INTERN: internal error\n-[profitbricks-build5-amd64:25470] *** MPI_ERRORS_ARE_FATAL (processes in this communicator will now abort,\n-[profitbricks-build5-amd64:25470] *** and potentially your MPI job)\n-[profitbricks-build5-amd64:25576] [[43006,2],1] ORTE_ERROR_LOG: Unreachable in file dpm/dpm.c at line 433\n-[profitbricks-build5-amd64:25472] [[43006,1],1] ORTE_ERROR_LOG: Unreachable in file dpm/dpm.c at line 433\n-[profitbricks-build5-amd64:25571] [[43006,2],0] ORTE_ERROR_LOG: Unreachable in file dpm/dpm.c at line 433\n+[i-capture-the-hostname:08524] [[44099,1],1] ORTE_ERROR_LOG: Unreachable in file dpm/dpm.c at line 433\n+[i-capture-the-hostname:08590] [[44099,2],0] ORTE_ERROR_LOG: Unreachable in file dpm/dpm.c at line 433\n+[i-capture-the-hostname:08593] [[44099,2],1] ORTE_ERROR_LOG: Unreachable in file dpm/dpm.c at line 433\n --------------------------------------------------------------------------\n It looks like MPI_INIT failed for some reason; your parallel process is\n likely to abort. There are many reasons that a parallel process can\n fail during MPI_INIT; some of which are due to configuration or environment\n problems. This failure appears to be an internal failure; here's some\n additional information (which may only be relevant to an Open MPI\n developer):\n \n ompi_dpm_dyn_init() failed\n --> Returned "Unreachable" (-12) instead of "Success" (0)\n --------------------------------------------------------------------------\n-[profitbricks-build5-amd64:25571] *** An error occurred in MPI_Init\n-[profitbricks-build5-amd64:25571] *** reported by process [2818441218,0]\n-[profitbricks-build5-amd64:25571] *** on a NULL communicator\n-[profitbricks-build5-amd64:25571] *** Unknown error\n-[profitbricks-build5-amd64:25571] *** MPI_ERRORS_ARE_FATAL (processes in this communicator will now abort,\n-[profitbricks-build5-amd64:25571] *** and potentially your MPI job)\n-[profitbricks-build5-amd64:25388] 3 more processes have sent help message help-mca-bml-r2.txt / unreachable proc\n-[profitbricks-build5-amd64:25388] Set MCA parameter "orte_base_help_aggregate" to 0 to see all help / error messages\n-[profitbricks-build5-amd64:25388] 1 more process has sent help message help-mpi-errors.txt / mpi_errors_are_fatal\n-[profitbricks-build5-amd64:25388] 1 more process has sent help message help-mpi-runtime.txt / mpi_init:startup:internal-failure\n-[profitbricks-build5-amd64:25388] 1 more process has sent help message help-mpi-errors.txt / mpi_errors_are_fatal unknown handle\n+[i-capture-the-hostname:08524] *** An error occurred in MPI_Comm_spawn_multiple\n+[i-capture-the-hostname:08524] *** reported by process [2890072065,1]\n+[i-capture-the-hostname:08524] *** on communicator MPI_COMM_WORLD\n+[i-capture-the-hostname:08524] *** MPI_ERR_INTERN: internal error\n+[i-capture-the-hostname:08524] *** MPI_ERRORS_ARE_FATAL (processes in this communicator will now abort,\n+[i-capture-the-hostname:08524] *** and potentially your MPI job)\n+[i-capture-the-hostname:08519] [[44099,1],0] ORTE_ERROR_LOG: Unreachable in file dpm/dpm.c at line 433\n+[i-capture-the-hostname:08590] *** An error occurred in MPI_Init\n+[i-capture-the-hostname:08590] *** reported by process [2890072066,0]\n+[i-capture-the-hostname:08590] *** on a NULL communicator\n+[i-capture-the-hostname:08590] *** Unknown error\n+[i-capture-the-hostname:08590] *** MPI_ERRORS_ARE_FATAL (processes in this communicator will now abort,\n+[i-capture-the-hostname:08590] *** and potentially your MPI job)\n+[i-capture-the-hostname:08466] 3 more processes have sent help message help-mca-bml-r2.txt / unreachable proc\n+[i-capture-the-hostname:08466] Set MCA parameter "orte_base_help_aggregate" to 0 to see all help / error messages\n+[i-capture-the-hostname:08466] 1 more process has sent help message help-mpi-runtime.txt / mpi_init:startup:internal-failure\n+[i-capture-the-hostname:08466] 1 more process has sent help message help-mpi-errors.txt / mpi_errors_are_fatal\n+[i-capture-the-hostname:08466] 1 more process has sent help message help-mpi-errors.txt / mpi_errors_are_fatal unknown handle\n \n \n spawnargvf03\n 1\n ./f90/spawn\n fail\n --------------------------------------------------------------------------\n At least one pair of MPI processes are unable to reach each other for\n MPI communications. This means that no Open MPI device has indicated\n that it can be used to communicate between these processes. This is\n an error; Open MPI requires that all MPI processes be able to reach\n each other. This error can sometimes be the result of forgetting to\n specify the "self" BTL.\n \n- Process 1 ([[41128,2],1]) is on host: profitbricks-build5-amd64\n- Process 2 ([[41128,1],0]) is on host: profitbricks-build5-amd64\n+ Process 1 ([[44881,2],0]) is on host: i-capture-the-hostname\n+ Process 2 ([[44881,1],0]) is on host: i-capture-the-hostname\n BTLs attempted: vader self\n \n Your MPI job is now going to abort; sorry.\n --------------------------------------------------------------------------\n-[profitbricks-build5-amd64:25943] [[41128,2],1] ORTE_ERROR_LOG: Unreachable in file dpm/dpm.c at line 433\n-[profitbricks-build5-amd64:25812] [[41128,1],0] ORTE_ERROR_LOG: Unreachable in file dpm/dpm.c at line 433\n+[i-capture-the-hostname:08894] [[44881,2],0] ORTE_ERROR_LOG: Unreachable in file dpm/dpm.c at line 433\n --------------------------------------------------------------------------\n It looks like MPI_INIT failed for some reason; your parallel process is\n likely to abort. There are many reasons that a parallel process can\n fail during MPI_INIT; some of which are due to configuration or environment\n problems. This failure appears to be an internal failure; here's some\n additional information (which may only be relevant to an Open MPI\n developer):\n \n ompi_dpm_dyn_init() failed\n --> Returned "Unreachable" (-12) instead of "Success" (0)\n --------------------------------------------------------------------------\n-[profitbricks-build5-amd64:25812] *** An error occurred in MPI_Comm_spawn\n-[profitbricks-build5-amd64:25812] *** reported by process [2695364609,0]\n-[profitbricks-build5-amd64:25812] *** on communicator MPI_COMM_WORLD\n-[profitbricks-build5-amd64:25812] *** MPI_ERR_INTERN: internal error\n-[profitbricks-build5-amd64:25812] *** MPI_ERRORS_ARE_FATAL (processes in this communicator will now abort,\n-[profitbricks-build5-amd64:25812] *** and potentially your MPI job)\n-[profitbricks-build5-amd64:25937] [[41128,2],0] ORTE_ERROR_LOG: Unreachable in file dpm/dpm.c at line 433\n-[profitbricks-build5-amd64:25937] *** An error occurred in MPI_Init\n-[profitbricks-build5-amd64:25937] *** reported by process [2695364610,0]\n-[profitbricks-build5-amd64:25937] *** on a NULL communicator\n-[profitbricks-build5-amd64:25937] *** Unknown error\n-[profitbricks-build5-amd64:25937] *** MPI_ERRORS_ARE_FATAL (processes in this communicator will now abort,\n-[profitbricks-build5-amd64:25937] *** and potentially your MPI job)\n-[profitbricks-build5-amd64:25722] 2 more processes have sent help message help-mca-bml-r2.txt / unreachable proc\n-[profitbricks-build5-amd64:25722] Set MCA parameter "orte_base_help_aggregate" to 0 to see all help / error messages\n-[profitbricks-build5-amd64:25722] 1 more process has sent help message help-mpi-runtime.txt / mpi_init:startup:internal-failure\n-[profitbricks-build5-amd64:25722] 1 more process has sent help message help-mpi-errors.txt / mpi_errors_are_fatal unknown handle\n+[i-capture-the-hostname:08792] [[44881,1],0] ORTE_ERROR_LOG: Unreachable in file dpm/dpm.c at line 433\n+[i-capture-the-hostname:08900] [[44881,2],1] ORTE_ERROR_LOG: Unreachable in file dpm/dpm.c at line 433\n+[i-capture-the-hostname:08792] *** An error occurred in MPI_Comm_spawn\n+[i-capture-the-hostname:08792] *** reported by process [2941321217,0]\n+[i-capture-the-hostname:08792] *** on communicator MPI_COMM_WORLD\n+[i-capture-the-hostname:08792] *** MPI_ERR_INTERN: internal error\n+[i-capture-the-hostname:08792] *** MPI_ERRORS_ARE_FATAL (processes in this communicator will now abort,\n+[i-capture-the-hostname:08792] *** and potentially your MPI job)\n+[i-capture-the-hostname:08894] *** An error occurred in MPI_Init\n+[i-capture-the-hostname:08894] *** reported by process [2941321218,0]\n+[i-capture-the-hostname:08894] *** on a NULL communicator\n+[i-capture-the-hostname:08894] *** Unknown error\n+[i-capture-the-hostname:08894] *** MPI_ERRORS_ARE_FATAL (processes in this communicator will now abort,\n+[i-capture-the-hostname:08894] *** and potentially your MPI job)\n+[i-capture-the-hostname:08704] 2 more processes have sent help message help-mca-bml-r2.txt / unreachable proc\n+[i-capture-the-hostname:08704] Set MCA parameter "orte_base_help_aggregate" to 0 to see all help / error messages\n+[i-capture-the-hostname:08704] 1 more process has sent help message help-mpi-runtime.txt / mpi_init:startup:internal-failure\n+[i-capture-the-hostname:08704] 1 more process has sent help message help-mpi-errors.txt / mpi_errors_are_fatal unknown handle\n+[warn] Epoll MOD(1) on fd 25 failed. Old events were 6; read change was 0 (none); write change was 2 (del); close change was 0 (none): Bad file descriptor\n \n \n spawnmultf03\n 1\n ./f90/spawn\n fail\n --------------------------------------------------------------------------\n At least one pair of MPI processes are unable to reach each other for\n MPI communications. This means that no Open MPI device has indicated\n that it can be used to communicate between these processes. This is\n an error; Open MPI requires that all MPI processes be able to reach\n each other. This error can sometimes be the result of forgetting to\n specify the "self" BTL.\n \n- Process 1 ([[41220,1],0]) is on host: profitbricks-build5-amd64\n- Process 2 ([[41220,2],0]) is on host: profitbricks-build5-amd64\n- BTLs attempted: self\n+ Process 1 ([[44617,2],1]) is on host: i-capture-the-hostname\n+ Process 2 ([[44617,1],0]) is on host: i-capture-the-hostname\n+ BTLs attempted: vader self\n \n Your MPI job is now going to abort; sorry.\n --------------------------------------------------------------------------\n-[profitbricks-build5-amd64:26121] [[41220,1],0] ORTE_ERROR_LOG: Unreachable in file dpm/dpm.c at line 433\n-[profitbricks-build5-amd64:26121] *** An error occurred in MPI_Comm_spawn_multiple\n-[profitbricks-build5-amd64:26121] *** reported by process [2701393921,0]\n-[profitbricks-build5-amd64:26121] *** on communicator MPI_COMM_WORLD\n-[profitbricks-build5-amd64:26121] *** MPI_ERR_INTERN: internal error\n-[profitbricks-build5-amd64:26121] *** MPI_ERRORS_ARE_FATAL (processes in this communicator will now abort,\n-[profitbricks-build5-amd64:26121] *** and potentially your MPI job)\n+[i-capture-the-hostname:09053] [[44617,2],0] ORTE_ERROR_LOG: Unreachable in file dpm/dpm.c at line 433\n+[i-capture-the-hostname:09054] [[44617,2],1] ORTE_ERROR_LOG: Unreachable in file dpm/dpm.c at line 433\n+[i-capture-the-hostname:09025] [[44617,1],0] ORTE_ERROR_LOG: Unreachable in file dpm/dpm.c at line 433\n --------------------------------------------------------------------------\n It looks like MPI_INIT failed for some reason; your parallel process is\n likely to abort. There are many reasons that a parallel process can\n fail during MPI_INIT; some of which are due to configuration or environment\n problems. This failure appears to be an internal failure; here's some\n additional information (which may only be relevant to an Open MPI\n developer):\n \n ompi_dpm_dyn_init() failed\n --> Returned "Unreachable" (-12) instead of "Success" (0)\n --------------------------------------------------------------------------\n-[profitbricks-build5-amd64:26209] [[41220,2],1] ORTE_ERROR_LOG: Unreachable in file dpm/dpm.c at line 433\n-[profitbricks-build5-amd64:26203] [[41220,2],0] ORTE_ERROR_LOG: Unreachable in file dpm/dpm.c at line 433\n-[profitbricks-build5-amd64:26209] *** An error occurred in MPI_Init\n-[profitbricks-build5-amd64:26209] *** reported by process [2701393922,1]\n-[profitbricks-build5-amd64:26209] *** on a NULL communicator\n-[profitbricks-build5-amd64:26209] *** Unknown error\n-[profitbricks-build5-amd64:26209] *** MPI_ERRORS_ARE_FATAL (processes in this communicator will now abort,\n-[profitbricks-build5-amd64:26209] *** and potentially your MPI job)\n-[profitbricks-build5-amd64:26070] 2 more processes have sent help message help-mca-bml-r2.txt / unreachable proc\n-[profitbricks-build5-amd64:26070] Set MCA parameter "orte_base_help_aggregate" to 0 to see all help / error messages\n-[profitbricks-build5-amd64:26070] 1 more process has sent help message help-mpi-runtime.txt / mpi_init:startup:internal-failure\n-[profitbricks-build5-amd64:26070] 1 more process has sent help message help-mpi-errors.txt / mpi_errors_are_fatal unknown handle\n+[i-capture-the-hostname:09025] *** An error occurred in MPI_Comm_spawn_multiple\n+[i-capture-the-hostname:09025] *** reported by process [2924019713,0]\n+[i-capture-the-hostname:09025] *** on communicator MPI_COMM_WORLD\n+[i-capture-the-hostname:09025] *** MPI_ERR_INTERN: internal error\n+[i-capture-the-hostname:09025] *** MPI_ERRORS_ARE_FATAL (processes in this communicator will now abort,\n+[i-capture-the-hostname:09025] *** and potentially your MPI job)\n+[i-capture-the-hostname:09053] *** An error occurred in MPI_Init\n+[i-capture-the-hostname:09053] *** reported by process [2924019714,0]\n+[i-capture-the-hostname:09053] *** on a NULL communicator\n+[i-capture-the-hostname:09053] *** Unknown error\n+[i-capture-the-hostname:09053] *** MPI_ERRORS_ARE_FATAL (processes in this communicator will now abort,\n+[i-capture-the-hostname:09053] *** and potentially your MPI job)\n+[i-capture-the-hostname:08984] 2 more processes have sent help message help-mca-bml-r2.txt / unreachable proc\n+[i-capture-the-hostname:08984] Set MCA parameter "orte_base_help_aggregate" to 0 to see all help / error messages\n+[i-capture-the-hostname:08984] 1 more process has sent help message help-mpi-runtime.txt / mpi_init:startup:internal-failure\n+[i-capture-the-hostname:08984] 1 more process has sent help message help-mpi-errors.txt / mpi_errors_are_fatal unknown handle\n+[warn] Epoll MOD(1) on fd 29 failed. Old events were 6; read change was 0 (none); write change was 2 (del); close change was 0 (none): Bad file descriptor\n \n \n wtimef90\n 1\n ./f90/timer\n pass\n \n@@ -1007183,80 +1007250,80 @@\n pass\n \n \n dgraph_wgtf90\n 2\n ./f90/topo\n fail\n- Error: Cannot partition 16 elements in 17 parts\n+ Error: Cannot partition 15 elements in 2 parts\n \n Program received signal SIGSEGV: Segmentation fault - invalid memory reference.\n \n Backtrace for this error:\n-#0 0x7f06aef7589d in ???\n-#1 0x7f06aef74ae3 in ???\n-#2 0x7f06aebaf8df in ???\n-#3 0x7f06ac0ede8b in ???\n-#4 0x7f06ac0ee422 in ???\n-#5 0x7f06ac0ee7c1 in ???\n-#6 0x7f06ac0eb28c in ???\n-#7 0x7f06ac0e19e8 in ???\n-#8 0x7f06af2313cc in ???\n-#9 0x7f06af31cbca in ???\n-#10 0x561cf1e2c9e6 in dgraph_unwgt\n+#0 0x7fe32072889d in ???\n+#1 0x7fe320727ae3 in ???\n+#2 0x7fe3203628df in ???\n+#3 0x7fe31cfa0e8b in ???\n+#4 0x7fe31cfa1422 in ???\n+#5 0x7fe31cfa17c1 in ???\n+#6 0x7fe31cf9e28c in ???\n+#7 0x7fe31cf949e8 in ???\n+#8 0x7fe3209e43cc in ???\n+#9 0x7fe320acfbca in ???\n+#10 0x56497adc79e6 in dgraph_unwgt\n \tat ../../../f90/topo/dgraph_wgtf90.f90:155\n-#11 0x561cf1e2c1fe in main\n+#11 0x56497adc71fe in main\n \tat ../../../f90/topo/dgraph_wgtf90.f90:127\n --------------------------------------------------------------------------\n Primary job terminated normally, but 1 process returned\n a non-zero exit code. Per user-direction, the job has been aborted.\n --------------------------------------------------------------------------\n --------------------------------------------------------------------------\n-mpiexec.openmpi noticed that process rank 0 with PID 0 on node profitbricks-build5-amd64 exited on signal 11 (Segmentation fault).\n+mpiexec.openmpi noticed that process rank 0 with PID 0 on node i-capture-the-hostname exited on signal 11 (Segmentation fault).\n --------------------------------------------------------------------------\n \n \n dgraph_unwgtf90\n 2\n ./f90/topo\n fail\n- Error: Cannot partition 16 elements in 17 parts\n+ Error: Cannot partition 15 elements in 2 parts\n \n Program received signal SIGSEGV: Segmentation fault - invalid memory reference.\n \n Backtrace for this error:\n-#0 0x7f961bc2589d in ???\n-#1 0x7f961bc24ae3 in ???\n-#2 0x7f961b85f8df in ???\n-#3 0x7f961849de8b in ???\n-#4 0x7f961849e422 in ???\n-#5 0x7f961849e7c1 in ???\n-#6 0x7f961849b28c in ???\n-#7 0x7f96184919e8 in ???\n-#8 0x7f961bee13cc in ???\n-#9 0x7f961bfccbca in ???\n-#10 0x55a829f7a8dd in dgraph_unwgt\n+#0 0x7fc31c3b089d in ???\n+#1 0x7fc31c3afae3 in ???\n+#2 0x7fc31bfea8df in ???\n+#3 0x7fc318c28e8b in ???\n+#4 0x7fc318c29422 in ???\n+#5 0x7fc318c297c1 in ???\n+#6 0x7fc318c2628c in ???\n+#7 0x7fc318c1c9e8 in ???\n+#8 0x7fc31c66c3cc in ???\n+#9 0x7fc31c757bca in ???\n+#10 0x55b5eed118dd in dgraph_unwgt\n \tat ../../../f90/topo/dgraph_unwgtf90.f90:139\n-#11 0x55a829f7a1fe in main\n+#11 0x55b5eed111fe in main\n \tat ../../../f90/topo/dgraph_unwgtf90.f90:114\n --------------------------------------------------------------------------\n Primary job terminated normally, but 1 process returned\n a non-zero exit code. Per user-direction, the job has been aborted.\n --------------------------------------------------------------------------\n --------------------------------------------------------------------------\n-mpiexec.openmpi noticed that process rank 0 with PID 0 on node profitbricks-build5-amd64 exited on signal 11 (Segmentation fault).\n+mpiexec.openmpi noticed that process rank 0 with PID 0 on node i-capture-the-hostname exited on signal 11 (Segmentation fault).\n --------------------------------------------------------------------------\n \n \n profile1f90\n \n ./f90/profile\n fail\n- Failed to build profile1f90; make[2]: Entering directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/f90/profile'\n-/usr/bin/mpif90.openmpi -g -O2 -ffile-prefix-map=/build/1st/mpi-testsuite-3.2+dfsg=. -fstack-protector-strong -fcray-pointer -c -o profile1f90.o ../../../f90/profile/profile1f90.f90\n+ Failed to build profile1f90; make[2]: Entering directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/f90/profile'\n+/usr/bin/mpif90.openmpi -g -O2 -ffile-prefix-map=/build/mpi-testsuite-3.2+dfsg/2nd=. -fstack-protector-strong -fcray-pointer -c -o profile1f90.o ../../../f90/profile/profile1f90.f90\n ../../../f90/profile/profile1f90.f90:68:11:\n \n use mpi\n 1\n Error: 'mpi_send' of module 'mpi', imported at (1), is also the name of the current program unit\n ../../../f90/profile/profile1f90.f90:68:11:\n \n@@ -1007270,27 +1007337,27 @@\n Error: 'mpi_recv' of module 'mpi', imported at (1), is also the name of the current program unit\n ../../../f90/profile/profile1f90.f90:82:11:\n \n use mpi\n 1\n Error: Name 'mpi_recv' at (1) is an ambiguous reference to 'mpi_recv' from current program unit\n make[2]: *** [Makefile:434: profile1f90.o] Error 1\n-make[2]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/f90/profile'\n+make[2]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/f90/profile'\n \n \n keyvalmis\n 1\n ./errors/attr\n fail\n- [profitbricks-build5-amd64:27846] *** An error occurred in MPI_Win_set_attr\n-[profitbricks-build5-amd64:27846] *** reported by process [2829385729,0]\n-[profitbricks-build5-amd64:27846] *** on win pt2pt window 3\n-[profitbricks-build5-amd64:27846] *** MPI_ERR_OTHER: known error not in list\n-[profitbricks-build5-amd64:27846] *** MPI_ERRORS_ARE_FATAL (processes in this win will now abort,\n-[profitbricks-build5-amd64:27846] *** and potentially your MPI job)\n+ [i-capture-the-hostname:09931] *** An error occurred in MPI_Win_set_attr\n+[i-capture-the-hostname:09931] *** reported by process [2885484545,0]\n+[i-capture-the-hostname:09931] *** on win pt2pt window 3\n+[i-capture-the-hostname:09931] *** MPI_ERR_OTHER: known error not in list\n+[i-capture-the-hostname:09931] *** MPI_ERRORS_ARE_FATAL (processes in this win will now abort,\n+[i-capture-the-hostname:09931] *** and potentially your MPI job)\n \n \n noalias\n 2\n ./errors/coll\n fail\n Did not detect aliased arguments in MPI_Allreduce\n@@ -1007324,50 +1007391,50 @@\n --------------------------------------------------------------------------\n \n \n noalias3\n 2\n ./errors/coll\n fail\n- [profitbricks-build5-amd64:10279] *** Process received signal ***\n-[profitbricks-build5-amd64:10279] Signal: Segmentation fault (11)\n-[profitbricks-build5-amd64:10279] Signal code: (128)\n-[profitbricks-build5-amd64:10279] Failing at address: (nil)\n-[profitbricks-build5-amd64:10279] [ 0] /lib/x86_64-linux-gnu/libpthread.so.0(+0x126b0)[0x7f4bf959a6b0]\n-[profitbricks-build5-amd64:10279] [ 1] /usr/lib/x86_64-linux-gnu/libmpi.so.40(PMPI_Ialltoallw+0x117)[0x7f4bf9613287]\n-[profitbricks-build5-amd64:10279] [ 2] ./noalias3(+0x388f)[0x55d8cc02688f]\n-[profitbricks-build5-amd64:10279] [ 3] /lib/x86_64-linux-gnu/libc.so.6(__libc_start_main+0xeb)[0x7f4bf93e409b]\n-[profitbricks-build5-amd64:10279] [ 4] ./noalias3(+0x3d9a)[0x55d8cc026d9a]\n-[profitbricks-build5-amd64:10279] *** End of error message ***\n+ [i-capture-the-hostname:16299] *** Process received signal ***\n+[i-capture-the-hostname:16299] Signal: Segmentation fault (11)\n+[i-capture-the-hostname:16299] Signal code: (128)\n+[i-capture-the-hostname:16299] Failing at address: (nil)\n+[i-capture-the-hostname:16298] *** Process received signal ***\n+[i-capture-the-hostname:16298] Signal: Segmentation fault (11)\n+[i-capture-the-hostname:16298] Signal code: (128)\n+[i-capture-the-hostname:16298] Failing at address: (nil)\n+[i-capture-the-hostname:16298] [ 0] /lib/x86_64-linux-gnu/libpthread.so.0(+0x126b0)[0x7f9c822f56b0]\n+[i-capture-the-hostname:16298] [ 1] /usr/lib/x86_64-linux-gnu/libmpi.so.40(PMPI_Ialltoallw+0x117)[0x7f9c8236e287]\n+[i-capture-the-hostname:16298] [ 2] ./noalias3(+0x388f)[0x5616c4ad488f]\n+[i-capture-the-hostname:16298] [ 3] /lib/x86_64-linux-gnu/libc.so.6(__libc_start_main+0xeb)[0x7f9c8213f09b]\n+[i-capture-the-hostname:16298] [ 4] ./noalias3(+0x3d9a)[0x5616c4ad4d9a]\n+[i-capture-the-hostname:16298] *** End of error message ***\n+[i-capture-the-hostname:16299] [ 0] /lib/x86_64-linux-gnu/libpthread.so.0(+0x126b0)[0x7f18c167d6b0]\n+[i-capture-the-hostname:16299] [ 1] /usr/lib/x86_64-linux-gnu/libmpi.so.40(PMPI_Ialltoallw+0x117)[0x7f18c16f6287]\n+[i-capture-the-hostname:16299] [ 2] ./noalias3(+0x388f)[0x55840866d88f]\n+[i-capture-the-hostname:16299] [ 3] /lib/x86_64-linux-gnu/libc.so.6(__libc_start_main+0xeb)[0x7f18c14c709b]\n+[i-capture-the-hostname:16299] [ 4] ./noalias3(+0x3d9a)[0x55840866dd9a]\n+[i-capture-the-hostname:16299] *** End of error message ***\n --------------------------------------------------------------------------\n Primary job terminated normally, but 1 process returned\n a non-zero exit code. Per user-direction, the job has been aborted.\n --------------------------------------------------------------------------\n-[profitbricks-build5-amd64:10300] *** Process received signal ***\n-[profitbricks-build5-amd64:10300] Signal: Segmentation fault (11)\n-[profitbricks-build5-amd64:10300] Signal code: (128)\n-[profitbricks-build5-amd64:10300] Failing at address: (nil)\n-[profitbricks-build5-amd64:10300] [ 0] /lib/x86_64-linux-gnu/libpthread.so.0(+0x126b0)[0x7f7eed4ba6b0]\n-[profitbricks-build5-amd64:10300] [ 1] /usr/lib/x86_64-linux-gnu/libmpi.so.40(PMPI_Ialltoallw+0x117)[0x7f7eed533287]\n-[profitbricks-build5-amd64:10300] [ 2] ./noalias3(+0x388f)[0x562e132d888f]\n-[profitbricks-build5-amd64:10300] [ 3] /lib/x86_64-linux-gnu/libc.so.6(__libc_start_main+0xeb)[0x7f7eed30409b]\n-[profitbricks-build5-amd64:10300] [ 4] ./noalias3(+0x3d9a)[0x562e132d8d9a]\n-[profitbricks-build5-amd64:10300] *** End of error message ***\n --------------------------------------------------------------------------\n-mpiexec.openmpi noticed that process rank 0 with PID 0 on node profitbricks-build5-amd64 exited on signal 11 (Segmentation fault).\n+mpiexec.openmpi noticed that process rank 1 with PID 0 on node i-capture-the-hostname exited on signal 11 (Segmentation fault).\n --------------------------------------------------------------------------\n \n \n rerr\n 2\n ./errors/coll\n fail\n Did not detect invalid type/op pair (byte,max) in Allreduce\n- Found 2 errors\n-Did not detect invalid type/op pair (byte,max) in Allreduce\n+Did not detect invalid type/op pair (byte,max) in Allreduce\n+ Found 2 errors\n \n \n nb_rerr\n 2\n ./errors/coll\n fail\n Did not detect invalid type/op pair (byte,max) in Allreduce\n@@ -1007375,82 +1007442,82 @@\n Found 2 errors\n \n \n reduce_local\n 1\n ./errors/coll\n fail\n- [profitbricks-build5-amd64:17027] *** Process received signal ***\n-[profitbricks-build5-amd64:17027] Signal: Segmentation fault (11)\n-[profitbricks-build5-amd64:17027] Signal code: Address not mapped (1)\n-[profitbricks-build5-amd64:17027] Failing at address: 0x1\n-[profitbricks-build5-amd64:17027] [ 0] /lib/x86_64-linux-gnu/libpthread.so.0(+0x126b0)[0x7f06717fa6b0]\n-[profitbricks-build5-amd64:17027] [ 1] /usr/lib/x86_64-linux-gnu/libmpi.so.40(+0xaf2c0)[0x7f06718bf2c0]\n-[profitbricks-build5-amd64:17027] [ 2] /usr/lib/x86_64-linux-gnu/libmpi.so.40(mca_coll_base_reduce_local+0xba)[0x7f06718b5bda]\n-[profitbricks-build5-amd64:17027] [ 3] /usr/lib/x86_64-linux-gnu/libmpi.so.40(MPI_Reduce_local+0x131)[0x7f0671893291]\n-[profitbricks-build5-amd64:17027] [ 4] ./reduce_local(+0x25ef)[0x5611902995ef]\n-[profitbricks-build5-amd64:17027] [ 5] /lib/x86_64-linux-gnu/libc.so.6(__libc_start_main+0xeb)[0x7f067164409b]\n-[profitbricks-build5-amd64:17027] [ 6] ./reduce_local(+0x27ca)[0x5611902997ca]\n-[profitbricks-build5-amd64:17027] *** End of error message ***\n+ [i-capture-the-hostname:16439] *** Process received signal ***\n+[i-capture-the-hostname:16439] Signal: Segmentation fault (11)\n+[i-capture-the-hostname:16439] Signal code: Address not mapped (1)\n+[i-capture-the-hostname:16439] Failing at address: 0x1\n+[i-capture-the-hostname:16439] [ 0] /lib/x86_64-linux-gnu/libpthread.so.0(+0x126b0)[0x7f52c1ce56b0]\n+[i-capture-the-hostname:16439] [ 1] /usr/lib/x86_64-linux-gnu/libmpi.so.40(+0xaf2c0)[0x7f52c1daa2c0]\n+[i-capture-the-hostname:16439] [ 2] /usr/lib/x86_64-linux-gnu/libmpi.so.40(mca_coll_base_reduce_local+0xba)[0x7f52c1da0bda]\n+[i-capture-the-hostname:16439] [ 3] /usr/lib/x86_64-linux-gnu/libmpi.so.40(MPI_Reduce_local+0x131)[0x7f52c1d7e291]\n+[i-capture-the-hostname:16439] [ 4] ./reduce_local(+0x25ef)[0x55d790b705ef]\n+[i-capture-the-hostname:16439] [ 5] /lib/x86_64-linux-gnu/libc.so.6(__libc_start_main+0xeb)[0x7f52c1b2f09b]\n+[i-capture-the-hostname:16439] [ 6] ./reduce_local(+0x27ca)[0x55d790b707ca]\n+[i-capture-the-hostname:16439] *** End of error message ***\n --------------------------------------------------------------------------\n Primary job terminated normally, but 1 process returned\n a non-zero exit code. Per user-direction, the job has been aborted.\n --------------------------------------------------------------------------\n --------------------------------------------------------------------------\n-mpiexec.openmpi noticed that process rank 0 with PID 0 on node profitbricks-build5-amd64 exited on signal 11 (Segmentation fault).\n+mpiexec.openmpi noticed that process rank 0 with PID 0 on node i-capture-the-hostname exited on signal 11 (Segmentation fault).\n --------------------------------------------------------------------------\n \n \n bcastlength\n 2\n ./errors/coll\n fail\n- Did not detect mismatched length (long) on process 1\n- Found 1 errors\n+ Found 1 errors\n+Did not detect mismatched length (long) on process 1\n \n \n ibcastlength\n 2\n ./errors/coll\n fail\n Did not detect mismatched length (long) on process 1\n-MPI Error in MPI_Testall() (req 0 = 15)\n- Found 1 errors\n+ Found 1 errors\n+MPI Error in MPI_Testall() (req 0 = 15)\n \n \n cfree\n 2\n ./errors/comm\n fail\n- [profitbricks-build5-amd64:31465] *** Process received signal ***\n-[profitbricks-build5-amd64:31465] Signal: Segmentation fault (11)\n-[profitbricks-build5-amd64:31465] Signal code: Address not mapped (1)\n-[profitbricks-build5-amd64:31465] Failing at address: 0x10\n-[profitbricks-build5-amd64:31488] *** Process received signal ***\n-[profitbricks-build5-amd64:31488] Signal: Segmentation fault (11)\n-[profitbricks-build5-amd64:31488] Signal code: Address not mapped (1)\n-[profitbricks-build5-amd64:31488] Failing at address: 0x10\n-[profitbricks-build5-amd64:31488] [ 0] /lib/x86_64-linux-gnu/libpthread.so.0(+0x126b0)[0x7fad54f426b0]\n-[profitbricks-build5-amd64:31488] [ 1] /usr/lib/x86_64-linux-gnu/libmpi.so.40(MPI_Barrier+0x91)[0x7fad54fbb961]\n-[profitbricks-build5-amd64:31488] [ 2] ./cfree(+0x255e)[0x564800fc455e]\n-[profitbricks-build5-amd64:31488] [ 3] /lib/x86_64-linux-gnu/libc.so.6(__libc_start_main+0xeb)[0x7fad54d8c09b]\n-[profitbricks-build5-amd64:31488] [ 4] ./cfree(+0x261a)[0x564800fc461a]\n-[profitbricks-build5-amd64:31488] *** End of error message ***\n-[profitbricks-build5-amd64:31465] [ 0] /lib/x86_64-linux-gnu/libpthread.so.0(+0x126b0)[0x7f0bd9eba6b0]\n-[profitbricks-build5-amd64:31465] [ 1] /usr/lib/x86_64-linux-gnu/libmpi.so.40(MPI_Barrier+0x91)[0x7f0bd9f33961]\n-[profitbricks-build5-amd64:31465] [ 2] ./cfree(+0x255e)[0x5623ec9e255e]\n-[profitbricks-build5-amd64:31465] [ 3] /lib/x86_64-linux-gnu/libc.so.6(__libc_start_main+0xeb)[0x7f0bd9d0409b]\n-[profitbricks-build5-amd64:31465] [ 4] ./cfree(+0x261a)[0x5623ec9e261a]\n-[profitbricks-build5-amd64:31465] *** End of error message ***\n+ [i-capture-the-hostname:16661] *** Process received signal ***\n+[i-capture-the-hostname:16661] Signal: Segmentation fault (11)\n+[i-capture-the-hostname:16661] Signal code: Address not mapped (1)\n+[i-capture-the-hostname:16661] Failing at address: 0x10\n+[i-capture-the-hostname:16661] [ 0] /lib/x86_64-linux-gnu/libpthread.so.0(+0x126b0)[0x7f6612cd56b0]\n+[i-capture-the-hostname:16661] [ 1] /usr/lib/x86_64-linux-gnu/libmpi.so.40(MPI_Barrier+0x91)[0x7f6612d4e961]\n+[i-capture-the-hostname:16661] [ 2] ./cfree(+0x255e)[0x55cdbcc4d55e]\n+[i-capture-the-hostname:16661] [ 3] /lib/x86_64-linux-gnu/libc.so.6(__libc_start_main+0xeb)[0x7f6612b1f09b]\n+[i-capture-the-hostname:16661] [ 4] ./cfree(+0x261a)[0x55cdbcc4d61a]\n+[i-capture-the-hostname:16661] *** End of error message ***\n+[i-capture-the-hostname:16662] *** Process received signal ***\n+[i-capture-the-hostname:16662] Signal: Segmentation fault (11)\n+[i-capture-the-hostname:16662] Signal code: Address not mapped (1)\n+[i-capture-the-hostname:16662] Failing at address: 0x10\n --------------------------------------------------------------------------\n Primary job terminated normally, but 1 process returned\n a non-zero exit code. Per user-direction, the job has been aborted.\n --------------------------------------------------------------------------\n+[i-capture-the-hostname:16662] [ 0] /lib/x86_64-linux-gnu/libpthread.so.0(+0x126b0)[0x7fb4186656b0]\n+[i-capture-the-hostname:16662] [ 1] /usr/lib/x86_64-linux-gnu/libmpi.so.40(MPI_Barrier+0x91)[0x7fb4186de961]\n+[i-capture-the-hostname:16662] [ 2] ./cfree(+0x255e)[0x55a7f6cbf55e]\n+[i-capture-the-hostname:16662] [ 3] /lib/x86_64-linux-gnu/libc.so.6(__libc_start_main+0xeb)[0x7fb4184af09b]\n+[i-capture-the-hostname:16662] [ 4] ./cfree(+0x261a)[0x55a7f6cbf61a]\n+[i-capture-the-hostname:16662] *** End of error message ***\n --------------------------------------------------------------------------\n-mpiexec.openmpi noticed that process rank 0 with PID 0 on node profitbricks-build5-amd64 exited on signal 11 (Segmentation fault).\n+mpiexec.openmpi noticed that process rank 0 with PID 0 on node i-capture-the-hostname exited on signal 11 (Segmentation fault).\n --------------------------------------------------------------------------\n \n \n ccreate1\n 2\n ./errors/comm\n fail\n@@ -1007458,29 +1007525,31 @@\n Did not detect group of high ranks in low comm\n \n \n userdup\n 2\n ./errors/comm\n fail\n- Found 2 errors\n+ Unexpected error class = 14, expected user-defined class 93\n Unexpected error class = 14, expected user-defined class 93\n-Unexpected error class = 14, expected user-defined class 93\n+ Found 2 errors\n \n \n manysplit\n 2\n ./errors/comm\n pass\n \n \n gerr\n 1\n ./errors/group\n- pass\n+ fail\n+ Did not detect invalid handle (comm) in group_incl\n+ Found 1 errors\n \n \n proberank\n 1\n ./errors/pt2pt\n pass\n \n@@ -1007491,42 +1007560,42 @@\n pass\n \n \n truncmsg2\n 2\n ./errors/pt2pt\n fail\n- [profitbricks-build5-amd64:60615] *** Process received signal ***\n-[profitbricks-build5-amd64:60615] Signal: Segmentation fault (11)\n-[profitbricks-build5-amd64:60615] Signal code: Address not mapped (1)\n-[profitbricks-build5-amd64:60615] Failing at address: 0x3017ea34\n-[profitbricks-build5-amd64:60615] [ 0] /lib/x86_64-linux-gnu/libpthread.so.0(+0x126b0)[0x7fb68d1fa6b0]\n-[profitbricks-build5-amd64:60615] [ 1] /usr/lib/x86_64-linux-gnu/libmpi.so.40(ompi_datatype_create_indexed+0x1f)[0x7fb68d26b15f]\n-[profitbricks-build5-amd64:60615] [ 2] /usr/lib/x86_64-linux-gnu/libmpi.so.40(MPI_Type_indexed+0xd2)[0x7fb68d29d592]\n-[profitbricks-build5-amd64:60615] [ 3] ./truncmsg2(+0x264d)[0x55a13017364d]\n-[profitbricks-build5-amd64:60615] [ 4] /lib/x86_64-linux-gnu/libc.so.6(__libc_start_main+0xeb)[0x7fb68d04409b]\n-[profitbricks-build5-amd64:60615] [ 5] ./truncmsg2(+0x28ba)[0x55a1301738ba]\n-[profitbricks-build5-amd64:60615] *** End of error message ***\n-[profitbricks-build5-amd64:60635] *** Process received signal ***\n-[profitbricks-build5-amd64:60635] Signal: Segmentation fault (11)\n-[profitbricks-build5-amd64:60635] Signal code: Address not mapped (1)\n-[profitbricks-build5-amd64:60635] Failing at address: 0xffffffffeba84a34\n-[profitbricks-build5-amd64:60635] [ 0] /lib/x86_64-linux-gnu/libpthread.so.0(+0x126b0)[0x7f567a86a6b0]\n-[profitbricks-build5-amd64:60635] [ 1] /usr/lib/x86_64-linux-gnu/libmpi.so.40(ompi_datatype_create_indexed+0x1f)[0x7f567a8db15f]\n-[profitbricks-build5-amd64:60635] [ 2] /usr/lib/x86_64-linux-gnu/libmpi.so.40(MPI_Type_indexed+0xd2)[0x7f567a90d592]\n-[profitbricks-build5-amd64:60635] [ 3] ./truncmsg2(+0x264d)[0x561beba7964d]\n-[profitbricks-build5-amd64:60635] [ 4] /lib/x86_64-linux-gnu/libc.so.6(__libc_start_main+0xeb)[0x7f567a6b409b]\n-[profitbricks-build5-amd64:60635] [ 5] ./truncmsg2(+0x28ba)[0x561beba798ba]\n-[profitbricks-build5-amd64:60635] *** End of error message ***\n+ [i-capture-the-hostname:17676] *** Process received signal ***\n+[i-capture-the-hostname:17676] Signal: Segmentation fault (11)\n+[i-capture-the-hostname:17676] Signal code: Address not mapped (1)\n+[i-capture-the-hostname:17676] Failing at address: 0xffffffff907a5a34\n+[i-capture-the-hostname:17676] [ 0] /lib/x86_64-linux-gnu/libpthread.so.0(+0x126b0)[0x7f0cac9356b0]\n+[i-capture-the-hostname:17676] [ 1] /usr/lib/x86_64-linux-gnu/libmpi.so.40(ompi_datatype_create_indexed+0x1f)[0x7f0cac9a615f]\n+[i-capture-the-hostname:17676] [ 2] /usr/lib/x86_64-linux-gnu/libmpi.so.40(MPI_Type_indexed+0xd2)[0x7f0cac9d8592]\n+[i-capture-the-hostname:17676] [ 3] ./truncmsg2(+0x264d)[0x55c09079a64d]\n+[i-capture-the-hostname:17676] [ 4] /lib/x86_64-linux-gnu/libc.so.6(__libc_start_main+0xeb)[0x7f0cac77f09b]\n+[i-capture-the-hostname:17676] [ 5] ./truncmsg2(+0x28ba)[0x55c09079a8ba]\n+[i-capture-the-hostname:17676] *** End of error message ***\n+[i-capture-the-hostname:17679] *** Process received signal ***\n+[i-capture-the-hostname:17679] Signal: Segmentation fault (11)\n+[i-capture-the-hostname:17679] Signal code: Address not mapped (1)\n+[i-capture-the-hostname:17679] Failing at address: 0x3fcf1a34\n+[i-capture-the-hostname:17679] [ 0] /lib/x86_64-linux-gnu/libpthread.so.0(+0x126b0)[0x7f8bec2b56b0]\n+[i-capture-the-hostname:17679] [ 1] /usr/lib/x86_64-linux-gnu/libmpi.so.40(ompi_datatype_create_indexed+0x1f)[0x7f8bec32615f]\n+[i-capture-the-hostname:17679] [ 2] /usr/lib/x86_64-linux-gnu/libmpi.so.40(MPI_Type_indexed+0xd2)[0x7f8bec358592]\n+[i-capture-the-hostname:17679] [ 3] ./truncmsg2(+0x264d)[0x555b3fce664d]\n+[i-capture-the-hostname:17679] [ 4] /lib/x86_64-linux-gnu/libc.so.6(__libc_start_main+0xeb)[0x7f8bec0ff09b]\n+[i-capture-the-hostname:17679] [ 5] ./truncmsg2(+0x28ba)[0x555b3fce68ba]\n+[i-capture-the-hostname:17679] *** End of error message ***\n --------------------------------------------------------------------------\n Primary job terminated normally, but 1 process returned\n a non-zero exit code. Per user-direction, the job has been aborted.\n --------------------------------------------------------------------------\n --------------------------------------------------------------------------\n-mpiexec.openmpi noticed that process rank 0 with PID 0 on node profitbricks-build5-amd64 exited on signal 11 (Segmentation fault).\n+mpiexec.openmpi noticed that process rank 0 with PID 0 on node i-capture-the-hostname exited on signal 11 (Segmentation fault).\n --------------------------------------------------------------------------\n \n \n errinstatts\n 2\n ./errors/pt2pt\n pass\n@@ -1007572,45 +1007641,45 @@\n \n \n cas_type_check\n 2\n ./errors/rma\n fail\n cas_type_check: ../../../errors/rma/cas_type_check.c:73: main: Assertion `err_class == 3' failed.\n-[profitbricks-build5-amd64:23319] *** Process received signal ***\n-[profitbricks-build5-amd64:23319] Signal: Aborted (6)\n-[profitbricks-build5-amd64:23319] Signal code: (-6)\n-[profitbricks-build5-amd64:23319] [ 0] /lib/x86_64-linux-gnu/libpthread.so.0(+0x126b0)[0x7f7e950426b0]\n-[profitbricks-build5-amd64:23319] [ 1] /lib/x86_64-linux-gnu/libc.so.6(gsignal+0x10b)[0x7f7e94e9f85b]\n-[profitbricks-build5-amd64:23319] [ 2] /lib/x86_64-linux-gnu/libc.so.6(abort+0x121)[0x7f7e94e8a535]\n-[profitbricks-build5-amd64:23319] [ 3] /lib/x86_64-linux-gnu/libc.so.6(+0x2240f)[0x7f7e94e8a40f]\n-[profitbricks-build5-amd64:23319] [ 4] /lib/x86_64-linux-gnu/libc.so.6(+0x300a2)[0x7f7e94e980a2]\n-[profitbricks-build5-amd64:23319] [ 5] ./cas_type_check(+0x4722)[0x55d869d03722]\n-[profitbricks-build5-amd64:23319] [ 6] /lib/x86_64-linux-gnu/libc.so.6(__libc_start_main+0xeb)[0x7f7e94e8c09b]\n-[profitbricks-build5-amd64:23319] [ 7] ./cas_type_check(+0x47fa)[0x55d869d037fa]\n-[profitbricks-build5-amd64:23319] *** End of error message ***\n+[i-capture-the-hostname:18545] *** Process received signal ***\n+[i-capture-the-hostname:18545] Signal: Aborted (6)\n+[i-capture-the-hostname:18545] Signal code: (-6)\n cas_type_check: ../../../errors/rma/cas_type_check.c:73: main: Assertion `err_class == 3' failed.\n-[profitbricks-build5-amd64:23305] *** Process received signal ***\n-[profitbricks-build5-amd64:23305] Signal: Aborted (6)\n-[profitbricks-build5-amd64:23305] Signal code: (-6)\n-[profitbricks-build5-amd64:23305] [ 0] /lib/x86_64-linux-gnu/libpthread.so.0(+0x126b0)[0x7f5f36bc26b0]\n-[profitbricks-build5-amd64:23305] [ 1] /lib/x86_64-linux-gnu/libc.so.6(gsignal+0x10b)[0x7f5f36a1f85b]\n-[profitbricks-build5-amd64:23305] [ 2] /lib/x86_64-linux-gnu/libc.so.6(abort+0x121)[0x7f5f36a0a535]\n-[profitbricks-build5-amd64:23305] [ 3] /lib/x86_64-linux-gnu/libc.so.6(+0x2240f)[0x7f5f36a0a40f]\n-[profitbricks-build5-amd64:23305] [ 4] /lib/x86_64-linux-gnu/libc.so.6(+0x300a2)[0x7f5f36a180a2]\n-[profitbricks-build5-amd64:23305] [ 5] ./cas_type_check(+0x4722)[0x5557f00f0722]\n-[profitbricks-build5-amd64:23305] [ 6] /lib/x86_64-linux-gnu/libc.so.6(__libc_start_main+0xeb)[0x7f5f36a0c09b]\n-[profitbricks-build5-amd64:23305] [ 7] ./cas_type_check(+0x47fa)[0x5557f00f07fa]\n-[profitbricks-build5-amd64:23305] *** End of error message ***\n+[i-capture-the-hostname:18544] *** Process received signal ***\n+[i-capture-the-hostname:18544] Signal: Aborted (6)\n+[i-capture-the-hostname:18544] Signal code: (-6)\n+[i-capture-the-hostname:18545] [ 0] /lib/x86_64-linux-gnu/libpthread.so.0(+0x126b0)[0x7ff4f84856b0]\n+[i-capture-the-hostname:18545] [ 1] /lib/x86_64-linux-gnu/libc.so.6(gsignal+0x10b)[0x7ff4f82e285b]\n+[i-capture-the-hostname:18545] [ 2] /lib/x86_64-linux-gnu/libc.so.6(abort+0x121)[0x7ff4f82cd535]\n+[i-capture-the-hostname:18545] [ 3] /lib/x86_64-linux-gnu/libc.so.6(+0x2240f)[0x7ff4f82cd40f]\n+[i-capture-the-hostname:18545] [ 4] /lib/x86_64-linux-gnu/libc.so.6(+0x300a2)[0x7ff4f82db0a2]\n+[i-capture-the-hostname:18545] [ 5] ./cas_type_check(+0x4722)[0x5579237e3722]\n+[i-capture-the-hostname:18545] [ 6] /lib/x86_64-linux-gnu/libc.so.6(__libc_start_main+0xeb)[0x7ff4f82cf09b]\n+[i-capture-the-hostname:18545] [ 7] ./cas_type_check(+0x47fa)[0x5579237e37fa]\n+[i-capture-the-hostname:18545] *** End of error message ***\n+[i-capture-the-hostname:18544] [ 0] /lib/x86_64-linux-gnu/libpthread.so.0(+0x126b0)[0x7fcd24de56b0]\n+[i-capture-the-hostname:18544] [ 1] /lib/x86_64-linux-gnu/libc.so.6(gsignal+0x10b)[0x7fcd24c4285b]\n+[i-capture-the-hostname:18544] [ 2] /lib/x86_64-linux-gnu/libc.so.6(abort+0x121)[0x7fcd24c2d535]\n+[i-capture-the-hostname:18544] [ 3] /lib/x86_64-linux-gnu/libc.so.6(+0x2240f)[0x7fcd24c2d40f]\n+[i-capture-the-hostname:18544] [ 4] /lib/x86_64-linux-gnu/libc.so.6(+0x300a2)[0x7fcd24c3b0a2]\n+[i-capture-the-hostname:18544] [ 5] ./cas_type_check(+0x4722)[0x55b4e0141722]\n+[i-capture-the-hostname:18544] [ 6] /lib/x86_64-linux-gnu/libc.so.6(__libc_start_main+0xeb)[0x7fcd24c2f09b]\n+[i-capture-the-hostname:18544] [ 7] ./cas_type_check(+0x47fa)[0x55b4e01417fa]\n+[i-capture-the-hostname:18544] *** End of error message ***\n --------------------------------------------------------------------------\n Primary job terminated normally, but 1 process returned\n a non-zero exit code. Per user-direction, the job has been aborted.\n --------------------------------------------------------------------------\n --------------------------------------------------------------------------\n-mpiexec.openmpi noticed that process rank 0 with PID 0 on node profitbricks-build5-amd64 exited on signal 6 (Aborted).\n+mpiexec.openmpi noticed that process rank 0 with PID 0 on node i-capture-the-hostname exited on signal 6 (Aborted).\n --------------------------------------------------------------------------\n \n \n win_sync_unlock\n 2\n ./errors/rma\n pass\n@@ -1007618,38 +1007687,38 @@\n \n win_sync_free_pt\n 2\n ./errors/rma\n fail\n 0: Operation succeeded, when it should have failed\n 1: Operation succeeded, when it should have failed\n-[profitbricks-build5-amd64:42526] *** An error occurred in MPI_Win_unlock\n-[profitbricks-build5-amd64:42526] *** reported by process [1724973057,1]\n-[profitbricks-build5-amd64:42526] *** on communicator MPI_COMM_WORLD\n-[profitbricks-build5-amd64:42526] *** MPI_ERR_WIN: invalid window\n-[profitbricks-build5-amd64:42526] *** MPI_ERRORS_ARE_FATAL (processes in this communicator will now abort,\n-[profitbricks-build5-amd64:42526] *** and potentially your MPI job)\n-[profitbricks-build5-amd64:41475] 1 more process has sent help message help-mpi-errors.txt / mpi_errors_are_fatal\n-[profitbricks-build5-amd64:41475] Set MCA parameter "orte_base_help_aggregate" to 0 to see all help / error messages\n+[i-capture-the-hostname:20184] *** An error occurred in MPI_Win_unlock\n+[i-capture-the-hostname:20184] *** reported by process [3281518593,1]\n+[i-capture-the-hostname:20184] *** on communicator MPI_COMM_WORLD\n+[i-capture-the-hostname:20184] *** MPI_ERR_WIN: invalid window\n+[i-capture-the-hostname:20184] *** MPI_ERRORS_ARE_FATAL (processes in this communicator will now abort,\n+[i-capture-the-hostname:20184] *** and potentially your MPI job)\n+[i-capture-the-hostname:20169] 1 more process has sent help message help-mpi-errors.txt / mpi_errors_are_fatal\n+[i-capture-the-hostname:20169] Set MCA parameter "orte_base_help_aggregate" to 0 to see all help / error messages\n \n \n win_sync_free_at\n 2\n ./errors/rma\n fail\n- 0: Operation succeeded, when it should have failed\n-1: Operation succeeded, when it should have failed\n-[profitbricks-build5-amd64:46026] *** An error occurred in MPI_Win_complete\n-[profitbricks-build5-amd64:46026] *** reported by process [1779564545,1]\n-[profitbricks-build5-amd64:46026] *** on communicator MPI_COMM_WORLD\n-[profitbricks-build5-amd64:46026] *** MPI_ERR_WIN: invalid window\n-[profitbricks-build5-amd64:46026] *** MPI_ERRORS_ARE_FATAL (processes in this communicator will now abort,\n-[profitbricks-build5-amd64:46026] *** and potentially your MPI job)\n-[profitbricks-build5-amd64:44736] 1 more process has sent help message help-mpi-errors.txt / mpi_errors_are_fatal\n-[profitbricks-build5-amd64:44736] Set MCA parameter "orte_base_help_aggregate" to 0 to see all help / error messages\n+ 1: Operation succeeded, when it should have failed\n+0: Operation succeeded, when it should have failed\n+[i-capture-the-hostname:20289] *** An error occurred in MPI_Win_complete\n+[i-capture-the-hostname:20289] *** reported by process [3260416001,1]\n+[i-capture-the-hostname:20289] *** on communicator MPI_COMM_WORLD\n+[i-capture-the-hostname:20289] *** MPI_ERR_WIN: invalid window\n+[i-capture-the-hostname:20289] *** MPI_ERRORS_ARE_FATAL (processes in this communicator will now abort,\n+[i-capture-the-hostname:20289] *** and potentially your MPI job)\n+[i-capture-the-hostname:20231] 1 more process has sent help message help-mpi-errors.txt / mpi_errors_are_fatal\n+[i-capture-the-hostname:20231] Set MCA parameter "orte_base_help_aggregate" to 0 to see all help / error messages\n \n \n win_sync_complete\n 2\n ./errors/rma\n pass\n \n@@ -1007660,17 +1007729,17 @@\n pass\n \n \n win_sync_lock_pt\n 2\n ./errors/rma\n fail\n- 0: Expected MPI_ERR_RMA_SYNC, got:\n+ 1: Expected MPI_ERR_RMA_SYNC, got:\n MPI_ERR_RMA_CONFLICT: rma conflict during operation\n-1: Expected MPI_ERR_RMA_SYNC, got:\n+0: Expected MPI_ERR_RMA_SYNC, got:\n MPI_ERR_RMA_CONFLICT: rma conflict during operation\n \n \n win_sync_lock_fence\n 2\n ./errors/rma\n pass\n@@ -1007688,15 +1007757,15 @@\n pass\n \n \n badport\n 2\n ./errors/spawn\n fail\n- [profitbricks-build5-amd64:63621] [[12841,1],0] ORTE_ERROR_LOG: Bad parameter in file util/name_fns.c at line 420\n+ [i-capture-the-hostname:21569] [[55670,1],0] ORTE_ERROR_LOG: Bad parameter in file util/name_fns.c at line 420\n --------------------------------------------------------------------------\n An operation involving MPI_Connect and/or MPI_Accept was called with\n an unrecognized port string. This typically happens when passing the\n string on a cmd line and failing to properly quote it to protect\n against the special characters it includes\n --------------------------------------------------------------------------\n --------------------------------------------------------------------------\n@@ -1007762,19 +1007831,19 @@\n mca_fbtl_posix_preadv: error in (p)readv:Invalid argument\n \n \n errgetx\n \n ./errors/cxx/errhan\n fail\n- Failed to build errgetx; make[2]: Entering directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/errors/cxx/errhan'\n+ Failed to build errgetx; make[2]: Entering directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/errors/cxx/errhan'\n (cd ../../../cxx/util && make mtest.o)\n-make[3]: Entering directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/util'\n+make[3]: Entering directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/util'\n depbase=`echo mtest.o | sed 's|[^/]*$|.deps/&|;s|\\.o$||'`;\\\n-/usr/bin/mpicxx.openmpi -DHAVE_CONFIG_H -I. -I../../../cxx/util -I../../include -I../../include -I../../../include -Wdate-time -D_FORTIFY_SOURCE=2 -g -O2 -ffile-prefix-map=/build/1st/mpi-testsuite-3.2+dfsg=. -fstack-protector-strong -Wformat -Werror=format-security -MT mtest.o -MD -MP -MF $depbase.Tpo -c -o mtest.o ../../../cxx/util/mtest.cxx &&\\\n+/usr/bin/mpicxx.openmpi -DHAVE_CONFIG_H -I. -I../../../cxx/util -I../../include -I../../include -I../../../include -Wdate-time -D_FORTIFY_SOURCE=2 -g -O2 -ffile-prefix-map=/build/mpi-testsuite-3.2+dfsg/2nd=. -fstack-protector-strong -Wformat -Werror=format-security -MT mtest.o -MD -MP -MF $depbase.Tpo -c -o mtest.o ../../../cxx/util/mtest.cxx &&\\\n mv -f $depbase.Tpo $depbase.Po\n ../../../cxx/util/mtest.cxx: In function 'void* MTestTypeContigInitRecv(MTestDatatype*)':\n ../../../cxx/util/mtest.cxx:189:42: warning: 'int MPI_Type_extent(MPI_Datatype, MPI_Aint*)' is deprecated: MPI_Type_extent is superseded by MPI_Type_get_extent in MPI-2.0 [-Wdeprecated-declarations]\n MPI_Type_extent( mtype->datatype, &size );\n ^\n In file included from ../../../cxx/util/mtest.cxx:7:\n /usr/lib/x86_64-linux-gnu/openmpi/include/mpi.h:1758:20: note: declared here\n@@ -1007806,28 +1007875,28 @@\n ../../../cxx/util/mtest.cxx:848:50: error: 'NULL_COPY_FN' is not a member of 'MPI::Win'\n mem_keyval = MPI::Win::Create_keyval( MPI::Win::NULL_COPY_FN,\n ^~~~~~~~~~~~\n ../../../cxx/util/mtest.cxx:849:22: error: 'NULL_DELETE_FN' is not a member of 'MPI::Win'\n MPI::Win::NULL_DELETE_FN, 0 );\n ^~~~~~~~~~~~~~\n make[3]: *** [Makefile:416: mtest.o] Error 1\n-make[3]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/util'\n+make[3]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/util'\n make[2]: *** [Makefile:670: ../../../cxx/util/mtest.o] Error 2\n-make[2]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/errors/cxx/errhan'\n+make[2]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/errors/cxx/errhan'\n \n \n errsetx\n \n ./errors/cxx/errhan\n fail\n- Failed to build errsetx; make[2]: Entering directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/errors/cxx/errhan'\n+ Failed to build errsetx; make[2]: Entering directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/errors/cxx/errhan'\n (cd ../../../cxx/util && make mtest.o)\n-make[3]: Entering directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/util'\n+make[3]: Entering directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/util'\n depbase=`echo mtest.o | sed 's|[^/]*$|.deps/&|;s|\\.o$||'`;\\\n-/usr/bin/mpicxx.openmpi -DHAVE_CONFIG_H -I. -I../../../cxx/util -I../../include -I../../include -I../../../include -Wdate-time -D_FORTIFY_SOURCE=2 -g -O2 -ffile-prefix-map=/build/1st/mpi-testsuite-3.2+dfsg=. -fstack-protector-strong -Wformat -Werror=format-security -MT mtest.o -MD -MP -MF $depbase.Tpo -c -o mtest.o ../../../cxx/util/mtest.cxx &&\\\n+/usr/bin/mpicxx.openmpi -DHAVE_CONFIG_H -I. -I../../../cxx/util -I../../include -I../../include -I../../../include -Wdate-time -D_FORTIFY_SOURCE=2 -g -O2 -ffile-prefix-map=/build/mpi-testsuite-3.2+dfsg/2nd=. -fstack-protector-strong -Wformat -Werror=format-security -MT mtest.o -MD -MP -MF $depbase.Tpo -c -o mtest.o ../../../cxx/util/mtest.cxx &&\\\n mv -f $depbase.Tpo $depbase.Po\n ../../../cxx/util/mtest.cxx: In function 'void* MTestTypeContigInitRecv(MTestDatatype*)':\n ../../../cxx/util/mtest.cxx:189:42: warning: 'int MPI_Type_extent(MPI_Datatype, MPI_Aint*)' is deprecated: MPI_Type_extent is superseded by MPI_Type_get_extent in MPI-2.0 [-Wdeprecated-declarations]\n MPI_Type_extent( mtype->datatype, &size );\n ^\n In file included from ../../../cxx/util/mtest.cxx:7:\n /usr/lib/x86_64-linux-gnu/openmpi/include/mpi.h:1758:20: note: declared here\n@@ -1007859,28 +1007928,28 @@\n ../../../cxx/util/mtest.cxx:848:50: error: 'NULL_COPY_FN' is not a member of 'MPI::Win'\n mem_keyval = MPI::Win::Create_keyval( MPI::Win::NULL_COPY_FN,\n ^~~~~~~~~~~~\n ../../../cxx/util/mtest.cxx:849:22: error: 'NULL_DELETE_FN' is not a member of 'MPI::Win'\n MPI::Win::NULL_DELETE_FN, 0 );\n ^~~~~~~~~~~~~~\n make[3]: *** [Makefile:416: mtest.o] Error 1\n-make[3]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/util'\n+make[3]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/util'\n make[2]: *** [Makefile:670: ../../../cxx/util/mtest.o] Error 2\n-make[2]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/errors/cxx/errhan'\n+make[2]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/errors/cxx/errhan'\n \n \n throwtest\n \n ./errors/cxx/errhan\n fail\n- Failed to build throwtest; make[2]: Entering directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/errors/cxx/errhan'\n+ Failed to build throwtest; make[2]: Entering directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/errors/cxx/errhan'\n (cd ../../../cxx/util && make mtest.o)\n-make[3]: Entering directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/util'\n+make[3]: Entering directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/util'\n depbase=`echo mtest.o | sed 's|[^/]*$|.deps/&|;s|\\.o$||'`;\\\n-/usr/bin/mpicxx.openmpi -DHAVE_CONFIG_H -I. -I../../../cxx/util -I../../include -I../../include -I../../../include -Wdate-time -D_FORTIFY_SOURCE=2 -g -O2 -ffile-prefix-map=/build/1st/mpi-testsuite-3.2+dfsg=. -fstack-protector-strong -Wformat -Werror=format-security -MT mtest.o -MD -MP -MF $depbase.Tpo -c -o mtest.o ../../../cxx/util/mtest.cxx &&\\\n+/usr/bin/mpicxx.openmpi -DHAVE_CONFIG_H -I. -I../../../cxx/util -I../../include -I../../include -I../../../include -Wdate-time -D_FORTIFY_SOURCE=2 -g -O2 -ffile-prefix-map=/build/mpi-testsuite-3.2+dfsg/2nd=. -fstack-protector-strong -Wformat -Werror=format-security -MT mtest.o -MD -MP -MF $depbase.Tpo -c -o mtest.o ../../../cxx/util/mtest.cxx &&\\\n mv -f $depbase.Tpo $depbase.Po\n ../../../cxx/util/mtest.cxx: In function 'void* MTestTypeContigInitRecv(MTestDatatype*)':\n ../../../cxx/util/mtest.cxx:189:42: warning: 'int MPI_Type_extent(MPI_Datatype, MPI_Aint*)' is deprecated: MPI_Type_extent is superseded by MPI_Type_get_extent in MPI-2.0 [-Wdeprecated-declarations]\n MPI_Type_extent( mtype->datatype, &size );\n ^\n In file included from ../../../cxx/util/mtest.cxx:7:\n /usr/lib/x86_64-linux-gnu/openmpi/include/mpi.h:1758:20: note: declared here\n@@ -1007912,26 +1007981,26 @@\n ../../../cxx/util/mtest.cxx:848:50: error: 'NULL_COPY_FN' is not a member of 'MPI::Win'\n mem_keyval = MPI::Win::Create_keyval( MPI::Win::NULL_COPY_FN,\n ^~~~~~~~~~~~\n ../../../cxx/util/mtest.cxx:849:22: error: 'NULL_DELETE_FN' is not a member of 'MPI::Win'\n MPI::Win::NULL_DELETE_FN, 0 );\n ^~~~~~~~~~~~~~\n make[3]: *** [Makefile:416: mtest.o] Error 1\n-make[3]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/util'\n+make[3]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/util'\n make[2]: *** [Makefile:670: ../../../cxx/util/mtest.o] Error 2\n-make[2]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/errors/cxx/errhan'\n+make[2]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/errors/cxx/errhan'\n \n \n commerrx\n \n ./errors/cxx/errhan\n fail\n- Failed to build commerrx; make[2]: Entering directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/errors/cxx/errhan'\n+ Failed to build commerrx; make[2]: Entering directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/errors/cxx/errhan'\n depbase=`echo commerrx.o | sed 's|[^/]*$|.deps/&|;s|\\.o$||'`;\\\n-/usr/bin/mpicxx.openmpi -DHAVE_CONFIG_H -I. -I../../../../errors/cxx/errhan -I../../../include -I../../../include -I../../../../include -Wdate-time -D_FORTIFY_SOURCE=2 -g -O2 -ffile-prefix-map=/build/1st/mpi-testsuite-3.2+dfsg=. -fstack-protector-strong -Wformat -Werror=format-security -MT commerrx.o -MD -MP -MF $depbase.Tpo -c -o commerrx.o ../../../../errors/cxx/errhan/commerrx.cxx &&\\\n+/usr/bin/mpicxx.openmpi -DHAVE_CONFIG_H -I. -I../../../../errors/cxx/errhan -I../../../include -I../../../include -I../../../../include -Wdate-time -D_FORTIFY_SOURCE=2 -g -O2 -ffile-prefix-map=/build/mpi-testsuite-3.2+dfsg/2nd=. -fstack-protector-strong -Wformat -Werror=format-security -MT commerrx.o -MD -MP -MF $depbase.Tpo -c -o commerrx.o ../../../../errors/cxx/errhan/commerrx.cxx &&\\\n mv -f $depbase.Tpo $depbase.Po\n ../../../../errors/cxx/errhan/commerrx.cxx: In function 'int testNullCommCall()':\n ../../../../errors/cxx/errhan/commerrx.cxx:48:34: error: cannot allocate an object of abstract type 'MPI::Comm'\n const MPI::Comm &comm = MPI::COMM_NULL;\n ^~~~~~~~~\n In file included from /usr/lib/x86_64-linux-gnu/openmpi/include/openmpi/ompi/mpi/cxx/mpicxx.h:222,\n from /usr/lib/x86_64-linux-gnu/openmpi/include/mpi.h:2684,\n@@ -1007942,26 +1008011,26 @@\n In file included from /usr/lib/x86_64-linux-gnu/openmpi/include/openmpi/ompi/mpi/cxx/mpicxx.h:222,\n from /usr/lib/x86_64-linux-gnu/openmpi/include/mpi.h:2684,\n from ../../../../errors/cxx/errhan/commerrx.cxx:13:\n /usr/lib/x86_64-linux-gnu/openmpi/include/openmpi/ompi/mpi/cxx/comm.h:262:17: note: \t'virtual MPI::Comm& MPI::Comm::Clone() const'\n virtual Comm& Clone() const = 0;\n ^~~~~\n make[2]: *** [Makefile:450: commerrx.o] Error 1\n-make[2]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/errors/cxx/errhan'\n+make[2]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/errors/cxx/errhan'\n \n \n fileerrretx\n \n ./errors/cxx/io\n fail\n- Failed to build fileerrretx; make[2]: Entering directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/errors/cxx/io'\n+ Failed to build fileerrretx; make[2]: Entering directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/errors/cxx/io'\n (cd ../../../cxx/util && make mtest.o)\n-make[3]: Entering directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/util'\n+make[3]: Entering directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/util'\n depbase=`echo mtest.o | sed 's|[^/]*$|.deps/&|;s|\\.o$||'`;\\\n-/usr/bin/mpicxx.openmpi -DHAVE_CONFIG_H -I. -I../../../cxx/util -I../../include -I../../include -I../../../include -Wdate-time -D_FORTIFY_SOURCE=2 -g -O2 -ffile-prefix-map=/build/1st/mpi-testsuite-3.2+dfsg=. -fstack-protector-strong -Wformat -Werror=format-security -MT mtest.o -MD -MP -MF $depbase.Tpo -c -o mtest.o ../../../cxx/util/mtest.cxx &&\\\n+/usr/bin/mpicxx.openmpi -DHAVE_CONFIG_H -I. -I../../../cxx/util -I../../include -I../../include -I../../../include -Wdate-time -D_FORTIFY_SOURCE=2 -g -O2 -ffile-prefix-map=/build/mpi-testsuite-3.2+dfsg/2nd=. -fstack-protector-strong -Wformat -Werror=format-security -MT mtest.o -MD -MP -MF $depbase.Tpo -c -o mtest.o ../../../cxx/util/mtest.cxx &&\\\n mv -f $depbase.Tpo $depbase.Po\n ../../../cxx/util/mtest.cxx: In function 'void* MTestTypeContigInitRecv(MTestDatatype*)':\n ../../../cxx/util/mtest.cxx:189:42: warning: 'int MPI_Type_extent(MPI_Datatype, MPI_Aint*)' is deprecated: MPI_Type_extent is superseded by MPI_Type_get_extent in MPI-2.0 [-Wdeprecated-declarations]\n MPI_Type_extent( mtype->datatype, &size );\n ^\n In file included from ../../../cxx/util/mtest.cxx:7:\n /usr/lib/x86_64-linux-gnu/openmpi/include/mpi.h:1758:20: note: declared here\n@@ -1007993,28 +1008062,28 @@\n ../../../cxx/util/mtest.cxx:848:50: error: 'NULL_COPY_FN' is not a member of 'MPI::Win'\n mem_keyval = MPI::Win::Create_keyval( MPI::Win::NULL_COPY_FN,\n ^~~~~~~~~~~~\n ../../../cxx/util/mtest.cxx:849:22: error: 'NULL_DELETE_FN' is not a member of 'MPI::Win'\n MPI::Win::NULL_DELETE_FN, 0 );\n ^~~~~~~~~~~~~~\n make[3]: *** [Makefile:416: mtest.o] Error 1\n-make[3]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/util'\n+make[3]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/util'\n make[2]: *** [Makefile:661: ../../../cxx/util/mtest.o] Error 2\n-make[2]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/errors/cxx/io'\n+make[2]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/errors/cxx/io'\n \n \n errgetfilex\n \n ./errors/cxx/io\n fail\n- Failed to build errgetfilex; make[2]: Entering directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/errors/cxx/io'\n+ Failed to build errgetfilex; make[2]: Entering directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/errors/cxx/io'\n (cd ../../../cxx/util && make mtest.o)\n-make[3]: Entering directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/util'\n+make[3]: Entering directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/util'\n depbase=`echo mtest.o | sed 's|[^/]*$|.deps/&|;s|\\.o$||'`;\\\n-/usr/bin/mpicxx.openmpi -DHAVE_CONFIG_H -I. -I../../../cxx/util -I../../include -I../../include -I../../../include -Wdate-time -D_FORTIFY_SOURCE=2 -g -O2 -ffile-prefix-map=/build/1st/mpi-testsuite-3.2+dfsg=. -fstack-protector-strong -Wformat -Werror=format-security -MT mtest.o -MD -MP -MF $depbase.Tpo -c -o mtest.o ../../../cxx/util/mtest.cxx &&\\\n+/usr/bin/mpicxx.openmpi -DHAVE_CONFIG_H -I. -I../../../cxx/util -I../../include -I../../include -I../../../include -Wdate-time -D_FORTIFY_SOURCE=2 -g -O2 -ffile-prefix-map=/build/mpi-testsuite-3.2+dfsg/2nd=. -fstack-protector-strong -Wformat -Werror=format-security -MT mtest.o -MD -MP -MF $depbase.Tpo -c -o mtest.o ../../../cxx/util/mtest.cxx &&\\\n mv -f $depbase.Tpo $depbase.Po\n ../../../cxx/util/mtest.cxx: In function 'void* MTestTypeContigInitRecv(MTestDatatype*)':\n ../../../cxx/util/mtest.cxx:189:42: warning: 'int MPI_Type_extent(MPI_Datatype, MPI_Aint*)' is deprecated: MPI_Type_extent is superseded by MPI_Type_get_extent in MPI-2.0 [-Wdeprecated-declarations]\n MPI_Type_extent( mtype->datatype, &size );\n ^\n In file included from ../../../cxx/util/mtest.cxx:7:\n /usr/lib/x86_64-linux-gnu/openmpi/include/mpi.h:1758:20: note: declared here\n@@ -1008046,28 +1008115,28 @@\n ../../../cxx/util/mtest.cxx:848:50: error: 'NULL_COPY_FN' is not a member of 'MPI::Win'\n mem_keyval = MPI::Win::Create_keyval( MPI::Win::NULL_COPY_FN,\n ^~~~~~~~~~~~\n ../../../cxx/util/mtest.cxx:849:22: error: 'NULL_DELETE_FN' is not a member of 'MPI::Win'\n MPI::Win::NULL_DELETE_FN, 0 );\n ^~~~~~~~~~~~~~\n make[3]: *** [Makefile:416: mtest.o] Error 1\n-make[3]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/util'\n+make[3]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/util'\n make[2]: *** [Makefile:661: ../../../cxx/util/mtest.o] Error 2\n-make[2]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/errors/cxx/io'\n+make[2]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/errors/cxx/io'\n \n \n throwtestfilex\n \n ./errors/cxx/io\n fail\n- Failed to build throwtestfilex; make[2]: Entering directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/errors/cxx/io'\n+ Failed to build throwtestfilex; make[2]: Entering directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/errors/cxx/io'\n (cd ../../../cxx/util && make mtest.o)\n-make[3]: Entering directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/util'\n+make[3]: Entering directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/util'\n depbase=`echo mtest.o | sed 's|[^/]*$|.deps/&|;s|\\.o$||'`;\\\n-/usr/bin/mpicxx.openmpi -DHAVE_CONFIG_H -I. -I../../../cxx/util -I../../include -I../../include -I../../../include -Wdate-time -D_FORTIFY_SOURCE=2 -g -O2 -ffile-prefix-map=/build/1st/mpi-testsuite-3.2+dfsg=. -fstack-protector-strong -Wformat -Werror=format-security -MT mtest.o -MD -MP -MF $depbase.Tpo -c -o mtest.o ../../../cxx/util/mtest.cxx &&\\\n+/usr/bin/mpicxx.openmpi -DHAVE_CONFIG_H -I. -I../../../cxx/util -I../../include -I../../include -I../../../include -Wdate-time -D_FORTIFY_SOURCE=2 -g -O2 -ffile-prefix-map=/build/mpi-testsuite-3.2+dfsg/2nd=. -fstack-protector-strong -Wformat -Werror=format-security -MT mtest.o -MD -MP -MF $depbase.Tpo -c -o mtest.o ../../../cxx/util/mtest.cxx &&\\\n mv -f $depbase.Tpo $depbase.Po\n ../../../cxx/util/mtest.cxx: In function 'void* MTestTypeContigInitRecv(MTestDatatype*)':\n ../../../cxx/util/mtest.cxx:189:42: warning: 'int MPI_Type_extent(MPI_Datatype, MPI_Aint*)' is deprecated: MPI_Type_extent is superseded by MPI_Type_get_extent in MPI-2.0 [-Wdeprecated-declarations]\n MPI_Type_extent( mtype->datatype, &size );\n ^\n In file included from ../../../cxx/util/mtest.cxx:7:\n /usr/lib/x86_64-linux-gnu/openmpi/include/mpi.h:1758:20: note: declared here\n@@ -1008099,17 +1008168,17 @@\n ../../../cxx/util/mtest.cxx:848:50: error: 'NULL_COPY_FN' is not a member of 'MPI::Win'\n mem_keyval = MPI::Win::Create_keyval( MPI::Win::NULL_COPY_FN,\n ^~~~~~~~~~~~\n ../../../cxx/util/mtest.cxx:849:22: error: 'NULL_DELETE_FN' is not a member of 'MPI::Win'\n MPI::Win::NULL_DELETE_FN, 0 );\n ^~~~~~~~~~~~~~\n make[3]: *** [Makefile:416: mtest.o] Error 1\n-make[3]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/cxx/util'\n+make[3]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/cxx/util'\n make[2]: *** [Makefile:661: ../../../cxx/util/mtest.o] Error 2\n-make[2]: Leaving directory '/build/1st/mpi-testsuite-3.2+dfsg/build-openmpi/errors/cxx/io'\n+make[2]: Leaving directory '/build/mpi-testsuite-3.2+dfsg/2nd/build-openmpi/errors/cxx/io'\n \n \n uerrhandf90\n 1\n ./errors/f90/io\n fail\n Failed to detect error in use of MPI_PUT\n" } ] } ] } ] } ] } ] }