Diff of the two buildlogs: -- --- b1/build.log 2025-03-10 02:11:12.688468564 +0000 +++ b2/build.log 2025-03-10 03:47:27.700946485 +0000 @@ -1,6 +1,6 @@ I: pbuilder: network access will be disabled during build -I: Current time: Sun Mar 9 11:10:58 -12 2025 -I: pbuilder-time-stamp: 1741561858 +I: Current time: Sun Apr 12 22:34:14 +14 2026 +I: pbuilder-time-stamp: 1775982854 I: Building the build Environment I: extracting base tarball [/var/cache/pbuilder/trixie-reproducible-base.tgz] I: copying local configuration @@ -28,52 +28,84 @@ dpkg-source: info: applying 0004-Add-support-for-USE_SYSTEM_JEMALLOC-flag.patch I: Not using root during the build. I: Installing the build-deps -I: user script /srv/workspace/pbuilder/1755253/tmp/hooks/D02_print_environment starting +I: user script /srv/workspace/pbuilder/4180030/tmp/hooks/D01_modify_environment starting +debug: Running on ionos15-amd64. +I: Changing host+domainname to test build reproducibility +I: Adding a custom variable just for the fun of it... +I: Changing /bin/sh to bash +'/bin/sh' -> '/bin/bash' +lrwxrwxrwx 1 root root 9 Apr 12 08:34 /bin/sh -> /bin/bash +I: Setting pbuilder2's login shell to /bin/bash +I: Setting pbuilder2's GECOS to second user,second room,second work-phone,second home-phone,second other +I: user script /srv/workspace/pbuilder/4180030/tmp/hooks/D01_modify_environment finished +I: user script /srv/workspace/pbuilder/4180030/tmp/hooks/D02_print_environment starting I: set - BUILDDIR='/build/reproducible-path' - BUILDUSERGECOS='first user,first room,first work-phone,first home-phone,first other' - BUILDUSERNAME='pbuilder1' - BUILD_ARCH='amd64' - DEBIAN_FRONTEND='noninteractive' - DEB_BUILD_OPTIONS='buildinfo=+all reproducible=+all parallel=20 ' - DISTRIBUTION='trixie' - HOME='/root' - HOST_ARCH='amd64' + BASH=/bin/sh + BASHOPTS=checkwinsize:cmdhist:complete_fullquote:extquote:force_fignore:globasciiranges:globskipdots:hostcomplete:interactive_comments:patsub_replacement:progcomp:promptvars:sourcepath + BASH_ALIASES=() + BASH_ARGC=() + BASH_ARGV=() + BASH_CMDS=() + BASH_LINENO=([0]="12" [1]="0") + BASH_LOADABLES_PATH=/usr/local/lib/bash:/usr/lib/bash:/opt/local/lib/bash:/usr/pkg/lib/bash:/opt/pkg/lib/bash:. + BASH_SOURCE=([0]="/tmp/hooks/D02_print_environment" [1]="/tmp/hooks/D02_print_environment") + BASH_VERSINFO=([0]="5" [1]="2" [2]="37" [3]="1" [4]="release" [5]="x86_64-pc-linux-gnu") + BASH_VERSION='5.2.37(1)-release' + BUILDDIR=/build/reproducible-path + BUILDUSERGECOS='second user,second room,second work-phone,second home-phone,second other' + BUILDUSERNAME=pbuilder2 + BUILD_ARCH=amd64 + DEBIAN_FRONTEND=noninteractive + DEB_BUILD_OPTIONS='buildinfo=+all reproducible=+all parallel=42 ' + DIRSTACK=() + DISTRIBUTION=trixie + EUID=0 + FUNCNAME=([0]="Echo" [1]="main") + GROUPS=() + HOME=/root + HOSTNAME=i-capture-the-hostname + HOSTTYPE=x86_64 + HOST_ARCH=amd64 IFS=' ' - INVOCATION_ID='0c9878be031d43a78d327ba31348eca6' - LANG='C' - LANGUAGE='en_US:en' - LC_ALL='C' - MAIL='/var/mail/root' - OPTIND='1' - PATH='/usr/sbin:/usr/bin:/sbin:/bin:/usr/games' - PBCURRENTCOMMANDLINEOPERATION='build' - PBUILDER_OPERATION='build' - PBUILDER_PKGDATADIR='/usr/share/pbuilder' - PBUILDER_PKGLIBDIR='/usr/lib/pbuilder' - PBUILDER_SYSCONFDIR='/etc' - PPID='1755253' - PS1='# ' - PS2='> ' + INVOCATION_ID=c47a8584d5744a57abc4a64db30d100f + LANG=C + LANGUAGE=et_EE:et + LC_ALL=C + MACHTYPE=x86_64-pc-linux-gnu + MAIL=/var/mail/root + OPTERR=1 + OPTIND=1 + OSTYPE=linux-gnu + PATH=/usr/sbin:/usr/bin:/sbin:/bin:/usr/games:/i/capture/the/path + PBCURRENTCOMMANDLINEOPERATION=build + PBUILDER_OPERATION=build + PBUILDER_PKGDATADIR=/usr/share/pbuilder + PBUILDER_PKGLIBDIR=/usr/lib/pbuilder + PBUILDER_SYSCONFDIR=/etc + PIPESTATUS=([0]="0") + POSIXLY_CORRECT=y + PPID=4180030 PS4='+ ' - PWD='/' - SHELL='/bin/bash' - SHLVL='2' - SUDO_COMMAND='/usr/bin/timeout -k 18.1h 18h /usr/bin/ionice -c 3 /usr/bin/nice /usr/sbin/pbuilder --build --configfile /srv/reproducible-results/rbuild-debian/r-b-build.PB8y6dpH/pbuilderrc_lfla --distribution trixie --hookdir /etc/pbuilder/first-build-hooks --debbuildopts -b --basetgz /var/cache/pbuilder/trixie-reproducible-base.tgz --buildresult /srv/reproducible-results/rbuild-debian/r-b-build.PB8y6dpH/b1 --logfile b1/build.log valkey_8.0.2+dfsg1-1.dsc' - SUDO_GID='110' - SUDO_UID='105' - SUDO_USER='jenkins' - TERM='unknown' - TZ='/usr/share/zoneinfo/Etc/GMT+12' - USER='root' - _='/usr/bin/systemd-run' - http_proxy='http://46.16.76.132:3128' + PWD=/ + SHELL=/bin/bash + SHELLOPTS=braceexpand:errexit:hashall:interactive-comments:posix + SHLVL=3 + SUDO_COMMAND='/usr/bin/timeout -k 24.1h 24h /usr/bin/ionice -c 3 /usr/bin/nice -n 11 /usr/bin/unshare --uts -- /usr/sbin/pbuilder --build --configfile /srv/reproducible-results/rbuild-debian/r-b-build.PB8y6dpH/pbuilderrc_SRSK --distribution trixie --hookdir /etc/pbuilder/rebuild-hooks --debbuildopts -b --basetgz /var/cache/pbuilder/trixie-reproducible-base.tgz --buildresult /srv/reproducible-results/rbuild-debian/r-b-build.PB8y6dpH/b2 --logfile b2/build.log valkey_8.0.2+dfsg1-1.dsc' + SUDO_GID=111 + SUDO_UID=106 + SUDO_USER=jenkins + TERM=unknown + TZ=/usr/share/zoneinfo/Etc/GMT-14 + UID=0 + USER=root + _='I: set' + http_proxy=http://213.165.73.152:3128 I: uname -a - Linux ionos1-amd64 6.1.0-31-amd64 #1 SMP PREEMPT_DYNAMIC Debian 6.1.128-1 (2025-02-07) x86_64 GNU/Linux + Linux i-capture-the-hostname 6.12.12+bpo-amd64 #1 SMP PREEMPT_DYNAMIC Debian 6.12.12-1~bpo12+1 (2025-02-23) x86_64 GNU/Linux I: ls -l /bin - lrwxrwxrwx 1 root root 7 Nov 22 14:40 /bin -> usr/bin -I: user script /srv/workspace/pbuilder/1755253/tmp/hooks/D02_print_environment finished + lrwxrwxrwx 1 root root 7 Mar 4 2025 /bin -> usr/bin +I: user script /srv/workspace/pbuilder/4180030/tmp/hooks/D02_print_environment finished -> Attempting to satisfy build-dependencies -> Creating pbuilder-satisfydepends-dummy package Package: pbuilder-satisfydepends-dummy @@ -196,7 +228,7 @@ Get: 59 http://deb.debian.org/debian trixie/main amd64 tcl8.6 amd64 8.6.16+dfsg-1 [121 kB] Get: 60 http://deb.debian.org/debian trixie/main amd64 tcl amd64 8.6.14 [4032 B] Get: 61 http://deb.debian.org/debian trixie/main amd64 tcl-tls amd64 1.7.22-3+b4 [67.5 kB] -Fetched 31.3 MB in 1s (29.4 MB/s) +Fetched 31.3 MB in 2s (18.8 MB/s) Preconfiguring packages ... Selecting previously unselected package tzdata. (Reading database ... (Reading database ... 5% (Reading database ... 10% (Reading database ... 15% (Reading database ... 20% (Reading database ... 25% (Reading database ... 30% (Reading database ... 35% (Reading database ... 40% (Reading database ... 45% (Reading database ... 50% (Reading database ... 55% (Reading database ... 60% (Reading database ... 65% (Reading database ... 70% (Reading database ... 75% (Reading database ... 80% (Reading database ... 85% (Reading database ... 90% (Reading database ... 95% (Reading database ... 100% (Reading database ... 19800 files and directories currently installed.) @@ -401,8 +433,8 @@ Setting up tzdata (2025a-2) ... Current default time zone: 'Etc/UTC' -Local time is now: Sun Mar 9 23:12:30 UTC 2025. -Universal Time is now: Sun Mar 9 23:12:30 UTC 2025. +Local time is now: Sun Apr 12 08:35:09 UTC 2026. +Universal Time is now: Sun Apr 12 08:35:09 UTC 2026. Run 'dpkg-reconfigure tzdata' if you wish to change it. Setting up libjemalloc-dev (5.3.0-3) ... @@ -465,7 +497,11 @@ Building tag database... -> Finished parsing the build-deps I: Building the package -I: Running cd /build/reproducible-path/valkey-8.0.2+dfsg1/ && env PATH="/usr/sbin:/usr/bin:/sbin:/bin:/usr/games" HOME="/nonexistent/first-build" dpkg-buildpackage -us -uc -b && env PATH="/usr/sbin:/usr/bin:/sbin:/bin:/usr/games" HOME="/nonexistent/first-build" dpkg-genchanges -S > ../valkey_8.0.2+dfsg1-1_source.changes +I: user script /srv/workspace/pbuilder/4180030/tmp/hooks/A99_set_merged_usr starting +Not re-configuring usrmerge for trixie +I: user script /srv/workspace/pbuilder/4180030/tmp/hooks/A99_set_merged_usr finished +hostname: Name or service not known +I: Running cd /build/reproducible-path/valkey-8.0.2+dfsg1/ && env PATH="/usr/sbin:/usr/bin:/sbin:/bin:/usr/games:/i/capture/the/path" HOME="/nonexistent/second-build" dpkg-buildpackage -us -uc -b && env PATH="/usr/sbin:/usr/bin:/sbin:/bin:/usr/games:/i/capture/the/path" HOME="/nonexistent/second-build" dpkg-genchanges -S > ../valkey_8.0.2+dfsg1-1_source.changes dpkg-buildpackage: info: source package valkey dpkg-buildpackage: info: source version 8.0.2+dfsg1-1 dpkg-buildpackage: info: source distribution unstable @@ -477,9 +513,9 @@ debian/rules override_dh_auto_clean make[1]: Entering directory '/build/reproducible-path/valkey-8.0.2+dfsg1' dh_auto_clean - make -j20 distclean + make -j42 distclean make[2]: Entering directory '/build/reproducible-path/valkey-8.0.2+dfsg1' -make[2]: warning: -j20 forced in submake: resetting jobserver mode. +make[2]: warning: -j42 forced in submake: resetting jobserver mode. cd src && make distclean make[3]: Entering directory '/build/reproducible-path/valkey-8.0.2+dfsg1/src' rm -rf valkey-server valkey-sentinel valkey-cli valkey-benchmark valkey-check-rdb valkey-check-aof valkey-unit-tests libvalkey.a unit/*.o unit/*.d *.o *.gcda *.gcno *.gcov valkey.info lcov-html Makefile.dep *.so @@ -516,9 +552,9 @@ debian/rules override_dh_auto_build make[1]: Entering directory '/build/reproducible-path/valkey-8.0.2+dfsg1' dh_auto_build -- V=1 USE_SYSTEM_JEMALLOC=yes USE_SYSTEMD=yes USE_JEMALLOC=yes - make -j20 "INSTALL=install --strip-program=true" V=1 USE_SYSTEM_JEMALLOC=yes USE_SYSTEMD=yes USE_JEMALLOC=yes + make -j42 "INSTALL=install --strip-program=true" V=1 USE_SYSTEM_JEMALLOC=yes USE_SYSTEMD=yes USE_JEMALLOC=yes make[2]: Entering directory '/build/reproducible-path/valkey-8.0.2+dfsg1' -make[2]: warning: -j20 forced in submake: resetting jobserver mode. +make[2]: warning: -j42 forced in submake: resetting jobserver mode. cd src && make all make[3]: Entering directory '/build/reproducible-path/valkey-8.0.2+dfsg1/src' cc -pedantic -std=gnu11 -Wall -W -Wno-missing-field-initializers -Werror=deprecated-declarations -Wstrict-prototypes -O3 -flto=auto -fno-omit-frame-pointer -g -ggdb -g -O2 -Werror=implicit-function-declaration -ffile-prefix-map=/build/reproducible-path/valkey-8.0.2+dfsg1=. -flto=auto -ffat-lto-objects -fstack-protector-strong -fstack-clash-protection -Wformat -Werror=format-security -fcf-protection -I/usr/include/liblzf -I../deps/hiredis -I../deps/linenoise -I../deps/lua/src -I../deps/hdr_histogram -I../deps/fpconv -DHAVE_LIBSYSTEMD -DUSE_JEMALLOC -I/usr/include/jemalloc/include -DUSE_OPENSSL=1 -DBUILD_TLS_MODULE=0 -Wdate-time -D_FORTIFY_SOURCE=2 -MM acl.c adlist.c ae.c anet.c aof.c bio.c bitops.c blocked.c call_reply.c childinfo.c cli_commands.c cli_common.c cluster.c cluster_legacy.c cluster_slot_stats.c commands.c config.c connection.c crc16.c crc64.c crccombine.c crcspeed.c db.c debug.c defrag.c dict.c endianconv.c eval.c evict.c expire.c function_lua.c functions.c geo.c geohash.c geohash_helper.c hyperloglog.c intset.c io_threads.c kvstore.c latency.c lazyfree.c listpack.c localtime.c logreqres.c lolwut.c lolwut5.c lolwut6.c lzf_c.c lzf_d.c memory_prefetch.c memtest.c module.c monotonic.c mt19937-64.c multi.c networking.c notify.c object.c pqsort.c pubsub.c quicklist.c rand.c rax.c rdb.c release.c replication.c resp_parser.c rio.c script.c script_lua.c sds.c sentinel.c server.c serverassert.c setcpuaffinity.c setproctitle.c sha1.c sha256.c siphash.c slowlog.c socket.c sort.c sparkline.c strl.c syncio.c syscheck.c t_hash.c t_list.c t_set.c t_stream.c t_string.c t_zset.c threads_mngr.c timeout.c tls.c tracking.c unix.c util.c valkey-benchmark.c valkey-check-aof.c valkey-check-rdb.c valkey-cli.c ziplist.c zipmap.c zmalloc.c > Makefile.dep 2> /dev/null || true @@ -569,49 +605,47 @@ MAKE linenoise cd hiredis && make static USE_SSL=1 cd linenoise && make -make[5]: Entering directory '/build/reproducible-path/valkey-8.0.2+dfsg1/deps/hiredis' MAKE lua cd lua/src && make all CFLAGS="-Wall -DLUA_ANSI -DENABLE_CJSON_GLOBAL -DLUA_USE_MKSTEMP -g -O2 -Werror=implicit-function-declaration -ffile-prefix-map=/build/reproducible-path/valkey-8.0.2+dfsg1=. -flto=auto -ffat-lto-objects -fstack-protector-strong -fstack-clash-protection -Wformat -Werror=format-security -fcf-protection -I/usr/include/liblzf -O2 " MYLDFLAGS="-flto=auto -ffat-lto-objects -Wl,-z,relro -Wl,-z,now -Wl,-no-as-needed -ldl -latomic -llzf" AR="ar rc" -make[5]: Entering directory '/build/reproducible-path/valkey-8.0.2+dfsg1/deps/linenoise' -cc -Wall -Os -g -g -O2 -Werror=implicit-function-declaration -ffile-prefix-map=/build/reproducible-path/valkey-8.0.2+dfsg1=. -flto=auto -ffat-lto-objects -fstack-protector-strong -fstack-clash-protection -Wformat -Werror=format-security -fcf-protection -I/usr/include/liblzf -Wdate-time -D_FORTIFY_SOURCE=2 -c linenoise.c -make[5]: Entering directory '/build/reproducible-path/valkey-8.0.2+dfsg1/deps/lua/src' -cc -Wall -DLUA_ANSI -DENABLE_CJSON_GLOBAL -DLUA_USE_MKSTEMP -g -O2 -Werror=implicit-function-declaration -ffile-prefix-map=/build/reproducible-path/valkey-8.0.2+dfsg1=. -flto=auto -ffat-lto-objects -fstack-protector-strong -fstack-clash-protection -Wformat -Werror=format-security -fcf-protection -I/usr/include/liblzf -O2 -Wdate-time -D_FORTIFY_SOURCE=2 -c -o lapi.o lapi.c MAKE hdr_histogram MAKE fpconv cd hdr_histogram && make +make[5]: Entering directory '/build/reproducible-path/valkey-8.0.2+dfsg1/deps/hiredis' cd fpconv && make +make[5]: Entering directory '/build/reproducible-path/valkey-8.0.2+dfsg1/deps/linenoise' +cc -Wall -Os -g -g -O2 -Werror=implicit-function-declaration -ffile-prefix-map=/build/reproducible-path/valkey-8.0.2+dfsg1=. -flto=auto -ffat-lto-objects -fstack-protector-strong -fstack-clash-protection -Wformat -Werror=format-security -fcf-protection -I/usr/include/liblzf -Wdate-time -D_FORTIFY_SOURCE=2 -c linenoise.c +make[5]: Entering directory '/build/reproducible-path/valkey-8.0.2+dfsg1/deps/lua/src' +cc -Wall -DLUA_ANSI -DENABLE_CJSON_GLOBAL -DLUA_USE_MKSTEMP -g -O2 -Werror=implicit-function-declaration -ffile-prefix-map=/build/reproducible-path/valkey-8.0.2+dfsg1=. -flto=auto -ffat-lto-objects -fstack-protector-strong -fstack-clash-protection -Wformat -Werror=format-security -fcf-protection -I/usr/include/liblzf -O2 -Wdate-time -D_FORTIFY_SOURCE=2 -c -o lapi.o lapi.c cc -Wall -DLUA_ANSI -DENABLE_CJSON_GLOBAL -DLUA_USE_MKSTEMP -g -O2 -Werror=implicit-function-declaration -ffile-prefix-map=/build/reproducible-path/valkey-8.0.2+dfsg1=. -flto=auto -ffat-lto-objects -fstack-protector-strong -fstack-clash-protection -Wformat -Werror=format-security -fcf-protection -I/usr/include/liblzf -O2 -Wdate-time -D_FORTIFY_SOURCE=2 -c -o lcode.o lcode.c -cc -Wall -DLUA_ANSI -DENABLE_CJSON_GLOBAL -DLUA_USE_MKSTEMP -g -O2 -Werror=implicit-function-declaration -ffile-prefix-map=/build/reproducible-path/valkey-8.0.2+dfsg1=. -flto=auto -ffat-lto-objects -fstack-protector-strong -fstack-clash-protection -Wformat -Werror=format-security -fcf-protection -I/usr/include/liblzf -O2 -Wdate-time -D_FORTIFY_SOURCE=2 -c -o ldebug.o ldebug.c make[5]: Entering directory '/build/reproducible-path/valkey-8.0.2+dfsg1/deps/hdr_histogram' cc -std=c99 -Wall -Os -g -g -O2 -Werror=implicit-function-declaration -ffile-prefix-map=/build/reproducible-path/valkey-8.0.2+dfsg1=. -flto=auto -ffat-lto-objects -fstack-protector-strong -fstack-clash-protection -Wformat -Werror=format-security -fcf-protection -I/usr/include/liblzf -Wdate-time -D_FORTIFY_SOURCE=2 -DHDR_MALLOC_INCLUDE=\"hdr_redis_malloc.h\" -c hdr_histogram.c -cc -std=c99 -c -O3 -fPIC -Wdate-time -D_FORTIFY_SOURCE=2 -g -O2 -Werror=implicit-function-declaration -ffile-prefix-map=/build/reproducible-path/valkey-8.0.2+dfsg1=. -flto=auto -ffat-lto-objects -fstack-protector-strong -fstack-clash-protection -Wformat -Werror=format-security -fcf-protection -I/usr/include/liblzf -DHIREDIS_TEST_SSL -Wall -Wextra -Werror -Wstrict-prototypes -Wwrite-strings -Wno-missing-field-initializers -g -ggdb -pedantic alloc.c make[5]: Entering directory '/build/reproducible-path/valkey-8.0.2+dfsg1/deps/fpconv' cc -Wall -Os -g -g -O2 -Werror=implicit-function-declaration -ffile-prefix-map=/build/reproducible-path/valkey-8.0.2+dfsg1=. -flto=auto -ffat-lto-objects -fstack-protector-strong -fstack-clash-protection -Wformat -Werror=format-security -fcf-protection -I/usr/include/liblzf -Wdate-time -D_FORTIFY_SOURCE=2 -c fpconv_dtoa.c +cc -Wall -DLUA_ANSI -DENABLE_CJSON_GLOBAL -DLUA_USE_MKSTEMP -g -O2 -Werror=implicit-function-declaration -ffile-prefix-map=/build/reproducible-path/valkey-8.0.2+dfsg1=. -flto=auto -ffat-lto-objects -fstack-protector-strong -fstack-clash-protection -Wformat -Werror=format-security -fcf-protection -I/usr/include/liblzf -O2 -Wdate-time -D_FORTIFY_SOURCE=2 -c -o ldebug.o ldebug.c cc -Wall -DLUA_ANSI -DENABLE_CJSON_GLOBAL -DLUA_USE_MKSTEMP -g -O2 -Werror=implicit-function-declaration -ffile-prefix-map=/build/reproducible-path/valkey-8.0.2+dfsg1=. -flto=auto -ffat-lto-objects -fstack-protector-strong -fstack-clash-protection -Wformat -Werror=format-security -fcf-protection -I/usr/include/liblzf -O2 -Wdate-time -D_FORTIFY_SOURCE=2 -c -o ldo.o ldo.c -cc -std=c99 -c -O3 -fPIC -Wdate-time -D_FORTIFY_SOURCE=2 -g -O2 -Werror=implicit-function-declaration -ffile-prefix-map=/build/reproducible-path/valkey-8.0.2+dfsg1=. -flto=auto -ffat-lto-objects -fstack-protector-strong -fstack-clash-protection -Wformat -Werror=format-security -fcf-protection -I/usr/include/liblzf -DHIREDIS_TEST_SSL -Wall -Wextra -Werror -Wstrict-prototypes -Wwrite-strings -Wno-missing-field-initializers -g -ggdb -pedantic net.c -cc -std=c99 -c -O3 -fPIC -Wdate-time -D_FORTIFY_SOURCE=2 -g -O2 -Werror=implicit-function-declaration -ffile-prefix-map=/build/reproducible-path/valkey-8.0.2+dfsg1=. -flto=auto -ffat-lto-objects -fstack-protector-strong -fstack-clash-protection -Wformat -Werror=format-security -fcf-protection -I/usr/include/liblzf -DHIREDIS_TEST_SSL -Wall -Wextra -Werror -Wstrict-prototypes -Wwrite-strings -Wno-missing-field-initializers -g -ggdb -pedantic hiredis.c cc -Wall -DLUA_ANSI -DENABLE_CJSON_GLOBAL -DLUA_USE_MKSTEMP -g -O2 -Werror=implicit-function-declaration -ffile-prefix-map=/build/reproducible-path/valkey-8.0.2+dfsg1=. -flto=auto -ffat-lto-objects -fstack-protector-strong -fstack-clash-protection -Wformat -Werror=format-security -fcf-protection -I/usr/include/liblzf -O2 -Wdate-time -D_FORTIFY_SOURCE=2 -c -o ldump.o ldump.c -cc -std=c99 -c -O3 -fPIC -Wdate-time -D_FORTIFY_SOURCE=2 -g -O2 -Werror=implicit-function-declaration -ffile-prefix-map=/build/reproducible-path/valkey-8.0.2+dfsg1=. -flto=auto -ffat-lto-objects -fstack-protector-strong -fstack-clash-protection -Wformat -Werror=format-security -fcf-protection -I/usr/include/liblzf -DHIREDIS_TEST_SSL -Wall -Wextra -Werror -Wstrict-prototypes -Wwrite-strings -Wno-missing-field-initializers -g -ggdb -pedantic sds.c cc -Wall -DLUA_ANSI -DENABLE_CJSON_GLOBAL -DLUA_USE_MKSTEMP -g -O2 -Werror=implicit-function-declaration -ffile-prefix-map=/build/reproducible-path/valkey-8.0.2+dfsg1=. -flto=auto -ffat-lto-objects -fstack-protector-strong -fstack-clash-protection -Wformat -Werror=format-security -fcf-protection -I/usr/include/liblzf -O2 -Wdate-time -D_FORTIFY_SOURCE=2 -c -o lfunc.o lfunc.c -cc -std=c99 -c -O3 -fPIC -Wdate-time -D_FORTIFY_SOURCE=2 -g -O2 -Werror=implicit-function-declaration -ffile-prefix-map=/build/reproducible-path/valkey-8.0.2+dfsg1=. -flto=auto -ffat-lto-objects -fstack-protector-strong -fstack-clash-protection -Wformat -Werror=format-security -fcf-protection -I/usr/include/liblzf -DHIREDIS_TEST_SSL -Wall -Wextra -Werror -Wstrict-prototypes -Wwrite-strings -Wno-missing-field-initializers -g -ggdb -pedantic async.c cc -Wall -DLUA_ANSI -DENABLE_CJSON_GLOBAL -DLUA_USE_MKSTEMP -g -O2 -Werror=implicit-function-declaration -ffile-prefix-map=/build/reproducible-path/valkey-8.0.2+dfsg1=. -flto=auto -ffat-lto-objects -fstack-protector-strong -fstack-clash-protection -Wformat -Werror=format-security -fcf-protection -I/usr/include/liblzf -O2 -Wdate-time -D_FORTIFY_SOURCE=2 -c -o lgc.o lgc.c -cc -std=c99 -c -O3 -fPIC -Wdate-time -D_FORTIFY_SOURCE=2 -g -O2 -Werror=implicit-function-declaration -ffile-prefix-map=/build/reproducible-path/valkey-8.0.2+dfsg1=. -flto=auto -ffat-lto-objects -fstack-protector-strong -fstack-clash-protection -Wformat -Werror=format-security -fcf-protection -I/usr/include/liblzf -DHIREDIS_TEST_SSL -Wall -Wextra -Werror -Wstrict-prototypes -Wwrite-strings -Wno-missing-field-initializers -g -ggdb -pedantic read.c cc -Wall -DLUA_ANSI -DENABLE_CJSON_GLOBAL -DLUA_USE_MKSTEMP -g -O2 -Werror=implicit-function-declaration -ffile-prefix-map=/build/reproducible-path/valkey-8.0.2+dfsg1=. -flto=auto -ffat-lto-objects -fstack-protector-strong -fstack-clash-protection -Wformat -Werror=format-security -fcf-protection -I/usr/include/liblzf -O2 -Wdate-time -D_FORTIFY_SOURCE=2 -c -o llex.o llex.c -cc -std=c99 -c -O3 -fPIC -Wdate-time -D_FORTIFY_SOURCE=2 -g -O2 -Werror=implicit-function-declaration -ffile-prefix-map=/build/reproducible-path/valkey-8.0.2+dfsg1=. -flto=auto -ffat-lto-objects -fstack-protector-strong -fstack-clash-protection -Wformat -Werror=format-security -fcf-protection -I/usr/include/liblzf -DHIREDIS_TEST_SSL -Wall -Wextra -Werror -Wstrict-prototypes -Wwrite-strings -Wno-missing-field-initializers -g -ggdb -pedantic sockcompat.c -cc -std=c99 -c -O3 -fPIC -Wdate-time -D_FORTIFY_SOURCE=2 -g -O2 -Werror=implicit-function-declaration -ffile-prefix-map=/build/reproducible-path/valkey-8.0.2+dfsg1=. -flto=auto -ffat-lto-objects -fstack-protector-strong -fstack-clash-protection -Wformat -Werror=format-security -fcf-protection -I/usr/include/liblzf -DHIREDIS_TEST_SSL -Wall -Wextra -Werror -Wstrict-prototypes -Wwrite-strings -Wno-missing-field-initializers -g -ggdb -pedantic ssl.c cc -Wall -DLUA_ANSI -DENABLE_CJSON_GLOBAL -DLUA_USE_MKSTEMP -g -O2 -Werror=implicit-function-declaration -ffile-prefix-map=/build/reproducible-path/valkey-8.0.2+dfsg1=. -flto=auto -ffat-lto-objects -fstack-protector-strong -fstack-clash-protection -Wformat -Werror=format-security -fcf-protection -I/usr/include/liblzf -O2 -Wdate-time -D_FORTIFY_SOURCE=2 -c -o lmem.o lmem.c cc -Wall -DLUA_ANSI -DENABLE_CJSON_GLOBAL -DLUA_USE_MKSTEMP -g -O2 -Werror=implicit-function-declaration -ffile-prefix-map=/build/reproducible-path/valkey-8.0.2+dfsg1=. -flto=auto -ffat-lto-objects -fstack-protector-strong -fstack-clash-protection -Wformat -Werror=format-security -fcf-protection -I/usr/include/liblzf -O2 -Wdate-time -D_FORTIFY_SOURCE=2 -c -o lobject.o lobject.c +cc -std=c99 -c -O3 -fPIC -Wdate-time -D_FORTIFY_SOURCE=2 -g -O2 -Werror=implicit-function-declaration -ffile-prefix-map=/build/reproducible-path/valkey-8.0.2+dfsg1=. -flto=auto -ffat-lto-objects -fstack-protector-strong -fstack-clash-protection -Wformat -Werror=format-security -fcf-protection -I/usr/include/liblzf -DHIREDIS_TEST_SSL -Wall -Wextra -Werror -Wstrict-prototypes -Wwrite-strings -Wno-missing-field-initializers -g -ggdb -pedantic alloc.c cc -Wall -DLUA_ANSI -DENABLE_CJSON_GLOBAL -DLUA_USE_MKSTEMP -g -O2 -Werror=implicit-function-declaration -ffile-prefix-map=/build/reproducible-path/valkey-8.0.2+dfsg1=. -flto=auto -ffat-lto-objects -fstack-protector-strong -fstack-clash-protection -Wformat -Werror=format-security -fcf-protection -I/usr/include/liblzf -O2 -Wdate-time -D_FORTIFY_SOURCE=2 -c -o lopcodes.o lopcodes.c +cc -std=c99 -c -O3 -fPIC -Wdate-time -D_FORTIFY_SOURCE=2 -g -O2 -Werror=implicit-function-declaration -ffile-prefix-map=/build/reproducible-path/valkey-8.0.2+dfsg1=. -flto=auto -ffat-lto-objects -fstack-protector-strong -fstack-clash-protection -Wformat -Werror=format-security -fcf-protection -I/usr/include/liblzf -DHIREDIS_TEST_SSL -Wall -Wextra -Werror -Wstrict-prototypes -Wwrite-strings -Wno-missing-field-initializers -g -ggdb -pedantic net.c +cc -std=c99 -c -O3 -fPIC -Wdate-time -D_FORTIFY_SOURCE=2 -g -O2 -Werror=implicit-function-declaration -ffile-prefix-map=/build/reproducible-path/valkey-8.0.2+dfsg1=. -flto=auto -ffat-lto-objects -fstack-protector-strong -fstack-clash-protection -Wformat -Werror=format-security -fcf-protection -I/usr/include/liblzf -DHIREDIS_TEST_SSL -Wall -Wextra -Werror -Wstrict-prototypes -Wwrite-strings -Wno-missing-field-initializers -g -ggdb -pedantic hiredis.c +cc -std=c99 -c -O3 -fPIC -Wdate-time -D_FORTIFY_SOURCE=2 -g -O2 -Werror=implicit-function-declaration -ffile-prefix-map=/build/reproducible-path/valkey-8.0.2+dfsg1=. -flto=auto -ffat-lto-objects -fstack-protector-strong -fstack-clash-protection -Wformat -Werror=format-security -fcf-protection -I/usr/include/liblzf -DHIREDIS_TEST_SSL -Wall -Wextra -Werror -Wstrict-prototypes -Wwrite-strings -Wno-missing-field-initializers -g -ggdb -pedantic sds.c cc -Wall -DLUA_ANSI -DENABLE_CJSON_GLOBAL -DLUA_USE_MKSTEMP -g -O2 -Werror=implicit-function-declaration -ffile-prefix-map=/build/reproducible-path/valkey-8.0.2+dfsg1=. -flto=auto -ffat-lto-objects -fstack-protector-strong -fstack-clash-protection -Wformat -Werror=format-security -fcf-protection -I/usr/include/liblzf -O2 -Wdate-time -D_FORTIFY_SOURCE=2 -c -o lparser.o lparser.c cc -Wall -DLUA_ANSI -DENABLE_CJSON_GLOBAL -DLUA_USE_MKSTEMP -g -O2 -Werror=implicit-function-declaration -ffile-prefix-map=/build/reproducible-path/valkey-8.0.2+dfsg1=. -flto=auto -ffat-lto-objects -fstack-protector-strong -fstack-clash-protection -Wformat -Werror=format-security -fcf-protection -I/usr/include/liblzf -O2 -Wdate-time -D_FORTIFY_SOURCE=2 -c -o lstate.o lstate.c cc -Wall -DLUA_ANSI -DENABLE_CJSON_GLOBAL -DLUA_USE_MKSTEMP -g -O2 -Werror=implicit-function-declaration -ffile-prefix-map=/build/reproducible-path/valkey-8.0.2+dfsg1=. -flto=auto -ffat-lto-objects -fstack-protector-strong -fstack-clash-protection -Wformat -Werror=format-security -fcf-protection -I/usr/include/liblzf -O2 -Wdate-time -D_FORTIFY_SOURCE=2 -c -o lstring.o lstring.c cc -Wall -DLUA_ANSI -DENABLE_CJSON_GLOBAL -DLUA_USE_MKSTEMP -g -O2 -Werror=implicit-function-declaration -ffile-prefix-map=/build/reproducible-path/valkey-8.0.2+dfsg1=. -flto=auto -ffat-lto-objects -fstack-protector-strong -fstack-clash-protection -Wformat -Werror=format-security -fcf-protection -I/usr/include/liblzf -O2 -Wdate-time -D_FORTIFY_SOURCE=2 -c -o ltable.o ltable.c +cc -std=c99 -c -O3 -fPIC -Wdate-time -D_FORTIFY_SOURCE=2 -g -O2 -Werror=implicit-function-declaration -ffile-prefix-map=/build/reproducible-path/valkey-8.0.2+dfsg1=. -flto=auto -ffat-lto-objects -fstack-protector-strong -fstack-clash-protection -Wformat -Werror=format-security -fcf-protection -I/usr/include/liblzf -DHIREDIS_TEST_SSL -Wall -Wextra -Werror -Wstrict-prototypes -Wwrite-strings -Wno-missing-field-initializers -g -ggdb -pedantic async.c +cc -std=c99 -c -O3 -fPIC -Wdate-time -D_FORTIFY_SOURCE=2 -g -O2 -Werror=implicit-function-declaration -ffile-prefix-map=/build/reproducible-path/valkey-8.0.2+dfsg1=. -flto=auto -ffat-lto-objects -fstack-protector-strong -fstack-clash-protection -Wformat -Werror=format-security -fcf-protection -I/usr/include/liblzf -DHIREDIS_TEST_SSL -Wall -Wextra -Werror -Wstrict-prototypes -Wwrite-strings -Wno-missing-field-initializers -g -ggdb -pedantic read.c +cc -std=c99 -c -O3 -fPIC -Wdate-time -D_FORTIFY_SOURCE=2 -g -O2 -Werror=implicit-function-declaration -ffile-prefix-map=/build/reproducible-path/valkey-8.0.2+dfsg1=. -flto=auto -ffat-lto-objects -fstack-protector-strong -fstack-clash-protection -Wformat -Werror=format-security -fcf-protection -I/usr/include/liblzf -DHIREDIS_TEST_SSL -Wall -Wextra -Werror -Wstrict-prototypes -Wwrite-strings -Wno-missing-field-initializers -g -ggdb -pedantic sockcompat.c cc -Wall -DLUA_ANSI -DENABLE_CJSON_GLOBAL -DLUA_USE_MKSTEMP -g -O2 -Werror=implicit-function-declaration -ffile-prefix-map=/build/reproducible-path/valkey-8.0.2+dfsg1=. -flto=auto -ffat-lto-objects -fstack-protector-strong -fstack-clash-protection -Wformat -Werror=format-security -fcf-protection -I/usr/include/liblzf -O2 -Wdate-time -D_FORTIFY_SOURCE=2 -c -o ltm.o ltm.c cc -Wall -DLUA_ANSI -DENABLE_CJSON_GLOBAL -DLUA_USE_MKSTEMP -g -O2 -Werror=implicit-function-declaration -ffile-prefix-map=/build/reproducible-path/valkey-8.0.2+dfsg1=. -flto=auto -ffat-lto-objects -fstack-protector-strong -fstack-clash-protection -Wformat -Werror=format-security -fcf-protection -I/usr/include/liblzf -O2 -Wdate-time -D_FORTIFY_SOURCE=2 -c -o lundump.o lundump.c cc -Wall -DLUA_ANSI -DENABLE_CJSON_GLOBAL -DLUA_USE_MKSTEMP -g -O2 -Werror=implicit-function-declaration -ffile-prefix-map=/build/reproducible-path/valkey-8.0.2+dfsg1=. -flto=auto -ffat-lto-objects -fstack-protector-strong -fstack-clash-protection -Wformat -Werror=format-security -fcf-protection -I/usr/include/liblzf -O2 -Wdate-time -D_FORTIFY_SOURCE=2 -c -o lvm.o lvm.c -ar rcs libfpconv.a fpconv_dtoa.o +cc -std=c99 -c -O3 -fPIC -Wdate-time -D_FORTIFY_SOURCE=2 -g -O2 -Werror=implicit-function-declaration -ffile-prefix-map=/build/reproducible-path/valkey-8.0.2+dfsg1=. -flto=auto -ffat-lto-objects -fstack-protector-strong -fstack-clash-protection -Wformat -Werror=format-security -fcf-protection -I/usr/include/liblzf -DHIREDIS_TEST_SSL -Wall -Wextra -Werror -Wstrict-prototypes -Wwrite-strings -Wno-missing-field-initializers -g -ggdb -pedantic ssl.c cc -Wall -DLUA_ANSI -DENABLE_CJSON_GLOBAL -DLUA_USE_MKSTEMP -g -O2 -Werror=implicit-function-declaration -ffile-prefix-map=/build/reproducible-path/valkey-8.0.2+dfsg1=. -flto=auto -ffat-lto-objects -fstack-protector-strong -fstack-clash-protection -Wformat -Werror=format-security -fcf-protection -I/usr/include/liblzf -O2 -Wdate-time -D_FORTIFY_SOURCE=2 -c -o lzio.o lzio.c -make[5]: Leaving directory '/build/reproducible-path/valkey-8.0.2+dfsg1/deps/fpconv' cc -Wall -DLUA_ANSI -DENABLE_CJSON_GLOBAL -DLUA_USE_MKSTEMP -g -O2 -Werror=implicit-function-declaration -ffile-prefix-map=/build/reproducible-path/valkey-8.0.2+dfsg1=. -flto=auto -ffat-lto-objects -fstack-protector-strong -fstack-clash-protection -Wformat -Werror=format-security -fcf-protection -I/usr/include/liblzf -O2 -Wdate-time -D_FORTIFY_SOURCE=2 -c -o strbuf.o strbuf.c cc -Wall -DLUA_ANSI -DENABLE_CJSON_GLOBAL -DLUA_USE_MKSTEMP -g -O2 -Werror=implicit-function-declaration -ffile-prefix-map=/build/reproducible-path/valkey-8.0.2+dfsg1=. -flto=auto -ffat-lto-objects -fstack-protector-strong -fstack-clash-protection -Wformat -Werror=format-security -fcf-protection -I/usr/include/liblzf -O2 -Wdate-time -D_FORTIFY_SOURCE=2 -c -o fpconv.o fpconv.c cc -Wall -DLUA_ANSI -DENABLE_CJSON_GLOBAL -DLUA_USE_MKSTEMP -g -O2 -Werror=implicit-function-declaration -ffile-prefix-map=/build/reproducible-path/valkey-8.0.2+dfsg1=. -flto=auto -ffat-lto-objects -fstack-protector-strong -fstack-clash-protection -Wformat -Werror=format-security -fcf-protection -I/usr/include/liblzf -O2 -Wdate-time -D_FORTIFY_SOURCE=2 -c -o lauxlib.o lauxlib.c @@ -621,7 +655,6 @@ cc -Wall -DLUA_ANSI -DENABLE_CJSON_GLOBAL -DLUA_USE_MKSTEMP -g -O2 -Werror=implicit-function-declaration -ffile-prefix-map=/build/reproducible-path/valkey-8.0.2+dfsg1=. -flto=auto -ffat-lto-objects -fstack-protector-strong -fstack-clash-protection -Wformat -Werror=format-security -fcf-protection -I/usr/include/liblzf -O2 -Wdate-time -D_FORTIFY_SOURCE=2 -c -o lmathlib.o lmathlib.c cc -Wall -DLUA_ANSI -DENABLE_CJSON_GLOBAL -DLUA_USE_MKSTEMP -g -O2 -Werror=implicit-function-declaration -ffile-prefix-map=/build/reproducible-path/valkey-8.0.2+dfsg1=. -flto=auto -ffat-lto-objects -fstack-protector-strong -fstack-clash-protection -Wformat -Werror=format-security -fcf-protection -I/usr/include/liblzf -O2 -Wdate-time -D_FORTIFY_SOURCE=2 -c -o loslib.o loslib.c cc -Wall -DLUA_ANSI -DENABLE_CJSON_GLOBAL -DLUA_USE_MKSTEMP -g -O2 -Werror=implicit-function-declaration -ffile-prefix-map=/build/reproducible-path/valkey-8.0.2+dfsg1=. -flto=auto -ffat-lto-objects -fstack-protector-strong -fstack-clash-protection -Wformat -Werror=format-security -fcf-protection -I/usr/include/liblzf -O2 -Wdate-time -D_FORTIFY_SOURCE=2 -c -o ltablib.o ltablib.c -ar rcs libhiredis_ssl.a ssl.o cc -Wall -DLUA_ANSI -DENABLE_CJSON_GLOBAL -DLUA_USE_MKSTEMP -g -O2 -Werror=implicit-function-declaration -ffile-prefix-map=/build/reproducible-path/valkey-8.0.2+dfsg1=. -flto=auto -ffat-lto-objects -fstack-protector-strong -fstack-clash-protection -Wformat -Werror=format-security -fcf-protection -I/usr/include/liblzf -O2 -Wdate-time -D_FORTIFY_SOURCE=2 -c -o lstrlib.o lstrlib.c cc -Wall -DLUA_ANSI -DENABLE_CJSON_GLOBAL -DLUA_USE_MKSTEMP -g -O2 -Werror=implicit-function-declaration -ffile-prefix-map=/build/reproducible-path/valkey-8.0.2+dfsg1=. -flto=auto -ffat-lto-objects -fstack-protector-strong -fstack-clash-protection -Wformat -Werror=format-security -fcf-protection -I/usr/include/liblzf -O2 -Wdate-time -D_FORTIFY_SOURCE=2 -c -o loadlib.o loadlib.c cc -Wall -DLUA_ANSI -DENABLE_CJSON_GLOBAL -DLUA_USE_MKSTEMP -g -O2 -Werror=implicit-function-declaration -ffile-prefix-map=/build/reproducible-path/valkey-8.0.2+dfsg1=. -flto=auto -ffat-lto-objects -fstack-protector-strong -fstack-clash-protection -Wformat -Werror=format-security -fcf-protection -I/usr/include/liblzf -O2 -Wdate-time -D_FORTIFY_SOURCE=2 -c -o linit.o linit.c @@ -631,14 +664,17 @@ cc -Wall -DLUA_ANSI -DENABLE_CJSON_GLOBAL -DLUA_USE_MKSTEMP -g -O2 -Werror=implicit-function-declaration -ffile-prefix-map=/build/reproducible-path/valkey-8.0.2+dfsg1=. -flto=auto -ffat-lto-objects -fstack-protector-strong -fstack-clash-protection -Wformat -Werror=format-security -fcf-protection -I/usr/include/liblzf -O2 -Wdate-time -D_FORTIFY_SOURCE=2 -c -o lua_bit.o lua_bit.c cc -Wall -DLUA_ANSI -DENABLE_CJSON_GLOBAL -DLUA_USE_MKSTEMP -g -O2 -Werror=implicit-function-declaration -ffile-prefix-map=/build/reproducible-path/valkey-8.0.2+dfsg1=. -flto=auto -ffat-lto-objects -fstack-protector-strong -fstack-clash-protection -Wformat -Werror=format-security -fcf-protection -I/usr/include/liblzf -O2 -Wdate-time -D_FORTIFY_SOURCE=2 -c -o lua.o lua.c cc -Wall -DLUA_ANSI -DENABLE_CJSON_GLOBAL -DLUA_USE_MKSTEMP -g -O2 -Werror=implicit-function-declaration -ffile-prefix-map=/build/reproducible-path/valkey-8.0.2+dfsg1=. -flto=auto -ffat-lto-objects -fstack-protector-strong -fstack-clash-protection -Wformat -Werror=format-security -fcf-protection -I/usr/include/liblzf -O2 -Wdate-time -D_FORTIFY_SOURCE=2 -c -o luac.o luac.c +cc -Wall -DLUA_ANSI -DENABLE_CJSON_GLOBAL -DLUA_USE_MKSTEMP -g -O2 -Werror=implicit-function-declaration -ffile-prefix-map=/build/reproducible-path/valkey-8.0.2+dfsg1=. -flto=auto -ffat-lto-objects -fstack-protector-strong -fstack-clash-protection -Wformat -Werror=format-security -fcf-protection -I/usr/include/liblzf -O2 -Wdate-time -D_FORTIFY_SOURCE=2 -c -o print.o print.c +ar rcs libfpconv.a fpconv_dtoa.o +make[5]: Leaving directory '/build/reproducible-path/valkey-8.0.2+dfsg1/deps/fpconv' +ar rcs libhiredis_ssl.a ssl.o ar rcs libhdrhistogram.a hdr_histogram.o make[5]: Leaving directory '/build/reproducible-path/valkey-8.0.2+dfsg1/deps/hdr_histogram' -cc -Wall -DLUA_ANSI -DENABLE_CJSON_GLOBAL -DLUA_USE_MKSTEMP -g -O2 -Werror=implicit-function-declaration -ffile-prefix-map=/build/reproducible-path/valkey-8.0.2+dfsg1=. -flto=auto -ffat-lto-objects -fstack-protector-strong -fstack-clash-protection -Wformat -Werror=format-security -fcf-protection -I/usr/include/liblzf -O2 -Wdate-time -D_FORTIFY_SOURCE=2 -c -o print.o print.c make[5]: Leaving directory '/build/reproducible-path/valkey-8.0.2+dfsg1/deps/linenoise' -ar rcs libhiredis.a alloc.o net.o hiredis.o sds.o async.o read.o sockcompat.o -make[5]: Leaving directory '/build/reproducible-path/valkey-8.0.2+dfsg1/deps/hiredis' ar rc liblua.a lapi.o lcode.o ldebug.o ldo.o ldump.o lfunc.o lgc.o llex.o lmem.o lobject.o lopcodes.o lparser.o lstate.o lstring.o ltable.o ltm.o lundump.o lvm.o lzio.o strbuf.o fpconv.o lauxlib.o lbaselib.o ldblib.o liolib.o lmathlib.o loslib.o ltablib.o lstrlib.o loadlib.o linit.o lua_cjson.o lua_struct.o lua_cmsgpack.o lua_bit.o # DLL needs all object files ranlib liblua.a +ar rcs libhiredis.a alloc.o net.o hiredis.o sds.o async.o read.o sockcompat.o +make[5]: Leaving directory '/build/reproducible-path/valkey-8.0.2+dfsg1/deps/hiredis' cc -o lua -flto=auto -ffat-lto-objects -Wl,-z,relro -Wl,-z,now -Wl,-no-as-needed -ldl -latomic -llzf lua.o liblua.a -lm cc -o luac -flto=auto -ffat-lto-objects -Wl,-z,relro -Wl,-z,now -Wl,-no-as-needed -ldl -latomic -llzf luac.o print.o liblua.a -lm make[5]: Leaving directory '/build/reproducible-path/valkey-8.0.2+dfsg1/deps/lua/src' @@ -722,13 +758,38 @@ cc -pedantic -std=gnu11 -Wall -W -Wno-missing-field-initializers -Werror=deprecated-declarations -Wstrict-prototypes -O3 -flto=auto -fno-omit-frame-pointer -g -ggdb -g -O2 -Werror=implicit-function-declaration -ffile-prefix-map=/build/reproducible-path/valkey-8.0.2+dfsg1=. -flto=auto -ffat-lto-objects -fstack-protector-strong -fstack-clash-protection -Wformat -Werror=format-security -fcf-protection -I/usr/include/liblzf -I../deps/hiredis -I../deps/linenoise -I../deps/lua/src -I../deps/hdr_histogram -I../deps/fpconv -DHAVE_LIBSYSTEMD -DUSE_JEMALLOC -I/usr/include/jemalloc/include -DUSE_OPENSSL=1 -DBUILD_TLS_MODULE=0 -Wdate-time -D_FORTIFY_SOURCE=2 -MMD -o localtime.o -c localtime.c cc -pedantic -std=gnu11 -Wall -W -Wno-missing-field-initializers -Werror=deprecated-declarations -Wstrict-prototypes -O3 -flto=auto -fno-omit-frame-pointer -g -ggdb -g -O2 -Werror=implicit-function-declaration -ffile-prefix-map=/build/reproducible-path/valkey-8.0.2+dfsg1=. -flto=auto -ffat-lto-objects -fstack-protector-strong -fstack-clash-protection -Wformat -Werror=format-security -fcf-protection -I/usr/include/liblzf -I../deps/hiredis -I../deps/linenoise -I../deps/lua/src -I../deps/hdr_histogram -I../deps/fpconv -DHAVE_LIBSYSTEMD -DUSE_JEMALLOC -I/usr/include/jemalloc/include -DUSE_OPENSSL=1 -DBUILD_TLS_MODULE=0 -Wdate-time -D_FORTIFY_SOURCE=2 -MMD -o lolwut.o -c lolwut.c cc -pedantic -std=gnu11 -Wall -W -Wno-missing-field-initializers -Werror=deprecated-declarations -Wstrict-prototypes -O3 -flto=auto -fno-omit-frame-pointer -g -ggdb -g -O2 -Werror=implicit-function-declaration -ffile-prefix-map=/build/reproducible-path/valkey-8.0.2+dfsg1=. -flto=auto -ffat-lto-objects -fstack-protector-strong -fstack-clash-protection -Wformat -Werror=format-security -fcf-protection -I/usr/include/liblzf -I../deps/hiredis -I../deps/linenoise -I../deps/lua/src -I../deps/hdr_histogram -I../deps/fpconv -DHAVE_LIBSYSTEMD -DUSE_JEMALLOC -I/usr/include/jemalloc/include -DUSE_OPENSSL=1 -DBUILD_TLS_MODULE=0 -Wdate-time -D_FORTIFY_SOURCE=2 -MMD -o lolwut5.o -c lolwut5.c +cc -pedantic -std=gnu11 -Wall -W -Wno-missing-field-initializers -Werror=deprecated-declarations -Wstrict-prototypes -O3 -flto=auto -fno-omit-frame-pointer -g -ggdb -g -O2 -Werror=implicit-function-declaration -ffile-prefix-map=/build/reproducible-path/valkey-8.0.2+dfsg1=. -flto=auto -ffat-lto-objects -fstack-protector-strong -fstack-clash-protection -Wformat -Werror=format-security -fcf-protection -I/usr/include/liblzf -I../deps/hiredis -I../deps/linenoise -I../deps/lua/src -I../deps/hdr_histogram -I../deps/fpconv -DHAVE_LIBSYSTEMD -DUSE_JEMALLOC -I/usr/include/jemalloc/include -DUSE_OPENSSL=1 -DBUILD_TLS_MODULE=0 -Wdate-time -D_FORTIFY_SOURCE=2 -MMD -o lolwut6.o -c lolwut6.c +cc -pedantic -std=gnu11 -Wall -W -Wno-missing-field-initializers -Werror=deprecated-declarations -Wstrict-prototypes -O3 -flto=auto -fno-omit-frame-pointer -g -ggdb -g -O2 -Werror=implicit-function-declaration -ffile-prefix-map=/build/reproducible-path/valkey-8.0.2+dfsg1=. -flto=auto -ffat-lto-objects -fstack-protector-strong -fstack-clash-protection -Wformat -Werror=format-security -fcf-protection -I/usr/include/liblzf -I../deps/hiredis -I../deps/linenoise -I../deps/lua/src -I../deps/hdr_histogram -I../deps/fpconv -DHAVE_LIBSYSTEMD -DUSE_JEMALLOC -I/usr/include/jemalloc/include -DUSE_OPENSSL=1 -DBUILD_TLS_MODULE=0 -Wdate-time -D_FORTIFY_SOURCE=2 -MMD -o acl.o -c acl.c +cc -pedantic -std=gnu11 -Wall -W -Wno-missing-field-initializers -Werror=deprecated-declarations -Wstrict-prototypes -O3 -flto=auto -fno-omit-frame-pointer -g -ggdb -g -O2 -Werror=implicit-function-declaration -ffile-prefix-map=/build/reproducible-path/valkey-8.0.2+dfsg1=. -flto=auto -ffat-lto-objects -fstack-protector-strong -fstack-clash-protection -Wformat -Werror=format-security -fcf-protection -I/usr/include/liblzf -I../deps/hiredis -I../deps/linenoise -I../deps/lua/src -I../deps/hdr_histogram -I../deps/fpconv -DHAVE_LIBSYSTEMD -DUSE_JEMALLOC -I/usr/include/jemalloc/include -DUSE_OPENSSL=1 -DBUILD_TLS_MODULE=0 -Wdate-time -D_FORTIFY_SOURCE=2 -MMD -o tracking.o -c tracking.c +cc -pedantic -std=gnu11 -Wall -W -Wno-missing-field-initializers -Werror=deprecated-declarations -Wstrict-prototypes -O3 -flto=auto -fno-omit-frame-pointer -g -ggdb -g -O2 -Werror=implicit-function-declaration -ffile-prefix-map=/build/reproducible-path/valkey-8.0.2+dfsg1=. -flto=auto -ffat-lto-objects -fstack-protector-strong -fstack-clash-protection -Wformat -Werror=format-security -fcf-protection -I/usr/include/liblzf -I../deps/hiredis -I../deps/linenoise -I../deps/lua/src -I../deps/hdr_histogram -I../deps/fpconv -DHAVE_LIBSYSTEMD -DUSE_JEMALLOC -I/usr/include/jemalloc/include -DUSE_OPENSSL=1 -DBUILD_TLS_MODULE=0 -Wdate-time -D_FORTIFY_SOURCE=2 -MMD -o socket.o -c socket.c +cc -pedantic -std=gnu11 -Wall -W -Wno-missing-field-initializers -Werror=deprecated-declarations -Wstrict-prototypes -O3 -flto=auto -fno-omit-frame-pointer -g -ggdb -g -O2 -Werror=implicit-function-declaration -ffile-prefix-map=/build/reproducible-path/valkey-8.0.2+dfsg1=. -flto=auto -ffat-lto-objects -fstack-protector-strong -fstack-clash-protection -Wformat -Werror=format-security -fcf-protection -I/usr/include/liblzf -I../deps/hiredis -I../deps/linenoise -I../deps/lua/src -I../deps/hdr_histogram -I../deps/fpconv -DHAVE_LIBSYSTEMD -DUSE_JEMALLOC -I/usr/include/jemalloc/include -DUSE_OPENSSL=1 -DBUILD_TLS_MODULE=0 -Wdate-time -D_FORTIFY_SOURCE=2 -MMD -o tls.o -c tls.c +cc -pedantic -std=gnu11 -Wall -W -Wno-missing-field-initializers -Werror=deprecated-declarations -Wstrict-prototypes -O3 -flto=auto -fno-omit-frame-pointer -g -ggdb -g -O2 -Werror=implicit-function-declaration -ffile-prefix-map=/build/reproducible-path/valkey-8.0.2+dfsg1=. -flto=auto -ffat-lto-objects -fstack-protector-strong -fstack-clash-protection -Wformat -Werror=format-security -fcf-protection -I/usr/include/liblzf -I../deps/hiredis -I../deps/linenoise -I../deps/lua/src -I../deps/hdr_histogram -I../deps/fpconv -DHAVE_LIBSYSTEMD -DUSE_JEMALLOC -I/usr/include/jemalloc/include -DUSE_OPENSSL=1 -DBUILD_TLS_MODULE=0 -Wdate-time -D_FORTIFY_SOURCE=2 -MMD -o sha256.o -c sha256.c +cc -pedantic -std=gnu11 -Wall -W -Wno-missing-field-initializers -Werror=deprecated-declarations -Wstrict-prototypes -O3 -flto=auto -fno-omit-frame-pointer -g -ggdb -g -O2 -Werror=implicit-function-declaration -ffile-prefix-map=/build/reproducible-path/valkey-8.0.2+dfsg1=. -flto=auto -ffat-lto-objects -fstack-protector-strong -fstack-clash-protection -Wformat -Werror=format-security -fcf-protection -I/usr/include/liblzf -I../deps/hiredis -I../deps/linenoise -I../deps/lua/src -I../deps/hdr_histogram -I../deps/fpconv -DHAVE_LIBSYSTEMD -DUSE_JEMALLOC -I/usr/include/jemalloc/include -DUSE_OPENSSL=1 -DBUILD_TLS_MODULE=0 -Wdate-time -D_FORTIFY_SOURCE=2 -MMD -o timeout.o -c timeout.c +cc -pedantic -std=gnu11 -Wall -W -Wno-missing-field-initializers -Werror=deprecated-declarations -Wstrict-prototypes -O3 -flto=auto -fno-omit-frame-pointer -g -ggdb -g -O2 -Werror=implicit-function-declaration -ffile-prefix-map=/build/reproducible-path/valkey-8.0.2+dfsg1=. -flto=auto -ffat-lto-objects -fstack-protector-strong -fstack-clash-protection -Wformat -Werror=format-security -fcf-protection -I/usr/include/liblzf -I../deps/hiredis -I../deps/linenoise -I../deps/lua/src -I../deps/hdr_histogram -I../deps/fpconv -DHAVE_LIBSYSTEMD -DUSE_JEMALLOC -I/usr/include/jemalloc/include -DUSE_OPENSSL=1 -DBUILD_TLS_MODULE=0 -Wdate-time -D_FORTIFY_SOURCE=2 -MMD -o setcpuaffinity.o -c setcpuaffinity.c +cc -pedantic -std=gnu11 -Wall -W -Wno-missing-field-initializers -Werror=deprecated-declarations -Wstrict-prototypes -O3 -flto=auto -fno-omit-frame-pointer -g -ggdb -g -O2 -Werror=implicit-function-declaration -ffile-prefix-map=/build/reproducible-path/valkey-8.0.2+dfsg1=. -flto=auto -ffat-lto-objects -fstack-protector-strong -fstack-clash-protection -Wformat -Werror=format-security -fcf-protection -I/usr/include/liblzf -I../deps/hiredis -I../deps/linenoise -I../deps/lua/src -I../deps/hdr_histogram -I../deps/fpconv -DHAVE_LIBSYSTEMD -DUSE_JEMALLOC -I/usr/include/jemalloc/include -DUSE_OPENSSL=1 -DBUILD_TLS_MODULE=0 -Wdate-time -D_FORTIFY_SOURCE=2 -MMD -o monotonic.o -c monotonic.c +cc -pedantic -std=gnu11 -Wall -W -Wno-missing-field-initializers -Werror=deprecated-declarations -Wstrict-prototypes -O3 -flto=auto -fno-omit-frame-pointer -g -ggdb -g -O2 -Werror=implicit-function-declaration -ffile-prefix-map=/build/reproducible-path/valkey-8.0.2+dfsg1=. -flto=auto -ffat-lto-objects -fstack-protector-strong -fstack-clash-protection -Wformat -Werror=format-security -fcf-protection -I/usr/include/liblzf -I../deps/hiredis -I../deps/linenoise -I../deps/lua/src -I../deps/hdr_histogram -I../deps/fpconv -DHAVE_LIBSYSTEMD -DUSE_JEMALLOC -I/usr/include/jemalloc/include -DUSE_OPENSSL=1 -DBUILD_TLS_MODULE=0 -Wdate-time -D_FORTIFY_SOURCE=2 -MMD -o mt19937-64.o -c mt19937-64.c +cc -pedantic -std=gnu11 -Wall -W -Wno-missing-field-initializers -Werror=deprecated-declarations -Wstrict-prototypes -O3 -flto=auto -fno-omit-frame-pointer -g -ggdb -g -O2 -Werror=implicit-function-declaration -ffile-prefix-map=/build/reproducible-path/valkey-8.0.2+dfsg1=. -flto=auto -ffat-lto-objects -fstack-protector-strong -fstack-clash-protection -Wformat -Werror=format-security -fcf-protection -I/usr/include/liblzf -I../deps/hiredis -I../deps/linenoise -I../deps/lua/src -I../deps/hdr_histogram -I../deps/fpconv -DHAVE_LIBSYSTEMD -DUSE_JEMALLOC -I/usr/include/jemalloc/include -DUSE_OPENSSL=1 -DBUILD_TLS_MODULE=0 -Wdate-time -D_FORTIFY_SOURCE=2 -MMD -o resp_parser.o -c resp_parser.c +cc -pedantic -std=gnu11 -Wall -W -Wno-missing-field-initializers -Werror=deprecated-declarations -Wstrict-prototypes -O3 -flto=auto -fno-omit-frame-pointer -g -ggdb -g -O2 -Werror=implicit-function-declaration -ffile-prefix-map=/build/reproducible-path/valkey-8.0.2+dfsg1=. -flto=auto -ffat-lto-objects -fstack-protector-strong -fstack-clash-protection -Wformat -Werror=format-security -fcf-protection -I/usr/include/liblzf -I../deps/hiredis -I../deps/linenoise -I../deps/lua/src -I../deps/hdr_histogram -I../deps/fpconv -DHAVE_LIBSYSTEMD -DUSE_JEMALLOC -I/usr/include/jemalloc/include -DUSE_OPENSSL=1 -DBUILD_TLS_MODULE=0 -Wdate-time -D_FORTIFY_SOURCE=2 -MMD -o call_reply.o -c call_reply.c +cc -pedantic -std=gnu11 -Wall -W -Wno-missing-field-initializers -Werror=deprecated-declarations -Wstrict-prototypes -O3 -flto=auto -fno-omit-frame-pointer -g -ggdb -g -O2 -Werror=implicit-function-declaration -ffile-prefix-map=/build/reproducible-path/valkey-8.0.2+dfsg1=. -flto=auto -ffat-lto-objects -fstack-protector-strong -fstack-clash-protection -Wformat -Werror=format-security -fcf-protection -I/usr/include/liblzf -I../deps/hiredis -I../deps/linenoise -I../deps/lua/src -I../deps/hdr_histogram -I../deps/fpconv -DHAVE_LIBSYSTEMD -DUSE_JEMALLOC -I/usr/include/jemalloc/include -DUSE_OPENSSL=1 -DBUILD_TLS_MODULE=0 -Wdate-time -D_FORTIFY_SOURCE=2 -MMD -o script_lua.o -c script_lua.c +cc -pedantic -std=gnu11 -Wall -W -Wno-missing-field-initializers -Werror=deprecated-declarations -Wstrict-prototypes -O3 -flto=auto -fno-omit-frame-pointer -g -ggdb -g -O2 -Werror=implicit-function-declaration -ffile-prefix-map=/build/reproducible-path/valkey-8.0.2+dfsg1=. -flto=auto -ffat-lto-objects -fstack-protector-strong -fstack-clash-protection -Wformat -Werror=format-security -fcf-protection -I/usr/include/liblzf -I../deps/hiredis -I../deps/linenoise -I../deps/lua/src -I../deps/hdr_histogram -I../deps/fpconv -DHAVE_LIBSYSTEMD -DUSE_JEMALLOC -I/usr/include/jemalloc/include -DUSE_OPENSSL=1 -DBUILD_TLS_MODULE=0 -Wdate-time -D_FORTIFY_SOURCE=2 -MMD -o script.o -c script.c +cc -pedantic -std=gnu11 -Wall -W -Wno-missing-field-initializers -Werror=deprecated-declarations -Wstrict-prototypes -O3 -flto=auto -fno-omit-frame-pointer -g -ggdb -g -O2 -Werror=implicit-function-declaration -ffile-prefix-map=/build/reproducible-path/valkey-8.0.2+dfsg1=. -flto=auto -ffat-lto-objects -fstack-protector-strong -fstack-clash-protection -Wformat -Werror=format-security -fcf-protection -I/usr/include/liblzf -I../deps/hiredis -I../deps/linenoise -I../deps/lua/src -I../deps/hdr_histogram -I../deps/fpconv -DHAVE_LIBSYSTEMD -DUSE_JEMALLOC -I/usr/include/jemalloc/include -DUSE_OPENSSL=1 -DBUILD_TLS_MODULE=0 -Wdate-time -D_FORTIFY_SOURCE=2 -MMD -o functions.o -c functions.c +cc -pedantic -std=gnu11 -Wall -W -Wno-missing-field-initializers -Werror=deprecated-declarations -Wstrict-prototypes -O3 -flto=auto -fno-omit-frame-pointer -g -ggdb -g -O2 -Werror=implicit-function-declaration -ffile-prefix-map=/build/reproducible-path/valkey-8.0.2+dfsg1=. -flto=auto -ffat-lto-objects -fstack-protector-strong -fstack-clash-protection -Wformat -Werror=format-security -fcf-protection -I/usr/include/liblzf -I../deps/hiredis -I../deps/linenoise -I../deps/lua/src -I../deps/hdr_histogram -I../deps/fpconv -DHAVE_LIBSYSTEMD -DUSE_JEMALLOC -I/usr/include/jemalloc/include -DUSE_OPENSSL=1 -DBUILD_TLS_MODULE=0 -Wdate-time -D_FORTIFY_SOURCE=2 -MMD -o function_lua.o -c function_lua.c +cc -pedantic -std=gnu11 -Wall -W -Wno-missing-field-initializers -Werror=deprecated-declarations -Wstrict-prototypes -O3 -flto=auto -fno-omit-frame-pointer -g -ggdb -g -O2 -Werror=implicit-function-declaration -ffile-prefix-map=/build/reproducible-path/valkey-8.0.2+dfsg1=. -flto=auto -ffat-lto-objects -fstack-protector-strong -fstack-clash-protection -Wformat -Werror=format-security -fcf-protection -I/usr/include/liblzf -I../deps/hiredis -I../deps/linenoise -I../deps/lua/src -I../deps/hdr_histogram -I../deps/fpconv -DHAVE_LIBSYSTEMD -DUSE_JEMALLOC -I/usr/include/jemalloc/include -DUSE_OPENSSL=1 -DBUILD_TLS_MODULE=0 -Wdate-time -D_FORTIFY_SOURCE=2 -MMD -o commands.o -c commands.c +cc -pedantic -std=gnu11 -Wall -W -Wno-missing-field-initializers -Werror=deprecated-declarations -Wstrict-prototypes -O3 -flto=auto -fno-omit-frame-pointer -g -ggdb -g -O2 -Werror=implicit-function-declaration -ffile-prefix-map=/build/reproducible-path/valkey-8.0.2+dfsg1=. -flto=auto -ffat-lto-objects -fstack-protector-strong -fstack-clash-protection -Wformat -Werror=format-security -fcf-protection -I/usr/include/liblzf -I../deps/hiredis -I../deps/linenoise -I../deps/lua/src -I../deps/hdr_histogram -I../deps/fpconv -DHAVE_LIBSYSTEMD -DUSE_JEMALLOC -I/usr/include/jemalloc/include -DUSE_OPENSSL=1 -DBUILD_TLS_MODULE=0 -Wdate-time -D_FORTIFY_SOURCE=2 -MMD -o strl.o -c strl.c +cc -pedantic -std=gnu11 -Wall -W -Wno-missing-field-initializers -Werror=deprecated-declarations -Wstrict-prototypes -O3 -flto=auto -fno-omit-frame-pointer -g -ggdb -g -O2 -Werror=implicit-function-declaration -ffile-prefix-map=/build/reproducible-path/valkey-8.0.2+dfsg1=. -flto=auto -ffat-lto-objects -fstack-protector-strong -fstack-clash-protection -Wformat -Werror=format-security -fcf-protection -I/usr/include/liblzf -I../deps/hiredis -I../deps/linenoise -I../deps/lua/src -I../deps/hdr_histogram -I../deps/fpconv -DHAVE_LIBSYSTEMD -DUSE_JEMALLOC -I/usr/include/jemalloc/include -DUSE_OPENSSL=1 -DBUILD_TLS_MODULE=0 -Wdate-time -D_FORTIFY_SOURCE=2 -MMD -o connection.o -c connection.c +cc -pedantic -std=gnu11 -Wall -W -Wno-missing-field-initializers -Werror=deprecated-declarations -Wstrict-prototypes -O3 -flto=auto -fno-omit-frame-pointer -g -ggdb -g -O2 -Werror=implicit-function-declaration -ffile-prefix-map=/build/reproducible-path/valkey-8.0.2+dfsg1=. -flto=auto -ffat-lto-objects -fstack-protector-strong -fstack-clash-protection -Wformat -Werror=format-security -fcf-protection -I/usr/include/liblzf -I../deps/hiredis -I../deps/linenoise -I../deps/lua/src -I../deps/hdr_histogram -I../deps/fpconv -DHAVE_LIBSYSTEMD -DUSE_JEMALLOC -I/usr/include/jemalloc/include -DUSE_OPENSSL=1 -DBUILD_TLS_MODULE=0 -Wdate-time -D_FORTIFY_SOURCE=2 -MMD -o unix.o -c unix.c +cc -pedantic -std=gnu11 -Wall -W -Wno-missing-field-initializers -Werror=deprecated-declarations -Wstrict-prototypes -O3 -flto=auto -fno-omit-frame-pointer -g -ggdb -g -O2 -Werror=implicit-function-declaration -ffile-prefix-map=/build/reproducible-path/valkey-8.0.2+dfsg1=. -flto=auto -ffat-lto-objects -fstack-protector-strong -fstack-clash-protection -Wformat -Werror=format-security -fcf-protection -I/usr/include/liblzf -I../deps/hiredis -I../deps/linenoise -I../deps/lua/src -I../deps/hdr_histogram -I../deps/fpconv -DHAVE_LIBSYSTEMD -DUSE_JEMALLOC -I/usr/include/jemalloc/include -DUSE_OPENSSL=1 -DBUILD_TLS_MODULE=0 -Wdate-time -D_FORTIFY_SOURCE=2 -MMD -o logreqres.o -c logreqres.c +cc -pedantic -std=gnu11 -Wall -W -Wno-missing-field-initializers -Werror=deprecated-declarations -Wstrict-prototypes -O3 -flto=auto -fno-omit-frame-pointer -g -ggdb -g -O2 -Werror=implicit-function-declaration -ffile-prefix-map=/build/reproducible-path/valkey-8.0.2+dfsg1=. -flto=auto -ffat-lto-objects -fstack-protector-strong -fstack-clash-protection -Wformat -Werror=format-security -fcf-protection -I/usr/include/liblzf -I../deps/hiredis -I../deps/linenoise -I../deps/lua/src -I../deps/hdr_histogram -I../deps/fpconv -DHAVE_LIBSYSTEMD -DUSE_JEMALLOC -I/usr/include/jemalloc/include -DUSE_OPENSSL=1 -DBUILD_TLS_MODULE=0 -Wdate-time -D_FORTIFY_SOURCE=2 -MMD -o valkey-cli.o -c valkey-cli.c +cc -pedantic -std=gnu11 -Wall -W -Wno-missing-field-initializers -Werror=deprecated-declarations -Wstrict-prototypes -O3 -flto=auto -fno-omit-frame-pointer -g -ggdb -g -O2 -Werror=implicit-function-declaration -ffile-prefix-map=/build/reproducible-path/valkey-8.0.2+dfsg1=. -flto=auto -ffat-lto-objects -fstack-protector-strong -fstack-clash-protection -Wformat -Werror=format-security -fcf-protection -I/usr/include/liblzf -I../deps/hiredis -I../deps/linenoise -I../deps/lua/src -I../deps/hdr_histogram -I../deps/fpconv -DHAVE_LIBSYSTEMD -DUSE_JEMALLOC -I/usr/include/jemalloc/include -DUSE_OPENSSL=1 -DBUILD_TLS_MODULE=0 -Wdate-time -D_FORTIFY_SOURCE=2 -MMD -o serverassert.o -c serverassert.c +cc -pedantic -std=gnu11 -Wall -W -Wno-missing-field-initializers -Werror=deprecated-declarations -Wstrict-prototypes -O3 -flto=auto -fno-omit-frame-pointer -g -ggdb -g -O2 -Werror=implicit-function-declaration -ffile-prefix-map=/build/reproducible-path/valkey-8.0.2+dfsg1=. -flto=auto -ffat-lto-objects -fstack-protector-strong -fstack-clash-protection -Wformat -Werror=format-security -fcf-protection -I/usr/include/liblzf -I../deps/hiredis -I../deps/linenoise -I../deps/lua/src -I../deps/hdr_histogram -I../deps/fpconv -DHAVE_LIBSYSTEMD -DUSE_JEMALLOC -I/usr/include/jemalloc/include -DUSE_OPENSSL=1 -DBUILD_TLS_MODULE=0 -Wdate-time -D_FORTIFY_SOURCE=2 -MMD -o cli_common.o -c cli_common.c +cc -pedantic -std=gnu11 -Wall -W -Wno-missing-field-initializers -Werror=deprecated-declarations -Wstrict-prototypes -O3 -flto=auto -fno-omit-frame-pointer -g -ggdb -g -O2 -Werror=implicit-function-declaration -ffile-prefix-map=/build/reproducible-path/valkey-8.0.2+dfsg1=. -flto=auto -ffat-lto-objects -fstack-protector-strong -fstack-clash-protection -Wformat -Werror=format-security -fcf-protection -I/usr/include/liblzf -I../deps/hiredis -I../deps/linenoise -I../deps/lua/src -I../deps/hdr_histogram -I../deps/fpconv -DHAVE_LIBSYSTEMD -DUSE_JEMALLOC -I/usr/include/jemalloc/include -DUSE_OPENSSL=1 -DBUILD_TLS_MODULE=0 -Wdate-time -D_FORTIFY_SOURCE=2 -MMD -o cli_commands.o -c cli_commands.c +cc -pedantic -std=gnu11 -Wall -W -Wno-missing-field-initializers -Werror=deprecated-declarations -Wstrict-prototypes -O3 -flto=auto -fno-omit-frame-pointer -g -ggdb -g -O2 -Werror=implicit-function-declaration -ffile-prefix-map=/build/reproducible-path/valkey-8.0.2+dfsg1=. -flto=auto -ffat-lto-objects -fstack-protector-strong -fstack-clash-protection -Wformat -Werror=format-security -fcf-protection -I/usr/include/liblzf -I../deps/hiredis -I../deps/linenoise -I../deps/lua/src -I../deps/hdr_histogram -I../deps/fpconv -DHAVE_LIBSYSTEMD -DUSE_JEMALLOC -I/usr/include/jemalloc/include -DUSE_OPENSSL=1 -DBUILD_TLS_MODULE=0 -Wdate-time -D_FORTIFY_SOURCE=2 -MMD -o valkey-benchmark.o -c valkey-benchmark.c In function 'createClusterMsgSendBlock', inlined from 'clusterSendFail' at cluster_legacy.c:4125:37, inlined from 'markNodeAsFailingIfNeeded' at cluster_legacy.c:2042:5: cluster_legacy.c:1293:24: warning: array subscript 'struct clusterMsgSendBlock[0]' is partly outside array bounds of 'unsigned char[2312]' [-Warray-bounds=] 1293 | msgblock->refcount = 1; | ~~~~~~~~~~~~~~~~~~~^~~ -cc -pedantic -std=gnu11 -Wall -W -Wno-missing-field-initializers -Werror=deprecated-declarations -Wstrict-prototypes -O3 -flto=auto -fno-omit-frame-pointer -g -ggdb -g -O2 -Werror=implicit-function-declaration -ffile-prefix-map=/build/reproducible-path/valkey-8.0.2+dfsg1=. -flto=auto -ffat-lto-objects -fstack-protector-strong -fstack-clash-protection -Wformat -Werror=format-security -fcf-protection -I/usr/include/liblzf -I../deps/hiredis -I../deps/linenoise -I../deps/lua/src -I../deps/hdr_histogram -I../deps/fpconv -DHAVE_LIBSYSTEMD -DUSE_JEMALLOC -I/usr/include/jemalloc/include -DUSE_OPENSSL=1 -DBUILD_TLS_MODULE=0 -Wdate-time -D_FORTIFY_SOURCE=2 -MMD -o lolwut6.o -c lolwut6.c cluster_legacy.c:1292:37: note: object of size 2312 allocated by 'zcalloc' 1292 | clusterMsgSendBlock *msgblock = zcalloc(blocklen); | ^~~~~~~~~~~~~~~~~ @@ -744,7 +805,6 @@ cluster_legacy.c:1306:13: warning: array subscript 'struct clusterMsgSendBlock[0]' is partly outside array bounds of 'unsigned char[2312]' [-Warray-bounds=] 1306 | msgblock->refcount--; | ~~~~~~~~^~~~~~~~~~ -cc -pedantic -std=gnu11 -Wall -W -Wno-missing-field-initializers -Werror=deprecated-declarations -Wstrict-prototypes -O3 -flto=auto -fno-omit-frame-pointer -g -ggdb -g -O2 -Werror=implicit-function-declaration -ffile-prefix-map=/build/reproducible-path/valkey-8.0.2+dfsg1=. -flto=auto -ffat-lto-objects -fstack-protector-strong -fstack-clash-protection -Wformat -Werror=format-security -fcf-protection -I/usr/include/liblzf -I../deps/hiredis -I../deps/linenoise -I../deps/lua/src -I../deps/hdr_histogram -I../deps/fpconv -DHAVE_LIBSYSTEMD -DUSE_JEMALLOC -I/usr/include/jemalloc/include -DUSE_OPENSSL=1 -DBUILD_TLS_MODULE=0 -Wdate-time -D_FORTIFY_SOURCE=2 -MMD -o acl.o -c acl.c In function 'createClusterMsgSendBlock', inlined from 'clusterSendFail' at cluster_legacy.c:4125:37, inlined from 'markNodeAsFailingIfNeeded' at cluster_legacy.c:2042:5: @@ -775,12 +835,6 @@ cluster_legacy.c:1292:37: note: object of size 2312 allocated by 'zcalloc' 1292 | clusterMsgSendBlock *msgblock = zcalloc(blocklen); | ^~~~~~~~~~~~~~~~~ -cc -pedantic -std=gnu11 -Wall -W -Wno-missing-field-initializers -Werror=deprecated-declarations -Wstrict-prototypes -O3 -flto=auto -fno-omit-frame-pointer -g -ggdb -g -O2 -Werror=implicit-function-declaration -ffile-prefix-map=/build/reproducible-path/valkey-8.0.2+dfsg1=. -flto=auto -ffat-lto-objects -fstack-protector-strong -fstack-clash-protection -Wformat -Werror=format-security -fcf-protection -I/usr/include/liblzf -I../deps/hiredis -I../deps/linenoise -I../deps/lua/src -I../deps/hdr_histogram -I../deps/fpconv -DHAVE_LIBSYSTEMD -DUSE_JEMALLOC -I/usr/include/jemalloc/include -DUSE_OPENSSL=1 -DBUILD_TLS_MODULE=0 -Wdate-time -D_FORTIFY_SOURCE=2 -MMD -o tracking.o -c tracking.c -cc -pedantic -std=gnu11 -Wall -W -Wno-missing-field-initializers -Werror=deprecated-declarations -Wstrict-prototypes -O3 -flto=auto -fno-omit-frame-pointer -g -ggdb -g -O2 -Werror=implicit-function-declaration -ffile-prefix-map=/build/reproducible-path/valkey-8.0.2+dfsg1=. -flto=auto -ffat-lto-objects -fstack-protector-strong -fstack-clash-protection -Wformat -Werror=format-security -fcf-protection -I/usr/include/liblzf -I../deps/hiredis -I../deps/linenoise -I../deps/lua/src -I../deps/hdr_histogram -I../deps/fpconv -DHAVE_LIBSYSTEMD -DUSE_JEMALLOC -I/usr/include/jemalloc/include -DUSE_OPENSSL=1 -DBUILD_TLS_MODULE=0 -Wdate-time -D_FORTIFY_SOURCE=2 -MMD -o socket.o -c socket.c -cc -pedantic -std=gnu11 -Wall -W -Wno-missing-field-initializers -Werror=deprecated-declarations -Wstrict-prototypes -O3 -flto=auto -fno-omit-frame-pointer -g -ggdb -g -O2 -Werror=implicit-function-declaration -ffile-prefix-map=/build/reproducible-path/valkey-8.0.2+dfsg1=. -flto=auto -ffat-lto-objects -fstack-protector-strong -fstack-clash-protection -Wformat -Werror=format-security -fcf-protection -I/usr/include/liblzf -I../deps/hiredis -I../deps/linenoise -I../deps/lua/src -I../deps/hdr_histogram -I../deps/fpconv -DHAVE_LIBSYSTEMD -DUSE_JEMALLOC -I/usr/include/jemalloc/include -DUSE_OPENSSL=1 -DBUILD_TLS_MODULE=0 -Wdate-time -D_FORTIFY_SOURCE=2 -MMD -o tls.o -c tls.c -cc -pedantic -std=gnu11 -Wall -W -Wno-missing-field-initializers -Werror=deprecated-declarations -Wstrict-prototypes -O3 -flto=auto -fno-omit-frame-pointer -g -ggdb -g -O2 -Werror=implicit-function-declaration -ffile-prefix-map=/build/reproducible-path/valkey-8.0.2+dfsg1=. -flto=auto -ffat-lto-objects -fstack-protector-strong -fstack-clash-protection -Wformat -Werror=format-security -fcf-protection -I/usr/include/liblzf -I../deps/hiredis -I../deps/linenoise -I../deps/lua/src -I../deps/hdr_histogram -I../deps/fpconv -DHAVE_LIBSYSTEMD -DUSE_JEMALLOC -I/usr/include/jemalloc/include -DUSE_OPENSSL=1 -DBUILD_TLS_MODULE=0 -Wdate-time -D_FORTIFY_SOURCE=2 -MMD -o sha256.o -c sha256.c -cc -pedantic -std=gnu11 -Wall -W -Wno-missing-field-initializers -Werror=deprecated-declarations -Wstrict-prototypes -O3 -flto=auto -fno-omit-frame-pointer -g -ggdb -g -O2 -Werror=implicit-function-declaration -ffile-prefix-map=/build/reproducible-path/valkey-8.0.2+dfsg1=. -flto=auto -ffat-lto-objects -fstack-protector-strong -fstack-clash-protection -Wformat -Werror=format-security -fcf-protection -I/usr/include/liblzf -I../deps/hiredis -I../deps/linenoise -I../deps/lua/src -I../deps/hdr_histogram -I../deps/fpconv -DHAVE_LIBSYSTEMD -DUSE_JEMALLOC -I/usr/include/jemalloc/include -DUSE_OPENSSL=1 -DBUILD_TLS_MODULE=0 -Wdate-time -D_FORTIFY_SOURCE=2 -MMD -o timeout.o -c timeout.c -cc -pedantic -std=gnu11 -Wall -W -Wno-missing-field-initializers -Werror=deprecated-declarations -Wstrict-prototypes -O3 -flto=auto -fno-omit-frame-pointer -g -ggdb -g -O2 -Werror=implicit-function-declaration -ffile-prefix-map=/build/reproducible-path/valkey-8.0.2+dfsg1=. -flto=auto -ffat-lto-objects -fstack-protector-strong -fstack-clash-protection -Wformat -Werror=format-security -fcf-protection -I/usr/include/liblzf -I../deps/hiredis -I../deps/linenoise -I../deps/lua/src -I../deps/hdr_histogram -I../deps/fpconv -DHAVE_LIBSYSTEMD -DUSE_JEMALLOC -I/usr/include/jemalloc/include -DUSE_OPENSSL=1 -DBUILD_TLS_MODULE=0 -Wdate-time -D_FORTIFY_SOURCE=2 -MMD -o setcpuaffinity.o -c setcpuaffinity.c In function 'createClusterMsgSendBlock', inlined from 'clusterSendFail' at cluster_legacy.c:4125:37: cluster_legacy.c:1293:24: warning: array subscript 'struct clusterMsgSendBlock[0]' is partly outside array bounds of 'unsigned char[2312]' [-Warray-bounds=] @@ -789,7 +843,6 @@ cluster_legacy.c:1292:37: note: object of size 2312 allocated by 'zcalloc' 1292 | clusterMsgSendBlock *msgblock = zcalloc(blocklen); | ^~~~~~~~~~~~~~~~~ -cc -pedantic -std=gnu11 -Wall -W -Wno-missing-field-initializers -Werror=deprecated-declarations -Wstrict-prototypes -O3 -flto=auto -fno-omit-frame-pointer -g -ggdb -g -O2 -Werror=implicit-function-declaration -ffile-prefix-map=/build/reproducible-path/valkey-8.0.2+dfsg1=. -flto=auto -ffat-lto-objects -fstack-protector-strong -fstack-clash-protection -Wformat -Werror=format-security -fcf-protection -I/usr/include/liblzf -I../deps/hiredis -I../deps/linenoise -I../deps/lua/src -I../deps/hdr_histogram -I../deps/fpconv -DHAVE_LIBSYSTEMD -DUSE_JEMALLOC -I/usr/include/jemalloc/include -DUSE_OPENSSL=1 -DBUILD_TLS_MODULE=0 -Wdate-time -D_FORTIFY_SOURCE=2 -MMD -o monotonic.o -c monotonic.c cluster_legacy.c:1294:22: warning: array subscript 'struct clusterMsgSendBlock[0]' is partly outside array bounds of 'unsigned char[2312]' [-Warray-bounds=] 1294 | msgblock->totlen = blocklen; | ~~~~~~~~~~~~~~~~~^~~~~~~~~~ @@ -826,8 +879,6 @@ cluster_legacy.c:1292:37: note: object of size 2312 allocated by 'zcalloc' 1292 | clusterMsgSendBlock *msgblock = zcalloc(blocklen); | ^~~~~~~~~~~~~~~~~ -cc -pedantic -std=gnu11 -Wall -W -Wno-missing-field-initializers -Werror=deprecated-declarations -Wstrict-prototypes -O3 -flto=auto -fno-omit-frame-pointer -g -ggdb -g -O2 -Werror=implicit-function-declaration -ffile-prefix-map=/build/reproducible-path/valkey-8.0.2+dfsg1=. -flto=auto -ffat-lto-objects -fstack-protector-strong -fstack-clash-protection -Wformat -Werror=format-security -fcf-protection -I/usr/include/liblzf -I../deps/hiredis -I../deps/linenoise -I../deps/lua/src -I../deps/hdr_histogram -I../deps/fpconv -DHAVE_LIBSYSTEMD -DUSE_JEMALLOC -I/usr/include/jemalloc/include -DUSE_OPENSSL=1 -DBUILD_TLS_MODULE=0 -Wdate-time -D_FORTIFY_SOURCE=2 -MMD -o mt19937-64.o -c mt19937-64.c -cc -pedantic -std=gnu11 -Wall -W -Wno-missing-field-initializers -Werror=deprecated-declarations -Wstrict-prototypes -O3 -flto=auto -fno-omit-frame-pointer -g -ggdb -g -O2 -Werror=implicit-function-declaration -ffile-prefix-map=/build/reproducible-path/valkey-8.0.2+dfsg1=. -flto=auto -ffat-lto-objects -fstack-protector-strong -fstack-clash-protection -Wformat -Werror=format-security -fcf-protection -I/usr/include/liblzf -I../deps/hiredis -I../deps/linenoise -I../deps/lua/src -I../deps/hdr_histogram -I../deps/fpconv -DHAVE_LIBSYSTEMD -DUSE_JEMALLOC -I/usr/include/jemalloc/include -DUSE_OPENSSL=1 -DBUILD_TLS_MODULE=0 -Wdate-time -D_FORTIFY_SOURCE=2 -MMD -o resp_parser.o -c resp_parser.c In function 'createClusterMsgSendBlock', inlined from 'clusterRequestFailoverAuth' at cluster_legacy.c:4252:37: cluster_legacy.c:1293:24: warning: array subscript 'struct clusterMsgSendBlock[0]' is partly outside array bounds of 'unsigned char[2272]' [-Warray-bounds=] @@ -1050,21 +1101,6 @@ cluster_legacy.c:1292:37: note: object of size 2272 allocated by 'zcalloc' 1292 | clusterMsgSendBlock *msgblock = zcalloc(blocklen); | ^~~~~~~~~~~~~~~~~ -cc -pedantic -std=gnu11 -Wall -W -Wno-missing-field-initializers -Werror=deprecated-declarations -Wstrict-prototypes -O3 -flto=auto -fno-omit-frame-pointer -g -ggdb -g -O2 -Werror=implicit-function-declaration -ffile-prefix-map=/build/reproducible-path/valkey-8.0.2+dfsg1=. -flto=auto -ffat-lto-objects -fstack-protector-strong -fstack-clash-protection -Wformat -Werror=format-security -fcf-protection -I/usr/include/liblzf -I../deps/hiredis -I../deps/linenoise -I../deps/lua/src -I../deps/hdr_histogram -I../deps/fpconv -DHAVE_LIBSYSTEMD -DUSE_JEMALLOC -I/usr/include/jemalloc/include -DUSE_OPENSSL=1 -DBUILD_TLS_MODULE=0 -Wdate-time -D_FORTIFY_SOURCE=2 -MMD -o call_reply.o -c call_reply.c -cc -pedantic -std=gnu11 -Wall -W -Wno-missing-field-initializers -Werror=deprecated-declarations -Wstrict-prototypes -O3 -flto=auto -fno-omit-frame-pointer -g -ggdb -g -O2 -Werror=implicit-function-declaration -ffile-prefix-map=/build/reproducible-path/valkey-8.0.2+dfsg1=. -flto=auto -ffat-lto-objects -fstack-protector-strong -fstack-clash-protection -Wformat -Werror=format-security -fcf-protection -I/usr/include/liblzf -I../deps/hiredis -I../deps/linenoise -I../deps/lua/src -I../deps/hdr_histogram -I../deps/fpconv -DHAVE_LIBSYSTEMD -DUSE_JEMALLOC -I/usr/include/jemalloc/include -DUSE_OPENSSL=1 -DBUILD_TLS_MODULE=0 -Wdate-time -D_FORTIFY_SOURCE=2 -MMD -o script_lua.o -c script_lua.c -cc -pedantic -std=gnu11 -Wall -W -Wno-missing-field-initializers -Werror=deprecated-declarations -Wstrict-prototypes -O3 -flto=auto -fno-omit-frame-pointer -g -ggdb -g -O2 -Werror=implicit-function-declaration -ffile-prefix-map=/build/reproducible-path/valkey-8.0.2+dfsg1=. -flto=auto -ffat-lto-objects -fstack-protector-strong -fstack-clash-protection -Wformat -Werror=format-security -fcf-protection -I/usr/include/liblzf -I../deps/hiredis -I../deps/linenoise -I../deps/lua/src -I../deps/hdr_histogram -I../deps/fpconv -DHAVE_LIBSYSTEMD -DUSE_JEMALLOC -I/usr/include/jemalloc/include -DUSE_OPENSSL=1 -DBUILD_TLS_MODULE=0 -Wdate-time -D_FORTIFY_SOURCE=2 -MMD -o script.o -c script.c -cc -pedantic -std=gnu11 -Wall -W -Wno-missing-field-initializers -Werror=deprecated-declarations -Wstrict-prototypes -O3 -flto=auto -fno-omit-frame-pointer -g -ggdb -g -O2 -Werror=implicit-function-declaration -ffile-prefix-map=/build/reproducible-path/valkey-8.0.2+dfsg1=. -flto=auto -ffat-lto-objects -fstack-protector-strong -fstack-clash-protection -Wformat -Werror=format-security -fcf-protection -I/usr/include/liblzf -I../deps/hiredis -I../deps/linenoise -I../deps/lua/src -I../deps/hdr_histogram -I../deps/fpconv -DHAVE_LIBSYSTEMD -DUSE_JEMALLOC -I/usr/include/jemalloc/include -DUSE_OPENSSL=1 -DBUILD_TLS_MODULE=0 -Wdate-time -D_FORTIFY_SOURCE=2 -MMD -o functions.o -c functions.c -cc -pedantic -std=gnu11 -Wall -W -Wno-missing-field-initializers -Werror=deprecated-declarations -Wstrict-prototypes -O3 -flto=auto -fno-omit-frame-pointer -g -ggdb -g -O2 -Werror=implicit-function-declaration -ffile-prefix-map=/build/reproducible-path/valkey-8.0.2+dfsg1=. -flto=auto -ffat-lto-objects -fstack-protector-strong -fstack-clash-protection -Wformat -Werror=format-security -fcf-protection -I/usr/include/liblzf -I../deps/hiredis -I../deps/linenoise -I../deps/lua/src -I../deps/hdr_histogram -I../deps/fpconv -DHAVE_LIBSYSTEMD -DUSE_JEMALLOC -I/usr/include/jemalloc/include -DUSE_OPENSSL=1 -DBUILD_TLS_MODULE=0 -Wdate-time -D_FORTIFY_SOURCE=2 -MMD -o function_lua.o -c function_lua.c -cc -pedantic -std=gnu11 -Wall -W -Wno-missing-field-initializers -Werror=deprecated-declarations -Wstrict-prototypes -O3 -flto=auto -fno-omit-frame-pointer -g -ggdb -g -O2 -Werror=implicit-function-declaration -ffile-prefix-map=/build/reproducible-path/valkey-8.0.2+dfsg1=. -flto=auto -ffat-lto-objects -fstack-protector-strong -fstack-clash-protection -Wformat -Werror=format-security -fcf-protection -I/usr/include/liblzf -I../deps/hiredis -I../deps/linenoise -I../deps/lua/src -I../deps/hdr_histogram -I../deps/fpconv -DHAVE_LIBSYSTEMD -DUSE_JEMALLOC -I/usr/include/jemalloc/include -DUSE_OPENSSL=1 -DBUILD_TLS_MODULE=0 -Wdate-time -D_FORTIFY_SOURCE=2 -MMD -o commands.o -c commands.c -cc -pedantic -std=gnu11 -Wall -W -Wno-missing-field-initializers -Werror=deprecated-declarations -Wstrict-prototypes -O3 -flto=auto -fno-omit-frame-pointer -g -ggdb -g -O2 -Werror=implicit-function-declaration -ffile-prefix-map=/build/reproducible-path/valkey-8.0.2+dfsg1=. -flto=auto -ffat-lto-objects -fstack-protector-strong -fstack-clash-protection -Wformat -Werror=format-security -fcf-protection -I/usr/include/liblzf -I../deps/hiredis -I../deps/linenoise -I../deps/lua/src -I../deps/hdr_histogram -I../deps/fpconv -DHAVE_LIBSYSTEMD -DUSE_JEMALLOC -I/usr/include/jemalloc/include -DUSE_OPENSSL=1 -DBUILD_TLS_MODULE=0 -Wdate-time -D_FORTIFY_SOURCE=2 -MMD -o strl.o -c strl.c -cc -pedantic -std=gnu11 -Wall -W -Wno-missing-field-initializers -Werror=deprecated-declarations -Wstrict-prototypes -O3 -flto=auto -fno-omit-frame-pointer -g -ggdb -g -O2 -Werror=implicit-function-declaration -ffile-prefix-map=/build/reproducible-path/valkey-8.0.2+dfsg1=. -flto=auto -ffat-lto-objects -fstack-protector-strong -fstack-clash-protection -Wformat -Werror=format-security -fcf-protection -I/usr/include/liblzf -I../deps/hiredis -I../deps/linenoise -I../deps/lua/src -I../deps/hdr_histogram -I../deps/fpconv -DHAVE_LIBSYSTEMD -DUSE_JEMALLOC -I/usr/include/jemalloc/include -DUSE_OPENSSL=1 -DBUILD_TLS_MODULE=0 -Wdate-time -D_FORTIFY_SOURCE=2 -MMD -o connection.o -c connection.c -cc -pedantic -std=gnu11 -Wall -W -Wno-missing-field-initializers -Werror=deprecated-declarations -Wstrict-prototypes -O3 -flto=auto -fno-omit-frame-pointer -g -ggdb -g -O2 -Werror=implicit-function-declaration -ffile-prefix-map=/build/reproducible-path/valkey-8.0.2+dfsg1=. -flto=auto -ffat-lto-objects -fstack-protector-strong -fstack-clash-protection -Wformat -Werror=format-security -fcf-protection -I/usr/include/liblzf -I../deps/hiredis -I../deps/linenoise -I../deps/lua/src -I../deps/hdr_histogram -I../deps/fpconv -DHAVE_LIBSYSTEMD -DUSE_JEMALLOC -I/usr/include/jemalloc/include -DUSE_OPENSSL=1 -DBUILD_TLS_MODULE=0 -Wdate-time -D_FORTIFY_SOURCE=2 -MMD -o unix.o -c unix.c -cc -pedantic -std=gnu11 -Wall -W -Wno-missing-field-initializers -Werror=deprecated-declarations -Wstrict-prototypes -O3 -flto=auto -fno-omit-frame-pointer -g -ggdb -g -O2 -Werror=implicit-function-declaration -ffile-prefix-map=/build/reproducible-path/valkey-8.0.2+dfsg1=. -flto=auto -ffat-lto-objects -fstack-protector-strong -fstack-clash-protection -Wformat -Werror=format-security -fcf-protection -I/usr/include/liblzf -I../deps/hiredis -I../deps/linenoise -I../deps/lua/src -I../deps/hdr_histogram -I../deps/fpconv -DHAVE_LIBSYSTEMD -DUSE_JEMALLOC -I/usr/include/jemalloc/include -DUSE_OPENSSL=1 -DBUILD_TLS_MODULE=0 -Wdate-time -D_FORTIFY_SOURCE=2 -MMD -o logreqres.o -c logreqres.c -cc -pedantic -std=gnu11 -Wall -W -Wno-missing-field-initializers -Werror=deprecated-declarations -Wstrict-prototypes -O3 -flto=auto -fno-omit-frame-pointer -g -ggdb -g -O2 -Werror=implicit-function-declaration -ffile-prefix-map=/build/reproducible-path/valkey-8.0.2+dfsg1=. -flto=auto -ffat-lto-objects -fstack-protector-strong -fstack-clash-protection -Wformat -Werror=format-security -fcf-protection -I/usr/include/liblzf -I../deps/hiredis -I../deps/linenoise -I../deps/lua/src -I../deps/hdr_histogram -I../deps/fpconv -DHAVE_LIBSYSTEMD -DUSE_JEMALLOC -I/usr/include/jemalloc/include -DUSE_OPENSSL=1 -DBUILD_TLS_MODULE=0 -Wdate-time -D_FORTIFY_SOURCE=2 -MMD -o valkey-cli.o -c valkey-cli.c -cc -pedantic -std=gnu11 -Wall -W -Wno-missing-field-initializers -Werror=deprecated-declarations -Wstrict-prototypes -O3 -flto=auto -fno-omit-frame-pointer -g -ggdb -g -O2 -Werror=implicit-function-declaration -ffile-prefix-map=/build/reproducible-path/valkey-8.0.2+dfsg1=. -flto=auto -ffat-lto-objects -fstack-protector-strong -fstack-clash-protection -Wformat -Werror=format-security -fcf-protection -I/usr/include/liblzf -I../deps/hiredis -I../deps/linenoise -I../deps/lua/src -I../deps/hdr_histogram -I../deps/fpconv -DHAVE_LIBSYSTEMD -DUSE_JEMALLOC -I/usr/include/jemalloc/include -DUSE_OPENSSL=1 -DBUILD_TLS_MODULE=0 -Wdate-time -D_FORTIFY_SOURCE=2 -MMD -o serverassert.o -c serverassert.c -cc -pedantic -std=gnu11 -Wall -W -Wno-missing-field-initializers -Werror=deprecated-declarations -Wstrict-prototypes -O3 -flto=auto -fno-omit-frame-pointer -g -ggdb -g -O2 -Werror=implicit-function-declaration -ffile-prefix-map=/build/reproducible-path/valkey-8.0.2+dfsg1=. -flto=auto -ffat-lto-objects -fstack-protector-strong -fstack-clash-protection -Wformat -Werror=format-security -fcf-protection -I/usr/include/liblzf -I../deps/hiredis -I../deps/linenoise -I../deps/lua/src -I../deps/hdr_histogram -I../deps/fpconv -DHAVE_LIBSYSTEMD -DUSE_JEMALLOC -I/usr/include/jemalloc/include -DUSE_OPENSSL=1 -DBUILD_TLS_MODULE=0 -Wdate-time -D_FORTIFY_SOURCE=2 -MMD -o cli_common.o -c cli_common.c -cc -pedantic -std=gnu11 -Wall -W -Wno-missing-field-initializers -Werror=deprecated-declarations -Wstrict-prototypes -O3 -flto=auto -fno-omit-frame-pointer -g -ggdb -g -O2 -Werror=implicit-function-declaration -ffile-prefix-map=/build/reproducible-path/valkey-8.0.2+dfsg1=. -flto=auto -ffat-lto-objects -fstack-protector-strong -fstack-clash-protection -Wformat -Werror=format-security -fcf-protection -I/usr/include/liblzf -I../deps/hiredis -I../deps/linenoise -I../deps/lua/src -I../deps/hdr_histogram -I../deps/fpconv -DHAVE_LIBSYSTEMD -DUSE_JEMALLOC -I/usr/include/jemalloc/include -DUSE_OPENSSL=1 -DBUILD_TLS_MODULE=0 -Wdate-time -D_FORTIFY_SOURCE=2 -MMD -o cli_commands.o -c cli_commands.c -cc -pedantic -std=gnu11 -Wall -W -Wno-missing-field-initializers -Werror=deprecated-declarations -Wstrict-prototypes -O3 -flto=auto -fno-omit-frame-pointer -g -ggdb -g -O2 -Werror=implicit-function-declaration -ffile-prefix-map=/build/reproducible-path/valkey-8.0.2+dfsg1=. -flto=auto -ffat-lto-objects -fstack-protector-strong -fstack-clash-protection -Wformat -Werror=format-security -fcf-protection -I/usr/include/liblzf -I../deps/hiredis -I../deps/linenoise -I../deps/lua/src -I../deps/hdr_histogram -I../deps/fpconv -DHAVE_LIBSYSTEMD -DUSE_JEMALLOC -I/usr/include/jemalloc/include -DUSE_OPENSSL=1 -DBUILD_TLS_MODULE=0 -Wdate-time -D_FORTIFY_SOURCE=2 -MMD -o valkey-benchmark.o -c valkey-benchmark.c cc -flto=auto -ffat-lto-objects -Wl,-z,relro -Wl,-z,now -Wl,-no-as-needed -ldl -latomic -llzf -O3 -flto=auto -fno-omit-frame-pointer -g -ggdb -rdynamic -o valkey-benchmark ae.o anet.o valkey-benchmark.o adlist.o dict.o zmalloc.o serverassert.o release.o crcspeed.o crccombine.o crc64.o siphash.o crc16.o monotonic.o cli_common.o mt19937-64.o strl.o ../deps/hiredis/libhiredis.a ../deps/hdr_histogram/libhdrhistogram.a -ljemalloc -lm -ldl -pthread -lrt -lsystemd ../deps/hiredis/libhiredis_ssl.a -lssl -lcrypto In function 'createClusterMsgSendBlock', inlined from 'clusterSendMFStart' at cluster_legacy.c:4279:37, @@ -1127,10 +1163,10 @@ cc -flto=auto -ffat-lto-objects -Wl,-z,relro -Wl,-z,now -Wl,-no-as-needed -ldl -latomic -llzf -O3 -flto=auto -fno-omit-frame-pointer -g -ggdb -rdynamic -o valkey-server threads_mngr.o adlist.o quicklist.o ae.o anet.o dict.o kvstore.o server.o sds.o zmalloc.o lzf_c.o lzf_d.o pqsort.o zipmap.o sha1.o ziplist.o release.o memory_prefetch.o io_threads.o networking.o util.o object.o db.o replication.o rdb.o t_string.o t_list.o t_set.o t_zset.o t_hash.o config.o aof.o pubsub.o multi.o debug.o sort.o intset.o syncio.o cluster.o cluster_legacy.o cluster_slot_stats.o crc16.o endianconv.o slowlog.o eval.o bio.o rio.o rand.o memtest.o syscheck.o crcspeed.o crccombine.o crc64.o bitops.o sentinel.o notify.o setproctitle.o blocked.o hyperloglog.o latency.o sparkline.o valkey-check-rdb.o valkey-check-aof.o geo.o lazyfree.o module.o evict.o expire.o geohash.o geohash_helper.o childinfo.o defrag.o siphash.o rax.o t_stream.o listpack.o localtime.o lolwut.o lolwut5.o lolwut6.o acl.o tracking.o socket.o tls.o sha256.o timeout.o setcpuaffinity.o monotonic.o mt19937-64.o resp_parser.o call_reply.o script_lua.o script.o functions.o function_lua.o commands.o strl.o connection.o unix.o logreqres.o ../deps/hiredis/libhiredis.a ../deps/lua/src/liblua.a ../deps/hdr_histogram/libhdrhistogram.a ../deps/fpconv/libfpconv.a -ljemalloc -lm -ldl -pthread -lrt -lsystemd ../deps/hiredis/libhiredis_ssl.a -lssl -lcrypto cc -flto=auto -ffat-lto-objects -Wl,-z,relro -Wl,-z,now -Wl,-no-as-needed -ldl -latomic -llzf -O3 -flto=auto -fno-omit-frame-pointer -g -ggdb -rdynamic -o valkey-cli anet.o adlist.o dict.o valkey-cli.o zmalloc.o release.o ae.o serverassert.o crcspeed.o crccombine.o crc64.o siphash.o crc16.o monotonic.o cli_common.o mt19937-64.o strl.o cli_commands.o ../deps/hiredis/libhiredis.a ../deps/linenoise/linenoise.o -ljemalloc -lm -ldl -pthread -lrt -lsystemd ../deps/hiredis/libhiredis_ssl.a -lssl -lcrypto install --strip-program=true valkey-server valkey-sentinel -install: WARNING: ignoring --strip-program option as -s option was not specified install --strip-program=true valkey-server valkey-check-rdb -install: install --strip-program=true valkey-server valkey-check-aof -WARNING: ignoring --strip-program option as -s option was not specified +install --strip-program=true valkey-server valkey-check-aof +install: WARNING: ignoring --strip-program option as -s option was not specified +install: WARNING: ignoring --strip-program option as -s option was not specified install: WARNING: ignoring --strip-program option as -s option was not specified Hint: It's a good idea to run 'make test' ;) @@ -1149,798 +1185,331 @@ Certificate request self-signature ok subject=O=Valkey Test, CN=Generic-cert Generating DH parameters, 2048 bit long safe prime -...................................................................................................................................................................+.......................................................+...................................................................................................+...........................................................+........+..........................................................................+........................................................................................................+..................................+.....................................................................................................................................................+..................................................................................................................................+........................................+...................................................................................................................................................................................................................................................................................................................................................................................................+......................................................+................................................................................................................................................................................+...............................................................................................+.........................................................................+.....................................................................................................................................................................+..................+........+................................................................+...................................................................................................................................................................................................................................................+....................................+..................+..............................................................................................+......................................................................................................................................................+..........................................................................+..........+.......................................................................................................................................................................................................................................................................................................+............................................................+............................................................+...............................................................+.................................................................................................................................+......................+...........................................................................................................................................................................................................................................................................................+..............+..............................................+...................................................................................................................+.............................................................+..........................+..........................................................................................................................................+.........................................+............................................................................................................................................................................................................................................................................................................................................................+.............+................+.............................................+.................................................................................................+...........+.................................................................+...............................+.................................................................................+.......................................................................................................+.........................................................................................................................................................................................................................................................................+........................................................................................+.................................................................................................................................+.........................................................+..........................+................................................................................................................................+...........................+.............+..................................................................................................+.........................+...........+................................................+.............................................................................................................................................................................................................................+.........................................................................................................................................................................................................................................................................................................+.................++*++*++*++*++*++*++*++*++*++*++*++*++*++*++*++*++*++*++*++*++*++*++*++*++*++*++*++*++*++*++*++*++*++*++*++*++*++*++*++*++*++*++*++*++*++*++*++*++*++*++*++*++*++*++*++*++*++*++*++*++*++*++*++* +......................................................................................+........................................................................................................................+................................................................................................................................................................................................................................................................................................................+.....................................................+........................................................................................................................................................................+.............................................................................+..+................................................................+......................................................................................................................................................+..........................................................................................................................................................................................+............................................................................+............................................................+..................................................................................+..............................................................................................................................+....+.....+.................+....................................................................+......................................................................................................................+.....+..............................+..............................................+....................................................................................................................................................................................+.............................................................................+....+..................................................................................+........................................................................................................................+.........................................................................+................................................................+........................................................................................................................................+...............+..............................+.......+...............................................................................................................................................................................................................................................+.............................................................................................................................................................+................................+..........................................................................................................+................................................................................+.....................................................................................................................................+..................................++*++*++*++*++*++*++*++*++*++*++*++*++*++*++*++*++*++*++*++*++*++*++*++*++*++*++*++*++*++*++*++*++*++*++*++*++*++*++*++*++*++*++*++*++*++*++*++*++*++*++*++*++*++*++*++*++*++*++*++*++*++*++*++* # Avoid race conditions in upstream testsuite ./runtest --clients 1 --verbose --dump-logs --tls || true Cleanup: may take some time... OK Starting test server at port 21079 -[ready]: 1790455 -Testing unit/oom-score-adj -=== (oom-score-adj external:skip) Starting server 127.0.0.1:21111 ok -[ok]: CONFIG SET oom-score-adj works as expected (83 ms) -[ok]: CONFIG SET oom-score-adj handles configuration failures (15 ms) -[ok]: CONFIG SET oom-score-adj-values doesn't touch proc when disabled (8 ms) -[ok]: CONFIG SET oom score restored on disable (2 ms) -[ok]: CONFIG SET oom score relative and absolute (4 ms) -[ok]: CONFIG SET out-of-range oom score (2 ms) -[1/124 done]: unit/oom-score-adj (0 seconds) -Testing unit/client-eviction -=== () Starting server 127.0.0.1:21113 ok -[ok]: client evicted due to large argv (92 ms) -[ok]: client evicted due to large query buf (69 ms) -[ok]: client evicted due to percentage of maxmemory (49 ms) -[ok]: client evicted due to large multi buf (223 ms) -[ok]: client evicted due to watched key list (680 ms) -[ok]: client evicted due to pubsub subscriptions (2662 ms) -[ok]: client evicted due to tracking redirection (136 ms) -[ok]: client evicted due to client tracking prefixes (2472 ms) -[ok]: client evicted due to output buf (63 ms) -[ok]: client no-evict on (119 ms) -[ok]: client no-evict off (74 ms) -=== () Starting server 127.0.0.1:21115 ok -[ok]: avoid client eviction when client is freed by output buffer limit (317 ms) -=== () Starting server 127.0.0.1:21117 ok -[ok]: decrease maxmemory-clients causes client eviction (699 ms) -=== () Starting server 127.0.0.1:21119 ok -[ok]: evict clients only until below limit (380 ms) -=== () Starting server 127.0.0.1:21121 ok -[ok]: evict clients in right order (large to small) (368 ms) -=== () Starting server 127.0.0.1:21123 ok -[ok]: client total memory grows during client no-evict (219 ms) -[ok]: client total memory grows during maxmemory-clients disabled (189 ms) -[2/124 done]: unit/client-eviction (11 seconds) -Testing unit/introspection-2 -=== (introspection) Starting server 127.0.0.1:21125 ok -[ok]: The microsecond part of the TIME command will not overflow (3 ms) -[ok]: TTL, TYPE and EXISTS do not alter the last access time of a key (3009 ms) -[ok]: TOUCH alters the last access time of a key (3031 ms) -[ok]: TOUCH alters the last access time of a key in no-touch mode (3043 ms) -[ok]: Operations in no-touch mode do not alter the last access time of a key (1148 ms) -[ok]: TOUCH returns the number of existing keys specified (2 ms) -[ok]: command stats for GEOADD (3 ms) -[ok]: errors stats for GEOADD (3 ms) -[ok]: command stats for EXPIRE (2 ms) -[ok]: command stats for BRPOP (4 ms) -[ok]: command stats for MULTI (7 ms) -[ok]: command stats for scripts (7 ms) -[ok]: COMMAND COUNT get total number of commands (1 ms) -[ok]: COMMAND GETKEYS GET (1 ms) -[ok]: COMMAND GETKEYSANDFLAGS (1 ms) -[ok]: COMMAND GETKEYS MEMORY USAGE (6 ms) -[ok]: COMMAND GETKEYS XGROUP (1 ms) -[ok]: COMMAND GETKEYS EVAL with keys (1 ms) -[ok]: COMMAND GETKEYS EVAL without keys (0 ms) -[ok]: COMMAND GETKEYS LCS (0 ms) -[ok]: COMMAND GETKEYS MORE THAN 256 KEYS (6 ms) -[ok]: COMMAND LIST syntax error (2 ms) -[ok]: COMMAND LIST WITHOUT FILTERBY (7 ms) -[ok]: COMMAND LIST FILTERBY ACLCAT against non existing category (2 ms) -[ok]: COMMAND LIST FILTERBY ACLCAT - list all commands/subcommands (1 ms) -[ok]: COMMAND LIST FILTERBY PATTERN - list all commands/subcommands (3 ms) -[ok]: COMMAND LIST FILTERBY MODULE against non existing module (0 ms) -[ok]: COMMAND INFO of invalid subcommands (0 ms) -[ok]: SET command will not be marked with movablekeys (1 ms) -[ok]: GET command will not be marked with movablekeys (1 ms) -[ok]: MSET command will not be marked with movablekeys (0 ms) -[ok]: BITFIELD command will not be marked with movablekeys (1 ms) -[ok]: LMOVE command will not be marked with movablekeys (1 ms) -[ok]: LPOP command will not be marked with movablekeys (1 ms) -[ok]: BLPOP command will not be marked with movablekeys (1 ms) -[ok]: PING command will not be marked with movablekeys (0 ms) -[ok]: MEMORY command will not be marked with movablekeys (1 ms) -[ok]: MEMORY|USAGE command will not be marked with movablekeys (0 ms) -[ok]: RENAME command will not be marked with movablekeys (1 ms) -[ok]: GEORADIUS_RO command will not be marked with movablekeys (0 ms) -[ok]: ZUNIONSTORE command is marked with movablekeys (1 ms) -[ok]: XREAD command is marked with movablekeys (1 ms) -[ok]: EVAL command is marked with movablekeys (1 ms) -[ok]: SORT command is marked with movablekeys (1 ms) -[ok]: SORT_RO command is marked with movablekeys (1 ms) -[ok]: MIGRATE command is marked with movablekeys (1 ms) -[ok]: GEORADIUS command is marked with movablekeys (1 ms) -[3/124 done]: unit/introspection-2 (11 seconds) -Testing unit/querybuf -=== (querybuf slow) Starting server 127.0.0.1:21127 ok -[ok]: query buffer resized correctly (2557 ms) -[ok]: query buffer resized correctly when not idle (60 ms) -[ok]: query buffer resized correctly with fat argv (2947 ms) -[4/124 done]: unit/querybuf (6 seconds) -Testing unit/wait -=== (wait network external:skip) Starting server 127.0.0.1:21129 ok -=== () Starting server 127.0.0.1:21131 ok -[ok]: Setup slave (110 ms) -[ok]: WAIT out of range timeout (milliseconds) (24 ms) -[ok]: WAIT should acknowledge 1 additional copy of the data (28 ms) -[ok]: WAIT should not acknowledge 2 additional copies of the data (1045 ms) -[ok]: WAIT should not acknowledge 1 additional copy if slave is blocked (1133 ms) -[ok]: WAIT implicitly blocks on client pause since ACKs aren't sent (1132 ms) -[ok]: WAIT replica multiple clients unblock - reuse last result (119 ms) -=== () Starting server 127.0.0.1:21133 ok -[ok]: Setup a new replica (118 ms) -[ok]: WAIT in script will work (841 ms) -=== () Starting server 127.0.0.1:21135 ok -[ok]: WAITAOF local copy before fsync (158 ms) -[ok]: WAITAOF local copy everysec (0 ms) -[ok]: WAITAOF local copy with appendfsync always (19 ms) -[ok]: WAITAOF local wait and then stop aof (28 ms) -[ok]: WAITAOF local on server with aof disabled (1 ms) -[ok]: WAITAOF local if AOFRW was postponed (378 ms) -=== () Starting server 127.0.0.1:21137 ok -[ok]: WAITAOF on demoted master gets unblocked with an error (20 ms) -[ok]: WAITAOF replica copy before fsync (100 ms) -[ok]: WAITAOF replica copy everysec (664 ms) - -Waiting for background AOF rewrite to finish... -Waiting for background AOF rewrite to finish... [ok]: WAITAOF replica copy everysec with AOFRW (1038 ms) - -Waiting for background AOF rewrite to finish... -Waiting for background AOF rewrite to finish... -Waiting for background AOF rewrite to finish... -Waiting for background AOF rewrite to finish... -Waiting for background AOF rewrite to finish... -Waiting for background AOF rewrite to finish... -Waiting for background AOF rewrite to finish... -Waiting for background AOF rewrite to finish... -Waiting for background AOF rewrite to finish... -Waiting for background AOF rewrite to finish... -Waiting for background AOF rewrite to finish... -Waiting for background AOF rewrite to finish... -Waiting for background AOF rewrite to finish... -Waiting for background AOF rewrite to finish... -Waiting for background AOF rewrite to finish... -Waiting for background AOF rewrite to finish... -Waiting for background AOF rewrite to finish... -Waiting for background AOF rewrite to finish... -Waiting for background AOF rewrite to finish... -Waiting for background AOF rewrite to finish... -Waiting for background AOF rewrite to finish... [ok]: WAITAOF replica copy everysec with slow AOFRW (2262 ms) - -Waiting for background AOF rewrite to finish... -Waiting for background AOF rewrite to finish... [ok]: WAITAOF replica copy everysec->always with AOFRW (172 ms) -[ok]: WAITAOF replica copy appendfsync always (27 ms) -[ok]: WAITAOF replica copy if replica is blocked (233 ms) -[ok]: WAITAOF replica multiple clients unblock - reuse last result (279 ms) -[ok]: WAITAOF on promoted replica (12 ms) -[ok]: WAITAOF master that loses a replica and backlog is dropped (2034 ms) -[ok]: WAITAOF master without backlog, wait is released when the replica finishes full-sync (567 ms) -[ok]: WAITAOF master isn't configured to do AOF (1018 ms) - -Waiting for background AOF rewrite to finish... -Waiting for background AOF rewrite to finish... [ok]: WAITAOF replica isn't configured to do AOF (126 ms) - -Waiting for background AOF rewrite to finish... -Waiting for background AOF rewrite to finish... [ok]: WAITAOF both local and replica got AOF enabled at runtime (935 ms) -[ok]: WAITAOF master sends PING after last write (1218 ms) -[ok]: WAITAOF master client didn't send any write command (1238 ms) -[ok]: WAITAOF master client didn't send any command (1218 ms) -=== () Starting server 127.0.0.1:21139 ok -=== () Starting server 127.0.0.1:21141 ok -[ok]: WAITAOF when replica switches between masters, fsync: no (5524 ms) -=== () Starting server 127.0.0.1:21143 ok -=== () Starting server 127.0.0.1:21145 ok -[ok]: WAITAOF when replica switches between masters, fsync: everysec (3104 ms) -=== () Starting server 127.0.0.1:21147 ok -=== () Starting server 127.0.0.1:21149 ok -[ok]: WAITAOF when replica switches between masters, fsync: always (2424 ms) -=== (failover external:skip) Starting server 127.0.0.1:21151 ok -=== () Starting server 127.0.0.1:21153 ok -=== () Starting server 127.0.0.1:21155 ok -[ok]: setup replication for following tests (136 ms) - -Waiting for background AOF rewrite to finish... -Waiting for background AOF rewrite to finish... [ok]: WAIT and WAITAOF replica multiple clients unblock - reuse last result (800 ms) -[5/124 done]: unit/wait (34 seconds) -Testing unit/scan -=== (scan network standalone) Starting server 127.0.0.1:21157 ok -[ok]: {standalone} SCAN basic (446 ms) -[ok]: {standalone} SCAN COUNT (304 ms) -[ok]: {standalone} SCAN MATCH (732 ms) -[ok]: {standalone} SCAN TYPE (322 ms) -[ok]: {standalone} SCAN unknown type (161 ms) -[ok]: {standalone} SCAN with expired keys (362 ms) -[ok]: {standalone} SCAN with expired keys with TYPE filter (199 ms) -[ok]: {standalone} SSCAN with encoding intset (4 ms) -[ok]: {standalone} SSCAN with encoding listpack (3 ms) -[ok]: {standalone} SSCAN with encoding hashtable (9 ms) -[ok]: {standalone} HSCAN with encoding listpack (3 ms) -[ok]: {standalone} HSCAN with encoding hashtable (148 ms) -[ok]: {standalone} ZSCAN with encoding listpack (13 ms) -[ok]: {standalone} ZSCAN with encoding skiplist (212 ms) -[ok]: {standalone} SCAN guarantees check under write load (272 ms) -[ok]: {standalone} SSCAN with integer encoded object (issue #1345) (2 ms) -[ok]: {standalone} SSCAN with PATTERN (1 ms) -[ok]: {standalone} HSCAN with PATTERN (1 ms) -[ok]: {standalone} HSCAN with NOVALUES (1 ms) -[ok]: {standalone} ZSCAN with PATTERN (0 ms) -[ok]: {standalone} ZSCAN with NOSCORES (0 ms) -[ok]: {standalone} ZSCAN scores: regression test for issue #2175 (237 ms) -[ok]: {standalone} SCAN regression test for issue #4906 (22225 ms) -[ok]: {standalone} SCAN MATCH pattern implies cluster slot (78 ms) -=== (external:skip cluster scan) Starting server 127.0.0.1:21159 ok -[ok]: {cluster} SCAN basic (387 ms) -[ok]: {cluster} SCAN COUNT (364 ms) -[ok]: {cluster} SCAN MATCH (376 ms) -[ok]: {cluster} SCAN TYPE (195 ms) -[ok]: {cluster} SCAN unknown type (112 ms) -[ok]: {cluster} SCAN with expired keys (322 ms) -[ok]: {cluster} SCAN with expired keys with TYPE filter (239 ms) -[ok]: {cluster} SSCAN with encoding intset (5 ms) -[ok]: {cluster} SSCAN with encoding listpack (3 ms) -[ok]: {cluster} SSCAN with encoding hashtable (17 ms) -[ok]: {cluster} HSCAN with encoding listpack (3 ms) -[ok]: {cluster} HSCAN with encoding hashtable (149 ms) -[ok]: {cluster} ZSCAN with encoding listpack (7 ms) -[ok]: {cluster} ZSCAN with encoding skiplist (149 ms) -[ok]: {cluster} SCAN guarantees check under write load (184 ms) -[ok]: {cluster} SSCAN with integer encoded object (issue #1345) (1 ms) -[ok]: {cluster} SSCAN with PATTERN (0 ms) -[ok]: {cluster} HSCAN with PATTERN (2 ms) -[ok]: {cluster} HSCAN with NOVALUES (4 ms) -[ok]: {cluster} ZSCAN with PATTERN (3 ms) -[ok]: {cluster} ZSCAN with NOSCORES (1 ms) -[ok]: {cluster} ZSCAN scores: regression test for issue #2175 (87 ms) -[ok]: {cluster} SCAN regression test for issue #4906 (18085 ms) -[ok]: {cluster} SCAN MATCH pattern implies cluster slot (63 ms) -[6/124 done]: unit/scan (50 seconds) -Testing unit/networking -=== () Starting server 127.0.0.1:21161 ok -[ok]: CONFIG SET port number (532 ms) -=== () Starting server 127.0.0.1:21164 ok -[ok]: CONFIG SET bind address (265 ms) -=== () Starting server 127.0.0.1:21166 ok -=== () Starting server 127.0.0.1:21168 ok -[ok]: CONFIG SET bind-source-addr (572 ms) -=== (external:skip) Starting server 127.0.0.1:21170 ok -[ok]: Default bind address configuration handling (256 ms) -[ok]: Protected mode works as expected (2 ms) -=== (external:skip) Starting server 127.0.0.1:21172 ok -[7/124 done]: unit/networking (2 seconds) -Testing unit/shutdown -=== (shutdown external:skip) Starting server 127.0.0.1:21174 ok -[ok]: Temp rdb will be deleted if we use bg_unlink when shutdown (421 ms) -=== (shutdown external:skip) Starting server 127.0.0.1:21176 ok -[ok]: SHUTDOWN ABORT can cancel SIGTERM (90 ms) -[ok]: Temp rdb will be deleted in signal handle (320 ms) -=== (shutdown external:skip) Starting server 127.0.0.1:21178 ok -[ok]: RDB save will be failed in shutdown (164 ms) -[ok]: SHUTDOWN will abort if rdb save failed on signal (120 ms) -[ok]: SHUTDOWN will abort if rdb save failed on shutdown command (23 ms) -[ok]: SHUTDOWN can proceed if shutdown command was with nosave (36 ms) -[ok]: Clean up rdb same named folder (16 ms) -=== (shutdown external:skip) Starting server 127.0.0.1:21180 ok -[ok]: SHUTDOWN SIGTERM will abort if there's an initial AOFRW - default (213 ms) -[8/124 done]: unit/shutdown (3 seconds) -Testing unit/dump -=== (dump) Starting server 127.0.0.1:21182 ok -[ok]: DUMP / RESTORE are able to serialize / unserialize a simple key (55 ms) -[ok]: RESTORE can set an arbitrary expire to the materialized key (44 ms) -[ok]: RESTORE can set an expire that overflows a 32 bit integer (47 ms) -[ok]: RESTORE can set an absolute expire (44 ms) -[ok]: RESTORE with ABSTTL in the past (48 ms) -[ok]: RESTORE can set LRU (60 ms) -[ok]: RESTORE can set LFU (60 ms) -[ok]: RESTORE returns an error of the key already exists (16 ms) -[ok]: RESTORE can overwrite an existing key with REPLACE (64 ms) -[ok]: RESTORE can detect a syntax error for unrecognized options (7 ms) -[ok]: RESTORE should not store key that are already expired, with REPLACE will propagate it as DEL or UNLINK (175 ms) -[ok]: DUMP of non existing key returns nil (8 ms) -=== (repl) Starting server 127.0.0.1:21184 ok -[ok]: MIGRATE is caching connections (552 ms) -[ok]: MIGRATE cached connections are released after some time (15013 ms) -=== (repl) Starting server 127.0.0.1:21186 ok -[ok]: MIGRATE is able to migrate a key between two instances (487 ms) -=== (repl) Starting server 127.0.0.1:21188 ok -[ok]: MIGRATE is able to copy a key between two instances (775 ms) -=== (repl) Starting server 127.0.0.1:21190 ok -[ok]: MIGRATE will not overwrite existing keys, unless REPLACE is used (972 ms) -=== (repl) Starting server 127.0.0.1:21192 ok -[ok]: MIGRATE propagates TTL correctly (1198 ms) -=== (repl) Starting server 127.0.0.1:21194 ok -[ok]: MIGRATE can correctly transfer large values (27567 ms) -=== (repl) Starting server 127.0.0.1:21196 ok -[ok]: MIGRATE can correctly transfer hashes (323 ms) -=== (repl) Starting server 127.0.0.1:21198 ok -[ok]: MIGRATE timeout actually works (933 ms) -=== (repl) Starting server 127.0.0.1:21200 ok -[ok]: MIGRATE can migrate multiple keys at once (329 ms) -[ok]: MIGRATE with multiple keys must have empty key arg (2 ms) -=== (repl) Starting server 127.0.0.1:21202 ok -[ok]: MIGRATE with multiple keys migrate just existing ones (444 ms) -=== (repl) Starting server 127.0.0.1:21204 ok -[ok]: MIGRATE with multiple keys: stress command rewriting (639 ms) -=== (repl) Starting server 127.0.0.1:21206 ok -[ok]: MIGRATE with multiple keys: delete just ack keys (973 ms) -=== (repl) Starting server 127.0.0.1:21208 ok -[ok]: MIGRATE AUTH: correct and wrong password cases (256 ms) -[9/124 done]: unit/dump (52 seconds) -Testing unit/bitfield -=== (bitops) Starting server 127.0.0.1:21210 ok -[ok]: BITFIELD signed SET and GET basics (1 ms) -[ok]: BITFIELD unsigned SET and GET basics (1 ms) -[ok]: BITFIELD signed SET and GET together (0 ms) -[ok]: BITFIELD unsigned with SET, GET and INCRBY arguments (1 ms) -[ok]: BITFIELD with only key as argument (0 ms) -[ok]: BITFIELD # form (1 ms) -[ok]: BITFIELD basic INCRBY form (1 ms) -[ok]: BITFIELD chaining of multiple commands (1 ms) -[ok]: BITFIELD unsigned overflow wrap (1 ms) -[ok]: BITFIELD unsigned overflow sat (1 ms) -[ok]: BITFIELD signed overflow wrap (1 ms) -[ok]: BITFIELD signed overflow sat (1 ms) -[ok]: BITFIELD overflow detection fuzzing (439 ms) -[ok]: BITFIELD overflow wrap fuzzing (680 ms) -[ok]: BITFIELD regression for #3221 (1 ms) -[ok]: BITFIELD regression for #3564 (4 ms) -[ok]: BITFIELD_RO with only key as argument (0 ms) -[ok]: BITFIELD_RO fails when write option is used (1 ms) -=== (repl external:skip) Starting server 127.0.0.1:21212 ok -=== () Starting server 127.0.0.1:21214 ok -[ok]: BITFIELD: setup slave (112 ms) -[ok]: BITFIELD: write on master, read on slave (14 ms) -[ok]: BITFIELD_RO with only key as argument on read-only replica (7 ms) -[ok]: BITFIELD_RO fails when write option is used on read-only replica (6 ms) -[10/124 done]: unit/bitfield (2 seconds) -Testing unit/pause -=== (pause network) Starting server 127.0.0.1:21216 ok -[ok]: Test read commands are not blocked by client pause (15 ms) -[ok]: Test old pause-all takes precedence over new pause-write (less restrictive) (316 ms) -[ok]: Test new pause time is smaller than old one, then old time preserved (129 ms) -[ok]: Test write commands are paused by RO (24 ms) -[ok]: Test special commands are paused by RO (35 ms) -[ok]: Test read/admin multi-execs are not blocked by pause RO (19 ms) -[ok]: Test write multi-execs are blocked by pause RO (17 ms) -[ok]: Test scripts are blocked by pause RO (52 ms) -[ok]: Test RO scripts are not blocked by pause RO (53 ms) -[ok]: Test read-only scripts in multi-exec are not blocked by pause RO (60 ms) -[ok]: Test write scripts in multi-exec are blocked by pause RO (53 ms) -[ok]: Test may-replicate commands are rejected in RO scripts (7 ms) -[ok]: Test multiple clients can be queued up and unblocked (55 ms) -[ok]: Test clients with syntax errors will get responses immediately (2 ms) -[ok]: Test eviction is skipped during client pause (4 ms) -[ok]: Test both active and passive expires are skipped during client pause (108 ms) -[ok]: Test that client pause starts at the end of a transaction (17 ms) -=== (needs:repl external:skip) Starting server 127.0.0.1:21218 ok -[ok]: Test when replica paused, offset would not grow (10 ms) -[ok]: Test replica offset would grow after unpause (2 ms) -[11/124 done]: unit/pause (2 seconds) -Testing unit/latency-monitor -=== (latency-monitor needs:latency) Starting server 127.0.0.1:21220 ok -[ok]: LATENCY HISTOGRAM with empty histogram (1 ms) -[ok]: LATENCY HISTOGRAM all commands (2 ms) -[ok]: LATENCY HISTOGRAM sub commands (3 ms) -[ok]: LATENCY HISTOGRAM with a subset of commands (18 ms) -[ok]: LATENCY HISTOGRAM command (6 ms) -[ok]: LATENCY HISTOGRAM with wrong command name skips the invalid one (30 ms) -[ok]: Test latency events logging (3419 ms) -LATENCY HISTORY data: -{1741562332 301} {1741562334 402} {1741562336 500} -[ok]: LATENCY HISTORY output is ok (1 ms) -LATENCY LATEST data: -{command 1741562336 500 500} -[ok]: LATENCY LATEST output is ok (2 ms) -LATENCY GRAPH data: -command - high 500 ms, low 301 ms (all time high 500 ms) --------------------------------------------------------------------------------- - # - _| - || -_|| - -420 -sss - -[ok]: LATENCY GRAPH can output the event graph (1 ms) -[ok]: LATENCY GRAPH can output the expire event graph (7 ms) -[ok]: LATENCY of expire events are correctly collected (6119 ms) -[ok]: LATENCY HISTORY / RESET with wrong event name is fine (11 ms) -[ok]: LATENCY DOCTOR produces some output (5 ms) -[ok]: LATENCY RESET is able to reset events (8 ms) -[ok]: LATENCY HELP should not have unexpected options (0 ms) -[12/124 done]: unit/latency-monitor (10 seconds) -Testing unit/replybufsize -=== (replybufsize) Starting server 127.0.0.1:21222 ok -[ok]: verify reply buffer limits (623 ms) -[13/124 done]: unit/replybufsize (1 seconds) -Testing unit/maxmemory -=== (maxmemory external:skip) Starting server 127.0.0.1:21224 ok -evicted keys: 14 -evicted clients: 0 -dbsize: 36 -[ok]: eviction due to output buffers of many MGET clients, client eviction: false (673 ms) -evicted keys: 28 -evicted clients: 0 -dbsize: 22 -[ok]: eviction due to input buffer of a dead client, client eviction: false (4431 ms) -evicted keys: 12 -evicted clients: 0 -dbsize: 38 -[ok]: eviction due to output buffers of pubsub, client eviction: false (937 ms) -evicted keys: 0 -evicted clients: 8 -dbsize: 50 -[ok]: eviction due to output buffers of many MGET clients, client eviction: true (470 ms) -evicted keys: 0 -evicted clients: 18 -dbsize: 50 -[ok]: eviction due to input buffer of a dead client, client eviction: true (4443 ms) -evicted keys: 0 -evicted clients: 6 -dbsize: 50 -[ok]: eviction due to output buffers of pubsub, client eviction: true (826 ms) -=== (maxmemory external:skip) Starting server 127.0.0.1:21226 ok -[ok]: Without maxmemory small integers are shared (8 ms) -[ok]: With maxmemory and non-LRU policy integers are still shared (1 ms) -[ok]: With maxmemory and LRU policy integers are not shared (26 ms) -[ok]: Shared integers are unshared with maxmemory and LRU policy (4 ms) -[ok]: maxmemory - is the memory limit honoured? (policy allkeys-random) (454 ms) -[ok]: maxmemory - is the memory limit honoured? (policy allkeys-lru) (1191 ms) -[ok]: maxmemory - is the memory limit honoured? (policy allkeys-lfu) (2140 ms) -[ok]: maxmemory - is the memory limit honoured? (policy volatile-lru) (2444 ms) -[ok]: maxmemory - is the memory limit honoured? (policy volatile-lfu) (3477 ms) -[ok]: maxmemory - is the memory limit honoured? (policy volatile-random) (734 ms) -[ok]: maxmemory - is the memory limit honoured? (policy volatile-ttl) (1753 ms) -[ok]: maxmemory - only allkeys-* should remove non-volatile keys (allkeys-random) (989 ms) -[ok]: maxmemory - only allkeys-* should remove non-volatile keys (allkeys-lru) (1770 ms) -[ok]: maxmemory - only allkeys-* should remove non-volatile keys (volatile-lru) (2339 ms) -[ok]: maxmemory - only allkeys-* should remove non-volatile keys (volatile-random) (2183 ms) -[ok]: maxmemory - only allkeys-* should remove non-volatile keys (volatile-ttl) (2003 ms) -[ok]: maxmemory - policy volatile-lru should only remove volatile keys. (4718 ms) -[ok]: maxmemory - policy volatile-lfu should only remove volatile keys. (1110 ms) -[ok]: maxmemory - policy volatile-random should only remove volatile keys. (1120 ms) -[ok]: maxmemory - policy volatile-ttl should only remove volatile keys. (780 ms) -=== (maxmemory external:skip) Starting server 127.0.0.1:21228 ok -=== () Starting server 127.0.0.1:21230 ok -[ok]: slave buffer are counted correctly (73200 ms) -=== (maxmemory external:skip) Starting server 127.0.0.1:21232 ok -=== () Starting server 127.0.0.1:21234 ok -[ok]: replica buffer don't induce eviction (19674 ms) -=== (maxmemory external:skip) Starting server 127.0.0.1:21236 ok -[ok]: Don't rehash if used memory exceeds maxmemory after rehash (596 ms) -=== (maxmemory external:skip io-threads:skip) Starting server 127.0.0.1:21238 ok -evicted: 30 -evicted: 30 -[ok]: client tracking don't cause eviction feedback loop (2433 ms) -=== (maxmemory external:skip) Starting server 127.0.0.1:21240 ok -[ok]: propagation with eviction (139 ms) -=== (maxmemory external:skip) Starting server 127.0.0.1:21242 ok -[ok]: propagation with eviction in MULTI (112 ms) -=== (maxmemory external:skip) Starting server 127.0.0.1:21244 ok -[ok]: lru/lfu value of the key just added (8 ms) -[14/124 done]: unit/maxmemory (142 seconds) -Testing unit/auth -=== (auth external:skip) Starting server 127.0.0.1:21246 ok -[ok]: AUTH fails if there is no password configured server side (7 ms) -[ok]: Arity check for auth command (4 ms) -=== (auth external:skip) Starting server 127.0.0.1:21248 ok -[ok]: AUTH fails when a wrong password is given (21 ms) -[ok]: Arbitrary command gives an error when AUTH is required (4 ms) -[ok]: AUTH succeeds when the right password is given (6 ms) -[ok]: Once AUTH succeeded we can actually send commands to the server (10 ms) -[ok]: For unauthenticated clients multibulk and bulk length are limited (66 ms) -=== (auth_binary_password external:skip) Starting server 127.0.0.1:21250 ok -[ok]: AUTH fails when binary password is wrong (1 ms) -[ok]: AUTH succeeds when binary password is correct (0 ms) -=== (primaryauth) Starting server 127.0.0.1:21252 ok -[ok]: primaryauth test with binary password dualchannel = yes (1027 ms) -=== (auth_binary_password external:skip) Starting server 127.0.0.1:21254 ok -[ok]: AUTH fails when binary password is wrong (1 ms) -[ok]: AUTH succeeds when binary password is correct (2 ms) -=== (primaryauth) Starting server 127.0.0.1:21256 ok -[ok]: primaryauth test with binary password dualchannel = no (989 ms) -[15/124 done]: unit/auth (4 seconds) -Testing unit/limits -=== (limits network external:skip) Starting server 127.0.0.1:21258 ok -[ok]: Check if maxclients works refusing connections (1205 ms) -[16/124 done]: unit/limits (2 seconds) -Testing unit/hyperloglog -=== (hll) Starting server 127.0.0.1:21260 ok -[ok]: HyperLogLog self test passes (1320 ms) -[ok]: PFADD without arguments creates an HLL value (1 ms) -[ok]: Approximated cardinality after creation is zero (0 ms) -[ok]: PFADD returns 1 when at least 1 reg was modified (1 ms) -[ok]: PFADD returns 0 when no reg was modified (0 ms) -[ok]: PFADD works with empty string (regression) (0 ms) -[ok]: PFCOUNT returns approximated cardinality of set (1 ms) -[ok]: HyperLogLogs are promote from sparse to dense (827 ms) -[ok]: Change hll-sparse-max-bytes (1 ms) -[ok]: Hyperloglog promote to dense well in different hll-sparse-max-bytes (112 ms) -[ok]: HyperLogLog sparse encoding stress test (2077 ms) -[ok]: Corrupted sparse HyperLogLogs are detected: Additional at tail (1 ms) -[ok]: Corrupted sparse HyperLogLogs are detected: Broken magic (1 ms) -[ok]: Corrupted sparse HyperLogLogs are detected: Invalid encoding (0 ms) -[ok]: Corrupted dense HyperLogLogs are detected: Wrong length (1 ms) -[ok]: Fuzzing dense/sparse encoding: Server should always detect errors (48021 ms) -[ok]: PFADD, PFCOUNT, PFMERGE type checking works (1 ms) -[ok]: PFMERGE results on the cardinality of union of sets (2 ms) -[ok]: PFMERGE on missing source keys will create an empty destkey (1 ms) -[ok]: PFMERGE with one empty input key, create an empty destkey (0 ms) -[ok]: PFMERGE with one non-empty input key, dest key is actually one of the source keys (1 ms) -[ok]: PFCOUNT multiple-keys merge returns cardinality of union #1 (9684 ms) -[ok]: PFCOUNT multiple-keys merge returns cardinality of union #2 (4370 ms) -[ok]: PFDEBUG GETREG returns the HyperLogLog raw registers (1064 ms) -[ok]: PFADD / PFCOUNT cache invalidation works (18 ms) -[17/124 done]: unit/hyperloglog (67 seconds) -Testing unit/info-command -=== (info and its relative command) Starting server 127.0.0.1:21262 ok -[ok]: info command with at most one sub command (3 ms) -[ok]: info command with one sub-section (1 ms) -[ok]: info command with multiple sub-sections (2 ms) -[18/124 done]: unit/info-command (1 seconds) -Testing unit/slowlog -=== (slowlog) Starting server 127.0.0.1:21264 ok -[ok]: SLOWLOG - check that it starts with an empty log (1 ms) -[ok]: SLOWLOG - only logs commands taking more time than specified (202 ms) -[ok]: SLOWLOG - zero max length is correctly handled (18 ms) -[ok]: SLOWLOG - max entries is correctly handled (27 ms) -[ok]: SLOWLOG - GET optional argument to limit output len works (2 ms) -[ok]: SLOWLOG - RESET subcommand works (2 ms) -[ok]: SLOWLOG - logged entry sanity check (202 ms) -[ok]: SLOWLOG - Certain commands are omitted that contain sensitive information (16 ms) -[ok]: SLOWLOG - Some commands can redact sensitive fields (6 ms) -[ok]: SLOWLOG - Rewritten commands are logged as their original command (23 ms) -[ok]: SLOWLOG - commands with too many arguments are trimmed (2 ms) -[ok]: SLOWLOG - too long arguments are trimmed (1 ms) -[ok]: SLOWLOG - EXEC is not logged, just executed commands (203 ms) -[ok]: SLOWLOG - can clean older entries (202 ms) -[ok]: SLOWLOG - can be disabled (404 ms) -[ok]: SLOWLOG - count must be >= -1 (1 ms) -[ok]: SLOWLOG - get all slow logs (5 ms) -[ok]: SLOWLOG - blocking command is reported only after unblocked (19 ms) -[ok]: SLOWLOG - the commands in script are recorded normally - is_eval: 0 (2 ms) -[ok]: SLOWLOG - the commands in script are recorded normally - is_eval: 1 (1 ms) -[19/124 done]: unit/slowlog (1 seconds) -Testing unit/info -=== (info external:skip) Starting server 127.0.0.1:21266 ok -=== () Starting server 127.0.0.1:21268 ok -[ok]: latencystats: disable/enable (2 ms) -[ok]: latencystats: configure percentiles (3 ms) -[ok]: latencystats: bad configure percentiles (4 ms) -[ok]: latencystats: blocking commands (21 ms) -[ok]: latencystats: subcommands (2 ms) -[ok]: latencystats: measure latency (53 ms) -[ok]: errorstats: failed call authentication error (3 ms) -[ok]: errorstats: failed call within MULTI/EXEC (3 ms) -[ok]: errorstats: failed call within LUA (2 ms) -[ok]: errorstats: failed call NOSCRIPT error (2 ms) -[ok]: errorstats: failed call NOGROUP error (2 ms) -[ok]: errorstats: rejected call unknown command (1 ms) -[ok]: errorstats: rejected call within MULTI/EXEC (10 ms) -[ok]: errorstats: rejected call due to wrong arity (3 ms) -[ok]: errorstats: rejected call by OOM error (22 ms) -[ok]: errorstats: rejected call by authorization error (3 ms) -[ok]: errorstats: blocking commands (18 ms) -[ok]: errorstats: limit errors will not increase indefinitely (293 ms) -eventloop metrics cycle1: 1118, cycle2: 1120 -eventloop metrics el_sum1: 80673, el_sum2: 81350 -eventloop metrics cmd_sum1: 19457, cmd_sum2: 19557 -[ok]: stats: eventloop metrics (117 ms) -instantaneous metrics instantaneous_eventloop_cycles_per_sec: 8 -instantaneous metrics instantaneous_eventloop_duration_usec: 732 -[ok]: stats: instantaneous metrics (1610 ms) -[ok]: stats: debug metrics (132 ms) -[ok]: stats: client input and output buffer limit disconnections (175 ms) -[ok]: clients: pubsub clients (131 ms) -[ok]: clients: watching clients (324 ms) -=== (info external:skip) Starting server 127.0.0.1:21270 ok -[ok]: memory: database and pubsub overhead and rehashing dict count (6 ms) -[20/124 done]: unit/info (4 seconds) +[ready]: 30294 +Testing unit/tracking +=== (tracking network logreqres:skip) Starting server 127.0.0.1:21111 ok +[ok]: Clients are able to enable tracking and redirect it (0 ms) +[ok]: The other connection is able to get invalidations (2 ms) +[ok]: The client is now able to disable tracking (0 ms) +[ok]: Clients can enable the BCAST mode with the empty prefix (1 ms) +[ok]: The connection gets invalidation messages about all the keys (0 ms) +[ok]: Clients can enable the BCAST mode with prefixes (2 ms) +[ok]: Adding prefixes to BCAST mode works (0 ms) +[ok]: Tracking NOLOOP mode in standard mode works (2 ms) +[ok]: Tracking NOLOOP mode in BCAST mode works (1 ms) +[ok]: Tracking gets notification of expired keys (1004 ms) +[ok]: Tracking gets notification of lazy expired keys (103 ms) +[ok]: HELLO 3 reply is correct (1 ms) +[ok]: HELLO without protover (2 ms) +[ok]: RESP3 based basic invalidation (1 ms) +[ok]: RESP3 tracking redirection (2 ms) +[ok]: Invalidations of previous keys can be redirected after switching to RESP3 (1 ms) +[ok]: Invalidations of new keys can be redirected after switching to RESP3 (2 ms) +[ok]: Invalid keys should not be tracked for scripts in NOLOOP mode (2 ms) +[ok]: Tracking only occurs for scripts when a command calls a read-only command (8 ms) +[ok]: RESP3 Client gets tracking-redir-broken push message after cached key changed when rediretion client is terminated (16 ms) +[ok]: Different clients can redirect to the same connection (2 ms) +[ok]: Different clients using different protocols can track the same key (2 ms) +[ok]: No invalidation message when using OPTIN option (2 ms) +[ok]: Invalidation message sent when using OPTIN option with CLIENT CACHING yes (2 ms) +[ok]: Invalidation message sent when using OPTOUT option (1 ms) +[ok]: No invalidation message when using OPTOUT option with CLIENT CACHING no (8 ms) +[ok]: Able to redirect to a RESP3 client (2 ms) +[ok]: After switching from normal tracking to BCAST mode, no invalidation message is produced for pre-BCAST keys (2 ms) +[ok]: BCAST with prefix collisions throw errors (15 ms) +[ok]: hdel deliver invalidate message after response in the same connection (2 ms) +[ok]: Tracking invalidation message is not interleaved with multiple keys response (114 ms) +[ok]: Tracking invalidation message is not interleaved with transaction response (2 ms) +[ok]: Tracking invalidation message of eviction keys should be before response (6 ms) +[ok]: Unblocked BLMOVE gets notification after response (10 ms) +[ok]: Tracking gets notification on tracking table key eviction (2496 ms) +[ok]: Invalidation message received for flushall (172 ms) +[ok]: Invalidation message received for flushdb (38 ms) +[ok]: Test ASYNC flushall (31 ms) +[ok]: flushdb tracking invalidation message is not interleaved with transaction response (31 ms) +[ok]: Server is able to evacuate enough keys when num of keys surpasses limit by more than defined initial effort (263 ms) +[ok]: Tracking info is correct (29 ms) +[ok]: CLIENT GETREDIR provides correct client id (1 ms) +[ok]: CLIENT TRACKINGINFO provides reasonable results when tracking off (0 ms) +[ok]: CLIENT TRACKINGINFO provides reasonable results when tracking on (1 ms) +[ok]: CLIENT TRACKINGINFO provides reasonable results when tracking on with options (1 ms) +[ok]: CLIENT TRACKINGINFO provides reasonable results when tracking optin (1 ms) +[ok]: CLIENT TRACKINGINFO provides reasonable results when tracking optout (2 ms) +[ok]: CLIENT TRACKINGINFO provides reasonable results when tracking bcast mode (1 ms) +[ok]: CLIENT TRACKINGINFO provides reasonable results when tracking redir broken (26 ms) +[ok]: Regression test for #11715 (27 ms) +[ok]: RESP3 based basic invalidation with client reply off (35 ms) +[ok]: RESP2 based basic invalidation with client reply off (73 ms) +[ok]: RESP3 based basic redirect invalidation with client reply off (44 ms) +[ok]: RESP3 based basic tracking-redir-broken with client reply off (162 ms) +=== (tracking network) Starting server 127.0.0.1:21113 ok +[ok]: Coverage: Basic CLIENT CACHING (11 ms) +[ok]: Coverage: Basic CLIENT REPLY (0 ms) +[ok]: Coverage: Basic CLIENT TRACKINGINFO (0 ms) +[ok]: Coverage: Basic CLIENT GETREDIR (0 ms) +[1/124 done]: unit/tracking (6 seconds) Testing unit/lazyfree -=== (lazyfree) Starting server 127.0.0.1:21272 ok -[ok]: UNLINK can reclaim memory in background (502 ms) -[ok]: FLUSHDB ASYNC can reclaim memory in background (463 ms) -[ok]: lazy free a stream with all types of metadata (273 ms) +=== (lazyfree) Starting server 127.0.0.1:21115 ok +[ok]: UNLINK can reclaim memory in background (371 ms) +[ok]: FLUSHDB ASYNC can reclaim memory in background (345 ms) +[ok]: lazy free a stream with all types of metadata (299 ms) [ok]: lazy free a stream with deleted cgroup (3 ms) -[21/124 done]: unit/lazyfree (2 seconds) -Testing unit/acl-v2 -=== (acl external:skip) Starting server 127.0.0.1:21274 ok -[ok]: Test basic multiple selectors (47 ms) -[ok]: Test ACL selectors by default have no permissions (12 ms) -[ok]: Test deleting selectors (11 ms) -[ok]: Test selector syntax error reports the error in the selector context (16 ms) -[ok]: Test flexible selector definition (8 ms) -[ok]: Test separate read permission (2 ms) -[ok]: Test separate write permission (1 ms) -[ok]: Test separate read and write permissions (17 ms) -[ok]: Validate read and write permissions format - empty permission (7 ms) -[ok]: Validate read and write permissions format - empty selector (7 ms) -[ok]: Validate read and write permissions format - empty pattern (14 ms) -[ok]: Validate read and write permissions format - no pattern (2 ms) -[ok]: Test separate read and write permissions on different selectors are not additive (21 ms) -[ok]: Test SET with separate read permission (14 ms) -[ok]: Test SET with separate write permission (11 ms) -[ok]: Test SET with read and write permissions (3 ms) -[ok]: Test BITFIELD with separate read permission (9 ms) -[ok]: Test BITFIELD with separate write permission (2 ms) -[ok]: Test BITFIELD with read and write permissions (7 ms) -[ok]: Test ACL log correctly identifies the relevant item when selectors are used (7 ms) -[ok]: Test ACL GETUSER response information (3 ms) -[ok]: Test ACL list idempotency (2 ms) -[ok]: Test R+W is the same as all permissions (3 ms) -[ok]: Test basic dry run functionality (1 ms) -[ok]: Test various commands for command permissions (1 ms) -[ok]: Test various odd commands for key permissions (9 ms) -[ok]: Existence test commands are not marked as access (2 ms) -[ok]: Intersection cardinaltiy commands are access commands (3 ms) -[ok]: Test general keyspace commands require some type of permission to execute (3 ms) -[ok]: Cardinality commands require some type of permission to execute (5 ms) -[ok]: Test sharded channel permissions (2 ms) -[ok]: Test sort with ACL permissions (4 ms) -[ok]: Test DRYRUN with wrong number of arguments (1 ms) -=== (external:skip) Starting server 127.0.0.1:21276 ok -[ok]: Test behavior of loading ACLs (3 ms) -[22/124 done]: unit/acl-v2 (1 seconds) -Testing unit/violations -=== () Starting server 127.0.0.1:21278 ok -[ignore]: XADD one huge field: large memory flag not provided -=== () Starting server 127.0.0.1:21280 ok -[ignore]: XADD one huge field - 1: large memory flag not provided -=== () Starting server 127.0.0.1:21282 ok -[ignore]: several XADD big fields: large memory flag not provided -=== () Starting server 127.0.0.1:21284 ok -[ignore]: single XADD big fields: large memory flag not provided -=== () Starting server 127.0.0.1:21286 ok -[ignore]: hash with many big fields: large memory flag not provided -=== () Starting server 127.0.0.1:21288 ok -[ignore]: hash with one huge field: large memory flag not provided -=== () Starting server 127.0.0.1:21290 ok -[ok]: SORT adds integer field to list (4 ms) -[23/124 done]: unit/violations (2 seconds) +[2/124 done]: unit/lazyfree (1 seconds) +Testing unit/oom-score-adj +=== (oom-score-adj external:skip) Starting server 127.0.0.1:21117 ok +[ok]: CONFIG SET oom-score-adj works as expected (168 ms) +[ok]: CONFIG SET oom-score-adj handles configuration failures (3 ms) +[ok]: CONFIG SET oom-score-adj-values doesn't touch proc when disabled (0 ms) +[ok]: CONFIG SET oom score restored on disable (0 ms) +[ok]: CONFIG SET oom score relative and absolute (1 ms) +[ok]: CONFIG SET out-of-range oom score (0 ms) +[3/124 done]: unit/oom-score-adj (1 seconds) Testing unit/bitops -=== (bitops) Starting server 127.0.0.1:21292 ok -[ok]: BITCOUNT against wrong type (2 ms) -[ok]: BITCOUNT returns 0 against non existing key (3 ms) -[ok]: BITCOUNT returns 0 with out of range indexes (3 ms) +=== (bitops) Starting server 127.0.0.1:21119 ok +[ok]: BITCOUNT against wrong type (3 ms) +[ok]: BITCOUNT returns 0 against non existing key (1 ms) +[ok]: BITCOUNT returns 0 with out of range indexes (1 ms) [ok]: BITCOUNT returns 0 with negative indexes where start > end (2 ms) [ok]: BITCOUNT against test vector #1 (1 ms) [ok]: BITCOUNT against test vector #2 (0 ms) -[ok]: BITCOUNT against test vector #3 (1 ms) +[ok]: BITCOUNT against test vector #3 (0 ms) [ok]: BITCOUNT against test vector #4 (1 ms) [ok]: BITCOUNT against test vector #5 (1 ms) -[ok]: BITCOUNT fuzzing without start/end (401 ms) -[ok]: BITCOUNT fuzzing with start/end (835 ms) -[ok]: BITCOUNT with just start (1 ms) -[ok]: BITCOUNT with start, end (3 ms) -[ok]: BITCOUNT with illegal arguments (1 ms) +[ok]: BITCOUNT fuzzing without start/end (391 ms) +[ok]: BITCOUNT fuzzing with start/end (805 ms) +[ok]: BITCOUNT with just start (2 ms) +[ok]: BITCOUNT with start, end (4 ms) +[ok]: BITCOUNT with illegal arguments (2 ms) [ok]: BITCOUNT against non-integer value (1 ms) [ok]: BITCOUNT regression test for github issue #582 (1 ms) [ok]: BITCOUNT misaligned prefix (1 ms) -[ok]: BITCOUNT misaligned prefix + full words + remainder (0 ms) +[ok]: BITCOUNT misaligned prefix + full words + remainder (1 ms) [ok]: BITOP NOT (empty string) (1 ms) [ok]: BITOP NOT (known string) (1 ms) [ok]: BITOP where dest and target are the same key (0 ms) -[ok]: BITOP AND|OR|XOR don't change the string with single input key (1 ms) -[ok]: BITOP missing key is considered a stream of zero (1 ms) +[ok]: BITOP AND|OR|XOR don't change the string with single input key (2 ms) +[ok]: BITOP missing key is considered a stream of zero (2 ms) [ok]: BITOP shorter keys are zero-padded to the key with max length (1 ms) -[ok]: BITOP and fuzzing (405 ms) -[ok]: BITOP or fuzzing (275 ms) -[ok]: BITOP xor fuzzing (325 ms) -[ok]: BITOP NOT fuzzing (47 ms) +[ok]: BITOP and fuzzing (262 ms) +[ok]: BITOP or fuzzing (389 ms) +[ok]: BITOP xor fuzzing (233 ms) +[ok]: BITOP NOT fuzzing (41 ms) [ok]: BITOP with integer encoded source objects (1 ms) [ok]: BITOP with non string source key (1 ms) -[ok]: BITOP with empty string after non empty string (issue #529) (1 ms) -[ok]: BITPOS against wrong type (1 ms) +[ok]: BITOP with empty string after non empty string (issue #529) (0 ms) +[ok]: BITPOS against wrong type (2 ms) [ok]: BITPOS will illegal arguments (1 ms) -[ok]: BITPOS against non-integer value (1 ms) +[ok]: BITPOS against non-integer value (3 ms) [ok]: BITPOS bit=0 with empty key returns 0 (1 ms) -[ok]: BITPOS bit=1 with empty key returns -1 (1 ms) +[ok]: BITPOS bit=1 with empty key returns -1 (0 ms) [ok]: BITPOS bit=0 with string less than 1 word works (0 ms) [ok]: BITPOS bit=1 with string less than 1 word works (1 ms) [ok]: BITPOS bit=0 starting at unaligned address (1 ms) -[ok]: BITPOS bit=1 starting at unaligned address (0 ms) +[ok]: BITPOS bit=1 starting at unaligned address (1 ms) [ok]: BITPOS bit=0 unaligned+full word+reminder (5 ms) -[ok]: BITPOS bit=1 unaligned+full word+reminder (4 ms) -[ok]: BITPOS bit=1 returns -1 if string is all 0 bits (10 ms) +[ok]: BITPOS bit=1 unaligned+full word+reminder (5 ms) +[ok]: BITPOS bit=1 returns -1 if string is all 0 bits (12 ms) [ok]: BITPOS bit=0 works with intervals (2 ms) [ok]: BITPOS bit=1 works with intervals (2 ms) -[ok]: BITPOS bit=0 changes behavior if end is given (0 ms) -[ok]: SETBIT/BITFIELD only increase dirty when the value changed (7 ms) -[ok]: BITPOS bit=1 fuzzy testing using SETBIT (529 ms) -[ok]: BITPOS bit=0 fuzzy testing using SETBIT (582 ms) -[ok]: BITPOS/BITCOUNT fuzzy testing using SETBIT (483 ms) -=== (bitops) Starting server 127.0.0.1:21294 ok +[ok]: BITPOS bit=0 changes behavior if end is given (1 ms) +[ok]: SETBIT/BITFIELD only increase dirty when the value changed (8 ms) +[ok]: BITPOS bit=1 fuzzy testing using SETBIT (513 ms) +[ok]: BITPOS bit=0 fuzzy testing using SETBIT (518 ms) +[ok]: BITPOS/BITCOUNT fuzzy testing using SETBIT (1660 ms) +=== (bitops) Starting server 127.0.0.1:21121 ok [ignore]: BIT pos larger than UINT_MAX: large memory flag not provided [ignore]: SETBIT values larger than UINT32_MAX and lzf_compress/lzf_decompress correctly: large memory flag not provided -[24/124 done]: unit/bitops (5 seconds) -Testing unit/tracking -=== (tracking network logreqres:skip) Starting server 127.0.0.1:21296 ok -[ok]: Clients are able to enable tracking and redirect it (0 ms) -[ok]: The other connection is able to get invalidations (4 ms) -[ok]: The client is now able to disable tracking (0 ms) -[ok]: Clients can enable the BCAST mode with the empty prefix (1 ms) -[ok]: The connection gets invalidation messages about all the keys (0 ms) -[ok]: Clients can enable the BCAST mode with prefixes (5 ms) -[ok]: Adding prefixes to BCAST mode works (3 ms) -[ok]: Tracking NOLOOP mode in standard mode works (1 ms) -[ok]: Tracking NOLOOP mode in BCAST mode works (1 ms) -[ok]: Tracking gets notification of expired keys (1008 ms) -[ok]: Tracking gets notification of lazy expired keys (103 ms) -[ok]: HELLO 3 reply is correct (1 ms) -[ok]: HELLO without protover (3 ms) -[ok]: RESP3 based basic invalidation (2 ms) -[ok]: RESP3 tracking redirection (2 ms) -[ok]: Invalidations of previous keys can be redirected after switching to RESP3 (7 ms) -[ok]: Invalidations of new keys can be redirected after switching to RESP3 (4 ms) -[ok]: Invalid keys should not be tracked for scripts in NOLOOP mode (2 ms) -[ok]: Tracking only occurs for scripts when a command calls a read-only command (17 ms) -[ok]: RESP3 Client gets tracking-redir-broken push message after cached key changed when rediretion client is terminated (20 ms) -[ok]: Different clients can redirect to the same connection (2 ms) -[ok]: Different clients using different protocols can track the same key (2 ms) -[ok]: No invalidation message when using OPTIN option (2 ms) -[ok]: Invalidation message sent when using OPTIN option with CLIENT CACHING yes (2 ms) -[ok]: Invalidation message sent when using OPTOUT option (2 ms) -[ok]: No invalidation message when using OPTOUT option with CLIENT CACHING no (2 ms) -[ok]: Able to redirect to a RESP3 client (2 ms) -[ok]: After switching from normal tracking to BCAST mode, no invalidation message is produced for pre-BCAST keys (4 ms) -[ok]: BCAST with prefix collisions throw errors (22 ms) -[ok]: hdel deliver invalidate message after response in the same connection (7 ms) -[ok]: Tracking invalidation message is not interleaved with multiple keys response (116 ms) -[ok]: Tracking invalidation message is not interleaved with transaction response (3 ms) -[ok]: Tracking invalidation message of eviction keys should be before response (7 ms) -[ok]: Unblocked BLMOVE gets notification after response (3 ms) -[ok]: Tracking gets notification on tracking table key eviction (65 ms) -[ok]: Invalidation message received for flushall (43 ms) -[ok]: Invalidation message received for flushdb (37 ms) -[ok]: Test ASYNC flushall (28 ms) -[ok]: flushdb tracking invalidation message is not interleaved with transaction response (27 ms) -[ok]: Server is able to evacuate enough keys when num of keys surpasses limit by more than defined initial effort (283 ms) -[ok]: Tracking info is correct (50 ms) -[ok]: CLIENT GETREDIR provides correct client id (1 ms) -[ok]: CLIENT TRACKINGINFO provides reasonable results when tracking off (0 ms) -[ok]: CLIENT TRACKINGINFO provides reasonable results when tracking on (0 ms) -[ok]: CLIENT TRACKINGINFO provides reasonable results when tracking on with options (1 ms) -[ok]: CLIENT TRACKINGINFO provides reasonable results when tracking optin (2 ms) -[ok]: CLIENT TRACKINGINFO provides reasonable results when tracking optout (1 ms) -[ok]: CLIENT TRACKINGINFO provides reasonable results when tracking bcast mode (2 ms) -[ok]: CLIENT TRACKINGINFO provides reasonable results when tracking redir broken (36 ms) -[ok]: Regression test for #11715 (33 ms) -[ok]: RESP3 based basic invalidation with client reply off (55 ms) -[ok]: RESP2 based basic invalidation with client reply off (88 ms) -[ok]: RESP3 based basic redirect invalidation with client reply off (48 ms) -[ok]: RESP3 based basic tracking-redir-broken with client reply off (33 ms) -=== (tracking network) Starting server 127.0.0.1:21298 ok -[ok]: Coverage: Basic CLIENT CACHING (27 ms) -[ok]: Coverage: Basic CLIENT REPLY (1 ms) -[ok]: Coverage: Basic CLIENT TRACKINGINFO (1 ms) -[ok]: Coverage: Basic CLIENT GETREDIR (0 ms) -[25/124 done]: unit/tracking (3 seconds) -Testing unit/protocol -=== (protocol network) Starting server 127.0.0.1:21300 ok -[ok]: Handle an empty query (66 ms) -[ok]: Negative multibulk length (68 ms) -[ok]: Out of range multibulk length (19 ms) -[ok]: Wrong multibulk payload header (15 ms) -[ok]: Negative multibulk payload length (16 ms) -[ok]: Out of range multibulk payload length (14 ms) -[ok]: Non-number multibulk payload length (14 ms) -[ok]: Multi bulk request not followed by bulk arguments (15 ms) -[ok]: Generic wrong number of args (15 ms) -[ok]: Unbalanced number of quotes (15 ms) -[ok]: Protocol desync regression test #1 (110 ms) -[ok]: Protocol desync regression test #2 (117 ms) -[ok]: Protocol desync regression test #3 (104 ms) -[ok]: raw protocol response (1 ms) -[ok]: raw protocol response - deferred (0 ms) -[ok]: raw protocol response - multiline (1 ms) -[ok]: RESP3 attributes (2 ms) -[ok]: RESP3 attributes readraw (1 ms) -[ok]: RESP3 attributes on RESP2 (1 ms) -[ok]: test big number parsing (0 ms) -[ok]: test bool parsing (2 ms) -[ok]: test verbatim str parsing (0 ms) -[ok]: test large number of args (31 ms) -[ok]: test argument rewriting - issue 9598 (2 ms) -=== (regression) Starting server 127.0.0.1:21302 ok -[ok]: Regression for a crash with blocking ops and pipelining (15 ms) -[26/124 done]: unit/protocol (1 seconds) +[4/124 done]: unit/bitops (5 seconds) +Testing unit/info +=== (info external:skip) Starting server 127.0.0.1:21123 ok +=== () Starting server 127.0.0.1:21125 ok +[ok]: latencystats: disable/enable (2 ms) +[ok]: latencystats: configure percentiles (4 ms) +[ok]: latencystats: bad configure percentiles (2 ms) +[ok]: latencystats: blocking commands (10 ms) +[ok]: latencystats: subcommands (1 ms) +[ok]: latencystats: measure latency (52 ms) +[ok]: errorstats: failed call authentication error (3 ms) +[ok]: errorstats: failed call within MULTI/EXEC (4 ms) +[ok]: errorstats: failed call within LUA (4 ms) +[ok]: errorstats: failed call NOSCRIPT error (2 ms) +[ok]: errorstats: failed call NOGROUP error (2 ms) +[ok]: errorstats: rejected call unknown command (2 ms) +[ok]: errorstats: rejected call within MULTI/EXEC (4 ms) +[ok]: errorstats: rejected call due to wrong arity (2 ms) +[ok]: errorstats: rejected call by OOM error (3 ms) +[ok]: errorstats: rejected call by authorization error (3 ms) +[ok]: errorstats: blocking commands (14 ms) +[ok]: errorstats: limit errors will not increase indefinitely (253 ms) +eventloop metrics cycle1: 1117, cycle2: 1120 +eventloop metrics el_sum1: 77916, el_sum2: 78571 +eventloop metrics cmd_sum1: 17637, cmd_sum2: 17665 +[ok]: stats: eventloop metrics (112 ms) +instantaneous metrics instantaneous_eventloop_cycles_per_sec: 7 +instantaneous metrics instantaneous_eventloop_duration_usec: 301 +[ok]: stats: instantaneous metrics (1604 ms) +[ok]: stats: debug metrics (114 ms) +[ok]: stats: client input and output buffer limit disconnections (89 ms) +[ok]: clients: pubsub clients (25 ms) +[ok]: clients: watching clients (23 ms) +=== (info external:skip) Starting server 127.0.0.1:21127 ok +[ok]: memory: database and pubsub overhead and rehashing dict count (4 ms) +[5/124 done]: unit/info (3 seconds) +Testing unit/scan +=== (scan network standalone) Starting server 127.0.0.1:21129 ok +[ok]: {standalone} SCAN basic (134 ms) +[ok]: {standalone} SCAN COUNT (154 ms) +[ok]: {standalone} SCAN MATCH (129 ms) +[ok]: {standalone} SCAN TYPE (178 ms) +[ok]: {standalone} SCAN unknown type (111 ms) +[ok]: {standalone} SCAN with expired keys (135 ms) +[ok]: {standalone} SCAN with expired keys with TYPE filter (135 ms) +[ok]: {standalone} SSCAN with encoding intset (2 ms) +[ok]: {standalone} SSCAN with encoding listpack (2 ms) +[ok]: {standalone} SSCAN with encoding hashtable (6 ms) +[ok]: {standalone} HSCAN with encoding listpack (2 ms) +[ok]: {standalone} HSCAN with encoding hashtable (83 ms) +[ok]: {standalone} ZSCAN with encoding listpack (3 ms) +[ok]: {standalone} ZSCAN with encoding skiplist (87 ms) +[ok]: {standalone} SCAN guarantees check under write load (43 ms) +[ok]: {standalone} SSCAN with integer encoded object (issue #1345) (1 ms) +[ok]: {standalone} SSCAN with PATTERN (1 ms) +[ok]: {standalone} HSCAN with PATTERN (1 ms) +[ok]: {standalone} HSCAN with NOVALUES (1 ms) +[ok]: {standalone} ZSCAN with PATTERN (0 ms) +[ok]: {standalone} ZSCAN with NOSCORES (1 ms) +[ok]: {standalone} ZSCAN scores: regression test for issue #2175 (97 ms) +[ok]: {standalone} SCAN regression test for issue #4906 (14001 ms) +[ok]: {standalone} SCAN MATCH pattern implies cluster slot (69 ms) +=== (external:skip cluster scan) Starting server 127.0.0.1:21131 ok +[ok]: {cluster} SCAN basic (170 ms) +[ok]: {cluster} SCAN COUNT (185 ms) +[ok]: {cluster} SCAN MATCH (173 ms) +[ok]: {cluster} SCAN TYPE (261 ms) +[ok]: {cluster} SCAN unknown type (156 ms) +[ok]: {cluster} SCAN with expired keys (172 ms) +[ok]: {cluster} SCAN with expired keys with TYPE filter (143 ms) +[ok]: {cluster} SSCAN with encoding intset (2 ms) +[ok]: {cluster} SSCAN with encoding listpack (1 ms) +[ok]: {cluster} SSCAN with encoding hashtable (8 ms) +[ok]: {cluster} HSCAN with encoding listpack (2 ms) +[ok]: {cluster} HSCAN with encoding hashtable (107 ms) +[ok]: {cluster} ZSCAN with encoding listpack (3 ms) +[ok]: {cluster} ZSCAN with encoding skiplist (114 ms) +[ok]: {cluster} SCAN guarantees check under write load (50 ms) +[ok]: {cluster} SSCAN with integer encoded object (issue #1345) (2 ms) +[ok]: {cluster} SSCAN with PATTERN (1 ms) +[ok]: {cluster} HSCAN with PATTERN (1 ms) +[ok]: {cluster} HSCAN with NOVALUES (2 ms) +[ok]: {cluster} ZSCAN with PATTERN (1 ms) +[ok]: {cluster} ZSCAN with NOSCORES (0 ms) +[ok]: {cluster} ZSCAN scores: regression test for issue #2175 (110 ms) +[ok]: {cluster} SCAN regression test for issue #4906 (11177 ms) +[ok]: {cluster} SCAN MATCH pattern implies cluster slot (51 ms) +[6/124 done]: unit/scan (31 seconds) +Testing unit/pause +=== (pause network) Starting server 127.0.0.1:21133 ok +[ok]: Test read commands are not blocked by client pause (8 ms) +[ok]: Test old pause-all takes precedence over new pause-write (less restrictive) (327 ms) +[ok]: Test new pause time is smaller than old one, then old time preserved (109 ms) +[ok]: Test write commands are paused by RO (7 ms) +[ok]: Test special commands are paused by RO (15 ms) +[ok]: Test read/admin multi-execs are not blocked by pause RO (8 ms) +[ok]: Test write multi-execs are blocked by pause RO (8 ms) +[ok]: Test scripts are blocked by pause RO (14 ms) +[ok]: Test RO scripts are not blocked by pause RO (9 ms) +[ok]: Test read-only scripts in multi-exec are not blocked by pause RO (9 ms) +[ok]: Test write scripts in multi-exec are blocked by pause RO (15 ms) +[ok]: Test may-replicate commands are rejected in RO scripts (1 ms) +[ok]: Test multiple clients can be queued up and unblocked (20 ms) +[ok]: Test clients with syntax errors will get responses immediately (0 ms) +[ok]: Test eviction is skipped during client pause (3 ms) +[ok]: Test both active and passive expires are skipped during client pause (105 ms) +[ok]: Test that client pause starts at the end of a transaction (11 ms) +=== (needs:repl external:skip) Starting server 127.0.0.1:21135 ok +[ok]: Test when replica paused, offset would not grow (3 ms) +[ok]: Test replica offset would grow after unpause (1 ms) +[7/124 done]: unit/pause (1 seconds) +Testing unit/sort +=== (sort) Starting server 127.0.0.1:21137 ok +[ok]: Listpack: SORT BY key (0 ms) +[ok]: Listpack: SORT BY key with limit (0 ms) +[ok]: Listpack: SORT BY hash field (0 ms) +[ok]: Quicklist: SORT BY key (8 ms) +[ok]: Quicklist: SORT BY key with limit (1 ms) +[ok]: Quicklist: SORT BY hash field (9 ms) +[ok]: Big Quicklist: SORT BY key (421 ms) +[ok]: Big Quicklist: SORT BY key with limit (6 ms) +[ok]: Big Quicklist: SORT BY hash field (419 ms) +[ok]: Intset: SORT BY key (1 ms) +[ok]: Intset: SORT BY key with limit (0 ms) +[ok]: Intset: SORT BY hash field (1 ms) +[ok]: Hash table: SORT BY key (9 ms) +[ok]: Hash table: SORT BY key with limit (1 ms) +[ok]: Hash table: SORT BY hash field (8 ms) +[ok]: Big Hash table: SORT BY key (439 ms) +[ok]: Big Hash table: SORT BY key with limit (7 ms) +[ok]: Big Hash table: SORT BY hash field (441 ms) +[ok]: SORT GET # (0 ms) +[ok]: SORT GET (1 ms) +[ok]: SORT_RO GET (0 ms) +[ok]: SORT GET (key and hash) with sanity check (6 ms) +[ok]: SORT BY key STORE (1 ms) +[ok]: SORT BY hash field STORE (2 ms) +[ok]: SORT extracts STORE correctly (0 ms) +[ok]: SORT_RO get keys (0 ms) +[ok]: SORT extracts multiple STORE correctly (0 ms) +[ok]: SORT DESC (0 ms) +[ok]: SORT ALPHA against integer encoded strings (1 ms) +[ok]: SORT sorted set (1 ms) +[ok]: SORT sorted set BY nosort should retain ordering (2 ms) +[ok]: SORT sorted set BY nosort + LIMIT (2 ms) +[ok]: SORT sorted set BY nosort works as expected from scripts (2 ms) +[ok]: SORT sorted set: +inf and -inf handling (2 ms) +[ok]: SORT regression for issue #19, sorting floats (4 ms) +[ok]: SORT with STORE returns zero if result is empty (github issue 224) (1 ms) +[ok]: SORT with STORE does not create empty lists (github issue 224) (0 ms) +[ok]: SORT with STORE removes key if result is empty (github issue 227) (0 ms) +[ok]: SORT with BY and STORE should still order output (1 ms) +[ok]: SORT will complain with numerical sorting and bad doubles (1) (1 ms) +[ok]: SORT will complain with numerical sorting and bad doubles (2) (1 ms) +[ok]: SORT BY sub-sorts lexicographically if score is the same (3 ms) +[ok]: SORT GET with pattern ending with just -> does not get hash field (1 ms) +[ok]: SORT by nosort retains native order for lists (0 ms) +[ok]: SORT by nosort plus store retains native order for lists (1 ms) +[ok]: SORT by nosort with limit returns based on original list order (0 ms) +[ok]: SORT_RO - Successful case (1 ms) +[ok]: SORT_RO - Cannot run with STORE arg (0 ms) + + Average time to sort: 0.20000000000000001 milliseconds [ok]: SORT speed, 100 element list BY key, 100 times (20 ms) + + Average time to sort: 0.25 milliseconds [ok]: SORT speed, 100 element list BY hash field, 100 times (25 ms) + + Average time to sort: 0.17999999999999999 milliseconds [ok]: SORT speed, 100 element list directly, 100 times (18 ms) + + Average time to sort: 0.16 milliseconds [ok]: SORT speed, 100 element list BY , 100 times (16 ms) +[ok]: SETRANGE with huge offset (1 ms) +[ok]: SORT STORE quicklist with the right options (53 ms) +=== (external:skip cluster sort) Starting server 127.0.0.1:21139 ok +[ok]: sort by in cluster mode (1 ms) +[ok]: sort get in cluster mode (1 ms) +[ok]: sort_ro by in cluster mode (1 ms) +[ok]: sort_ro get in cluster mode (1 ms) +[8/124 done]: unit/sort (17 seconds) Testing unit/aofrw -=== (aofrw external:skip logreqres:skip) Starting server 127.0.0.1:21304 ok +=== (aofrw external:skip logreqres:skip) Starting server 127.0.0.1:21141 ok Waiting for background AOF rewrite to finish... Waiting for background AOF rewrite to finish... Waiting for background AOF rewrite to finish... Waiting for background AOF rewrite to finish... Waiting for background AOF rewrite to finish... -Waiting for background AOF rewrite to finish... [ok]: AOF rewrite during write load: RDB preamble=yes (6845 ms) +Waiting for background AOF rewrite to finish... [ok]: AOF rewrite during write load: RDB preamble=yes (7494 ms) Waiting for background AOF rewrite to finish... Waiting for background AOF rewrite to finish... @@ -1948,295 +1517,682 @@ Waiting for background AOF rewrite to finish... Waiting for background AOF rewrite to finish... Waiting for background AOF rewrite to finish... -Waiting for background AOF rewrite to finish... [ok]: AOF rewrite during write load: RDB preamble=no (9498 ms) -=== (aofrw external:skip) Starting server 127.0.0.1:21306 ok +Waiting for background AOF rewrite to finish... +Waiting for background AOF rewrite to finish... +Waiting for background AOF rewrite to finish... +Waiting for background AOF rewrite to finish... [ok]: AOF rewrite during write load: RDB preamble=no (12440 ms) +=== (aofrw external:skip) Starting server 127.0.0.1:21143 ok Waiting for background AOF rewrite to finish... -Waiting for background AOF rewrite to finish... [ok]: Turning off AOF kills the background writing child if any (132 ms) +Waiting for background AOF rewrite to finish... [ok]: Turning off AOF kills the background writing child if any (112 ms) -Waiting for background AOF rewrite to finish... [ok]: AOF rewrite of list with listpack encoding, string data (60 ms) +Waiting for background AOF rewrite to finish... +Waiting for background AOF rewrite to finish... [ok]: AOF rewrite of list with listpack encoding, string data (110 ms) -Waiting for background AOF rewrite to finish... [ok]: AOF rewrite of list with quicklist encoding, string data (315 ms) +Waiting for background AOF rewrite to finish... +Waiting for background AOF rewrite to finish... [ok]: AOF rewrite of list with quicklist encoding, string data (292 ms) Waiting for background AOF rewrite to finish... -Waiting for background AOF rewrite to finish... [ok]: AOF rewrite of list with listpack encoding, int data (124 ms) +Waiting for background AOF rewrite to finish... [ok]: AOF rewrite of list with listpack encoding, int data (533 ms) -Waiting for background AOF rewrite to finish... [ok]: AOF rewrite of list with quicklist encoding, int data (222 ms) +Waiting for background AOF rewrite to finish... [ok]: AOF rewrite of list with quicklist encoding, int data (233 ms) Waiting for background AOF rewrite to finish... -Waiting for background AOF rewrite to finish... [ok]: AOF rewrite of set with intset encoding, string data (375 ms) +Waiting for background AOF rewrite to finish... [ok]: AOF rewrite of set with intset encoding, string data (108 ms) Waiting for background AOF rewrite to finish... +Waiting for background AOF rewrite to finish... [ok]: AOF rewrite of set with hashtable encoding, string data (308 ms) + Waiting for background AOF rewrite to finish... -Waiting for background AOF rewrite to finish... [ok]: AOF rewrite of set with hashtable encoding, string data (935 ms) +Waiting for background AOF rewrite to finish... [ok]: AOF rewrite of set with intset encoding, int data (108 ms) Waiting for background AOF rewrite to finish... Waiting for background AOF rewrite to finish... +Waiting for background AOF rewrite to finish... [ok]: AOF rewrite of set with hashtable encoding, int data (342 ms) + Waiting for background AOF rewrite to finish... +Waiting for background AOF rewrite to finish... [ok]: AOF rewrite of hash with listpack encoding, string data (512 ms) + +Waiting for background AOF rewrite to finish... [ok]: AOF rewrite of hash with hashtable encoding, string data (242 ms) + Waiting for background AOF rewrite to finish... +Waiting for background AOF rewrite to finish... [ok]: AOF rewrite of hash with listpack encoding, int data (107 ms) + +Waiting for background AOF rewrite to finish... +Waiting for background AOF rewrite to finish... [ok]: AOF rewrite of hash with hashtable encoding, int data (284 ms) + Waiting for background AOF rewrite to finish... +Waiting for background AOF rewrite to finish... [ok]: AOF rewrite of zset with listpack encoding, string data (108 ms) + Waiting for background AOF rewrite to finish... Waiting for background AOF rewrite to finish... -Waiting for background AOF rewrite to finish... [ok]: AOF rewrite of set with intset encoding, int data (1204 ms) +Waiting for background AOF rewrite to finish... [ok]: AOF rewrite of zset with skiplist encoding, string data (354 ms) + +Waiting for background AOF rewrite to finish... [ok]: AOF rewrite of zset with listpack encoding, int data (77 ms) + +Waiting for background AOF rewrite to finish... [ok]: AOF rewrite of zset with skiplist encoding, int data (233 ms) Waiting for background AOF rewrite to finish... +Waiting for background AOF rewrite to finish... [ok]: AOF rewrite functions (450 ms) +[ok]: BGREWRITEAOF is delayed if BGSAVE is in progress (293 ms) +[ok]: BGREWRITEAOF is refused if already in progress (46 ms) +[9/124 done]: unit/aofrw (25 seconds) +Testing unit/auth +=== (auth external:skip) Starting server 127.0.0.1:21145 ok +[ok]: AUTH fails if there is no password configured server side (1 ms) +[ok]: Arity check for auth command (0 ms) +=== (auth external:skip) Starting server 127.0.0.1:21147 ok +[ok]: AUTH fails when a wrong password is given (6 ms) +[ok]: Arbitrary command gives an error when AUTH is required (0 ms) +[ok]: AUTH succeeds when the right password is given (0 ms) +[ok]: Once AUTH succeeded we can actually send commands to the server (0 ms) +[ok]: For unauthenticated clients multibulk and bulk length are limited (14 ms) +=== (auth_binary_password external:skip) Starting server 127.0.0.1:21149 ok +[ok]: AUTH fails when binary password is wrong (1 ms) +[ok]: AUTH succeeds when binary password is correct (0 ms) +=== (primaryauth) Starting server 127.0.0.1:21151 ok +[ok]: primaryauth test with binary password dualchannel = yes (928 ms) +=== (auth_binary_password external:skip) Starting server 127.0.0.1:21153 ok +[ok]: AUTH fails when binary password is wrong (1 ms) +[ok]: AUTH succeeds when binary password is correct (0 ms) +=== (primaryauth) Starting server 127.0.0.1:21155 ok +[ok]: primaryauth test with binary password dualchannel = no (931 ms) +[10/124 done]: unit/auth (3 seconds) +Testing unit/multi +=== (multi) Starting server 127.0.0.1:21157 ok +[ok]: MULTI / EXEC basics (1 ms) +[ok]: DISCARD (2 ms) +[ok]: Nested MULTI are not allowed (0 ms) +[ok]: MULTI where commands alter argc/argv (1 ms) +[ok]: WATCH inside MULTI is not allowed (1 ms) +[ok]: EXEC fails if there are errors while queueing commands #1 (1 ms) +[ok]: EXEC fails if there are errors while queueing commands #2 (9 ms) +[ok]: If EXEC aborts, the client MULTI state is cleared (1 ms) +[ok]: EXEC works on WATCHed key not modified (1 ms) +[ok]: EXEC fail on WATCHed key modified (1 key of 1 watched) (1 ms) +[ok]: EXEC fail on WATCHed key modified (1 key of 5 watched) (1 ms) +[ok]: EXEC fail on WATCHed key modified by SORT with STORE even if the result is empty (1 ms) +EXEC fail on lazy expired WATCHed key attempts: 0 +[ok]: EXEC fail on lazy expired WATCHed key (104 ms) +[ok]: WATCH stale keys should not fail EXEC (3 ms) +[ok]: Delete WATCHed stale keys should not fail EXEC (3 ms) +[ok]: FLUSHDB while watching stale keys should not fail EXEC (4 ms) +[ok]: After successful EXEC key is no longer watched (2 ms) +[ok]: After failed EXEC key is no longer watched (1 ms) +[ok]: It is possible to UNWATCH (1 ms) +[ok]: UNWATCH when there is nothing watched works as expected (0 ms) +[ok]: FLUSHALL is able to touch the watched keys (1 ms) +[ok]: FLUSHALL does not touch non affected keys (2 ms) +[ok]: FLUSHDB is able to touch the watched keys (1 ms) +[ok]: FLUSHDB does not touch non affected keys (1 ms) +[ok]: SWAPDB is able to touch the watched keys that exist (1 ms) +[ok]: SWAPDB is able to touch the watched keys that do not exist (1 ms) +[ok]: SWAPDB does not touch watched stale keys (4 ms) +[ok]: SWAPDB does not touch non-existing key replaced with stale key (4 ms) +[ok]: SWAPDB does not touch stale key replaced with another stale key (4 ms) +[ok]: WATCH is able to remember the DB a key belongs to (2 ms) +[ok]: WATCH will consider touched keys target of EXPIRE (2 ms) +[ok]: WATCH will consider touched expired keys (1127 ms) +[ok]: DISCARD should clear the WATCH dirty flag on the client (1 ms) +[ok]: DISCARD should UNWATCH all the keys (1 ms) +[ok]: MULTI / EXEC is not propagated (single write command) (707 ms) +[ok]: MULTI / EXEC is propagated correctly (multiple commands) (305 ms) +[ok]: MULTI / EXEC is propagated correctly (multiple commands with SELECT) (102 ms) +[ok]: MULTI / EXEC is propagated correctly (empty transaction) (100 ms) +[ok]: MULTI / EXEC is propagated correctly (read-only commands) (107 ms) +[ok]: MULTI / EXEC is propagated correctly (write command, no effect) (96 ms) +[ok]: MULTI / EXEC with REPLICAOF (103 ms) +[ok]: DISCARD should not fail during OOM (13 ms) +[ok]: MULTI and script timeout (426 ms) +[ok]: EXEC and script timeout (419 ms) +[ok]: MULTI-EXEC body and script timeout (423 ms) +[ok]: just EXEC and script timeout (426 ms) +[ok]: exec with write commands and state change (10 ms) +[ok]: exec with read commands and stale replica state change (11 ms) +[ok]: EXEC with only read commands should not be rejected when OOM (11 ms) +[ok]: EXEC with at least one use-memory command should fail (12 ms) +[ok]: Blocking commands ignores the timeout (2 ms) +[ok]: MULTI propagation of PUBLISH (481 ms) +[ok]: MULTI propagation of SCRIPT LOAD (102 ms) +[ok]: MULTI propagation of EVAL (104 ms) +[ok]: MULTI propagation of SCRIPT FLUSH (99 ms) +[ok]: MULTI propagation of XREADGROUP (102 ms) +[ok]: MULTI with SAVE (1 ms) +[ok]: MULTI with SHUTDOWN (1 ms) + +Waiting for background AOF rewrite to finish... +Waiting for background AOF rewrite to finish... [ok]: MULTI with BGREWRITEAOF (206 ms) + +Waiting for background AOF rewrite to finish... +Waiting for background AOF rewrite to finish... [ok]: MULTI with config set appendonly (209 ms) +[ok]: MULTI with config error (1 ms) +[ok]: Flushall while watching several keys by one client (1 ms) +=== (external:skip) Starting server 127.0.0.1:21159 ok +[ok]: MULTI with FLUSHALL and AOF (4 ms) +=== (external:skip cluster) Starting server 127.0.0.1:21161 ok +[ok]: Regression test for multi-exec with RANDOMKEY accessing the wrong per-slot dictionary (3 ms) +[11/124 done]: unit/multi (9 seconds) +Testing unit/other +=== (other) Starting server 127.0.0.1:21163 ok +[ok]: Coverage: HELP commands (3 ms) +[ok]: Coverage: MEMORY MALLOC-STATS (2 ms) +[ok]: Coverage: MEMORY PURGE (1 ms) +[ok]: SAVE - make sure there are all the types as values (7 ms) +[ok]: FUZZ stresser with data model binary (599 ms) +[ok]: FUZZ stresser with data model alpha (619 ms) +[ok]: FUZZ stresser with data model compr (563 ms) +=== (external:skip) Starting server 127.0.0.1:21165 ok +[ok]: FLUSHALL should not reset the dirty counter if we disable save (1 ms) +[ok]: FLUSHALL should reset the dirty counter to 0 if we enable save (12 ms) + +Waiting for background save to finish... +Waiting for background save to finish... [ok]: BGSAVE (118 ms) +[ok]: SELECT an out of range DB (1 ms) +[ok]: Check consistency of different data types after a reload (1391 ms) + +Waiting for background AOF rewrite to finish... +Waiting for background AOF rewrite to finish... +Waiting for background AOF rewrite to finish... +Waiting for background AOF rewrite to finish... +Waiting for background AOF rewrite to finish... +Waiting for background AOF rewrite to finish... +Waiting for background AOF rewrite to finish... +Waiting for background AOF rewrite to finish... +Waiting for background AOF rewrite to finish... +Waiting for background AOF rewrite to finish... +Waiting for background AOF rewrite to finish... +Waiting for background AOF rewrite to finish... Waiting for background AOF rewrite to finish... Waiting for background AOF rewrite to finish... Waiting for background AOF rewrite to finish... Waiting for background AOF rewrite to finish... Waiting for background AOF rewrite to finish... Waiting for background AOF rewrite to finish... -Waiting for background AOF rewrite to finish... [ok]: AOF rewrite of set with hashtable encoding, int data (1131 ms) - Waiting for background AOF rewrite to finish... Waiting for background AOF rewrite to finish... Waiting for background AOF rewrite to finish... Waiting for background AOF rewrite to finish... Waiting for background AOF rewrite to finish... -Waiting for background AOF rewrite to finish... [ok]: AOF rewrite of hash with listpack encoding, string data (790 ms) - Waiting for background AOF rewrite to finish... Waiting for background AOF rewrite to finish... Waiting for background AOF rewrite to finish... -Waiting for background AOF rewrite to finish... [ok]: AOF rewrite of hash with hashtable encoding, string data (479 ms) - -Waiting for background AOF rewrite to finish... [ok]: AOF rewrite of hash with listpack encoding, int data (58 ms) - -Waiting for background AOF rewrite to finish... [ok]: AOF rewrite of hash with hashtable encoding, int data (237 ms) - Waiting for background AOF rewrite to finish... -Waiting for background AOF rewrite to finish... [ok]: AOF rewrite of zset with listpack encoding, string data (109 ms) - Waiting for background AOF rewrite to finish... -Waiting for background AOF rewrite to finish... [ok]: AOF rewrite of zset with skiplist encoding, string data (310 ms) - Waiting for background AOF rewrite to finish... -Waiting for background AOF rewrite to finish... [ok]: AOF rewrite of zset with listpack encoding, int data (108 ms) - -Waiting for background AOF rewrite to finish... [ok]: AOF rewrite of zset with skiplist encoding, int data (217 ms) - Waiting for background AOF rewrite to finish... -Waiting for background AOF rewrite to finish... [ok]: AOF rewrite functions (109 ms) -[ok]: BGREWRITEAOF is delayed if BGSAVE is in progress (278 ms) -[ok]: BGREWRITEAOF is refused if already in progress (76 ms) -[27/124 done]: unit/aofrw (24 seconds) -Testing unit/other -=== (other) Starting server 127.0.0.1:21308 ok -[ok]: Coverage: HELP commands (4 ms) -[ok]: Coverage: MEMORY MALLOC-STATS (4 ms) -[ok]: Coverage: MEMORY PURGE (1 ms) -[ok]: SAVE - make sure there are all the types as values (8 ms) -[ok]: FUZZ stresser with data model binary (1031 ms) -[ok]: FUZZ stresser with data model alpha (1034 ms) -[ok]: FUZZ stresser with data model compr (1008 ms) -=== (external:skip) Starting server 127.0.0.1:21310 ok -[ok]: FLUSHALL should not reset the dirty counter if we disable save (1 ms) -[ok]: FLUSHALL should reset the dirty counter to 0 if we enable save (7 ms) - -Waiting for background save to finish... -Waiting for background save to finish... [ok]: BGSAVE (127 ms) -[ok]: SELECT an out of range DB (1 ms) -[ok]: Check consistency of different data types after a reload (1354 ms) - Waiting for background AOF rewrite to finish... -Waiting for background AOF rewrite to finish... [ok]: Same dataset digest if saving/reloading as AOF? (353 ms) +Waiting for background AOF rewrite to finish... +Waiting for background AOF rewrite to finish... +Waiting for background AOF rewrite to finish... +Waiting for background AOF rewrite to finish... [ok]: Same dataset digest if saving/reloading as AOF? (2423 ms) -Waiting for background AOF rewrite to finish... [ok]: EXPIRES after a reload (snapshot + append only file rewrite) (67 ms) +Waiting for background AOF rewrite to finish... +Waiting for background AOF rewrite to finish... +Waiting for background AOF rewrite to finish... +Waiting for background AOF rewrite to finish... +Waiting for background AOF rewrite to finish... +Waiting for background AOF rewrite to finish... +Waiting for background AOF rewrite to finish... +Waiting for background AOF rewrite to finish... [ok]: EXPIRES after a reload (snapshot + append only file rewrite) (2234 ms) Waiting for background AOF rewrite to finish... -Waiting for background AOF rewrite to finish... [ok]: EXPIRES after AOF reload (without rewrite) (2121 ms) -[ok]: PIPELINING stresser (also a regression for the old epoll bug) (823 ms) -[ok]: APPEND basics (2 ms) +Waiting for background AOF rewrite to finish... [ok]: EXPIRES after AOF reload (without rewrite) (2119 ms) +[ok]: PIPELINING stresser (also a regression for the old epoll bug) (515 ms) +[ok]: APPEND basics (1 ms) [ok]: APPEND basics, integer encoded values (1 ms) -[ok]: APPEND fuzzing (489 ms) +[ok]: APPEND fuzzing (478 ms) [ok]: FLUSHDB (1 ms) -[ok]: Perform a final SAVE to leave a clean DB on disk (9 ms) +[ok]: Perform a final SAVE to leave a clean DB on disk (8 ms) [ok]: RESET clears client state (1 ms) -[ok]: RESET clears MONITOR state (15 ms) +[ok]: RESET clears MONITOR state (7 ms) [ok]: RESET clears and discards MULTI state (1 ms) [ok]: RESET clears Pub/Sub state (0 ms) -[ok]: RESET clears authenticated state (0 ms) +[ok]: RESET clears authenticated state (2 ms) [ok]: Subcommand syntax error crash (issue #10070) (1 ms) -[ok]: Extended Redis Compatibility config (1 ms) -=== (other external:skip) Starting server 127.0.0.1:21312 ok +[ok]: Extended Redis Compatibility config (2 ms) +=== (other external:skip) Starting server 127.0.0.1:21167 ok -Waiting for background save to finish... [ok]: Don't rehash if server has child process (588 ms) -=== (other external:skip) Starting server 127.0.0.1:21314 ok -[ok]: Process title set as expected (3 ms) -=== (other external:skip cluster slow) Starting server 127.0.0.1:21316 ok -[ok]: Server can trigger resizing (1929 ms) -[ok]: Server can rewind and trigger smaller slot resizing (703 ms) -=== (other external:skip) Starting server 127.0.0.1:21318 ok -[ok]: Server can resize empty dict (40 ms) -Creating temp file /build/reproducible-path/valkey-8.0.2+dfsg1/1790455 -=== (repl external:skip) Starting server 127.0.0.1:21320 ok -[ok]: test unixsocket options are set correctly (1 ms) -Deleting temp file: /build/reproducible-path/valkey-8.0.2+dfsg1/1790455 -[28/124 done]: unit/other (15 seconds) -Testing unit/expire -=== (expire) Starting server 127.0.0.1:21322 ok -[ok]: EXPIRE - set timeouts multiple times (15 ms) -[ok]: EXPIRE - It should be still possible to read 'x' (1 ms) -[ok]: EXPIRE - After 2.1 seconds the key should no longer be here (2108 ms) -[ok]: EXPIRE - write on expire should work (5 ms) -[ok]: EXPIREAT - Check for EXPIRE alike behavior (1 ms) -[ok]: SETEX - Set + Expire combo operation. Check for TTL (1 ms) -[ok]: SETEX - Check value (1 ms) -[ok]: SETEX - Overwrite old key (1 ms) -[ok]: SETEX - Wait for the key to expire (1103 ms) -[ok]: SETEX - Wrong time parameter (1 ms) -[ok]: PERSIST can undo an EXPIRE (1 ms) -[ok]: PERSIST returns 0 against non existing or non volatile keys (1 ms) -millisecond expire test attempts: 0 -[ok]: EXPIRE precision is now the millisecond (1112 ms) -PSETEX sub-second expire test attempts: 0 -[ok]: PSETEX can set sub-second expires (104 ms) -PEXPIRE sub-second expire test attempts: 0 -[ok]: PEXPIRE can set sub-second expires (104 ms) -PEXPIREAT sub-second expire test attempts: 0 -[ok]: PEXPIREAT can set sub-second expires (209 ms) -[ok]: TTL returns time to live in seconds (24 ms) -[ok]: PTTL returns time to live in milliseconds (16 ms) -[ok]: TTL / PTTL / EXPIRETIME / PEXPIRETIME return -1 if key has no expire (25 ms) -[ok]: TTL / PTTL / EXPIRETIME / PEXPIRETIME return -2 if key does not exit (26 ms) -[ok]: EXPIRETIME returns absolute expiration time in seconds (21 ms) -[ok]: PEXPIRETIME returns absolute expiration time in milliseconds (20 ms) -[ok]: Server should actively expire keys incrementally (603 ms) -[ok]: Server should lazy expire keys (1058 ms) -[ok]: EXPIRE should not resurrect keys (issue #1026) (1017 ms) -[ok]: 5 keys in, 5 keys out (2 ms) -[ok]: EXPIRE with empty string as TTL should report an error (1 ms) -[ok]: SET with EX with big integer should report an error (0 ms) -[ok]: SET with EX with smallest integer should report an error (0 ms) -[ok]: GETEX with big integer should report an error (0 ms) -[ok]: GETEX with smallest integer should report an error (1 ms) -[ok]: EXPIRE with big integer overflows when converted to milliseconds (6 ms) -[ok]: PEXPIRE with big integer overflow when basetime is added (2 ms) -[ok]: EXPIRE with big negative integer (1 ms) -[ok]: PEXPIREAT with big integer works (1 ms) -[ok]: PEXPIREAT with big negative integer works (1 ms) -=== (external:skip) Starting server 127.0.0.1:21324 ok -[ok]: All time-to-live(TTL) in commands are propagated as absolute timestamp in milliseconds in AOF (2322 ms) -[ok]: All TTL in commands are propagated as absolute timestamp in replication stream (69 ms) -=== (needs:repl external:skip) Starting server 127.0.0.1:21326 ok -[ok]: First server should have role slave after REPLICAOF (2 ms) -[ok]: For all replicated TTL-related commands, absolute expire times are identical on primary and replica (37 ms) +Waiting for background save to finish... +Waiting for background save to finish... [ok]: Don't rehash if server has child process (732 ms) +=== (other external:skip) Starting server 127.0.0.1:21169 ok +[ok]: Process title set as expected (2 ms) +=== (other external:skip cluster slow) Starting server 127.0.0.1:21171 ok +[ok]: Server can trigger resizing (1810 ms) +[ok]: Server can rewind and trigger smaller slot resizing (695 ms) +=== (other external:skip) Starting server 127.0.0.1:21173 ok +[ok]: Server can resize empty dict (43 ms) +Creating temp file /build/reproducible-path/valkey-8.0.2+dfsg1/30294 +=== (repl external:skip) Starting server 127.0.0.1:21175 ok +[ok]: test unixsocket options are set correctly (0 ms) +Deleting temp file: /build/reproducible-path/valkey-8.0.2+dfsg1/30294 +[12/124 done]: unit/other (19 seconds) +Testing unit/obuf-limits +=== (obuf-limits external:skip logreqres:skip) Starting server 127.0.0.1:21177 ok +[ok]: CONFIG SET client-output-buffer-limit (3 ms) +[ok]: Client output buffer hard limit is enforced (11135 ms) +[ok]: Client output buffer soft limit is enforced if time is overreached (4267 ms) +[ok]: Client output buffer soft limit is not enforced too early and is enforced when no traffic (6066 ms) +[ok]: No response for single command if client output buffer hard limit is enforced (445 ms) +[ok]: No response for multi commands in pipeline if client output buffer limit is enforced (1125 ms) +[ok]: Execute transactions completely even if client output buffer limit is enforced (1483 ms) +[ok]: Obuf limit, HRANDFIELD with huge count stopped mid-run (24 ms) +[ok]: Obuf limit, KEYS stopped mid-run (166 ms) +[13/124 done]: unit/obuf-limits (25 seconds) +Testing unit/wait +=== (wait network external:skip) Starting server 127.0.0.1:21179 ok +=== () Starting server 127.0.0.1:21181 ok +[ok]: Setup slave (403 ms) +[ok]: WAIT out of range timeout (milliseconds) (1 ms) +[ok]: WAIT should acknowledge 1 additional copy of the data (1 ms) +[ok]: WAIT should not acknowledge 2 additional copies of the data (1050 ms) +[ok]: WAIT should not acknowledge 1 additional copy if slave is blocked (1117 ms) +[ok]: WAIT implicitly blocks on client pause since ACKs aren't sent (1125 ms) +[ok]: WAIT replica multiple clients unblock - reuse last result (170 ms) +=== () Starting server 127.0.0.1:21183 ok +[ok]: Setup a new replica (112 ms) +[ok]: WAIT in script will work (805 ms) +=== () Starting server 127.0.0.1:21185 ok +[ok]: WAITAOF local copy before fsync (96 ms) +[ok]: WAITAOF local copy everysec (0 ms) +[ok]: WAITAOF local copy with appendfsync always (234 ms) +[ok]: WAITAOF local wait and then stop aof (87 ms) +[ok]: WAITAOF local on server with aof disabled (5 ms) +[ok]: WAITAOF local if AOFRW was postponed (960 ms) +=== () Starting server 127.0.0.1:21187 ok +[ok]: WAITAOF on demoted master gets unblocked with an error (48 ms) +[ok]: WAITAOF replica copy before fsync (58 ms) +[ok]: WAITAOF replica copy everysec (2212 ms) -Waiting for background AOF rewrite to finish... [ok]: expired key which is created in writeable replicas should be deleted by active expiry (258 ms) -[ok]: SET command will remove expire (1 ms) -[ok]: SET - use KEEPTTL option, TTL should not be removed (0 ms) -[ok]: SET - use KEEPTTL option, TTL should not be removed after loadaof (2014 ms) -[ok]: GETEX use of PERSIST option should remove TTL (1 ms) -[ok]: GETEX use of PERSIST option should remove TTL after loadaof (1 ms) -[ok]: GETEX propagate as to replica as PERSIST, UNLINK, or nothing (54 ms) -[ok]: EXPIRE with NX option on a key with ttl (1 ms) -[ok]: EXPIRE with NX option on a key without ttl (0 ms) -[ok]: EXPIRE with XX option on a key with ttl (0 ms) -[ok]: EXPIRE with XX option on a key without ttl (0 ms) -[ok]: EXPIRE with GT option on a key with lower ttl (0 ms) -[ok]: EXPIRE with GT option on a key with higher ttl (0 ms) -[ok]: EXPIRE with GT option on a key without ttl (0 ms) -[ok]: EXPIRE with LT option on a key with higher ttl (1 ms) -[ok]: EXPIRE with LT option on a key with lower ttl (0 ms) -[ok]: EXPIRE with LT option on a key without ttl (1 ms) -[ok]: EXPIRE with LT and XX option on a key with ttl (0 ms) -[ok]: EXPIRE with LT and XX option on a key without ttl (1 ms) -[ok]: EXPIRE with conflicting options: LT GT (0 ms) -[ok]: EXPIRE with conflicting options: NX GT (1 ms) -[ok]: EXPIRE with conflicting options: NX LT (0 ms) -[ok]: EXPIRE with conflicting options: NX XX (0 ms) -[ok]: EXPIRE with unsupported options (0 ms) -[ok]: EXPIRE with unsupported options (1 ms) -[ok]: EXPIRE with negative expiry (0 ms) -[ok]: EXPIRE with negative expiry on a non-valitale key (0 ms) -[ok]: EXPIRE with non-existed key (1 ms) -[ok]: Server should not propagate the read command on lazy expire (84 ms) -[ok]: SCAN: Lazy-expire should not be wrapped in MULTI/EXEC (102 ms) -[ok]: RANDOMKEY: Lazy-expire should not be wrapped in MULTI/EXEC (102 ms) -=== (expire external:skip cluster) Starting server 127.0.0.1:21328 ok -[ok]: expire scan should skip dictionaries with lot's of empty buckets (2001 ms) -[29/124 done]: unit/expire (19 seconds) -Testing unit/printver -=== () Starting server 127.0.0.1:21330 ok -Testing Valkey version 7.2.4 (00000000) -[30/124 done]: unit/printver (0 seconds) +Waiting for background AOF rewrite to finish... +Waiting for background AOF rewrite to finish... [ok]: WAITAOF replica copy everysec with AOFRW (1045 ms) + +Waiting for background AOF rewrite to finish... +Waiting for background AOF rewrite to finish... +Waiting for background AOF rewrite to finish... +Waiting for background AOF rewrite to finish... +Waiting for background AOF rewrite to finish... +Waiting for background AOF rewrite to finish... +Waiting for background AOF rewrite to finish... +Waiting for background AOF rewrite to finish... +Waiting for background AOF rewrite to finish... +Waiting for background AOF rewrite to finish... +Waiting for background AOF rewrite to finish... +Waiting for background AOF rewrite to finish... +Waiting for background AOF rewrite to finish... +Waiting for background AOF rewrite to finish... +Waiting for background AOF rewrite to finish... +Waiting for background AOF rewrite to finish... +Waiting for background AOF rewrite to finish... +Waiting for background AOF rewrite to finish... +Waiting for background AOF rewrite to finish... +Waiting for background AOF rewrite to finish... [ok]: WAITAOF replica copy everysec with slow AOFRW (2110 ms) + +Waiting for background AOF rewrite to finish... [ok]: WAITAOF replica copy everysec->always with AOFRW (125 ms) +[ok]: WAITAOF replica copy appendfsync always (14 ms) +[ok]: WAITAOF replica copy if replica is blocked (673 ms) +[ok]: WAITAOF replica multiple clients unblock - reuse last result (1067 ms) +[ok]: WAITAOF on promoted replica (25 ms) +[ok]: WAITAOF master that loses a replica and backlog is dropped (4499 ms) +[ok]: WAITAOF master without backlog, wait is released when the replica finishes full-sync (955 ms) +[ok]: WAITAOF master isn't configured to do AOF (1030 ms) + +Waiting for background AOF rewrite to finish... +Waiting for background AOF rewrite to finish... [ok]: WAITAOF replica isn't configured to do AOF (174 ms) + +Waiting for background AOF rewrite to finish... [ok]: WAITAOF both local and replica got AOF enabled at runtime (900 ms) +[ok]: WAITAOF master sends PING after last write (1231 ms) +[ok]: WAITAOF master client didn't send any write command (1249 ms) +[ok]: WAITAOF master client didn't send any command (2994 ms) +=== () Starting server 127.0.0.1:21189 ok +=== () Starting server 127.0.0.1:21191 ok +[ok]: WAITAOF when replica switches between masters, fsync: no (9857 ms) +=== () Starting server 127.0.0.1:21193 ok +=== () Starting server 127.0.0.1:21195 ok +[ok]: WAITAOF when replica switches between masters, fsync: everysec (2770 ms) +=== () Starting server 127.0.0.1:21197 ok +=== () Starting server 127.0.0.1:21199 ok +[ok]: WAITAOF when replica switches between masters, fsync: always (1519 ms) +=== (failover external:skip) Starting server 127.0.0.1:21201 ok +=== () Starting server 127.0.0.1:21203 ok +=== () Starting server 127.0.0.1:21205 ok +[ok]: setup replication for following tests (114 ms) + +Waiting for background AOF rewrite to finish... +Waiting for background AOF rewrite to finish... [ok]: WAIT and WAITAOF replica multiple clients unblock - reuse last result (2758 ms) +[14/124 done]: unit/wait (47 seconds) +Testing unit/info-command +=== (info and its relative command) Starting server 127.0.0.1:21207 ok +[ok]: info command with at most one sub command (2 ms) +[ok]: info command with one sub-section (1 ms) +[ok]: info command with multiple sub-sections (1 ms) +[15/124 done]: unit/info-command (0 seconds) +Testing unit/pubsubshard +=== (pubsubshard external:skip) Starting server 127.0.0.1:21209 ok +[ok]: SPUBLISH/SSUBSCRIBE basics (51 ms) +[ok]: SPUBLISH/SSUBSCRIBE with two clients (50 ms) +[ok]: SPUBLISH/SSUBSCRIBE after UNSUBSCRIBE without arguments (17 ms) +[ok]: SSUBSCRIBE to one channel more than once (17 ms) +[ok]: SUNSUBSCRIBE from non-subscribed channels (29 ms) +[ok]: PUBSUB command basics (1 ms) +[ok]: SPUBLISH/SSUBSCRIBE with two clients (64 ms) +[ok]: SPUBLISH/SSUBSCRIBE with PUBLISH/SUBSCRIBE (55 ms) +[ok]: PubSubShard with CLIENT REPLY OFF (104 ms) +=== (pubsubshard external:skip) Starting server 127.0.0.1:21211 ok +=== (pubsubshard external:skip) Starting server 127.0.0.1:21213 ok +[ok]: setup replication for following tests (250 ms) +[ok]: publish message to master and receive on replica (81 ms) +[16/124 done]: unit/pubsubshard (2 seconds) +Testing unit/client-eviction +=== () Starting server 127.0.0.1:21215 ok +[ok]: client evicted due to large argv (154 ms) +[ok]: client evicted due to large query buf (94 ms) +[ok]: client evicted due to percentage of maxmemory (78 ms) +[ok]: client evicted due to large multi buf (232 ms) +[ok]: client evicted due to watched key list (972 ms) +[ok]: client evicted due to pubsub subscriptions (6031 ms) +[ok]: client evicted due to tracking redirection (200 ms) +[ok]: client evicted due to client tracking prefixes (9862 ms) +[ok]: client evicted due to output buf (140 ms) +[ok]: client no-evict on (223 ms) +[ok]: client no-evict off (170 ms) +=== () Starting server 127.0.0.1:21217 ok +[ok]: avoid client eviction when client is freed by output buffer limit (567 ms) +=== () Starting server 127.0.0.1:21219 ok +[ok]: decrease maxmemory-clients causes client eviction (507 ms) +=== () Starting server 127.0.0.1:21221 ok +[ok]: evict clients only until below limit (348 ms) +=== () Starting server 127.0.0.1:21223 ok +[ok]: evict clients in right order (large to small) (281 ms) +=== () Starting server 127.0.0.1:21225 ok +[ok]: client total memory grows during client no-evict (140 ms) +[ok]: client total memory grows during maxmemory-clients disabled (137 ms) +[17/124 done]: unit/client-eviction (25 seconds) +Testing unit/limits +=== (limits network external:skip) Starting server 127.0.0.1:21227 ok +[ok]: Check if maxclients works refusing connections (1074 ms) +[18/124 done]: unit/limits (1 seconds) +Testing unit/replybufsize +=== (replybufsize) Starting server 127.0.0.1:21229 ok +[ok]: verify reply buffer limits (695 ms) +[19/124 done]: unit/replybufsize (1 seconds) +Testing unit/maxmemory +=== (maxmemory external:skip) Starting server 127.0.0.1:21231 ok +evicted keys: 23 +evicted clients: 0 +dbsize: 0 +[ok]: eviction due to output buffers of many MGET clients, client eviction: false (446 ms) +evicted keys: 28 +evicted clients: 0 +dbsize: 22 +[ok]: eviction due to input buffer of a dead client, client eviction: false (6143 ms) +evicted keys: 15 +evicted clients: 0 +dbsize: 35 +[ok]: eviction due to output buffers of pubsub, client eviction: false (1240 ms) +evicted keys: 0 +evicted clients: 7 +dbsize: 50 +[ok]: eviction due to output buffers of many MGET clients, client eviction: true (1028 ms) +evicted keys: 0 +evicted clients: 18 +dbsize: 50 +[ok]: eviction due to input buffer of a dead client, client eviction: true (6823 ms) +evicted keys: 0 +evicted clients: 6 +dbsize: 50 +[ok]: eviction due to output buffers of pubsub, client eviction: true (1066 ms) +=== (maxmemory external:skip) Starting server 127.0.0.1:21233 ok +[ok]: Without maxmemory small integers are shared (1 ms) +[ok]: With maxmemory and non-LRU policy integers are still shared (1 ms) +[ok]: With maxmemory and LRU policy integers are not shared (1 ms) +[ok]: Shared integers are unshared with maxmemory and LRU policy (2 ms) +[ok]: maxmemory - is the memory limit honoured? (policy allkeys-random) (361 ms) +[ok]: maxmemory - is the memory limit honoured? (policy allkeys-lru) (716 ms) +[ok]: maxmemory - is the memory limit honoured? (policy allkeys-lfu) (686 ms) +[ok]: maxmemory - is the memory limit honoured? (policy volatile-lru) (896 ms) +[ok]: maxmemory - is the memory limit honoured? (policy volatile-lfu) (1162 ms) +[ok]: maxmemory - is the memory limit honoured? (policy volatile-random) (3473 ms) +[ok]: maxmemory - is the memory limit honoured? (policy volatile-ttl) (4074 ms) +[ok]: maxmemory - only allkeys-* should remove non-volatile keys (allkeys-random) (1188 ms) +[ok]: maxmemory - only allkeys-* should remove non-volatile keys (allkeys-lru) (4304 ms) +[ok]: maxmemory - only allkeys-* should remove non-volatile keys (volatile-lru) (3521 ms) +[ok]: maxmemory - only allkeys-* should remove non-volatile keys (volatile-random) (1359 ms) +[ok]: maxmemory - only allkeys-* should remove non-volatile keys (volatile-ttl) (3546 ms) +[ok]: maxmemory - policy volatile-lru should only remove volatile keys. (5909 ms) +[ok]: maxmemory - policy volatile-lfu should only remove volatile keys. (13631 ms) +[ok]: maxmemory - policy volatile-random should only remove volatile keys. (4766 ms) +[ok]: maxmemory - policy volatile-ttl should only remove volatile keys. (8992 ms) +=== (maxmemory external:skip) Starting server 127.0.0.1:21235 ok +=== () Starting server 127.0.0.1:21237 ok +[ok]: slave buffer are counted correctly (199640 ms) +=== (maxmemory external:skip) Starting server 127.0.0.1:21239 ok +=== () Starting server 127.0.0.1:21241 ok +[ok]: replica buffer don't induce eviction (55513 ms) +=== (maxmemory external:skip) Starting server 127.0.0.1:21243 ok +[ok]: Don't rehash if used memory exceeds maxmemory after rehash (3000 ms) +=== (maxmemory external:skip io-threads:skip) Starting server 127.0.0.1:21245 ok +evicted: 30 +evicted: 30 +[ok]: client tracking don't cause eviction feedback loop (2054 ms) +=== (maxmemory external:skip) Starting server 127.0.0.1:21247 ok +[ok]: propagation with eviction (91 ms) +=== (maxmemory external:skip) Starting server 127.0.0.1:21249 ok +[ok]: propagation with eviction in MULTI (73 ms) +=== (maxmemory external:skip) Starting server 127.0.0.1:21251 ok +[ok]: lru/lfu value of the key just added (12 ms) +[20/124 done]: unit/maxmemory (341 seconds) +Testing unit/tls +=== (tls) Starting server 127.0.0.1:21253 ok +[ok]: TLS: Not accepting non-TLS connections on a TLS port (7 ms) +[ok]: TLS: Verify tls-auth-clients behaves as expected (39 ms) +[ok]: TLS: Verify tls-protocols behaves as expected (19 ms) +[ok]: TLS: Verify tls-ciphers behaves as expected (32 ms) +[ok]: TLS: Verify tls-prefer-server-ciphers behaves as expected (35 ms) +=== () Starting server 127.0.0.1:21255 ok +[ok]: TLS: Verify tls-cert-file is also used as a client cert if none specified (368 ms) +[ok]: TLS: switch between tcp and tls ports (20 ms) +[ok]: TLS: Working with an encrypted keyfile (32 ms) +[21/124 done]: unit/tls (1 seconds) +Testing unit/latency-monitor +=== (latency-monitor needs:latency) Starting server 127.0.0.1:21257 ok +[ok]: LATENCY HISTOGRAM with empty histogram (1 ms) +[ok]: LATENCY HISTOGRAM all commands (2 ms) +[ok]: LATENCY HISTOGRAM sub commands (2 ms) +[ok]: LATENCY HISTOGRAM with a subset of commands (3 ms) +[ok]: LATENCY HISTOGRAM command (1 ms) +[ok]: LATENCY HISTOGRAM with wrong command name skips the invalid one (2 ms) +[ok]: Test latency events logging (5204 ms) +LATENCY HISTORY data: +{1775983554 300} {1775983555 400} {1775983559 500} +[ok]: LATENCY HISTORY output is ok (0 ms) +LATENCY LATEST data: +{command 1775983559 500 500} +[ok]: LATENCY LATEST output is ok (1 ms) +LATENCY GRAPH data: +command - high 500 ms, low 300 ms (all time high 500 ms) +-------------------------------------------------------------------------------- + # + _| + || +_|| + +540 +sss + +[ok]: LATENCY GRAPH can output the event graph (1 ms) +[ok]: LATENCY GRAPH can output the expire event graph (5 ms) +[ok]: LATENCY of expire events are correctly collected (10317 ms) +[ok]: LATENCY HISTORY / RESET with wrong event name is fine (2 ms) +[ok]: LATENCY DOCTOR produces some output (4 ms) +[ok]: LATENCY RESET is able to reset events (11 ms) +[ok]: LATENCY HELP should not have unexpected options (1 ms) +[22/124 done]: unit/latency-monitor (16 seconds) +Testing unit/slowlog +=== (slowlog) Starting server 127.0.0.1:21259 ok +[ok]: SLOWLOG - check that it starts with an empty log (0 ms) +[ok]: SLOWLOG - only logs commands taking more time than specified (202 ms) +[ok]: SLOWLOG - zero max length is correctly handled (56 ms) +[ok]: SLOWLOG - max entries is correctly handled (57 ms) +[ok]: SLOWLOG - GET optional argument to limit output len works (5 ms) +[ok]: SLOWLOG - RESET subcommand works (1 ms) +[ok]: SLOWLOG - logged entry sanity check (211 ms) +[ok]: SLOWLOG - Certain commands are omitted that contain sensitive information (12 ms) +[ok]: SLOWLOG - Some commands can redact sensitive fields (10 ms) +[ok]: SLOWLOG - Rewritten commands are logged as their original command (58 ms) +[ok]: SLOWLOG - commands with too many arguments are trimmed (17 ms) +[ok]: SLOWLOG - too long arguments are trimmed (7 ms) +[ok]: SLOWLOG - EXEC is not logged, just executed commands (209 ms) +[ok]: SLOWLOG - can clean older entries (211 ms) +[ok]: SLOWLOG - can be disabled (419 ms) +[ok]: SLOWLOG - count must be >= -1 (5 ms) +[ok]: SLOWLOG - get all slow logs (11 ms) +[ok]: SLOWLOG - blocking command is reported only after unblocked (77 ms) +[ok]: SLOWLOG - the commands in script are recorded normally - is_eval: 0 (15 ms) +[ok]: SLOWLOG - the commands in script are recorded normally - is_eval: 1 (11 ms) +[23/124 done]: unit/slowlog (2 seconds) +Testing unit/violations +=== () Starting server 127.0.0.1:21261 ok +[ignore]: XADD one huge field: large memory flag not provided +=== () Starting server 127.0.0.1:21263 ok +[ignore]: XADD one huge field - 1: large memory flag not provided +=== () Starting server 127.0.0.1:21265 ok +[ignore]: several XADD big fields: large memory flag not provided +=== () Starting server 127.0.0.1:21267 ok +[ignore]: single XADD big fields: large memory flag not provided +=== () Starting server 127.0.0.1:21269 ok +[ignore]: hash with many big fields: large memory flag not provided +=== () Starting server 127.0.0.1:21271 ok +[ignore]: hash with one huge field: large memory flag not provided +=== () Starting server 127.0.0.1:21273 ok +[ok]: SORT adds integer field to list (5 ms) +[24/124 done]: unit/violations (5 seconds) +Testing unit/introspection-2 +=== (introspection) Starting server 127.0.0.1:21275 ok +[ok]: The microsecond part of the TIME command will not overflow (11 ms) +[ok]: TTL, TYPE and EXISTS do not alter the last access time of a key (3006 ms) +[ok]: TOUCH alters the last access time of a key (3051 ms) +[ok]: TOUCH alters the last access time of a key in no-touch mode (4924 ms) +[ok]: Operations in no-touch mode do not alter the last access time of a key (1106 ms) +[ok]: TOUCH returns the number of existing keys specified (2 ms) +[ok]: command stats for GEOADD (10 ms) +[ok]: errors stats for GEOADD (10 ms) +[ok]: command stats for EXPIRE (5 ms) +[ok]: command stats for BRPOP (4 ms) +[ok]: command stats for MULTI (14 ms) +[ok]: command stats for scripts (9 ms) +[ok]: COMMAND COUNT get total number of commands (0 ms) +[ok]: COMMAND GETKEYS GET (0 ms) +[ok]: COMMAND GETKEYSANDFLAGS (2 ms) +[ok]: COMMAND GETKEYS MEMORY USAGE (1 ms) +[ok]: COMMAND GETKEYS XGROUP (0 ms) +[ok]: COMMAND GETKEYS EVAL with keys (1 ms) +[ok]: COMMAND GETKEYS EVAL without keys (0 ms) +[ok]: COMMAND GETKEYS LCS (1 ms) +[ok]: COMMAND GETKEYS MORE THAN 256 KEYS (5 ms) +[ok]: COMMAND LIST syntax error (0 ms) +[ok]: COMMAND LIST WITHOUT FILTERBY (5 ms) +[ok]: COMMAND LIST FILTERBY ACLCAT against non existing category (0 ms) +[ok]: COMMAND LIST FILTERBY ACLCAT - list all commands/subcommands (1 ms) +[ok]: COMMAND LIST FILTERBY PATTERN - list all commands/subcommands (2 ms) +[ok]: COMMAND LIST FILTERBY MODULE against non existing module (0 ms) +[ok]: COMMAND INFO of invalid subcommands (1 ms) +[ok]: SET command will not be marked with movablekeys (0 ms) +[ok]: GET command will not be marked with movablekeys (0 ms) +[ok]: MSET command will not be marked with movablekeys (0 ms) +[ok]: BITFIELD command will not be marked with movablekeys (1 ms) +[ok]: LMOVE command will not be marked with movablekeys (3 ms) +[ok]: LPOP command will not be marked with movablekeys (2 ms) +[ok]: BLPOP command will not be marked with movablekeys (4 ms) +[ok]: PING command will not be marked with movablekeys (1 ms) +[ok]: MEMORY command will not be marked with movablekeys (9 ms) +[ok]: MEMORY|USAGE command will not be marked with movablekeys (1 ms) +[ok]: RENAME command will not be marked with movablekeys (2 ms) +[ok]: GEORADIUS_RO command will not be marked with movablekeys (6 ms) +[ok]: ZUNIONSTORE command is marked with movablekeys (1 ms) +[ok]: XREAD command is marked with movablekeys (1 ms) +[ok]: EVAL command is marked with movablekeys (1 ms) +[ok]: SORT command is marked with movablekeys (4 ms) +[ok]: SORT_RO command is marked with movablekeys (1 ms) +[ok]: MIGRATE command is marked with movablekeys (2 ms) +[ok]: GEORADIUS command is marked with movablekeys (1 ms) +[25/124 done]: unit/introspection-2 (13 seconds) Testing unit/memefficiency -=== (memefficiency external:skip) Starting server 127.0.0.1:21332 ok -[ok]: Memory efficiency with values in range 32 (1388 ms) -[ok]: Memory efficiency with values in range 64 (728 ms) -[ok]: Memory efficiency with values in range 128 (882 ms) -[ok]: Memory efficiency with values in range 1024 (794 ms) -[ok]: Memory efficiency with values in range 16384 (2429 ms) -=== (defrag external:skip cluster) Starting server 127.0.0.1:21334 ok +=== (memefficiency external:skip) Starting server 127.0.0.1:21277 ok +[ok]: Memory efficiency with values in range 32 (5008 ms) +[ok]: Memory efficiency with values in range 64 (3232 ms) +[ok]: Memory efficiency with values in range 128 (2627 ms) +[ok]: Memory efficiency with values in range 1024 (5351 ms) +[ok]: Memory efficiency with values in range 16384 (9492 ms) +=== (defrag external:skip cluster) Starting server 127.0.0.1:21279 ok frag 1.53 -[ok]: Active defrag main dictionary: cluster (134402 ms) -used 80247024 -rss 81338368 -frag 1.01 -frag_bytes 935696 -used 53559424 -rss 80982016 -frag 1.51 -frag_bytes 27295616 -[ok]: Active defrag eval scripts: cluster (21568 ms) -frag 1.80 -[ok]: Active defrag big keys: cluster (95953 ms) -used 98923920 -rss 99721216 +[ok]: Active defrag main dictionary: cluster (384965 ms) +used 80365472 +rss 81760256 +frag 1.02 +frag_bytes 1222752 +used 53554528 +rss 81281024 +frag 1.52 +frag_bytes 27588768 +[ok]: Active defrag eval scripts: cluster (23814 ms) +frag 1.81 +[ok]: Active defrag big keys: cluster (198476 ms) +used 98936352 +rss 99663872 frag 1.01 -frag_bytes 649840 -used 72412024 -rss 99745792 -frag 1.38 -frag_bytes 27178120 -[ok]: Active defrag pubsub: cluster (167458 ms) -=== (defrag external:skip standalone) Starting server 127.0.0.1:21336 ok +frag_bytes 580064 +used 72427576 +rss 99708928 +frag 1.37 +frag_bytes 27125704 +[ok]: Active defrag pubsub: cluster (285089 ms) +=== (defrag external:skip standalone) Starting server 127.0.0.1:21281 ok frag 1.52 -[ok]: Active defrag main dictionary: standalone (350492 ms) -used 77338088 -rss 78462976 +[ok]: Active defrag main dictionary: standalone (160104 ms) +used 77369096 +rss 78503936 frag 1.01 -frag_bytes 965144 -used 51775464 -rss 77582336 +frag_bytes 966904 +used 51761304 +rss 77594624 frag 1.50 -frag_bytes 25708568 -[ok]: Active defrag eval scripts: standalone (19368 ms) +frag_bytes 25735016 +[ok]: Active defrag eval scripts: standalone (9127 ms) frag 1.86 -[ok]: Active defrag big keys: standalone (85659 ms) -used 94504160 -rss 95739904 +[ok]: Active defrag big keys: standalone (60028 ms) +used 94520016 +rss 95719424 frag 1.01 -frag_bytes 1104672 -used 69282848 -rss 95248384 +frag_bytes 1068336 +used 69268888 +rss 95227904 frag 1.37 -frag_bytes 25825504 -[ok]: Active defrag pubsub: standalone (216912 ms) -frag 1.39 +frag_bytes 25823848 +[ok]: Active defrag pubsub: standalone (204448 ms) +frag 1.46 [err]: Active defrag big list: standalone in tests/unit/memefficiency.tcl -Expected 1.39 >= 1.7 (context: type eval line 40 cmd {assert {$frag >= $expected_frag}} proc ::test) -=== (defrag) Starting server 127.0.0.1:21338 ok -frag 1.95 +Expected 1.46 >= 1.7 (context: type eval line 40 cmd {assert {$frag >= $expected_frag}} proc ::test) +=== (defrag) Starting server 127.0.0.1:21283 ok +frag 1.96 -===== Start of server log (pid 1941717) ===== +===== Start of server log (pid 991151) ===== -[ok]: Active defrag edge case: standalone (101975 ms) +[ok]: Active defrag edge case: standalone (226108 ms) ### Starting server for test -1941717:C 09 Mar 2025 11:31:05.033 # WARNING Memory overcommit must be enabled! Without it, a background save or replication may fail under low memory condition. Being disabled, it can also cause failures without low memory condition, see https://github.com/jemalloc/jemalloc/issues/1328. To fix this issue add 'vm.overcommit_memory = 1' to /etc/sysctl.conf and then reboot or run the command 'sysctl vm.overcommit_memory=1' for this to take effect. -1941717:C 09 Mar 2025 11:31:05.033 * oO0OoO0OoO0Oo Valkey is starting oO0OoO0OoO0Oo -1941717:C 09 Mar 2025 11:31:05.033 * Valkey version=8.0.2, bits=64, commit=00000000, modified=0, pid=1941717, just started -1941717:C 09 Mar 2025 11:31:05.033 * Configuration loaded -1941717:M 09 Mar 2025 11:31:05.034 * Increased maximum number of open files to 10032 (it was originally set to 1024). -1941717:M 09 Mar 2025 11:31:05.034 * monotonic clock: POSIX clock_gettime -1941717:M 09 Mar 2025 11:31:05.034 # Failed to write PID file: Permission denied +991151:C 12 Apr 2026 23:01:50.003 # WARNING Memory overcommit must be enabled! Without it, a background save or replication may fail under low memory condition. Being disabled, it can also cause failures without low memory condition, see https://github.com/jemalloc/jemalloc/issues/1328. To fix this issue add 'vm.overcommit_memory = 1' to /etc/sysctl.conf and then reboot or run the command 'sysctl vm.overcommit_memory=1' for this to take effect. +991151:C 12 Apr 2026 23:01:50.004 * oO0OoO0OoO0Oo Valkey is starting oO0OoO0OoO0Oo +991151:C 12 Apr 2026 23:01:50.004 * Valkey version=8.0.2, bits=64, commit=00000000, modified=0, pid=991151, just started +991151:C 12 Apr 2026 23:01:50.004 * Configuration loaded +991151:M 12 Apr 2026 23:01:50.004 * Increased maximum number of open files to 10032 (it was originally set to 1024). +991151:M 12 Apr 2026 23:01:50.004 * monotonic clock: POSIX clock_gettime +991151:M 12 Apr 2026 23:01:50.005 # Failed to write PID file: Permission denied .+^+. .+#########+. .+########+########+. Valkey 8.0.2 (00000000/0) 64 bit .+########+' '+########+. .########+' .+. '+########. Running in standalone mode - |####+' .+#######+. '+####| Port: 21337 - |###| .+###############+. |###| PID: 1941717 + |####+' .+#######+. '+####| Port: 21282 + |###| .+###############+. |###| PID: 991151 |###| |#####*'' ''*#####| |###| |###| |####' .-. '####| |###| |###| |###( (@@@) )###| |###| https://valkey.io @@ -2250,716 +2206,1245 @@ +#########+' '+v+' -1941717:M 09 Mar 2025 11:31:05.040 * Server initialized -1941717:M 09 Mar 2025 11:31:05.046 * Creating AOF base file appendonly.aof.1.base.rdb on server start -1941717:M 09 Mar 2025 11:31:05.050 * Creating AOF incr file appendonly.aof.1.incr.aof on server start -1941717:M 09 Mar 2025 11:31:05.050 * Ready to accept connections tcp -1941717:M 09 Mar 2025 11:31:05.050 * Ready to accept connections unix -1941717:M 09 Mar 2025 11:31:05.050 * Ready to accept connections tls -1941717:M 09 Mar 2025 11:31:05.166 - Accepted 127.0.0.1:36547 -1941717:M 09 Mar 2025 11:31:05.188 - Reading from client: error:0A000126:SSL routines::unexpected eof while reading -1941717:M 09 Mar 2025 11:31:05.197 - Accepted 127.0.0.1:44515 +991151:M 12 Apr 2026 23:01:50.009 * Server initialized +991151:M 12 Apr 2026 23:01:50.014 * Creating AOF base file appendonly.aof.1.base.rdb on server start +991151:M 12 Apr 2026 23:01:50.032 * Creating AOF incr file appendonly.aof.1.incr.aof on server start +991151:M 12 Apr 2026 23:01:50.032 * Ready to accept connections tcp +991151:M 12 Apr 2026 23:01:50.032 * Ready to accept connections unix +991151:M 12 Apr 2026 23:01:50.032 * Ready to accept connections tls +991151:M 12 Apr 2026 23:01:50.107 - Accepted 127.0.0.1:33057 +991151:M 12 Apr 2026 23:01:50.117 - Reading from client: error:0A000126:SSL routines::unexpected eof while reading +991151:M 12 Apr 2026 23:01:50.122 - Accepted 127.0.0.1:36807 ### Starting test Active defrag main dictionary: standalone in tests/unit/memefficiency.tcl -1941717:M 09 Mar 2025 11:31:10.783 - DB 9: 31024 keys (0 volatile) in 32768 slots HT. -1941717:M 09 Mar 2025 11:31:17.578 - DB 9: 48809 keys (0 volatile) in 65536 slots HT. -1941717:M 09 Mar 2025 11:31:24.735 - DB 9: 65060 keys (0 volatile) in 65536 slots HT. -1941717:M 09 Mar 2025 11:31:31.091 - DB 9: 86593 keys (0 volatile) in 131072 slots HT. -1941717:M 09 Mar 2025 11:31:38.219 - DB 9: 103909 keys (0 volatile) in 131072 slots HT. -1941717:M 09 Mar 2025 11:31:45.127 - DB 9: 118287 keys (0 volatile) in 131072 slots HT. -1941717:M 09 Mar 2025 11:31:51.568 - DB 9: 138523 keys (0 volatile) in 262144 slots HT. -1941717:M 09 Mar 2025 11:31:58.511 - DB 9: 154509 keys (0 volatile) in 262144 slots HT. -1941717:M 09 Mar 2025 11:32:05.371 - DB 9: 168444 keys (0 volatile) in 262144 slots HT. -1941717:M 09 Mar 2025 11:32:12.251 - DB 9: 182274 keys (0 volatile) in 262144 slots HT. -1941717:M 09 Mar 2025 11:32:18.633 - DB 9: 197589 keys (0 volatile) in 262144 slots HT. -1941717:M 09 Mar 2025 11:32:24.911 - DB 9: 215580 keys (0 volatile) in 262144 slots HT. -1941717:M 09 Mar 2025 11:32:31.772 - DB 9: 230509 keys (0 volatile) in 262144 slots HT. -1941717:M 09 Mar 2025 11:32:39.442 - DB 9: 251848 keys (0 volatile) in 262144 slots HT. -1941717:M 09 Mar 2025 11:32:46.719 - DB 9: 266330 keys (0 volatile) in 524288 slots HT. -1941717:M 09 Mar 2025 11:32:53.559 - DB 9: 285455 keys (0 volatile) in 524288 slots HT. -1941717:M 09 Mar 2025 11:33:00.501 - DB 9: 301353 keys (0 volatile) in 524288 slots HT. -1941717:M 09 Mar 2025 11:33:07.511 - DB 9: 320858 keys (0 volatile) in 524288 slots HT. -1941717:M 09 Mar 2025 11:33:14.206 - DB 9: 337754 keys (0 volatile) in 524288 slots HT. -1941717:M 09 Mar 2025 11:33:21.023 - DB 9: 352815 keys (0 volatile) in 524288 slots HT. -1941717:M 09 Mar 2025 11:33:27.835 - DB 9: 369833 keys (0 volatile) in 524288 slots HT. -1941717:M 09 Mar 2025 11:33:34.699 - DB 9: 387033 keys (0 volatile) in 524288 slots HT. -1941717:M 09 Mar 2025 11:33:41.587 - DB 9: 405123 keys (0 volatile) in 524288 slots HT. -1941717:M 09 Mar 2025 11:33:48.919 - DB 9: 420385 keys (0 volatile) in 524288 slots HT. -1941717:M 09 Mar 2025 11:33:55.699 - DB 9: 436440 keys (0 volatile) in 524288 slots HT. -1941717:M 09 Mar 2025 11:34:02.373 - DB 9: 451380 keys (0 volatile) in 524288 slots HT. -1941717:M 09 Mar 2025 11:34:09.506 - DB 9: 466682 keys (0 volatile) in 524288 slots HT. -1941717:M 09 Mar 2025 11:34:16.287 - DB 9: 478717 keys (0 volatile) in 524288 slots HT. -1941717:M 09 Mar 2025 11:34:23.571 - DB 9: 478717 keys (0 volatile) in 524288 slots HT. -1941717:M 09 Mar 2025 11:34:30.455 - DB 9: 478717 keys (0 volatile) in 524288 slots HT. -1941717:M 09 Mar 2025 11:34:37.067 - DB 9: 478717 keys (0 volatile) in 524288 slots HT. -1941717:M 09 Mar 2025 11:34:44.103 - DB 9: 478717 keys (0 volatile) in 524288 slots HT. -1941717:M 09 Mar 2025 11:34:50.692 - DB 9: 478717 keys (0 volatile) in 524288 slots HT. -1941717:M 09 Mar 2025 11:34:57.439 - DB 9: 478717 keys (0 volatile) in 524288 slots HT. -1941717:M 09 Mar 2025 11:35:04.221 - DB 9: 478717 keys (0 volatile) in 524288 slots HT. -1941717:M 09 Mar 2025 11:35:11.060 - DB 9: 478717 keys (0 volatile) in 524288 slots HT. -1941717:M 09 Mar 2025 11:35:17.735 - DB 9: 478717 keys (0 volatile) in 524288 slots HT. -1941717:M 09 Mar 2025 11:35:24.143 - DB 9: 478717 keys (0 volatile) in 524288 slots HT. -1941717:M 09 Mar 2025 11:35:30.677 - DB 9: 478717 keys (0 volatile) in 524288 slots HT. -1941717:M 09 Mar 2025 11:35:36.841 - DB 9: 478717 keys (0 volatile) in 524288 slots HT. -1941717:M 09 Mar 2025 11:35:43.172 - DB 9: 468210 keys (100 volatile) in 524288 slots HT. -1941717:M 09 Mar 2025 11:35:49.435 - DB 9: 451330 keys (100 volatile) in 524288 slots HT. -1941717:M 09 Mar 2025 11:35:56.384 - DB 9: 438724 keys (100 volatile) in 524288 slots HT. -1941717:M 09 Mar 2025 11:36:03.503 - DB 9: 426495 keys (100 volatile) in 524288 slots HT. -1941717:M 09 Mar 2025 11:36:10.407 - DB 9: 413176 keys (100 volatile) in 524288 slots HT. -1941717:M 09 Mar 2025 11:36:17.575 - DB 9: 400634 keys (100 volatile) in 524288 slots HT. -1941717:M 09 Mar 2025 11:36:24.326 - DB 9: 387423 keys (100 volatile) in 524288 slots HT. -1941717:M 09 Mar 2025 11:36:30.835 - DB 9: 373936 keys (100 volatile) in 524288 slots HT. -1941717:M 09 Mar 2025 11:36:37.867 - DB 9: 360632 keys (100 volatile) in 524288 slots HT. -1941717:M 09 Mar 2025 11:36:45.151 - DB 9: 348513 keys (100 volatile) in 524288 slots HT. -1941717:M 09 Mar 2025 11:36:55.713 * DB saved on disk +991151:M 12 Apr 2026 23:01:54.524 * Asynchronous AOF fsync is taking too long (disk is busy?). Writing the AOF buffer without waiting for fsync to complete, this may slow down the server. +991151:M 12 Apr 2026 23:01:57.385 - DB 9: 44725 keys (0 volatile) in 65536 slots HT. +991151:M 12 Apr 2026 23:02:04.924 - DB 9: 88318 keys (0 volatile) in 131072 slots HT. +991151:M 12 Apr 2026 23:02:09.987 - DB 9: 133638 keys (0 volatile) in 393216 slots HT. +991151:M 12 Apr 2026 23:02:16.419 - DB 9: 181468 keys (0 volatile) in 262144 slots HT. +991151:M 12 Apr 2026 23:02:21.490 - DB 9: 227543 keys (0 volatile) in 262144 slots HT. +991151:M 12 Apr 2026 23:02:27.747 - DB 9: 274355 keys (0 volatile) in 524288 slots HT. +991151:M 12 Apr 2026 23:02:34.612 - DB 9: 315513 keys (0 volatile) in 524288 slots HT. +991151:M 12 Apr 2026 23:02:39.663 - DB 9: 361538 keys (0 volatile) in 524288 slots HT. +991151:M 12 Apr 2026 23:02:46.796 - DB 9: 404267 keys (0 volatile) in 524288 slots HT. +991151:M 12 Apr 2026 23:02:52.741 - DB 9: 439781 keys (0 volatile) in 524288 slots HT. +991151:M 12 Apr 2026 23:02:57.869 - DB 9: 478706 keys (0 volatile) in 524288 slots HT. +991151:M 12 Apr 2026 23:03:05.979 - DB 9: 478706 keys (0 volatile) in 524288 slots HT. +991151:M 12 Apr 2026 23:03:13.818 - DB 9: 478706 keys (0 volatile) in 524288 slots HT. +991151:M 12 Apr 2026 23:03:21.961 - DB 9: 478706 keys (0 volatile) in 524288 slots HT. +991151:M 12 Apr 2026 23:03:30.024 - DB 9: 478706 keys (0 volatile) in 524288 slots HT. +991151:M 12 Apr 2026 23:03:35.205 - DB 9: 478706 keys (0 volatile) in 524288 slots HT. +991151:M 12 Apr 2026 23:03:42.811 - DB 9: 478706 keys (0 volatile) in 524288 slots HT. +991151:M 12 Apr 2026 23:03:49.834 - DB 9: 478706 keys (0 volatile) in 524288 slots HT. +991151:M 12 Apr 2026 23:03:57.548 - DB 9: 460263 keys (100 volatile) in 524288 slots HT. +991151:M 12 Apr 2026 23:04:02.825 - DB 9: 436068 keys (100 volatile) in 524288 slots HT. +991151:M 12 Apr 2026 23:04:07.009 * Asynchronous AOF fsync is taking too long (disk is busy?). Writing the AOF buffer without waiting for fsync to complete, this may slow down the server. +991151:M 12 Apr 2026 23:04:09.137 - DB 9: 416722 keys (100 volatile) in 524288 slots HT. +991151:M 12 Apr 2026 23:04:18.194 - DB 9: 385723 keys (100 volatile) in 524288 slots HT. +991151:M 12 Apr 2026 23:04:25.372 - DB 9: 348850 keys (100 volatile) in 524288 slots HT. +991151:M 12 Apr 2026 23:04:30.235 * DB saved on disk ### Starting test Active defrag eval scripts: standalone in tests/unit/memefficiency.tcl -1941717:M 09 Mar 2025 11:36:57.038 - Accepted 127.0.0.1:35125 -1941717:M 09 Mar 2025 11:37:02.675 - DB 9: 35186 keys (0 volatile) in 65536 slots HT. -1941717:M 09 Mar 2025 11:37:08.712 - DB 9: 50000 keys (0 volatile) in 65536 slots HT. -1941717:M 09 Mar 2025 11:37:14.359 - Reading from client: error:0A000126:SSL routines::unexpected eof while reading +991151:M 12 Apr 2026 23:04:30.498 - Accepted 127.0.0.1:38339 +991151:M 12 Apr 2026 23:04:37.429 - DB 9: 32740 keys (0 volatile) in 65536 slots HT. +991151:M 12 Apr 2026 23:04:39.138 - Reading from client: error:0A000126:SSL routines::unexpected eof while reading ### Starting test Active defrag big keys: standalone in tests/unit/memefficiency.tcl -1941717:M 09 Mar 2025 11:37:15.283 - DB 9: 5 keys (0 volatile) in 8 slots HT. -1941717:M 09 Mar 2025 11:37:15.290 - Accepted 127.0.0.1:42081 -1941717:M 09 Mar 2025 11:37:21.467 - DB 9: 10 keys (0 volatile) in 16 slots HT. -1941717:M 09 Mar 2025 11:37:27.583 - DB 9: 55979 keys (0 volatile) in 65536 slots HT. -1941717:M 09 Mar 2025 11:37:33.770 - DB 9: 135082 keys (0 volatile) in 393216 slots HT. -1941717:M 09 Mar 2025 11:37:39.927 - DB 9: 230285 keys (0 volatile) in 262144 slots HT. -1941717:M 09 Mar 2025 11:37:46.115 - DB 9: 356323 keys (0 volatile) in 524288 slots HT. -1941717:M 09 Mar 2025 11:37:52.181 - DB 9: 474095 keys (0 volatile) in 524288 slots HT. -1941717:M 09 Mar 2025 11:37:58.328 - DB 9: 500015 keys (0 volatile) in 524288 slots HT. -1941717:M 09 Mar 2025 11:38:04.407 - DB 9: 500015 keys (0 volatile) in 524288 slots HT. -1941717:M 09 Mar 2025 11:38:09.941 - DB 9: 410391 keys (0 volatile) in 524288 slots HT. -1941717:M 09 Mar 2025 11:38:15.639 - DB 9: 317964 keys (0 volatile) in 524288 slots HT. -1941717:M 09 Mar 2025 11:38:21.579 - DB 9: 250015 keys (0 volatile) in 524288 slots HT. -1941717:M 09 Mar 2025 11:38:27.668 - DB 9: 250015 keys (0 volatile) in 524288 slots HT. -1941717:M 09 Mar 2025 11:38:40.750 * DB saved on disk +991151:M 12 Apr 2026 23:04:39.379 - Accepted 127.0.0.1:33051 +991151:M 12 Apr 2026 23:04:42.576 - DB 9: 12512 keys (0 volatile) in 16384 slots HT. +991151:M 12 Apr 2026 23:04:48.855 - DB 9: 183749 keys (0 volatile) in 262144 slots HT. +991151:M 12 Apr 2026 23:04:54.018 - DB 9: 387869 keys (0 volatile) in 524288 slots HT. +991151:M 12 Apr 2026 23:05:00.217 - DB 9: 500015 keys (0 volatile) in 524288 slots HT. +991151:M 12 Apr 2026 23:05:05.372 - DB 9: 500015 keys (0 volatile) in 524288 slots HT. +991151:M 12 Apr 2026 23:05:11.986 - DB 9: 409424 keys (0 volatile) in 524288 slots HT. +991151:M 12 Apr 2026 23:05:19.716 - DB 9: 291844 keys (0 volatile) in 524288 slots HT. +991151:M 12 Apr 2026 23:05:27.435 - DB 9: 250015 keys (0 volatile) in 524288 slots HT. +991151:M 12 Apr 2026 23:05:39.390 * DB saved on disk ### Starting test Active defrag pubsub: standalone in tests/unit/memefficiency.tcl -1941717:M 09 Mar 2025 11:38:42.330 - Accepted 127.0.0.1:40159 -1941717:M 09 Mar 2025 11:38:42.362 - Accepted 127.0.0.1:36431 -1941717:M 09 Mar 2025 11:38:46.703 - DB 9: 617 keys (0 volatile) in 1024 slots HT. -1941717:M 09 Mar 2025 11:38:52.715 - DB 9: 2977 keys (0 volatile) in 4096 slots HT. -1941717:M 09 Mar 2025 11:38:58.413 - DB 9: 5306 keys (0 volatile) in 8192 slots HT. -1941717:M 09 Mar 2025 11:39:04.050 - DB 9: 9163 keys (0 volatile) in 16384 slots HT. -1941717:M 09 Mar 2025 11:39:09.811 - DB 9: 12094 keys (0 volatile) in 16384 slots HT. -1941717:M 09 Mar 2025 11:39:15.699 - DB 9: 14451 keys (0 volatile) in 16384 slots HT. -1941717:M 09 Mar 2025 11:39:21.334 - DB 9: 18607 keys (0 volatile) in 32768 slots HT. -1941717:M 09 Mar 2025 11:39:26.911 - DB 9: 22403 keys (0 volatile) in 32768 slots HT. -1941717:M 09 Mar 2025 11:39:32.494 - DB 9: 26839 keys (0 volatile) in 32768 slots HT. -1941717:M 09 Mar 2025 11:39:37.954 - DB 9: 31261 keys (0 volatile) in 32768 slots HT. -1941717:M 09 Mar 2025 11:39:43.430 - DB 9: 35097 keys (0 volatile) in 65536 slots HT. -1941717:M 09 Mar 2025 11:39:49.058 - DB 9: 39229 keys (0 volatile) in 65536 slots HT. -1941717:M 09 Mar 2025 11:39:54.696 - DB 9: 43592 keys (0 volatile) in 65536 slots HT. -1941717:M 09 Mar 2025 11:40:00.365 - DB 9: 47536 keys (0 volatile) in 65536 slots HT. -1941717:M 09 Mar 2025 11:40:05.991 - DB 9: 17584 keys (0 volatile) in 65536 slots HT. -1941717:M 09 Mar 2025 11:40:07.947 - Reading from client: error:0A000126:SSL routines::unexpected eof while reading +991151:M 12 Apr 2026 23:05:39.591 - Accepted 127.0.0.1:38599 +991151:M 12 Apr 2026 23:05:39.597 - Accepted 127.0.0.1:38649 +991151:M 12 Apr 2026 23:05:40.403 - DB 9: 1202 keys (0 volatile) in 2048 slots HT. +991151:M 12 Apr 2026 23:05:46.373 - DB 9: 9585 keys (0 volatile) in 16384 slots HT. +991151:M 12 Apr 2026 23:05:51.466 - DB 9: 18704 keys (0 volatile) in 32768 slots HT. +991151:M 12 Apr 2026 23:05:59.183 - DB 9: 25727 keys (0 volatile) in 32768 slots HT. +991151:M 12 Apr 2026 23:06:06.198 - DB 9: 31927 keys (0 volatile) in 32768 slots HT. +991151:M 12 Apr 2026 23:06:14.086 - DB 9: 36955 keys (0 volatile) in 65536 slots HT. +991151:M 12 Apr 2026 23:06:21.716 - DB 9: 44214 keys (0 volatile) in 65536 slots HT. +991151:M 12 Apr 2026 23:06:27.323 - DB 9: 47991 keys (0 volatile) in 65536 slots HT. +991151:M 12 Apr 2026 23:06:35.288 - DB 9: 31850 keys (0 volatile) in 65536 slots HT. +991151:M 12 Apr 2026 23:06:43.993 - Reading from client: error:0A000126:SSL routines::unexpected eof while reading +991151:M 12 Apr 2026 23:09:03.839 - Reading from client: error:0A000126:SSL routines::unexpected eof while reading ### Starting test Active defrag big list: standalone in tests/unit/memefficiency.tcl -1941717:M 09 Mar 2025 11:42:17.670 - Reading from client: error:0A000126:SSL routines::unexpected eof while reading -1941717:M 09 Mar 2025 11:42:17.677 - Accepted 127.0.0.1:39253 -1941717:M 09 Mar 2025 11:42:23.035 - DB 9: 2 keys (0 volatile) in 4 slots HT. -1941717:M 09 Mar 2025 11:42:29.351 - DB 9: 2 keys (0 volatile) in 4 slots HT. -1941717:M 09 Mar 2025 11:42:35.651 - DB 9: 2 keys (0 volatile) in 4 slots HT. -1941717:M 09 Mar 2025 11:42:41.858 - DB 9: 2 keys (0 volatile) in 4 slots HT. -1941717:M 09 Mar 2025 11:42:47.867 - DB 9: 2 keys (0 volatile) in 4 slots HT. -1941717:M 09 Mar 2025 11:42:53.989 - DB 9: 2 keys (0 volatile) in 4 slots HT. -1941717:M 09 Mar 2025 11:43:00.067 - DB 9: 2 keys (0 volatile) in 4 slots HT. -1941717:M 09 Mar 2025 11:43:06.407 - DB 9: 2 keys (0 volatile) in 4 slots HT. -1941717:M 09 Mar 2025 11:43:12.387 - DB 9: 2 keys (0 volatile) in 4 slots HT. -1941717:M 09 Mar 2025 11:43:18.447 - DB 9: 2 keys (0 volatile) in 4 slots HT. -1941717:M 09 Mar 2025 11:43:24.527 - DB 9: 2 keys (0 volatile) in 4 slots HT. -1941717:M 09 Mar 2025 11:43:30.611 - DB 9: 2 keys (0 volatile) in 4 slots HT. -1941717:M 09 Mar 2025 11:43:36.663 - DB 9: 2 keys (0 volatile) in 4 slots HT. +991151:M 12 Apr 2026 23:09:03.845 - Accepted 127.0.0.1:39435 +991151:M 12 Apr 2026 23:09:07.424 - DB 9: 2 keys (0 volatile) in 4 slots HT. +991151:M 12 Apr 2026 23:09:16.549 - DB 9: 2 keys (0 volatile) in 4 slots HT. +991151:M 12 Apr 2026 23:09:25.345 - DB 9: 2 keys (0 volatile) in 4 slots HT. +991151:M 12 Apr 2026 23:09:35.105 - DB 9: 2 keys (0 volatile) in 4 slots HT. +991151:M 12 Apr 2026 23:09:44.661 - DB 9: 2 keys (0 volatile) in 4 slots HT. +991151:M 12 Apr 2026 23:09:53.537 - DB 9: 2 keys (0 volatile) in 4 slots HT. +991151:M 12 Apr 2026 23:10:02.389 - DB 9: 2 keys (0 volatile) in 4 slots HT. +991151:M 12 Apr 2026 23:10:12.137 - DB 9: 2 keys (0 volatile) in 4 slots HT. +991151:M 12 Apr 2026 23:10:23.033 - DB 9: 2 keys (0 volatile) in 4 slots HT. +991151:M 12 Apr 2026 23:10:31.673 - DB 9: 2 keys (0 volatile) in 4 slots HT. +991151:M 12 Apr 2026 23:10:39.373 - DB 9: 2 keys (0 volatile) in 4 slots HT. +991151:M 12 Apr 2026 23:10:48.554 - DB 9: 2 keys (0 volatile) in 4 slots HT. +991151:M 12 Apr 2026 23:10:59.061 - DB 9: 2 keys (0 volatile) in 4 slots HT. +991151:M 12 Apr 2026 23:11:08.329 - DB 9: 2 keys (0 volatile) in 4 slots HT. +991151:M 12 Apr 2026 23:11:20.641 - DB 9: 2 keys (0 volatile) in 4 slots HT. +991151:M 12 Apr 2026 23:11:31.395 - DB 9: 2 keys (0 volatile) in 4 slots HT. +991151:M 12 Apr 2026 23:11:40.702 - DB 9: 2 keys (0 volatile) in 4 slots HT. +991151:M 12 Apr 2026 23:11:47.769 - DB 9: 2 keys (0 volatile) in 4 slots HT. +991151:M 12 Apr 2026 23:11:56.601 - DB 9: 2 keys (0 volatile) in 4 slots HT. +991151:M 12 Apr 2026 23:12:05.516 - DB 9: 2 keys (0 volatile) in 4 slots HT. +991151:M 12 Apr 2026 23:12:14.565 - DB 9: 2 keys (0 volatile) in 4 slots HT. +991151:M 12 Apr 2026 23:12:25.261 - DB 9: 2 keys (0 volatile) in 4 slots HT. +991151:M 12 Apr 2026 23:12:32.273 - DB 9: 2 keys (0 volatile) in 4 slots HT. +991151:M 12 Apr 2026 23:12:39.420 - DB 9: 2 keys (0 volatile) in 4 slots HT. +991151:M 12 Apr 2026 23:12:54.337 - DB 9: 2 keys (0 volatile) in 4 slots HT. +991151:M 12 Apr 2026 23:13:12.039 - DB 9: 2 keys (0 volatile) in 4 slots HT. +991151:M 12 Apr 2026 23:13:18.557 - DB 9: 2 keys (0 volatile) in 4 slots HT. +991151:M 12 Apr 2026 23:13:23.953 - DB 9: 2 keys (0 volatile) in 4 slots HT. +991151:M 12 Apr 2026 23:13:29.096 - DB 9: 2 keys (0 volatile) in 4 slots HT. ### Starting test Active defrag edge case: standalone in tests/unit/memefficiency.tcl -1941717:M 09 Mar 2025 11:43:42.719 - DB 9: 1 keys (0 volatile) in 4 slots HT. -1941717:M 09 Mar 2025 11:43:48.797 - DB 9: 1 keys (0 volatile) in 4 slots HT. -1941717:M 09 Mar 2025 11:43:54.903 - DB 9: 1 keys (0 volatile) in 4 slots HT. -1941717:M 09 Mar 2025 11:44:00.975 - DB 9: 1 keys (0 volatile) in 4 slots HT. -1941717:M 09 Mar 2025 11:44:07.023 - DB 9: 1 keys (0 volatile) in 4 slots HT. -1941717:M 09 Mar 2025 11:44:13.043 - DB 9: 1 keys (0 volatile) in 4 slots HT. -1941717:M 09 Mar 2025 11:44:19.059 - DB 9: 1 keys (0 volatile) in 4 slots HT. -1941717:M 09 Mar 2025 11:44:25.147 - DB 9: 1 keys (0 volatile) in 4 slots HT. -1941717:M 09 Mar 2025 11:44:31.223 - DB 9: 1 keys (0 volatile) in 4 slots HT. -1941717:M 09 Mar 2025 11:44:37.335 - DB 9: 1 keys (0 volatile) in 4 slots HT. -1941717:M 09 Mar 2025 11:44:43.399 - DB 9: 1 keys (0 volatile) in 4 slots HT. -1941717:M 09 Mar 2025 11:44:49.464 - DB 9: 1 keys (0 volatile) in 4 slots HT. -1941717:M 09 Mar 2025 11:44:55.515 - DB 9: 1 keys (0 volatile) in 4 slots HT. -1941717:M 09 Mar 2025 11:45:01.591 - DB 9: 1 keys (0 volatile) in 4 slots HT. -1941717:M 09 Mar 2025 11:45:07.533 - DB 9: 1 keys (0 volatile) in 4 slots HT. -1941717:M 09 Mar 2025 11:45:13.619 - DB 9: 1 keys (0 volatile) in 4 slots HT. -1941717:M 09 Mar 2025 11:45:19.640 - DB 9: 1 keys (0 volatile) in 4 slots HT. -===== End of server log (pid 1941717) ===== +991151:M 12 Apr 2026 23:13:35.486 - DB 9: 1 keys (0 volatile) in 4 slots HT. +991151:M 12 Apr 2026 23:13:42.806 - DB 9: 1 keys (0 volatile) in 4 slots HT. +991151:M 12 Apr 2026 23:13:47.899 - DB 9: 1 keys (0 volatile) in 4 slots HT. +991151:M 12 Apr 2026 23:13:55.187 - DB 9: 1 keys (0 volatile) in 4 slots HT. +991151:M 12 Apr 2026 23:14:00.289 - DB 9: 1 keys (0 volatile) in 4 slots HT. +991151:M 12 Apr 2026 23:14:05.349 - DB 9: 1 keys (0 volatile) in 4 slots HT. +991151:M 12 Apr 2026 23:14:11.637 - DB 9: 1 keys (0 volatile) in 4 slots HT. +991151:M 12 Apr 2026 23:14:16.690 - DB 9: 1 keys (0 volatile) in 4 slots HT. +991151:M 12 Apr 2026 23:14:21.740 - DB 9: 1 keys (0 volatile) in 4 slots HT. +991151:M 12 Apr 2026 23:14:26.794 - DB 9: 1 keys (0 volatile) in 4 slots HT. +991151:M 12 Apr 2026 23:14:31.846 - DB 9: 1 keys (0 volatile) in 4 slots HT. +991151:M 12 Apr 2026 23:14:36.937 - DB 9: 1 keys (0 volatile) in 4 slots HT. +991151:M 12 Apr 2026 23:14:42.227 - DB 9: 1 keys (0 volatile) in 4 slots HT. +991151:M 12 Apr 2026 23:14:47.387 - DB 9: 1 keys (0 volatile) in 4 slots HT. +991151:M 12 Apr 2026 23:14:52.440 - DB 9: 1 keys (0 volatile) in 4 slots HT. +991151:M 12 Apr 2026 23:14:57.490 - DB 9: 1 keys (0 volatile) in 4 slots HT. +991151:M 12 Apr 2026 23:15:02.542 - DB 9: 1 keys (0 volatile) in 4 slots HT. +991151:M 12 Apr 2026 23:15:07.591 - DB 9: 1 keys (0 volatile) in 4 slots HT. +991151:M 12 Apr 2026 23:15:12.642 - DB 9: 1 keys (0 volatile) in 4 slots HT. +991151:M 12 Apr 2026 23:15:17.699 - DB 9: 1 keys (0 volatile) in 4 slots HT. +991151:M 12 Apr 2026 23:15:23.030 - DB 9: 1 keys (0 volatile) in 4 slots HT. +991151:M 12 Apr 2026 23:15:28.210 - DB 9: 1 keys (0 volatile) in 4 slots HT. +991151:M 12 Apr 2026 23:15:33.662 - DB 9: 1 keys (0 volatile) in 4 slots HT. +991151:M 12 Apr 2026 23:15:39.148 - DB 9: 1 keys (0 volatile) in 4 slots HT. +991151:M 12 Apr 2026 23:15:44.354 - DB 9: 1 keys (0 volatile) in 4 slots HT. +991151:M 12 Apr 2026 23:15:49.600 - DB 9: 1 keys (0 volatile) in 4 slots HT. +991151:M 12 Apr 2026 23:15:58.755 - DB 9: 1 keys (0 volatile) in 4 slots HT. +991151:M 12 Apr 2026 23:16:03.805 - DB 9: 1 keys (0 volatile) in 4 slots HT. +991151:M 12 Apr 2026 23:16:08.854 - DB 9: 1 keys (0 volatile) in 4 slots HT. +991151:M 12 Apr 2026 23:16:15.954 - DB 9: 1 keys (0 volatile) in 4 slots HT. +991151:M 12 Apr 2026 23:16:21.272 - DB 9: 1 keys (0 volatile) in 4 slots HT. +991151:M 12 Apr 2026 23:16:26.597 - DB 9: 1 keys (0 volatile) in 4 slots HT. +991151:M 12 Apr 2026 23:16:31.921 - DB 9: 1 keys (0 volatile) in 4 slots HT. +991151:M 12 Apr 2026 23:16:37.314 - DB 9: 1 keys (0 volatile) in 4 slots HT. +991151:M 12 Apr 2026 23:16:42.512 - DB 9: 1 keys (0 volatile) in 4 slots HT. +991151:M 12 Apr 2026 23:16:47.566 - DB 9: 1 keys (0 volatile) in 4 slots HT. +991151:M 12 Apr 2026 23:16:52.617 - DB 9: 1 keys (0 volatile) in 4 slots HT. +991151:M 12 Apr 2026 23:16:57.669 - DB 9: 1 keys (0 volatile) in 4 slots HT. +991151:M 12 Apr 2026 23:17:02.721 - DB 9: 1 keys (0 volatile) in 4 slots HT. +991151:M 12 Apr 2026 23:17:10.779 - DB 9: 1 keys (0 volatile) in 4 slots HT. +991151:M 12 Apr 2026 23:17:16.693 - DB 9: 1 keys (0 volatile) in 4 slots HT. +===== End of server log (pid 991151) ===== -===== Start of server stderr log (pid 1941717) ===== +===== Start of server stderr log (pid 991151) ===== -===== End of server stderr log (pid 1941717) ===== +===== End of server stderr log (pid 991151) ===== -[31/124 done]: unit/memefficiency (1287 seconds) +[26/124 done]: unit/memefficiency (1848 seconds) +Testing unit/protocol +=== (protocol network) Starting server 127.0.0.1:21285 ok +[ok]: Handle an empty query (75 ms) +[ok]: Negative multibulk length (72 ms) +[ok]: Out of range multibulk length (43 ms) +[ok]: Wrong multibulk payload header (37 ms) +[ok]: Negative multibulk payload length (63 ms) +[ok]: Out of range multibulk payload length (57 ms) +[ok]: Non-number multibulk payload length (58 ms) +[ok]: Multi bulk request not followed by bulk arguments (55 ms) +[ok]: Generic wrong number of args (27 ms) +[ok]: Unbalanced number of quotes (25 ms) +[ok]: Protocol desync regression test #1 (327 ms) +[ok]: Protocol desync regression test #2 (306 ms) +[ok]: Protocol desync regression test #3 (255 ms) +[ok]: raw protocol response (5 ms) +[ok]: raw protocol response - deferred (1 ms) +[ok]: raw protocol response - multiline (2 ms) +[ok]: RESP3 attributes (3 ms) +[ok]: RESP3 attributes readraw (1 ms) +[ok]: RESP3 attributes on RESP2 (1 ms) +[ok]: test big number parsing (1 ms) +[ok]: test bool parsing (3 ms) +[ok]: test verbatim str parsing (4 ms) +[ok]: test large number of args (118 ms) +[ok]: test argument rewriting - issue 9598 (3 ms) +=== (regression) Starting server 127.0.0.1:21287 ok +[ok]: Regression for a crash with blocking ops and pipelining (58 ms) +[27/124 done]: unit/protocol (3 seconds) +Testing unit/acl-v2 +=== (acl external:skip) Starting server 127.0.0.1:21289 ok +[ok]: Test basic multiple selectors (3 ms) +[ok]: Test ACL selectors by default have no permissions (3 ms) +[ok]: Test deleting selectors (2 ms) +[ok]: Test selector syntax error reports the error in the selector context (3 ms) +[ok]: Test flexible selector definition (3 ms) +[ok]: Test separate read permission (6 ms) +[ok]: Test separate write permission (2 ms) +[ok]: Test separate read and write permissions (1 ms) +[ok]: Validate read and write permissions format - empty permission (1 ms) +[ok]: Validate read and write permissions format - empty selector (0 ms) +[ok]: Validate read and write permissions format - empty pattern (1 ms) +[ok]: Validate read and write permissions format - no pattern (1 ms) +[ok]: Test separate read and write permissions on different selectors are not additive (3 ms) +[ok]: Test SET with separate read permission (5 ms) +[ok]: Test SET with separate write permission (3 ms) +[ok]: Test SET with read and write permissions (5 ms) +[ok]: Test BITFIELD with separate read permission (3 ms) +[ok]: Test BITFIELD with separate write permission (1 ms) +[ok]: Test BITFIELD with read and write permissions (2 ms) +[ok]: Test ACL log correctly identifies the relevant item when selectors are used (4 ms) +[ok]: Test ACL GETUSER response information (2 ms) +[ok]: Test ACL list idempotency (7 ms) +[ok]: Test R+W is the same as all permissions (1 ms) +[ok]: Test basic dry run functionality (1 ms) +[ok]: Test various commands for command permissions (0 ms) +[ok]: Test various odd commands for key permissions (15 ms) +[ok]: Existence test commands are not marked as access (2 ms) +[ok]: Intersection cardinaltiy commands are access commands (2 ms) +[ok]: Test general keyspace commands require some type of permission to execute (3 ms) +[ok]: Cardinality commands require some type of permission to execute (5 ms) +[ok]: Test sharded channel permissions (2 ms) +[ok]: Test sort with ACL permissions (4 ms) +[ok]: Test DRYRUN with wrong number of arguments (1 ms) +=== (external:skip) Starting server 127.0.0.1:21291 ok +[ok]: Test behavior of loading ACLs (1 ms) +[28/124 done]: unit/acl-v2 (1 seconds) Testing unit/acl -=== (acl external:skip) Starting server 127.0.0.1:21340 ok -[ok]: Connections start with the default user (6 ms) -[ok]: It is possible to create new users (4 ms) -[ok]: Coverage: ACL USERS (8 ms) -[ok]: Usernames can not contain spaces or null characters (6 ms) -[ok]: New users start disabled (11 ms) -[ok]: Enabling the user allows the login (15 ms) -[ok]: Only the set of correct passwords work (32 ms) -[ok]: It is possible to remove passwords from the set of valid ones (20 ms) -[ok]: Test password hashes can be added (11 ms) -[ok]: Test password hashes validate input (10 ms) -[ok]: ACL GETUSER returns the password hash instead of the actual password (14 ms) -[ok]: Test hashed passwords removal (16 ms) -[ok]: By default users are not able to access any command (4 ms) -[ok]: By default users are not able to access any key (17 ms) -[ok]: It's possible to allow the access of a subset of keys (39 ms) -[ok]: By default, only default user is able to publish to any channel (36 ms) -[ok]: By default, only default user is not able to publish to any shard channel (32 ms) -[ok]: By default, only default user is able to subscribe to any channel (64 ms) -[ok]: By default, only default user is able to subscribe to any shard channel (61 ms) -[ok]: By default, only default user is able to subscribe to any pattern (62 ms) -[ok]: It's possible to allow publishing to a subset of channels (30 ms) -[ok]: It's possible to allow publishing to a subset of shard channels (24 ms) -[ok]: Validate subset of channels is prefixed with resetchannels flag (55 ms) -[ok]: In transaction queue publish/subscribe/psubscribe to unauthorized channel will fail (79 ms) -[ok]: It's possible to allow subscribing to a subset of channels (64 ms) -[ok]: It's possible to allow subscribing to a subset of shard channels (67 ms) -[ok]: It's possible to allow subscribing to a subset of channel patterns (68 ms) -[ok]: Subscribers are killed when revoked of channel permission (80 ms) -[ok]: Subscribers are killed when revoked of channel permission (64 ms) -[ok]: Subscribers are killed when revoked of pattern permission (68 ms) -[ok]: Subscribers are killed when revoked of allchannels permission (68 ms) -[ok]: Subscribers are pardoned if literal permissions are retained and/or gaining allchannels (92 ms) -[ok]: blocked command gets rejected when reprocessed after permission change (96 ms) -[ok]: Users can be configured to authenticate with any password (15 ms) -[ok]: ACLs can exclude single commands (20 ms) -[ok]: ACLs can include or exclude whole classes of commands (44 ms) -[ok]: ACLs can include single subcommands (40 ms) -[ok]: ACLs can exclude single subcommands, case 1 (36 ms) -[ok]: ACLs can exclude single subcommands, case 2 (13 ms) -[ok]: ACLs cannot include a subcommand with a specific arg (11 ms) +=== (acl external:skip) Starting server 127.0.0.1:21293 ok +[ok]: Connections start with the default user (0 ms) +[ok]: It is possible to create new users (0 ms) +[ok]: Coverage: ACL USERS (0 ms) +[ok]: Usernames can not contain spaces or null characters (1 ms) +[ok]: New users start disabled (0 ms) +[ok]: Enabling the user allows the login (4 ms) +[ok]: Only the set of correct passwords work (1 ms) +[ok]: It is possible to remove passwords from the set of valid ones (0 ms) +[ok]: Test password hashes can be added (2 ms) +[ok]: Test password hashes validate input (11 ms) +[ok]: ACL GETUSER returns the password hash instead of the actual password (8 ms) +[ok]: Test hashed passwords removal (6 ms) +[ok]: By default users are not able to access any command (1 ms) +[ok]: By default users are not able to access any key (1 ms) +[ok]: It's possible to allow the access of a subset of keys (1 ms) +[ok]: By default, only default user is able to publish to any channel (3 ms) +[ok]: By default, only default user is not able to publish to any shard channel (11 ms) +[ok]: By default, only default user is able to subscribe to any channel (29 ms) +[ok]: By default, only default user is able to subscribe to any shard channel (37 ms) +[ok]: By default, only default user is able to subscribe to any pattern (32 ms) +[ok]: It's possible to allow publishing to a subset of channels (2 ms) +[ok]: It's possible to allow publishing to a subset of shard channels (1 ms) +[ok]: Validate subset of channels is prefixed with resetchannels flag (6 ms) +[ok]: In transaction queue publish/subscribe/psubscribe to unauthorized channel will fail (8 ms) +[ok]: It's possible to allow subscribing to a subset of channels (25 ms) +[ok]: It's possible to allow subscribing to a subset of shard channels (18 ms) +[ok]: It's possible to allow subscribing to a subset of channel patterns (18 ms) +[ok]: Subscribers are killed when revoked of channel permission (22 ms) +[ok]: Subscribers are killed when revoked of channel permission (28 ms) +[ok]: Subscribers are killed when revoked of pattern permission (16 ms) +[ok]: Subscribers are killed when revoked of allchannels permission (14 ms) +[ok]: Subscribers are pardoned if literal permissions are retained and/or gaining allchannels (26 ms) +[ok]: blocked command gets rejected when reprocessed after permission change (47 ms) +[ok]: Users can be configured to authenticate with any password (2 ms) +[ok]: ACLs can exclude single commands (1 ms) +[ok]: ACLs can include or exclude whole classes of commands (1 ms) +[ok]: ACLs can include single subcommands (2 ms) +[ok]: ACLs can exclude single subcommands, case 1 (1 ms) +[ok]: ACLs can exclude single subcommands, case 2 (2 ms) +[ok]: ACLs cannot include a subcommand with a specific arg (1 ms) [ok]: ACLs cannot exclude or include a container commands with a specific arg (0 ms) [ok]: ACLs cannot exclude or include a container command with two args (1 ms) [ok]: ACLs including of a type includes also subcommands (2 ms) -[ok]: ACLs can block SELECT of all but a specific DB (2 ms) -[ok]: ACLs can block all DEBUG subcommands except one (2 ms) -[ok]: ACLs set can include subcommands, if already full command exists (2 ms) -[ok]: ACLs set can exclude subcommands, if already full command exists (8 ms) -[ok]: ACL SETUSER RESET reverting to default newly created user (2 ms) -[ok]: ACL GETUSER is able to translate back command permissions (2 ms) +[ok]: ACLs can block SELECT of all but a specific DB (1 ms) +[ok]: ACLs can block all DEBUG subcommands except one (1 ms) +[ok]: ACLs set can include subcommands, if already full command exists (13 ms) +[ok]: ACLs set can exclude subcommands, if already full command exists (13 ms) +[ok]: ACL SETUSER RESET reverting to default newly created user (4 ms) +[ok]: ACL GETUSER is able to translate back command permissions (5 ms) [ok]: ACL GETUSER provides reasonable results (41 ms) -[ok]: ACL GETUSER provides correct results (15 ms) +[ok]: ACL GETUSER provides correct results (9 ms) [ok]: ACL CAT with illegal arguments (1 ms) -[ok]: ACL CAT without category - list all categories (0 ms) -[ok]: ACL CAT category - list all commands/subcommands that belong to category (3 ms) -[ok]: ACL requires explicit permission for scripting for EVAL_RO, EVALSHA_RO and FCALL_RO (3 ms) +[ok]: ACL CAT without category - list all categories (1 ms) +[ok]: ACL CAT category - list all commands/subcommands that belong to category (2 ms) +[ok]: ACL requires explicit permission for scripting for EVAL_RO, EVALSHA_RO and FCALL_RO (1 ms) [ok]: ACL #5998 regression: memory leaks adding / removing subcommands (1 ms) -[ok]: ACL LOG aggregates similar errors together and assigns unique entry-id to new errors (12 ms) +[ok]: ACL LOG aggregates similar errors together and assigns unique entry-id to new errors (16 ms) [ok]: ACL LOG shows failed command executions at toplevel (3 ms) -[ok]: ACL LOG shows failed subcommand executions at toplevel (5 ms) -[ok]: ACL LOG is able to test similar events (52 ms) -[ok]: ACL LOG is able to log keys access violations and key name (35 ms) -[ok]: ACL LOG is able to log channel access violations and channel name (13 ms) -[ok]: ACL LOG RESET is able to flush the entries in the log (9 ms) -[ok]: ACL LOG can distinguish the transaction context (1) (24 ms) -[ok]: ACL LOG can distinguish the transaction context (2) (93 ms) -[ok]: ACL can log errors in the context of Lua scripting (26 ms) -[ok]: ACL LOG can accept a numerical argument to show less entries (62 ms) -[ok]: ACL LOG can log failed auth attempts (11 ms) -[ok]: ACLLOG - zero max length is correctly handled (83 ms) -[ok]: ACL LOG entries are limited to a maximum amount (106 ms) -[ok]: ACL LOG entries are still present on update of max len config (14 ms) -[ok]: When default user is off, new connections are not authenticated (37 ms) -[ok]: When default user has no command permission, hello command still works for other users (34 ms) -[ok]: When an authentication chain is used in the HELLO cmd, the last auth cmd has precedence (31 ms) -[ok]: When a setname chain is used in the HELLO cmd, the last setname cmd has precedence (29 ms) -[ok]: When authentication fails in the HELLO cmd, the client setname should not be applied (22 ms) -[ok]: ACL HELP should not have unexpected options (8 ms) -[ok]: Delete a user that the client doesn't use (19 ms) -[ok]: Delete a user that the client is using (18 ms) -[ok]: ACL GENPASS command failed test (34 ms) -[ok]: Default user can not be removed (4 ms) -[ok]: ACL load non-existing configured ACL file (7 ms) -[ok]: ACL-Metrics user AUTH failure (86 ms) -[ok]: ACL-Metrics invalid command accesses (71 ms) -[ok]: ACL-Metrics invalid key accesses (82 ms) -[ok]: ACL-Metrics invalid channels accesses (90 ms) -=== (external:skip) Starting server 127.0.0.1:21342 ok -[ok]: default: load from include file, can access any channels (36 ms) -[ok]: default: with config acl-pubsub-default allchannels after reset, can access any channels (33 ms) -[ok]: default: with config acl-pubsub-default resetchannels after reset, can not access any channels (31 ms) -[ok]: Alice: can execute all command (13 ms) -[ok]: Bob: just execute @set and acl command (17 ms) -[ok]: ACL LOAD only disconnects affected clients (125 ms) -[ok]: ACL LOAD disconnects clients of deleted users (134 ms) -[ok]: ACL load and save (87 ms) -[ok]: ACL load and save with restricted channels (83 ms) -=== (external:skip) Starting server 127.0.0.1:21344 ok -[ok]: Default user has access to all channels irrespective of flag (9 ms) -[ok]: Update acl-pubsub-default, existing users shouldn't get affected (44 ms) -[ok]: Single channel is valid (21 ms) -[ok]: Single channel is not valid with allchannels (12 ms) -=== (external:skip) Starting server 127.0.0.1:21346 ok -[ok]: Only default user has access to all channels irrespective of flag (19 ms) -=== (external:skip) Starting server 127.0.0.1:21348 ok -[ok]: default: load from config file, without channel permission default user can't access any channels (8 ms) -=== (external:skip) Starting server 127.0.0.1:21350 ok -[ok]: default: load from config file with all channels permissions (23 ms) -=== (external:skip) Starting server 127.0.0.1:21352 ok -[ok]: Test loading an ACL file with duplicate users (49 ms) -[ok]: Test loading an ACL file with duplicate default user (33 ms) -[ok]: Test loading duplicate users in config on startup (123 ms) -=== (acl external:skip) Starting server 127.0.0.1:21354 ok -[ok]: ACL from config file and config rewrite (424 ms) -[32/124 done]: unit/acl (9 seconds) -Testing unit/tls -=== (tls) Starting server 127.0.0.1:21356 ok -[ok]: TLS: Not accepting non-TLS connections on a TLS port (4 ms) -[ok]: TLS: Verify tls-auth-clients behaves as expected (99 ms) -[ok]: TLS: Verify tls-protocols behaves as expected (61 ms) -[ok]: TLS: Verify tls-ciphers behaves as expected (136 ms) -[ok]: TLS: Verify tls-prefer-server-ciphers behaves as expected (163 ms) -=== () Starting server 127.0.0.1:21358 ok -[ok]: TLS: Verify tls-cert-file is also used as a client cert if none specified (640 ms) -[ok]: TLS: switch between tcp and tls ports (80 ms) -[ok]: TLS: Working with an encrypted keyfile (79 ms) -[33/124 done]: unit/tls (2 seconds) -Testing unit/multi -=== (multi) Starting server 127.0.0.1:21360 ok -[ok]: MULTI / EXEC basics (35 ms) -[ok]: DISCARD (45 ms) -[ok]: Nested MULTI are not allowed (16 ms) -[ok]: MULTI where commands alter argc/argv (20 ms) -[ok]: WATCH inside MULTI is not allowed (12 ms) -[ok]: EXEC fails if there are errors while queueing commands #1 (29 ms) -[ok]: EXEC fails if there are errors while queueing commands #2 (107 ms) -[ok]: If EXEC aborts, the client MULTI state is cleared (39 ms) -[ok]: EXEC works on WATCHed key not modified (23 ms) -[ok]: EXEC fail on WATCHed key modified (1 key of 1 watched) (32 ms) -[ok]: EXEC fail on WATCHed key modified (1 key of 5 watched) (43 ms) -[ok]: EXEC fail on WATCHed key modified by SORT with STORE even if the result is empty (35 ms) -EXEC fail on lazy expired WATCHed key attempts: 0 -[ok]: EXEC fail on lazy expired WATCHed key (163 ms) -[ok]: WATCH stale keys should not fail EXEC (69 ms) -[ok]: Delete WATCHed stale keys should not fail EXEC (44 ms) -[ok]: FLUSHDB while watching stale keys should not fail EXEC (57 ms) -[ok]: After successful EXEC key is no longer watched (48 ms) -[ok]: After failed EXEC key is no longer watched (71 ms) -[ok]: It is possible to UNWATCH (51 ms) -[ok]: UNWATCH when there is nothing watched works as expected (4 ms) -[ok]: FLUSHALL is able to touch the watched keys (32 ms) -[ok]: FLUSHALL does not touch non affected keys (57 ms) -[ok]: FLUSHDB is able to touch the watched keys (43 ms) -[ok]: FLUSHDB does not touch non affected keys (35 ms) -[ok]: SWAPDB is able to touch the watched keys that exist (59 ms) -[ok]: SWAPDB is able to touch the watched keys that do not exist (59 ms) -[ok]: SWAPDB does not touch watched stale keys (76 ms) -[ok]: SWAPDB does not touch non-existing key replaced with stale key (76 ms) -[ok]: SWAPDB does not touch stale key replaced with another stale key (97 ms) -[ok]: WATCH is able to remember the DB a key belongs to (58 ms) -[ok]: WATCH will consider touched keys target of EXPIRE (53 ms) -[ok]: WATCH will consider touched expired keys (1276 ms) -[ok]: DISCARD should clear the WATCH dirty flag on the client (49 ms) -[ok]: DISCARD should UNWATCH all the keys (55 ms) -[ok]: MULTI / EXEC is not propagated (single write command) (125 ms) -[ok]: MULTI / EXEC is propagated correctly (multiple commands) (213 ms) -[ok]: MULTI / EXEC is propagated correctly (multiple commands with SELECT) (215 ms) -[ok]: MULTI / EXEC is propagated correctly (empty transaction) (167 ms) -[ok]: MULTI / EXEC is propagated correctly (read-only commands) (112 ms) -[ok]: MULTI / EXEC is propagated correctly (write command, no effect) (206 ms) -[ok]: MULTI / EXEC with REPLICAOF (131 ms) -[ok]: DISCARD should not fail during OOM (63 ms) -[ok]: MULTI and script timeout (520 ms) -[ok]: EXEC and script timeout (503 ms) -[ok]: MULTI-EXEC body and script timeout (516 ms) -[ok]: just EXEC and script timeout (519 ms) -[ok]: exec with write commands and state change (83 ms) -[ok]: exec with read commands and stale replica state change (152 ms) -[ok]: EXEC with only read commands should not be rejected when OOM (82 ms) -[ok]: EXEC with at least one use-memory command should fail (76 ms) -[ok]: Blocking commands ignores the timeout (76 ms) -[ok]: MULTI propagation of PUBLISH (168 ms) -[ok]: MULTI propagation of SCRIPT LOAD (220 ms) -[ok]: MULTI propagation of EVAL (101 ms) -[ok]: MULTI propagation of SCRIPT FLUSH (178 ms) -[ok]: MULTI propagation of XREADGROUP (124 ms) -[ok]: MULTI with SAVE (1 ms) -[ok]: MULTI with SHUTDOWN (1 ms) +[ok]: ACL LOG shows failed subcommand executions at toplevel (2 ms) +[ok]: ACL LOG is able to test similar events (2 ms) +[ok]: ACL LOG is able to log keys access violations and key name (1 ms) +[ok]: ACL LOG is able to log channel access violations and channel name (1 ms) +[ok]: ACL LOG RESET is able to flush the entries in the log (0 ms) +[ok]: ACL LOG can distinguish the transaction context (1) (2 ms) +[ok]: ACL LOG can distinguish the transaction context (2) (45 ms) +[ok]: ACL can log errors in the context of Lua scripting (1 ms) +[ok]: ACL LOG can accept a numerical argument to show less entries (3 ms) +[ok]: ACL LOG can log failed auth attempts (2 ms) +[ok]: ACLLOG - zero max length is correctly handled (4 ms) +[ok]: ACL LOG entries are limited to a maximum amount (4 ms) +[ok]: ACL LOG entries are still present on update of max len config (1 ms) +[ok]: When default user is off, new connections are not authenticated (9 ms) +[ok]: When default user has no command permission, hello command still works for other users (2 ms) +[ok]: When an authentication chain is used in the HELLO cmd, the last auth cmd has precedence (2 ms) +[ok]: When a setname chain is used in the HELLO cmd, the last setname cmd has precedence (2 ms) +[ok]: When authentication fails in the HELLO cmd, the client setname should not be applied (1 ms) +[ok]: ACL HELP should not have unexpected options (0 ms) +[ok]: Delete a user that the client doesn't use (1 ms) +[ok]: Delete a user that the client is using (2 ms) +[ok]: ACL GENPASS command failed test (19 ms) +[ok]: Default user can not be removed (1 ms) +[ok]: ACL load non-existing configured ACL file (1 ms) +[ok]: ACL-Metrics user AUTH failure (10 ms) +[ok]: ACL-Metrics invalid command accesses (5 ms) +[ok]: ACL-Metrics invalid key accesses (11 ms) +[ok]: ACL-Metrics invalid channels accesses (9 ms) +=== (external:skip) Starting server 127.0.0.1:21295 ok +[ok]: default: load from include file, can access any channels (2 ms) +[ok]: default: with config acl-pubsub-default allchannels after reset, can access any channels (6 ms) +[ok]: default: with config acl-pubsub-default resetchannels after reset, can not access any channels (7 ms) +[ok]: Alice: can execute all command (1 ms) +[ok]: Bob: just execute @set and acl command (6 ms) +[ok]: ACL LOAD only disconnects affected clients (91 ms) +[ok]: ACL LOAD disconnects clients of deleted users (82 ms) +[ok]: ACL load and save (14 ms) +[ok]: ACL load and save with restricted channels (10 ms) +=== (external:skip) Starting server 127.0.0.1:21297 ok +[ok]: Default user has access to all channels irrespective of flag (1 ms) +[ok]: Update acl-pubsub-default, existing users shouldn't get affected (2 ms) +[ok]: Single channel is valid (1 ms) +[ok]: Single channel is not valid with allchannels (1 ms) +=== (external:skip) Starting server 127.0.0.1:21299 ok +[ok]: Only default user has access to all channels irrespective of flag (1 ms) +=== (external:skip) Starting server 127.0.0.1:21301 ok +[ok]: default: load from config file, without channel permission default user can't access any channels (12 ms) +=== (external:skip) Starting server 127.0.0.1:21303 ok +[ok]: default: load from config file with all channels permissions (1 ms) +=== (external:skip) Starting server 127.0.0.1:21305 ok +[ok]: Test loading an ACL file with duplicate users (25 ms) +[ok]: Test loading an ACL file with duplicate default user (23 ms) +[ok]: Test loading duplicate users in config on startup (100 ms) +=== (acl external:skip) Starting server 127.0.0.1:21307 ok +[ok]: ACL from config file and config rewrite (303 ms) +[29/124 done]: unit/acl (4 seconds) +Testing unit/expire +=== (expire) Starting server 127.0.0.1:21309 ok +[ok]: EXPIRE - set timeouts multiple times (5 ms) +[ok]: EXPIRE - It should be still possible to read 'x' (0 ms) +[ok]: EXPIRE - After 2.1 seconds the key should no longer be here (2104 ms) +[ok]: EXPIRE - write on expire should work (6 ms) +[ok]: EXPIREAT - Check for EXPIRE alike behavior (6 ms) +[ok]: SETEX - Set + Expire combo operation. Check for TTL (1 ms) +[ok]: SETEX - Check value (1 ms) +[ok]: SETEX - Overwrite old key (1 ms) +[ok]: SETEX - Wait for the key to expire (1109 ms) +[ok]: SETEX - Wrong time parameter (1 ms) +[ok]: PERSIST can undo an EXPIRE (9 ms) +[ok]: PERSIST returns 0 against non existing or non volatile keys (1 ms) +millisecond expire test attempts: 0 +[ok]: EXPIRE precision is now the millisecond (1119 ms) +PSETEX sub-second expire test attempts: 0 +[ok]: PSETEX can set sub-second expires (105 ms) +PEXPIRE sub-second expire test attempts: 0 +[ok]: PEXPIRE can set sub-second expires (109 ms) +PEXPIREAT sub-second expire test attempts: 0 +[ok]: PEXPIREAT can set sub-second expires (227 ms) +[ok]: TTL returns time to live in seconds (14 ms) +[ok]: PTTL returns time to live in milliseconds (19 ms) +[ok]: TTL / PTTL / EXPIRETIME / PEXPIRETIME return -1 if key has no expire (48 ms) +[ok]: TTL / PTTL / EXPIRETIME / PEXPIRETIME return -2 if key does not exit (13 ms) +[ok]: EXPIRETIME returns absolute expiration time in seconds (11 ms) +[ok]: PEXPIRETIME returns absolute expiration time in milliseconds (26 ms) +[ok]: Server should actively expire keys incrementally (542 ms) +[ok]: Server should lazy expire keys (1039 ms) +[ok]: EXPIRE should not resurrect keys (issue #1026) (1008 ms) +[ok]: 5 keys in, 5 keys out (16 ms) +[ok]: EXPIRE with empty string as TTL should report an error (1 ms) +[ok]: SET with EX with big integer should report an error (6 ms) +[ok]: SET with EX with smallest integer should report an error (2 ms) +[ok]: GETEX with big integer should report an error (0 ms) +[ok]: GETEX with smallest integer should report an error (0 ms) +[ok]: EXPIRE with big integer overflows when converted to milliseconds (2 ms) +[ok]: PEXPIRE with big integer overflow when basetime is added (0 ms) +[ok]: EXPIRE with big negative integer (1 ms) +[ok]: PEXPIREAT with big integer works (1 ms) +[ok]: PEXPIREAT with big negative integer works (1 ms) +=== (external:skip) Starting server 127.0.0.1:21311 ok +[ok]: All time-to-live(TTL) in commands are propagated as absolute timestamp in milliseconds in AOF (2354 ms) +[ok]: All TTL in commands are propagated as absolute timestamp in replication stream (237 ms) +=== (needs:repl external:skip) Starting server 127.0.0.1:21313 ok +[ok]: First server should have role slave after REPLICAOF (29 ms) +[ok]: For all replicated TTL-related commands, absolute expire times are identical on primary and replica (153 ms) -Waiting for background AOF rewrite to finish... -Waiting for background AOF rewrite to finish... [ok]: MULTI with BGREWRITEAOF (241 ms) +Waiting for background AOF rewrite to finish... [ok]: expired key which is created in writeable replicas should be deleted by active expiry (280 ms) +[ok]: SET command will remove expire (1 ms) +[ok]: SET - use KEEPTTL option, TTL should not be removed (2 ms) +[ok]: SET - use KEEPTTL option, TTL should not be removed after loadaof (2016 ms) +[ok]: GETEX use of PERSIST option should remove TTL (1 ms) +[ok]: GETEX use of PERSIST option should remove TTL after loadaof (3 ms) +[ok]: GETEX propagate as to replica as PERSIST, UNLINK, or nothing (563 ms) +[ok]: EXPIRE with NX option on a key with ttl (1 ms) +[ok]: EXPIRE with NX option on a key without ttl (3 ms) +[ok]: EXPIRE with XX option on a key with ttl (3 ms) +[ok]: EXPIRE with XX option on a key without ttl (6 ms) +[ok]: EXPIRE with GT option on a key with lower ttl (2 ms) +[ok]: EXPIRE with GT option on a key with higher ttl (1 ms) +[ok]: EXPIRE with GT option on a key without ttl (2 ms) +[ok]: EXPIRE with LT option on a key with higher ttl (1 ms) +[ok]: EXPIRE with LT option on a key with lower ttl (2 ms) +[ok]: EXPIRE with LT option on a key without ttl (14 ms) +[ok]: EXPIRE with LT and XX option on a key with ttl (2 ms) +[ok]: EXPIRE with LT and XX option on a key without ttl (2 ms) +[ok]: EXPIRE with conflicting options: LT GT (1 ms) +[ok]: EXPIRE with conflicting options: NX GT (0 ms) +[ok]: EXPIRE with conflicting options: NX LT (7 ms) +[ok]: EXPIRE with conflicting options: NX XX (4 ms) +[ok]: EXPIRE with unsupported options (2 ms) +[ok]: EXPIRE with unsupported options (8 ms) +[ok]: EXPIRE with negative expiry (38 ms) +[ok]: EXPIRE with negative expiry on a non-valitale key (7 ms) +[ok]: EXPIRE with non-existed key (19 ms) +[ok]: Server should not propagate the read command on lazy expire (242 ms) +[ok]: SCAN: Lazy-expire should not be wrapped in MULTI/EXEC (93 ms) +[ok]: RANDOMKEY: Lazy-expire should not be wrapped in MULTI/EXEC (121 ms) +=== (expire external:skip cluster) Starting server 127.0.0.1:21315 ok +[ok]: expire scan should skip dictionaries with lot's of empty buckets (1126 ms) +[30/124 done]: unit/expire (18 seconds) +Testing unit/introspection +=== (introspection) Starting server 127.0.0.1:21317 ok +[ok]: PING (1 ms) +[ok]: CLIENT LIST (0 ms) +[ok]: CLIENT LIST with IDs (0 ms) +[ok]: CLIENT INFO (0 ms) +[ok]: client input output and command process statistics (96 ms) +[ok]: CLIENT KILL with illegal arguments (22 ms) +CLIENT KILL maxAGE will kill old clients test attempts: 0 +[ok]: CLIENT KILL maxAGE will kill old clients (2097 ms) +[ok]: CLIENT KILL SKIPME YES/NO will kill all clients (109 ms) +[ok]: CLIENT command unhappy path coverage (14 ms) +[ok]: CLIENT KILL close the client connection during bgsave (56 ms) +[ok]: CLIENT REPLY OFF/ON: disable all commands reply (81 ms) +[ok]: CLIENT REPLY SKIP: skip the next command reply (90 ms) +[ok]: CLIENT REPLY ON: unset SKIP flag (78 ms) +[ok]: MONITOR can log executed commands (56 ms) +[ok]: MONITOR can log commands issued by the scripting engine (30 ms) +[ok]: MONITOR can log commands issued by functions (28 ms) +[ok]: MONITOR supports redacting command arguments (63 ms) +[ok]: MONITOR correctly handles multi-exec cases (42 ms) +[ok]: MONITOR log blocked command only once (93 ms) +[ok]: CLIENT GETNAME should return NIL if name is not assigned (7 ms) +[ok]: CLIENT GETNAME check if name set correctly (6 ms) +[ok]: CLIENT LIST shows empty fields for unassigned names (0 ms) +[ok]: CLIENT SETNAME does not accept spaces (0 ms) +[ok]: CLIENT SETNAME can assign a name to this connection (0 ms) +[ok]: CLIENT SETNAME can change the name of an existing connection (1 ms) +[ok]: After CLIENT SETNAME, connection can still be closed (79 ms) +[ok]: CLIENT SETINFO can set a library name to this connection (1 ms) +[ok]: CLIENT SETINFO invalid args (1 ms) +[ok]: RESET does NOT clean library name (1 ms) +[ok]: CLIENT SETINFO can clear library name (0 ms) +=== () Starting server 127.0.0.1:21319 ok +=== () Starting server 127.0.0.1:21321 ok +=== () Starting server 127.0.0.1:21323 ok +=== () Starting server 127.0.0.1:21325 ok +[ok]: CONFIG save params special case handled properly (4847 ms) +[ok]: CONFIG sanity (253 ms) +[ok]: CONFIG REWRITE sanity (549 ms) +=== () Starting server 127.0.0.1:21327 ok +[ok]: CONFIG REWRITE handles save and shutdown properly (1295 ms) +[ok]: CONFIG SET with multiple args (17 ms) +[ok]: CONFIG SET rollback on set error (15 ms) +dummy_accept: sock56377294ad20 +[ok]: CONFIG SET rollback on apply error (60 ms) +[ok]: CONFIG SET duplicate configs (3 ms) +[ok]: CONFIG SET set immutable (9 ms) +[ok]: CONFIG GET hidden configs (21 ms) +[ok]: CONFIG GET multiple args (1 ms) +[ok]: valkey-server command line arguments - error cases (319 ms) +=== () Starting server 127.0.0.1:21330 ok +=== () Starting server 127.0.0.1:21332 ok +[ok]: tot-net-out for replica client (822 ms) +=== () Starting server 127.0.0.1:21334 ok +[ok]: valkey-server command line arguments - allow passing option name and option value in the same arg (359 ms) +=== () Starting server 127.0.0.1:21336 ok +[ok]: valkey-server command line arguments - wrong usage that we support anyway (405 ms) +=== () Starting server 127.0.0.1:21338 ok +[ok]: valkey-server command line arguments - allow option value to use the `--` prefix (362 ms) +=== () Starting server 127.0.0.1:21340 ok +[ok]: valkey-server command line arguments - option name and option value in the same arg and `--` prefix (281 ms) +=== () Starting server 127.0.0.1:21342 ok +=== () Starting server 127.0.0.1:21344 ok +=== () Starting server 127.0.0.1:21346 ok +=== () Starting server 127.0.0.1:21348 ok +=== () Starting server 127.0.0.1:21350 ok +[ok]: valkey-server command line arguments - save with empty input (1952 ms) +=== () Starting server 127.0.0.1:21352 ok +[ok]: valkey-server command line arguments - take one bulk string with spaces for MULTI_ARG configs parsing (555 ms) +=== (introspection external:skip) Starting server 127.0.0.1:21354 ok +[ok]: cannot modify protected configuration - no (14 ms) +=== (introspection external:skip) Starting server 127.0.0.1:21356 ok +[ok]: cannot modify protected configuration - local (11 ms) +=== () Starting server 127.0.0.1:21358 ok +Waiting for process 1428493 to exit... +[ok]: config during loading (3868 ms) +=== (introspection) Starting server 127.0.0.1:21360 ok +[ok]: CONFIG REWRITE handles rename-command properly (734 ms) +=== (introspection) Starting server 127.0.0.1:21362 ok +=== (introspection) Starting server 127.0.0.1:21364 ok +[ok]: CONFIG REWRITE handles alias config properly (1465 ms) +[31/124 done]: unit/introspection (23 seconds) +Testing unit/geo +=== (geo) Starting server 127.0.0.1:21366 ok +[ok]: GEO with wrong type src key (27 ms) +[ok]: GEO with non existing src key (3 ms) +[ok]: GEO BYLONLAT with empty search (1 ms) +[ok]: GEO BYMEMBER with non existing member (2 ms) +[ok]: GEOADD create (1 ms) +[ok]: GEOADD update (0 ms) +[ok]: GEOADD update with CH option (0 ms) +[ok]: GEOADD update with NX option (0 ms) +[ok]: GEOADD update with XX option (1 ms) +[ok]: GEOADD update with CH NX option (1 ms) +[ok]: GEOADD update with CH XX option (0 ms) +[ok]: GEOADD update with XX NX option will return syntax error (0 ms) +[ok]: GEOADD update with invalid option (0 ms) +[ok]: GEOADD invalid coordinates (0 ms) +[ok]: GEOADD multi add (1 ms) +[ok]: Check geoset values (1 ms) +[ok]: GEORADIUS simple (sorted) (0 ms) +[ok]: GEORADIUS_RO simple (sorted) (0 ms) +[ok]: GEOSEARCH simple (sorted) (0 ms) +[ok]: GEOSEARCH FROMLONLAT and FROMMEMBER cannot exist at the same time (0 ms) +[ok]: GEOSEARCH FROMLONLAT and FROMMEMBER one must exist (0 ms) +[ok]: GEOSEARCH BYRADIUS and BYBOX cannot exist at the same time (0 ms) +[ok]: GEOSEARCH BYRADIUS and BYBOX one must exist (1 ms) +[ok]: GEOSEARCH with STOREDIST option (1 ms) +[ok]: GEORADIUS withdist (sorted) (3 ms) +[ok]: GEOSEARCH withdist (sorted) (1 ms) +[ok]: GEORADIUS with COUNT (3 ms) +[ok]: GEORADIUS with multiple WITH* tokens (6 ms) +[ok]: GEORADIUS with ANY not sorted by default (0 ms) +[ok]: GEORADIUS with ANY sorted by ASC (13 ms) +[ok]: GEORADIUS with ANY but no COUNT (4 ms) +[ok]: GEORADIUS with COUNT but missing integer argument (5 ms) +[ok]: GEORADIUS with COUNT DESC (15 ms) +[ok]: GEORADIUS HUGE, issue #2767 (11 ms) +[ok]: GEORADIUSBYMEMBER simple (sorted) (13 ms) +[ok]: GEORADIUSBYMEMBER_RO simple (sorted) (8 ms) +[ok]: GEORADIUSBYMEMBER search areas contain satisfied points in oblique direction (6 ms) +[ok]: GEORADIUSBYMEMBER crossing pole search (2 ms) +[ok]: GEOSEARCH FROMMEMBER simple (sorted) (0 ms) +[ok]: GEOSEARCH vs GEORADIUS (9 ms) +[ok]: GEOSEARCH non square, long and narrow (2 ms) +[ok]: GEOSEARCH corner point test (6 ms) +[ok]: GEORADIUSBYMEMBER withdist (sorted) (1 ms) +[ok]: GEOHASH is able to return geohash strings (1 ms) +[ok]: GEOHASH with only key as argument (1 ms) +[ok]: GEOPOS simple (1 ms) +[ok]: GEOPOS missing element (1 ms) +[ok]: GEOPOS with only key as argument (1 ms) +[ok]: GEODIST simple & unit (1 ms) +[ok]: GEODIST missing elements (2 ms) +[ok]: GEORADIUS STORE option: syntax error (1 ms) +[ok]: GEOSEARCHSTORE STORE option: syntax error (1 ms) +[ok]: GEORANGE STORE option: incompatible options (2 ms) +[ok]: GEORANGE STORE option: plain usage (1 ms) +[ok]: GEORADIUSBYMEMBER STORE/STOREDIST option: plain usage (1 ms) +[ok]: GEOSEARCHSTORE STORE option: plain usage (1 ms) +[ok]: GEORANGE STOREDIST option: plain usage (1 ms) +[ok]: GEOSEARCHSTORE STOREDIST option: plain usage (1 ms) +[ok]: GEORANGE STOREDIST option: COUNT ASC and DESC (2 ms) +[ok]: GEOSEARCH the box spans -180° or 180° (15 ms) +[ok]: GEOSEARCH with small distance (2 ms) +[ok]: GEOSEARCH fuzzy test - byradius (81251 ms) +[ok]: GEOSEARCH fuzzy test - bybox (75206 ms) +[ok]: GEOSEARCH box edges fuzzy test (298 ms) +[32/124 done]: unit/geo (157 seconds) +Testing unit/functions +=== (scripting) Starting server 127.0.0.1:21368 ok +[ok]: FUNCTION - Basic usage (1 ms) +[ok]: FUNCTION - Load with unknown argument (2 ms) +[ok]: FUNCTION - Create an already exiting library raise error (0 ms) +[ok]: FUNCTION - Create an already exiting library raise error (case insensitive) (0 ms) +[ok]: FUNCTION - Create a library with wrong name format (0 ms) +[ok]: FUNCTION - Create library with unexisting engine (1 ms) +[ok]: FUNCTION - Test uncompiled script (1 ms) +[ok]: FUNCTION - test replace argument (2 ms) +[ok]: FUNCTION - test function case insensitive (0 ms) +[ok]: FUNCTION - test replace argument with failure keeps old libraries (2 ms) +[ok]: FUNCTION - test function delete (1 ms) +[ok]: FUNCTION - test fcall bad arguments (2 ms) +[ok]: FUNCTION - test fcall bad number of keys arguments (1 ms) +[ok]: FUNCTION - test fcall negative number of keys (1 ms) +[ok]: FUNCTION - test delete on not exiting library (0 ms) +[ok]: FUNCTION - test function kill when function is not running (1 ms) +[ok]: FUNCTION - test wrong subcommand (0 ms) +[ok]: FUNCTION - test loading from rdb (15 ms) +[ok]: FUNCTION - test debug reload different options (28 ms) +[ok]: FUNCTION - test debug reload with nosave and noflush (22 ms) +[ok]: FUNCTION - test flushall and flushdb do not clean functions (2 ms) +[ok]: FUNCTION - test function dump and restore (2 ms) +[ok]: FUNCTION - test function dump and restore with flush argument (1 ms) +[ok]: FUNCTION - test function dump and restore with append argument (3 ms) +[ok]: FUNCTION - test function dump and restore with replace argument (2 ms) +[ok]: FUNCTION - test function restore with bad payload do not drop existing functions (1 ms) +[ok]: FUNCTION - test function restore with wrong number of arguments (0 ms) +[ok]: FUNCTION - test fcall_ro with write command (1 ms) +[ok]: FUNCTION - test fcall_ro with read only commands (1 ms) +[ok]: FUNCTION - test keys and argv (0 ms) +[ok]: FUNCTION - test command get keys on fcall (0 ms) +[ok]: FUNCTION - test command get keys on fcall_ro (0 ms) +[ok]: FUNCTION - test function kill (432 ms) +[ok]: FUNCTION - test script kill not working on function (475 ms) +[ok]: FUNCTION - test function kill not working on eval (481 ms) +[ok]: FUNCTION - test function flush (16 ms) +[ok]: FUNCTION - test function wrong argument (1 ms) +=== (scripting repl external:skip) Starting server 127.0.0.1:21370 ok +=== () Starting server 127.0.0.1:21372 ok +[ok]: Connect a replica to the master instance (2067 ms) +[ok]: FUNCTION - creation is replicated to replica (1 ms) +[ok]: FUNCTION - call on replica (1 ms) +[ok]: FUNCTION - restore is replicated to replica (3 ms) +[ok]: FUNCTION - delete is replicated to replica (0 ms) +[ok]: FUNCTION - flush is replicated to replica (1 ms) +[ok]: Disconnecting the replica from master instance (110 ms) +[ok]: FUNCTION - test replication to replica on rdb phase (0 ms) +[ok]: FUNCTION - test replication to replica on rdb phase info command (1 ms) +[ok]: FUNCTION - create on read only replica (3 ms) +[ok]: FUNCTION - delete on read only replica (0 ms) +[ok]: FUNCTION - function effect is replicated to replica (2 ms) +[ok]: FUNCTION - modify key space of read only replica (0 ms) +=== () Starting server 127.0.0.1:21374 ok Waiting for background AOF rewrite to finish... -Waiting for background AOF rewrite to finish... [ok]: MULTI with config set appendonly (176 ms) -[ok]: MULTI with config error (1 ms) -[ok]: Flushall while watching several keys by one client (2 ms) -=== (external:skip) Starting server 127.0.0.1:21362 ok -[ok]: MULTI with FLUSHALL and AOF (72 ms) -=== (external:skip cluster) Starting server 127.0.0.1:21364 ok -[ok]: Regression test for multi-exec with RANDOMKEY accessing the wrong per-slot dictionary (5 ms) -[34/124 done]: unit/multi (11 seconds) -Testing unit/pubsubshard -=== (pubsubshard external:skip) Starting server 127.0.0.1:21366 ok -[ok]: SPUBLISH/SSUBSCRIBE basics (17 ms) -[ok]: SPUBLISH/SSUBSCRIBE with two clients (29 ms) -[ok]: SPUBLISH/SSUBSCRIBE after UNSUBSCRIBE without arguments (19 ms) -[ok]: SSUBSCRIBE to one channel more than once (15 ms) -[ok]: SUNSUBSCRIBE from non-subscribed channels (14 ms) -[ok]: PUBSUB command basics (0 ms) -[ok]: SPUBLISH/SSUBSCRIBE with two clients (32 ms) -[ok]: SPUBLISH/SSUBSCRIBE with PUBLISH/SUBSCRIBE (30 ms) -[ok]: PubSubShard with CLIENT REPLY OFF (111 ms) -=== (pubsubshard external:skip) Starting server 127.0.0.1:21368 ok -=== (pubsubshard external:skip) Starting server 127.0.0.1:21370 ok -[ok]: setup replication for following tests (124 ms) -[ok]: publish message to master and receive on replica (79 ms) -[35/124 done]: unit/pubsubshard (2 seconds) -Testing unit/obuf-limits -=== (obuf-limits external:skip logreqres:skip) Starting server 127.0.0.1:21372 ok -[ok]: CONFIG SET client-output-buffer-limit (64 ms) -[ok]: Client output buffer hard limit is enforced (17194 ms) -[ok]: Client output buffer soft limit is enforced if time is overreached (7522 ms) -[ok]: Client output buffer soft limit is not enforced too early and is enforced when no traffic (6983 ms) -[ok]: No response for single command if client output buffer hard limit is enforced (316 ms) -[ok]: No response for multi commands in pipeline if client output buffer limit is enforced (1139 ms) -[ok]: Execute transactions completely even if client output buffer limit is enforced (1215 ms) -[ok]: Obuf limit, HRANDFIELD with huge count stopped mid-run (73 ms) -[ok]: Obuf limit, KEYS stopped mid-run (252 ms) -[36/124 done]: unit/obuf-limits (36 seconds) +Waiting for background AOF rewrite to finish... [ok]: FUNCTION can processes create, delete and flush commands in AOF when doing "debug loadaof" in read-only slaves (430 ms) +=== (scripting) Starting server 127.0.0.1:21376 ok +[ok]: LIBRARIES - test shared function can access default globals (1 ms) +[ok]: LIBRARIES - usage and code sharing (1 ms) +[ok]: LIBRARIES - test registration failure revert the entire load (1 ms) +[ok]: LIBRARIES - test registration function name collision (1 ms) +[ok]: LIBRARIES - test registration function name collision on same library (1 ms) +[ok]: LIBRARIES - test registration with no argument (0 ms) +[ok]: LIBRARIES - test registration with only name (1 ms) +[ok]: LIBRARIES - test registration with to many arguments (1 ms) +[ok]: LIBRARIES - test registration with no string name (0 ms) +[ok]: LIBRARIES - test registration with wrong name format (1 ms) +[ok]: LIBRARIES - test registration with empty name (0 ms) +[ok]: LIBRARIES - math.random from function load (1 ms) +[ok]: LIBRARIES - redis.call from function load (1 ms) +[ok]: LIBRARIES - redis.setresp from function load (0 ms) +[ok]: LIBRARIES - redis.set_repl from function load (0 ms) +[ok]: LIBRARIES - redis.acl_check_cmd from function load (0 ms) +[ok]: LIBRARIES - malicious access test (1 ms) +[ok]: LIBRARIES - delete removed all functions on library (0 ms) +[ok]: LIBRARIES - register function inside a function (1 ms) +[ok]: LIBRARIES - register library with no functions (0 ms) +[ok]: LIBRARIES - load timeout (502 ms) +[ok]: LIBRARIES - verify global protection on the load run (1 ms) +[ok]: LIBRARIES - named arguments (1 ms) +[ok]: LIBRARIES - named arguments, bad function name (0 ms) +[ok]: LIBRARIES - named arguments, bad callback type (0 ms) +[ok]: LIBRARIES - named arguments, bad description (1 ms) +[ok]: LIBRARIES - named arguments, unknown argument (1 ms) +[ok]: LIBRARIES - named arguments, missing function name (1 ms) +[ok]: LIBRARIES - named arguments, missing callback (1 ms) +[ok]: FUNCTION - test function restore with function name collision (5 ms) +[ok]: FUNCTION - test function list with code (1 ms) +[ok]: FUNCTION - test function list with pattern (1 ms) +[ok]: FUNCTION - test function list wrong argument (0 ms) +[ok]: FUNCTION - test function list with bad argument to library name (0 ms) +[ok]: FUNCTION - test function list withcode multiple times (0 ms) +[ok]: FUNCTION - test function list libraryname multiple times (1 ms) +[ok]: FUNCTION - verify OOM on function load and function restore (3 ms) +[ok]: FUNCTION - verify allow-omm allows running any command (2 ms) +=== (scripting) Starting server 127.0.0.1:21378 ok +[ok]: FUNCTION - wrong flags type named arguments (1 ms) +[ok]: FUNCTION - wrong flag type (6 ms) +[ok]: FUNCTION - unknown flag (1 ms) +[ok]: FUNCTION - write script on fcall_ro (1 ms) +[ok]: FUNCTION - write script with no-writes flag (0 ms) +[ok]: FUNCTION - deny oom (5 ms) +[ok]: FUNCTION - deny oom on no-writes function (1 ms) +[ok]: FUNCTION - allow stale (3 ms) +[ok]: FUNCTION - valkey version api (1 ms) +[ok]: FUNCTION - function stats (4 ms) +[ok]: FUNCTION - function stats reloaded correctly from rdb (24 ms) +[ok]: FUNCTION - function stats delete library (1 ms) +[ok]: FUNCTION - test function stats on loading failure (2 ms) +[ok]: FUNCTION - function stats cleaned after flush (1 ms) +[ok]: FUNCTION - function test empty engine (0 ms) +[ok]: FUNCTION - function test unknown metadata value (7 ms) +[ok]: FUNCTION - function test no name (0 ms) +[ok]: FUNCTION - function test multiple names (1 ms) +[ok]: FUNCTION - function test name with quotes (0 ms) +[ok]: FUNCTION - trick global protection 1 (2 ms) +[ok]: FUNCTION - test getmetatable on script load (1 ms) +[33/124 done]: unit/functions (7 seconds) +Testing unit/networking +=== () Starting server 127.0.0.1:21380 ok +[ok]: CONFIG SET port number (426 ms) +=== () Starting server 127.0.0.1:21383 ok +[ok]: CONFIG SET bind address (297 ms) +=== () Starting server 127.0.0.1:21385 ok +=== () Starting server 127.0.0.1:21387 ok +[ok]: CONFIG SET bind-source-addr (757 ms) +=== (external:skip) Starting server 127.0.0.1:21389 ok +[ok]: Default bind address configuration handling (372 ms) +[ok]: Protected mode works as expected (15 ms) +=== (external:skip) Starting server 127.0.0.1:21391 ok +[34/124 done]: unit/networking (3 seconds) +Testing unit/bitfield +=== (bitops) Starting server 127.0.0.1:21393 ok +[ok]: BITFIELD signed SET and GET basics (1 ms) +[ok]: BITFIELD unsigned SET and GET basics (4 ms) +[ok]: BITFIELD signed SET and GET together (1 ms) +[ok]: BITFIELD unsigned with SET, GET and INCRBY arguments (1 ms) +[ok]: BITFIELD with only key as argument (0 ms) +[ok]: BITFIELD # form (1 ms) +[ok]: BITFIELD basic INCRBY form (5 ms) +[ok]: BITFIELD chaining of multiple commands (1 ms) +[ok]: BITFIELD unsigned overflow wrap (8 ms) +[ok]: BITFIELD unsigned overflow sat (1 ms) +[ok]: BITFIELD signed overflow wrap (2 ms) +[ok]: BITFIELD signed overflow sat (1 ms) +[ok]: BITFIELD overflow detection fuzzing (932 ms) +[ok]: BITFIELD overflow wrap fuzzing (985 ms) +[ok]: BITFIELD regression for #3221 (1 ms) +[ok]: BITFIELD regression for #3564 (5 ms) +[ok]: BITFIELD_RO with only key as argument (0 ms) +[ok]: BITFIELD_RO fails when write option is used (0 ms) +=== (repl external:skip) Starting server 127.0.0.1:21395 ok +=== () Starting server 127.0.0.1:21397 ok +[ok]: BITFIELD: setup slave (127 ms) +[ok]: BITFIELD: write on master, read on slave (7 ms) +[ok]: BITFIELD_RO with only key as argument on read-only replica (5 ms) +[ok]: BITFIELD_RO fails when write option is used on read-only replica (0 ms) +[35/124 done]: unit/bitfield (5 seconds) +Testing unit/quit +=== (quit) Starting server 127.0.0.1:21399 ok +[ok]: QUIT returns OK (29 ms) +[ok]: Pipelined commands after QUIT must not be executed (63 ms) +[ok]: Pipelined commands after QUIT that exceed read buffer size (50 ms) +[36/124 done]: unit/quit (0 seconds) +Testing unit/pubsub +=== (pubsub network) Starting server 127.0.0.1:21401 ok +[ok]: Pub/Sub PING on RESP2 (10 ms) +[ok]: Pub/Sub PING on RESP3 (1 ms) +[ok]: PUBLISH/SUBSCRIBE basics (15 ms) +[ok]: PUBLISH/SUBSCRIBE with two clients (26 ms) +[ok]: PUBLISH/SUBSCRIBE after UNSUBSCRIBE without arguments (14 ms) +[ok]: SUBSCRIBE to one channel more than once (13 ms) +[ok]: UNSUBSCRIBE from non-subscribed channels (13 ms) +[ok]: PUBLISH/PSUBSCRIBE basics (15 ms) +[ok]: PUBLISH/PSUBSCRIBE with two clients (26 ms) +[ok]: PUBLISH/PSUBSCRIBE after PUNSUBSCRIBE without arguments (15 ms) +[ok]: PubSub messages with CLIENT REPLY OFF (103 ms) +[ok]: PUNSUBSCRIBE from non-subscribed channels (28 ms) +[ok]: NUMSUB returns numbers, not strings (#1561) (1 ms) +[ok]: NUMPATs returns the number of unique patterns (55 ms) +[ok]: Mix SUBSCRIBE and PSUBSCRIBE (31 ms) +[ok]: PUNSUBSCRIBE and UNSUBSCRIBE should always reply (11 ms) +[ok]: Keyspace notifications: we receive keyspace notifications (93 ms) +[ok]: Keyspace notifications: we receive keyevent notifications (76 ms) +[ok]: Keyspace notifications: we can receive both kind of events (25 ms) +[ok]: Keyspace notifications: we are able to mask events (29 ms) +[ok]: Keyspace notifications: general events test (32 ms) +[ok]: Keyspace notifications: list events test (37 ms) +[ok]: Keyspace notifications: set events test (32 ms) +[ok]: Keyspace notifications: zset events test (41 ms) +[ok]: Keyspace notifications: hash events test (51 ms) +[ok]: Keyspace notifications: stream events test (48 ms) +[ok]: Keyspace notifications: expired events (triggered expire) (154 ms) +[ok]: Keyspace notifications: expired events (background expire) (226 ms) +[ok]: Keyspace notifications: evicted events (38 ms) +[ok]: Keyspace notifications: test CONFIG GET/SET of event flags (2 ms) +[ok]: Keyspace notifications: new key test (21 ms) +[ok]: publish to self inside multi (5 ms) +[ok]: publish to self inside script (1 ms) +[ok]: unsubscribe inside multi, and publish to self (10 ms) +[37/124 done]: unit/pubsub (2 seconds) +Testing unit/printver +=== () Starting server 127.0.0.1:21403 ok +Testing Valkey version 7.2.4 (00000000) +[38/124 done]: unit/printver (0 seconds) +Testing unit/keyspace +=== (keyspace) Starting server 127.0.0.1:21405 ok +[ok]: DEL against a single item (8 ms) +[ok]: Vararg DEL (6 ms) +[ok]: Untagged multi-key commands (9 ms) +[ok]: KEYS with pattern (2 ms) +[ok]: KEYS to get all keys (0 ms) +[ok]: DBSIZE (0 ms) +[ok]: KEYS with hashtag (6 ms) +[ok]: DEL all keys (4 ms) +[ok]: DEL against expired key (1106 ms) +[ok]: EXISTS (10 ms) +[ok]: Zero length value in key. SET/GET/EXISTS (2 ms) +[ok]: Commands pipelining (0 ms) +[ok]: Non existing command (0 ms) +[ok]: RENAME basic usage (3 ms) +[ok]: RENAME source key should no longer exist (0 ms) +[ok]: RENAME against already existing key (1 ms) +[ok]: RENAMENX basic usage (18 ms) +[ok]: RENAMENX against already existing key (1 ms) +[ok]: RENAMENX against already existing key (2) (1 ms) +[ok]: RENAME against non existing source key (1 ms) +[ok]: RENAME where source and dest key are the same (existing) (1 ms) +[ok]: RENAMENX where source and dest key are the same (existing) (1 ms) +[ok]: RENAME where source and dest key are the same (non existing) (1 ms) +[ok]: RENAME with volatile key, should move the TTL as well (2 ms) +[ok]: RENAME with volatile key, should not inherit TTL of target key (3 ms) +[ok]: DEL all keys again (DB 0) (1 ms) +[ok]: DEL all keys again (DB 1) (1 ms) +[ok]: COPY basic usage for string (28 ms) +[ok]: COPY for string does not replace an existing key without REPLACE option (1 ms) +[ok]: COPY for string can replace an existing key with REPLACE option (1 ms) +[ok]: COPY for string ensures that copied data is independent of copying data (14 ms) +[ok]: COPY for string does not copy data to no-integer DB (7 ms) +[ok]: COPY can copy key expire metadata as well (2 ms) +[ok]: COPY does not create an expire if it does not exist (1 ms) +[ok]: COPY basic usage for list - quicklist (46 ms) +[ok]: COPY basic usage for list - listpack (4 ms) +[ok]: COPY basic usage for $type set (3 ms) +[ok]: COPY basic usage for $type set (9 ms) +[ok]: COPY basic usage for $type set (105 ms) +[ok]: COPY basic usage for listpack sorted set (6 ms) +[ok]: COPY basic usage for skiplist sorted set (39 ms) +[ok]: COPY basic usage for listpack hash (2 ms) +[ok]: COPY basic usage for hashtable hash (67 ms) +[ok]: COPY basic usage for stream (281 ms) +[ok]: COPY basic usage for stream-cgroups (14 ms) +[ok]: MOVE basic usage (16 ms) +[ok]: MOVE against key existing in the target DB (0 ms) +[ok]: MOVE against non-integer DB (#1428) (1 ms) +[ok]: MOVE can move key expire metadata as well (3 ms) +[ok]: MOVE does not create an expire if it does not exist (13 ms) +[ok]: SET/GET keys in different DBs (5 ms) +[ok]: RANDOMKEY (77 ms) +[ok]: RANDOMKEY against empty DB (1 ms) +[ok]: RANDOMKEY regression 1 (1 ms) +[ok]: KEYS * two times with long key, Github issue #1208 (4 ms) +[ok]: Regression for pattern matching long nested loops (0 ms) +[ok]: Coverage: basic SWAPDB test and unhappy path (5 ms) +[ok]: Coverage: SWAPDB and FLUSHDB (5 ms) +[ok]: Regression for pattern matching very long nested loops (96 ms) +[39/124 done]: unit/keyspace (3 seconds) +Testing unit/dump +=== (dump) Starting server 127.0.0.1:21407 ok +[ok]: DUMP / RESTORE are able to serialize / unserialize a simple key (1 ms) +[ok]: RESTORE can set an arbitrary expire to the materialized key (2 ms) +[ok]: RESTORE can set an expire that overflows a 32 bit integer (1 ms) +[ok]: RESTORE can set an absolute expire (1 ms) +[ok]: RESTORE with ABSTTL in the past (1 ms) +[ok]: RESTORE can set LRU (1 ms) +[ok]: RESTORE can set LFU (2 ms) +[ok]: RESTORE returns an error of the key already exists (1 ms) +[ok]: RESTORE can overwrite an existing key with REPLACE (1 ms) +[ok]: RESTORE can detect a syntax error for unrecognized options (0 ms) +[ok]: RESTORE should not store key that are already expired, with REPLACE will propagate it as DEL or UNLINK (478 ms) +[ok]: DUMP of non existing key returns nil (0 ms) +=== (repl) Starting server 127.0.0.1:21409 ok +[ok]: MIGRATE is caching connections (278 ms) +[ok]: MIGRATE cached connections are released after some time (15006 ms) +=== (repl) Starting server 127.0.0.1:21411 ok +[ok]: MIGRATE is able to migrate a key between two instances (370 ms) +=== (repl) Starting server 127.0.0.1:21413 ok +[ok]: MIGRATE is able to copy a key between two instances (308 ms) +=== (repl) Starting server 127.0.0.1:21415 ok +[ok]: MIGRATE will not overwrite existing keys, unless REPLACE is used (382 ms) +=== (repl) Starting server 127.0.0.1:21417 ok +[ok]: MIGRATE propagates TTL correctly (267 ms) +=== (repl) Starting server 127.0.0.1:21419 ok +[ok]: MIGRATE can correctly transfer large values (23479 ms) +=== (repl) Starting server 127.0.0.1:21421 ok +[ok]: MIGRATE can correctly transfer hashes (287 ms) +=== (repl) Starting server 127.0.0.1:21423 ok +[ok]: MIGRATE timeout actually works (860 ms) +=== (repl) Starting server 127.0.0.1:21425 ok +[ok]: MIGRATE can migrate multiple keys at once (277 ms) +[ok]: MIGRATE with multiple keys must have empty key arg (0 ms) +=== (repl) Starting server 127.0.0.1:21427 ok +[ok]: MIGRATE with multiple keys migrate just existing ones (291 ms) +=== (repl) Starting server 127.0.0.1:21429 ok +[ok]: MIGRATE with multiple keys: stress command rewriting (276 ms) +=== (repl) Starting server 127.0.0.1:21431 ok +[ok]: MIGRATE with multiple keys: delete just ack keys (257 ms) +=== (repl) Starting server 127.0.0.1:21433 ok +[ok]: MIGRATE AUTH: correct and wrong password cases (270 ms) +[40/124 done]: unit/dump (43 seconds) +Testing unit/hyperloglog +=== (hll) Starting server 127.0.0.1:21435 ok +[ok]: HyperLogLog self test passes (951 ms) +[ok]: PFADD without arguments creates an HLL value (1 ms) +[ok]: Approximated cardinality after creation is zero (0 ms) +[ok]: PFADD returns 1 when at least 1 reg was modified (0 ms) +[ok]: PFADD returns 0 when no reg was modified (0 ms) +[ok]: PFADD works with empty string (regression) (0 ms) +[ok]: PFCOUNT returns approximated cardinality of set (2 ms) +[ok]: HyperLogLogs are promote from sparse to dense (664 ms) +[ok]: Change hll-sparse-max-bytes (2 ms) +[ok]: Hyperloglog promote to dense well in different hll-sparse-max-bytes (104 ms) +[ok]: HyperLogLog sparse encoding stress test (1822 ms) +[ok]: Corrupted sparse HyperLogLogs are detected: Additional at tail (1 ms) +[ok]: Corrupted sparse HyperLogLogs are detected: Broken magic (0 ms) +[ok]: Corrupted sparse HyperLogLogs are detected: Invalid encoding (0 ms) +[ok]: Corrupted dense HyperLogLogs are detected: Wrong length (1 ms) +[ok]: Fuzzing dense/sparse encoding: Server should always detect errors (81593 ms) +[ok]: PFADD, PFCOUNT, PFMERGE type checking works (9 ms) +[ok]: PFMERGE results on the cardinality of union of sets (12 ms) +[ok]: PFMERGE on missing source keys will create an empty destkey (10 ms) +[ok]: PFMERGE with one empty input key, create an empty destkey (1 ms) +[ok]: PFMERGE with one non-empty input key, dest key is actually one of the source keys (20 ms) +[ok]: PFCOUNT multiple-keys merge returns cardinality of union #1 (43838 ms) +[ok]: PFCOUNT multiple-keys merge returns cardinality of union #2 (5049 ms) +[ok]: PFDEBUG GETREG returns the HyperLogLog raw registers (1079 ms) +[ok]: PFADD / PFCOUNT cache invalidation works (2 ms) +[41/124 done]: unit/hyperloglog (135 seconds) Testing unit/scripting -=== (scripting) Starting server 127.0.0.1:21374 ok -[ok]: EVAL - Does Lua interpreter replies to our requests? (12 ms) -[ok]: EVAL - Return _G (16 ms) -[ok]: EVAL - Return table with a metatable that raise error (13 ms) -[ok]: EVAL - Return table with a metatable that call server (18 ms) -[ok]: EVAL - Lua integer -> Redis protocol type conversion (11 ms) -[ok]: EVAL - Lua string -> Redis protocol type conversion (12 ms) -[ok]: EVAL - Lua true boolean -> Redis protocol type conversion (16 ms) -[ok]: EVAL - Lua false boolean -> Redis protocol type conversion (16 ms) -[ok]: EVAL - Lua status code reply -> Redis protocol type conversion (15 ms) -[ok]: EVAL - Lua error reply -> Redis protocol type conversion (16 ms) -[ok]: EVAL - Lua table -> Redis protocol type conversion (15 ms) -[ok]: EVAL - Are the KEYS and ARGV arrays populated correctly? (16 ms) -[ok]: EVAL - is Lua able to call Redis API? (15 ms) -[ok]: EVAL - Redis integer -> Lua type conversion (24 ms) -[ok]: EVAL - Lua number -> Redis integer conversion (24 ms) -[ok]: EVAL - Redis bulk -> Lua type conversion (28 ms) -[ok]: EVAL - Redis multi bulk -> Lua type conversion (40 ms) -[ok]: EVAL - Redis status reply -> Lua type conversion (12 ms) -[ok]: EVAL - Redis error reply -> Lua type conversion (29 ms) -[ok]: EVAL - Redis nil bulk reply -> Lua type conversion (14 ms) -[ok]: EVAL - Is the Lua client using the currently selected DB? (32 ms) -[ok]: EVAL - SELECT inside Lua should not affect the caller (28 ms) -[ok]: EVAL - Scripts do not block on blpop command (15 ms) -[ok]: EVAL - Scripts do not block on brpop command (27 ms) -[ok]: EVAL - Scripts do not block on brpoplpush command (16 ms) -[ok]: EVAL - Scripts do not block on blmove command (21 ms) -[ok]: EVAL - Scripts do not block on bzpopmin command (10 ms) -[ok]: EVAL - Scripts do not block on bzpopmax command (14 ms) -[ok]: EVAL - Scripts do not block on wait (9 ms) -[ok]: EVAL - Scripts do not block on waitaof (16 ms) -[ok]: EVAL - Scripts do not block on XREAD with BLOCK option (23 ms) -[ok]: EVAL - Scripts do not block on XREADGROUP with BLOCK option (27 ms) -[ok]: EVAL - Scripts do not block on XREAD with BLOCK option -- non empty stream (24 ms) -[ok]: EVAL - Scripts do not block on XREADGROUP with BLOCK option -- non empty stream (10 ms) -[ok]: EVAL - Scripts can run non-deterministic commands (12 ms) -[ok]: EVAL - No arguments to redis.call/pcall is considered an error (10 ms) -[ok]: EVAL - redis.call variant raises a Lua error on Redis cmd error (1) (4 ms) -[ok]: EVAL - redis.call variant raises a Lua error on Redis cmd error (1) (18 ms) -[ok]: EVAL - redis.call variant raises a Lua error on Redis cmd error (1) (13 ms) +=== (scripting) Starting server 127.0.0.1:21437 ok +[ok]: EVAL - Does Lua interpreter replies to our requests? (1 ms) +[ok]: EVAL - Return _G (1 ms) +[ok]: EVAL - Return table with a metatable that raise error (0 ms) +[ok]: EVAL - Return table with a metatable that call server (1 ms) +[ok]: EVAL - Lua integer -> Redis protocol type conversion (1 ms) +[ok]: EVAL - Lua string -> Redis protocol type conversion (0 ms) +[ok]: EVAL - Lua true boolean -> Redis protocol type conversion (1 ms) +[ok]: EVAL - Lua false boolean -> Redis protocol type conversion (1 ms) +[ok]: EVAL - Lua status code reply -> Redis protocol type conversion (0 ms) +[ok]: EVAL - Lua error reply -> Redis protocol type conversion (1 ms) +[ok]: EVAL - Lua table -> Redis protocol type conversion (1 ms) +[ok]: EVAL - Are the KEYS and ARGV arrays populated correctly? (0 ms) +[ok]: EVAL - is Lua able to call Redis API? (1 ms) +[ok]: EVAL - Redis integer -> Lua type conversion (1 ms) +[ok]: EVAL - Lua number -> Redis integer conversion (1 ms) +[ok]: EVAL - Redis bulk -> Lua type conversion (1 ms) +[ok]: EVAL - Redis multi bulk -> Lua type conversion (1 ms) +[ok]: EVAL - Redis status reply -> Lua type conversion (0 ms) +[ok]: EVAL - Redis error reply -> Lua type conversion (1 ms) +[ok]: EVAL - Redis nil bulk reply -> Lua type conversion (1 ms) +[ok]: EVAL - Is the Lua client using the currently selected DB? (1 ms) +[ok]: EVAL - SELECT inside Lua should not affect the caller (1 ms) +[ok]: EVAL - Scripts do not block on blpop command (1 ms) +[ok]: EVAL - Scripts do not block on brpop command (1 ms) +[ok]: EVAL - Scripts do not block on brpoplpush command (1 ms) +[ok]: EVAL - Scripts do not block on blmove command (1 ms) +[ok]: EVAL - Scripts do not block on bzpopmin command (1 ms) +[ok]: EVAL - Scripts do not block on bzpopmax command (1 ms) +[ok]: EVAL - Scripts do not block on wait (1 ms) +[ok]: EVAL - Scripts do not block on waitaof (1 ms) +[ok]: EVAL - Scripts do not block on XREAD with BLOCK option (1 ms) +[ok]: EVAL - Scripts do not block on XREADGROUP with BLOCK option (1 ms) +[ok]: EVAL - Scripts do not block on XREAD with BLOCK option -- non empty stream (1 ms) +[ok]: EVAL - Scripts do not block on XREADGROUP with BLOCK option -- non empty stream (1 ms) +[ok]: EVAL - Scripts can run non-deterministic commands (1 ms) +[ok]: EVAL - No arguments to redis.call/pcall is considered an error (1 ms) +[ok]: EVAL - redis.call variant raises a Lua error on Redis cmd error (1) (0 ms) +[ok]: EVAL - redis.call variant raises a Lua error on Redis cmd error (1) (1 ms) +[ok]: EVAL - redis.call variant raises a Lua error on Redis cmd error (1) (1 ms) [ignore]: EVAL - JSON string encoding a string larger than 2GB: large memory flag not provided -[ok]: EVAL - JSON numeric decoding (3 ms) +[ok]: EVAL - JSON numeric decoding (1 ms) [ok]: EVAL - JSON string decoding (1 ms) [ok]: EVAL - JSON smoke test (1 ms) -[ok]: EVAL - cmsgpack can pack double? (2 ms) -[ok]: EVAL - cmsgpack can pack negative int64? (1 ms) -[ok]: EVAL - cmsgpack pack/unpack smoke test (1403 ms) -[ok]: EVAL - cmsgpack can pack and unpack circular references? (12 ms) -[ok]: EVAL - Numerical sanity check from bitop (14 ms) -[ok]: EVAL - Verify minimal bitop functionality (12 ms) -[ok]: EVAL - Able to parse trailing comments (16 ms) -[ok]: EVAL_RO - Successful case (20 ms) -[ok]: EVAL_RO - Cannot run write commands (20 ms) -[ok]: redis.sha1hex() implementation (19 ms) -[ok]: Measures elapsed time os.clock() (1779 ms) -[ok]: Prohibit dangerous lua methods in sandbox (17 ms) -[ok]: Verify execution of prohibit dangerous Lua methods will fail (104 ms) -[ok]: Globals protection reading an undeclared global variable (9 ms) -[ok]: Globals protection setting an undeclared global* (18 ms) -[ok]: lua bit.tohex bug (24 ms) -[ok]: Test an example script DECR_IF_GT (84 ms) -[ok]: EVAL does not leak in the Lua stack (2787 ms) -[ok]: Call Redis command with many args from Lua (issue #1764) (17 ms) -[ok]: Number conversion precision test (issue #1118) (14 ms) -[ok]: String containing number precision test (regression of issue #1118) (15 ms) -[ok]: Verify negative arg count is error instead of crash (issue #1842) (15 ms) -[ok]: Scripts can handle commands with incorrect arity (23 ms) -[ok]: Correct handling of reused argv (issue #1939) (15 ms) -[ok]: Functions in the Redis namespace are able to report errors (11 ms) -[ok]: CLUSTER RESET can not be invoke from within a script (11 ms) -[ok]: Script with RESP3 map (88 ms) -[ok]: Script return recursive object (363 ms) -[ok]: Script check unpack with massive arguments (27 ms) -[ok]: Script read key with expiration set (24 ms) -[ok]: Script del key with expiration set (19 ms) -[ok]: Script ACL check (84 ms) -[ok]: Binary code loading failed (16 ms) -[ok]: Try trick global protection 1 (16 ms) -[ok]: Try trick global protection 2 (11 ms) -[ok]: Try trick global protection 3 (16 ms) -[ok]: Try trick global protection 4 (16 ms) -[ok]: Try trick readonly table on valkey table (12 ms) -[ok]: Try trick readonly table on json table (16 ms) -[ok]: Try trick readonly table on cmsgpack table (17 ms) -[ok]: Try trick readonly table on bit table (16 ms) -[ok]: Test loadfile are not available (16 ms) -[ok]: Test dofile are not available (16 ms) -[ok]: Test print are not available (16 ms) -=== (scripting) Starting server 127.0.0.1:21376 ok -[ok]: Timedout read-only scripts can be killed by SCRIPT KILL (460 ms) -[ok]: Timedout read-only scripts can be killed by SCRIPT KILL even when use pcall (236 ms) -[ok]: Timedout script does not cause a false dead client (164 ms) -[ok]: Timedout script link is still usable after Lua returns (358 ms) -[ok]: Timedout scripts and unblocked command (759 ms) -[ok]: Timedout scripts that modified data can't be killed by SCRIPT KILL (275 ms) -[ok]: SHUTDOWN NOSAVE can kill a timedout script anyway (25 ms) -=== (scripting repl needs:debug external:skip) Starting server 127.0.0.1:21378 ok -=== () Starting server 127.0.0.1:21380 ok -[ok]: Before the replica connects we issue two EVAL commands (22 ms) -[ok]: Connect a replica to the master instance (206 ms) -[ok]: Replication of script multiple pushes to list with BLPOP (34 ms) +[ok]: EVAL - cmsgpack can pack double? (0 ms) +[ok]: EVAL - cmsgpack can pack negative int64? (0 ms) +[ok]: EVAL - cmsgpack pack/unpack smoke test (184 ms) +[ok]: EVAL - cmsgpack can pack and unpack circular references? (1 ms) +[ok]: EVAL - Numerical sanity check from bitop (1 ms) +[ok]: EVAL - Verify minimal bitop functionality (0 ms) +[ok]: EVAL - Able to parse trailing comments (1 ms) +[ok]: EVAL_RO - Successful case (1 ms) +[ok]: EVAL_RO - Cannot run write commands (1 ms) +[ok]: redis.sha1hex() implementation (1 ms) +[ok]: Measures elapsed time os.clock() (1001 ms) +[ok]: Prohibit dangerous lua methods in sandbox (1 ms) +[ok]: Verify execution of prohibit dangerous Lua methods will fail (4 ms) +[ok]: Globals protection reading an undeclared global variable (1 ms) +[ok]: Globals protection setting an undeclared global* (1 ms) +[ok]: lua bit.tohex bug (1 ms) +[ok]: Test an example script DECR_IF_GT (4 ms) +[ok]: EVAL does not leak in the Lua stack (934 ms) +[ok]: Call Redis command with many args from Lua (issue #1764) (1 ms) +[ok]: Number conversion precision test (issue #1118) (0 ms) +[ok]: String containing number precision test (regression of issue #1118) (0 ms) +[ok]: Verify negative arg count is error instead of crash (issue #1842) (0 ms) +[ok]: Scripts can handle commands with incorrect arity (1 ms) +[ok]: Correct handling of reused argv (issue #1939) (1 ms) +[ok]: Functions in the Redis namespace are able to report errors (0 ms) +[ok]: CLUSTER RESET can not be invoke from within a script (1 ms) +[ok]: Script with RESP3 map (2 ms) +[ok]: Script return recursive object (81 ms) +[ok]: Script check unpack with massive arguments (3 ms) +[ok]: Script read key with expiration set (1 ms) +[ok]: Script del key with expiration set (0 ms) +[ok]: Script ACL check (1 ms) +[ok]: Binary code loading failed (1 ms) +[ok]: Try trick global protection 1 (0 ms) +[ok]: Try trick global protection 2 (1 ms) +[ok]: Try trick global protection 3 (0 ms) +[ok]: Try trick global protection 4 (1 ms) +[ok]: Try trick readonly table on valkey table (0 ms) +[ok]: Try trick readonly table on json table (1 ms) +[ok]: Try trick readonly table on cmsgpack table (0 ms) +[ok]: Try trick readonly table on bit table (0 ms) +[ok]: Test loadfile are not available (0 ms) +[ok]: Test dofile are not available (0 ms) +[ok]: Test print are not available (1 ms) +=== (scripting) Starting server 127.0.0.1:21439 ok +[ok]: Timedout read-only scripts can be killed by SCRIPT KILL (411 ms) +[ok]: Timedout read-only scripts can be killed by SCRIPT KILL even when use pcall (44 ms) +[ok]: Timedout script does not cause a false dead client (43 ms) +[ok]: Timedout script link is still usable after Lua returns (60 ms) +[ok]: Timedout scripts and unblocked command (380 ms) +[ok]: Timedout scripts that modified data can't be killed by SCRIPT KILL (210 ms) +[ok]: SHUTDOWN NOSAVE can kill a timedout script anyway (6 ms) +=== (scripting repl needs:debug external:skip) Starting server 127.0.0.1:21441 ok +=== () Starting server 127.0.0.1:21443 ok +[ok]: Before the replica connects we issue two EVAL commands (2 ms) +[ok]: Connect a replica to the master instance (103 ms) +[ok]: Replication of script multiple pushes to list with BLPOP (10 ms) [ok]: Lua scripts using SELECT are replicated correctly (3 ms) -=== (scripting repl external:skip) Starting server 127.0.0.1:21382 ok -=== () Starting server 127.0.0.1:21384 ok -[ok]: Connect a replica to the master instance (136 ms) -[ok]: Redis.replicate_commands() can be issued anywhere now (7 ms) -[ok]: Redis.set_repl() can be issued before replicate_commands() now (7 ms) -[ok]: Redis.set_repl() don't accept invalid values (24 ms) -[ok]: Test selective replication of certain commands from Lua (39 ms) -[ok]: PRNG is seeded randomly for command replication (10 ms) -[ok]: Using side effects is not a problem with command replication (22 ms) -=== (scripting needs:debug) Starting server 127.0.0.1:21386 ok -[ok]: test RESP2/2 big number protocol parsing (12 ms) -[ok]: test RESP2/2 malformed big number protocol parsing (9 ms) -[ok]: test RESP2/2 map protocol parsing (6 ms) -[ok]: test RESP2/2 set protocol parsing (7 ms) -[ok]: test RESP2/2 double protocol parsing (12 ms) -[ok]: test RESP2/2 null protocol parsing (10 ms) -[ok]: test RESP2/2 verbatim protocol parsing (14 ms) -[ok]: test RESP2/2 true protocol parsing (12 ms) -[ok]: test RESP2/2 false protocol parsing (11 ms) -[ok]: test RESP2/3 big number protocol parsing (14 ms) -[ok]: test RESP2/3 malformed big number protocol parsing (14 ms) -[ok]: test RESP2/3 map protocol parsing (16 ms) -[ok]: test RESP2/3 set protocol parsing (16 ms) -[ok]: test RESP2/3 double protocol parsing (12 ms) -[ok]: test RESP2/3 null protocol parsing (11 ms) -[ok]: test RESP2/3 verbatim protocol parsing (16 ms) -[ok]: test RESP2/3 true protocol parsing (15 ms) -[ok]: test RESP2/3 false protocol parsing (15 ms) -[ok]: test RESP3/2 big number protocol parsing (20 ms) -[ok]: test RESP3/2 malformed big number protocol parsing (16 ms) -[ok]: test RESP3/2 map protocol parsing (16 ms) -[ok]: test RESP3/2 set protocol parsing (16 ms) -[ok]: test RESP3/2 double protocol parsing (11 ms) -[ok]: test RESP3/2 null protocol parsing (15 ms) -[ok]: test RESP3/2 verbatim protocol parsing (16 ms) -[ok]: test RESP3/2 true protocol parsing (11 ms) -[ok]: test RESP3/2 false protocol parsing (19 ms) -[ok]: test RESP3/3 big number protocol parsing (8 ms) -[ok]: test RESP3/3 malformed big number protocol parsing (1 ms) -[ok]: test RESP3/3 map protocol parsing (2 ms) -[ok]: test RESP3/3 set protocol parsing (3 ms) -[ok]: test RESP3/3 double protocol parsing (2 ms) -[ok]: test RESP3/3 null protocol parsing (9 ms) -[ok]: test RESP3/3 verbatim protocol parsing (2 ms) -[ok]: test RESP3/3 true protocol parsing (6 ms) +=== (scripting repl external:skip) Starting server 127.0.0.1:21445 ok +=== () Starting server 127.0.0.1:21447 ok +[ok]: Connect a replica to the master instance (102 ms) +[ok]: Redis.replicate_commands() can be issued anywhere now (0 ms) +[ok]: Redis.set_repl() can be issued before replicate_commands() now (1 ms) +[ok]: Redis.set_repl() don't accept invalid values (1 ms) +[ok]: Test selective replication of certain commands from Lua (5 ms) +[ok]: PRNG is seeded randomly for command replication (1 ms) +[ok]: Using side effects is not a problem with command replication (2 ms) +=== (scripting needs:debug) Starting server 127.0.0.1:21449 ok +[ok]: test RESP2/2 big number protocol parsing (0 ms) +[ok]: test RESP2/2 malformed big number protocol parsing (1 ms) +[ok]: test RESP2/2 map protocol parsing (0 ms) +[ok]: test RESP2/2 set protocol parsing (0 ms) +[ok]: test RESP2/2 double protocol parsing (1 ms) +[ok]: test RESP2/2 null protocol parsing (1 ms) +[ok]: test RESP2/2 verbatim protocol parsing (0 ms) +[ok]: test RESP2/2 true protocol parsing (1 ms) +[ok]: test RESP2/2 false protocol parsing (1 ms) +[ok]: test RESP2/3 big number protocol parsing (1 ms) +[ok]: test RESP2/3 malformed big number protocol parsing (0 ms) +[ok]: test RESP2/3 map protocol parsing (1 ms) +[ok]: test RESP2/3 set protocol parsing (1 ms) +[ok]: test RESP2/3 double protocol parsing (0 ms) +[ok]: test RESP2/3 null protocol parsing (1 ms) +[ok]: test RESP2/3 verbatim protocol parsing (0 ms) +[ok]: test RESP2/3 true protocol parsing (0 ms) +[ok]: test RESP2/3 false protocol parsing (1 ms) +[ok]: test RESP3/2 big number protocol parsing (0 ms) +[ok]: test RESP3/2 malformed big number protocol parsing (1 ms) +[ok]: test RESP3/2 map protocol parsing (0 ms) +[ok]: test RESP3/2 set protocol parsing (1 ms) +[ok]: test RESP3/2 double protocol parsing (1 ms) +[ok]: test RESP3/2 null protocol parsing (1 ms) +[ok]: test RESP3/2 verbatim protocol parsing (1 ms) +[ok]: test RESP3/2 true protocol parsing (0 ms) +[ok]: test RESP3/2 false protocol parsing (0 ms) +[ok]: test RESP3/3 big number protocol parsing (1 ms) +[ok]: test RESP3/3 malformed big number protocol parsing (0 ms) +[ok]: test RESP3/3 map protocol parsing (1 ms) +[ok]: test RESP3/3 set protocol parsing (1 ms) +[ok]: test RESP3/3 double protocol parsing (0 ms) +[ok]: test RESP3/3 null protocol parsing (1 ms) +[ok]: test RESP3/3 verbatim protocol parsing (1 ms) +[ok]: test RESP3/3 true protocol parsing (0 ms) [ok]: test RESP3/3 false protocol parsing (0 ms) [ok]: test resp3 attribute protocol parsing (0 ms) -[ok]: Script block the time during execution (21 ms) -[ok]: Script delete the expired key (5 ms) -[ok]: TIME command using cached time (15 ms) -[ok]: Script block the time in some expiration related commands (108 ms) -[ok]: RESTORE expired keys with expiration time (17 ms) -=== (scripting) Starting server 127.0.0.1:21388 ok -[ok]: EVAL - Does Lua interpreter replies to our requests? (15 ms) -[ok]: EVAL - Return _G (16 ms) -[ok]: EVAL - Return table with a metatable that raise error (8 ms) -[ok]: EVAL - Return table with a metatable that call server (24 ms) -[ok]: EVAL - Lua integer -> Redis protocol type conversion (11 ms) -[ok]: EVAL - Lua string -> Redis protocol type conversion (15 ms) -[ok]: EVAL - Lua true boolean -> Redis protocol type conversion (19 ms) -[ok]: EVAL - Lua false boolean -> Redis protocol type conversion (10 ms) -[ok]: EVAL - Lua status code reply -> Redis protocol type conversion (8 ms) -[ok]: EVAL - Lua error reply -> Redis protocol type conversion (16 ms) -[ok]: EVAL - Lua table -> Redis protocol type conversion (15 ms) -[ok]: EVAL - Are the KEYS and ARGV arrays populated correctly? (16 ms) -[ok]: EVAL - is Lua able to call Redis API? (23 ms) -[ok]: EVAL - Redis integer -> Lua type conversion (24 ms) -[ok]: EVAL - Lua number -> Redis integer conversion (20 ms) -[ok]: EVAL - Redis bulk -> Lua type conversion (23 ms) -[ok]: EVAL - Redis multi bulk -> Lua type conversion (48 ms) -[ok]: EVAL - Redis status reply -> Lua type conversion (20 ms) -[ok]: EVAL - Redis error reply -> Lua type conversion (19 ms) -[ok]: EVAL - Redis nil bulk reply -> Lua type conversion (16 ms) -[ok]: EVAL - Is the Lua client using the currently selected DB? (32 ms) -[ok]: EVAL - SELECT inside Lua should not affect the caller (28 ms) -[ok]: EVAL - Scripts do not block on blpop command (28 ms) -[ok]: EVAL - Scripts do not block on brpop command (24 ms) -[ok]: EVAL - Scripts do not block on brpoplpush command (31 ms) -[ok]: EVAL - Scripts do not block on blmove command (27 ms) -[ok]: EVAL - Scripts do not block on bzpopmin command (31 ms) -[ok]: EVAL - Scripts do not block on bzpopmax command (27 ms) -[ok]: EVAL - Scripts do not block on wait (15 ms) -[ok]: EVAL - Scripts do not block on waitaof (15 ms) -[ok]: EVAL - Scripts do not block on XREAD with BLOCK option (46 ms) -[ok]: EVAL - Scripts do not block on XREADGROUP with BLOCK option (28 ms) -[ok]: EVAL - Scripts do not block on XREAD with BLOCK option -- non empty stream (32 ms) -[ok]: EVAL - Scripts do not block on XREADGROUP with BLOCK option -- non empty stream (20 ms) -[ok]: EVAL - Scripts can run non-deterministic commands (15 ms) -[ok]: EVAL - No arguments to redis.call/pcall is considered an error (15 ms) -[ok]: EVAL - redis.call variant raises a Lua error on Redis cmd error (1) (16 ms) -[ok]: EVAL - redis.call variant raises a Lua error on Redis cmd error (1) (12 ms) -[ok]: EVAL - redis.call variant raises a Lua error on Redis cmd error (1) (23 ms) +[ok]: Script block the time during execution (15 ms) +[ok]: Script delete the expired key (4 ms) +[ok]: TIME command using cached time (11 ms) +[ok]: Script block the time in some expiration related commands (84 ms) +[ok]: RESTORE expired keys with expiration time (11 ms) +=== (scripting) Starting server 127.0.0.1:21451 ok +[ok]: EVAL - Does Lua interpreter replies to our requests? (1 ms) +[ok]: EVAL - Return _G (0 ms) +[ok]: EVAL - Return table with a metatable that raise error (1 ms) +[ok]: EVAL - Return table with a metatable that call server (1 ms) +[ok]: EVAL - Lua integer -> Redis protocol type conversion (0 ms) +[ok]: EVAL - Lua string -> Redis protocol type conversion (1 ms) +[ok]: EVAL - Lua true boolean -> Redis protocol type conversion (1 ms) +[ok]: EVAL - Lua false boolean -> Redis protocol type conversion (0 ms) +[ok]: EVAL - Lua status code reply -> Redis protocol type conversion (1 ms) +[ok]: EVAL - Lua error reply -> Redis protocol type conversion (0 ms) +[ok]: EVAL - Lua table -> Redis protocol type conversion (0 ms) +[ok]: EVAL - Are the KEYS and ARGV arrays populated correctly? (1 ms) +[ok]: EVAL - is Lua able to call Redis API? (1 ms) +[ok]: EVAL - Redis integer -> Lua type conversion (1 ms) +[ok]: EVAL - Lua number -> Redis integer conversion (0 ms) +[ok]: EVAL - Redis bulk -> Lua type conversion (0 ms) +[ok]: EVAL - Redis multi bulk -> Lua type conversion (2 ms) +[ok]: EVAL - Redis status reply -> Lua type conversion (0 ms) +[ok]: EVAL - Redis error reply -> Lua type conversion (0 ms) +[ok]: EVAL - Redis nil bulk reply -> Lua type conversion (1 ms) +[ok]: EVAL - Is the Lua client using the currently selected DB? (1 ms) +[ok]: EVAL - SELECT inside Lua should not affect the caller (1 ms) +[ok]: EVAL - Scripts do not block on blpop command (1 ms) +[ok]: EVAL - Scripts do not block on brpop command (1 ms) +[ok]: EVAL - Scripts do not block on brpoplpush command (1 ms) +[ok]: EVAL - Scripts do not block on blmove command (1 ms) +[ok]: EVAL - Scripts do not block on bzpopmin command (1 ms) +[ok]: EVAL - Scripts do not block on bzpopmax command (1 ms) +[ok]: EVAL - Scripts do not block on wait (1 ms) +[ok]: EVAL - Scripts do not block on waitaof (1 ms) +[ok]: EVAL - Scripts do not block on XREAD with BLOCK option (1 ms) +[ok]: EVAL - Scripts do not block on XREADGROUP with BLOCK option (1 ms) +[ok]: EVAL - Scripts do not block on XREAD with BLOCK option -- non empty stream (1 ms) +[ok]: EVAL - Scripts do not block on XREADGROUP with BLOCK option -- non empty stream (0 ms) +[ok]: EVAL - Scripts can run non-deterministic commands (1 ms) +[ok]: EVAL - No arguments to redis.call/pcall is considered an error (1 ms) +[ok]: EVAL - redis.call variant raises a Lua error on Redis cmd error (1) (0 ms) +[ok]: EVAL - redis.call variant raises a Lua error on Redis cmd error (1) (1 ms) +[ok]: EVAL - redis.call variant raises a Lua error on Redis cmd error (1) (1 ms) [ignore]: EVAL - JSON string encoding a string larger than 2GB: large memory flag not provided -[ok]: EVAL - JSON numeric decoding (19 ms) -[ok]: EVAL - JSON string decoding (14 ms) -[ok]: EVAL - JSON smoke test (12 ms) -[ok]: EVAL - cmsgpack can pack double? (9 ms) -[ok]: EVAL - cmsgpack can pack negative int64? (15 ms) -[ok]: EVAL - cmsgpack pack/unpack smoke test (1040 ms) -[ok]: EVAL - cmsgpack can pack and unpack circular references? (11 ms) -[ok]: EVAL - Numerical sanity check from bitop (14 ms) -[ok]: EVAL - Verify minimal bitop functionality (16 ms) -[ok]: EVAL - Able to parse trailing comments (7 ms) -[ok]: EVAL_RO - Successful case (24 ms) -[ok]: EVAL_RO - Cannot run write commands (23 ms) -[ok]: redis.sha1hex() implementation (32 ms) -[ok]: Measures elapsed time os.clock() (1575 ms) -[ok]: Prohibit dangerous lua methods in sandbox (6 ms) -[ok]: Verify execution of prohibit dangerous Lua methods will fail (10 ms) +[ok]: EVAL - JSON numeric decoding (0 ms) +[ok]: EVAL - JSON string decoding (0 ms) +[ok]: EVAL - JSON smoke test (1 ms) +[ok]: EVAL - cmsgpack can pack double? (1 ms) +[ok]: EVAL - cmsgpack can pack negative int64? (0 ms) +[ok]: EVAL - cmsgpack pack/unpack smoke test (229 ms) +[ok]: EVAL - cmsgpack can pack and unpack circular references? (1 ms) +[ok]: EVAL - Numerical sanity check from bitop (2 ms) +[ok]: EVAL - Verify minimal bitop functionality (0 ms) +[ok]: EVAL - Able to parse trailing comments (0 ms) +[ok]: EVAL_RO - Successful case (1 ms) +[ok]: EVAL_RO - Cannot run write commands (1 ms) +[ok]: redis.sha1hex() implementation (1 ms) +[ok]: Measures elapsed time os.clock() (1002 ms) +[ok]: Prohibit dangerous lua methods in sandbox (1 ms) +[ok]: Verify execution of prohibit dangerous Lua methods will fail (5 ms) [ok]: Globals protection reading an undeclared global variable (1 ms) [ok]: Globals protection setting an undeclared global* (1 ms) -[ok]: lua bit.tohex bug (0 ms) -[ok]: Test an example script DECR_IF_GT (3 ms) -[ok]: EVAL does not leak in the Lua stack (2810 ms) -[ok]: Call Redis command with many args from Lua (issue #1764) (17 ms) -[ok]: Number conversion precision test (issue #1118) (14 ms) -[ok]: String containing number precision test (regression of issue #1118) (16 ms) -[ok]: Verify negative arg count is error instead of crash (issue #1842) (16 ms) -[ok]: Scripts can handle commands with incorrect arity (28 ms) -[ok]: Correct handling of reused argv (issue #1939) (15 ms) -[ok]: Functions in the Redis namespace are able to report errors (16 ms) -[ok]: CLUSTER RESET can not be invoke from within a script (12 ms) -[ok]: Script with RESP3 map (92 ms) -[ok]: Script return recursive object (389 ms) -[ok]: Script check unpack with massive arguments (11 ms) +[ok]: lua bit.tohex bug (1 ms) +[ok]: Test an example script DECR_IF_GT (4 ms) +[ok]: EVAL does not leak in the Lua stack (2785 ms) +[ok]: Call Redis command with many args from Lua (issue #1764) (2 ms) +[ok]: Number conversion precision test (issue #1118) (0 ms) +[ok]: String containing number precision test (regression of issue #1118) (0 ms) +[ok]: Verify negative arg count is error instead of crash (issue #1842) (0 ms) +[ok]: Scripts can handle commands with incorrect arity (2 ms) +[ok]: Correct handling of reused argv (issue #1939) (1 ms) +[ok]: Functions in the Redis namespace are able to report errors (1 ms) +[ok]: CLUSTER RESET can not be invoke from within a script (1 ms) +[ok]: Script with RESP3 map (3 ms) +[ok]: Script return recursive object (100 ms) +[ok]: Script check unpack with massive arguments (4 ms) [ok]: Script read key with expiration set (2 ms) -[ok]: Script del key with expiration set (1 ms) +[ok]: Script del key with expiration set (2 ms) [ok]: Script ACL check (3 ms) -[ok]: Binary code loading failed (2 ms) +[ok]: Binary code loading failed (1 ms) [ok]: Try trick global protection 1 (1 ms) -[ok]: Try trick global protection 2 (1 ms) +[ok]: Try trick global protection 2 (2 ms) [ok]: Try trick global protection 3 (1 ms) -[ok]: Try trick global protection 4 (1 ms) +[ok]: Try trick global protection 4 (2 ms) [ok]: Try trick readonly table on valkey table (1 ms) -[ok]: Try trick readonly table on json table (1 ms) -[ok]: Try trick readonly table on cmsgpack table (1 ms) +[ok]: Try trick readonly table on json table (0 ms) +[ok]: Try trick readonly table on cmsgpack table (0 ms) [ok]: Try trick readonly table on bit table (1 ms) [ok]: Test loadfile are not available (1 ms) -[ok]: Test dofile are not available (1 ms) +[ok]: Test dofile are not available (0 ms) [ok]: Test print are not available (1 ms) -=== (scripting) Starting server 127.0.0.1:21390 ok -[ok]: Timedout read-only scripts can be killed by SCRIPT KILL (476 ms) -[ok]: Timedout read-only scripts can be killed by SCRIPT KILL even when use pcall (216 ms) -[ok]: Timedout script does not cause a false dead client (195 ms) -[ok]: Timedout script link is still usable after Lua returns (274 ms) -[ok]: Timedout scripts and unblocked command (803 ms) -[ok]: Timedout scripts that modified data can't be killed by SCRIPT KILL (260 ms) -[ok]: SHUTDOWN NOSAVE can kill a timedout script anyway (33 ms) -=== (scripting repl needs:debug external:skip) Starting server 127.0.0.1:21392 ok -=== () Starting server 127.0.0.1:21394 ok -[ok]: Before the replica connects we issue two EVAL commands (33 ms) -[ok]: Connect a replica to the master instance (212 ms) -[ok]: Replication of script multiple pushes to list with BLPOP (48 ms) -[ok]: Lua scripts using SELECT are replicated correctly (36 ms) -=== (scripting repl external:skip) Starting server 127.0.0.1:21396 ok -=== () Starting server 127.0.0.1:21398 ok -[ok]: Connect a replica to the master instance (125 ms) +=== (scripting) Starting server 127.0.0.1:21453 ok +[ok]: Timedout read-only scripts can be killed by SCRIPT KILL (413 ms) +[ok]: Timedout read-only scripts can be killed by SCRIPT KILL even when use pcall (160 ms) +[ok]: Timedout script does not cause a false dead client (141 ms) +[ok]: Timedout script link is still usable after Lua returns (92 ms) +[ok]: Timedout scripts and unblocked command (369 ms) +[ok]: Timedout scripts that modified data can't be killed by SCRIPT KILL (215 ms) +[ok]: SHUTDOWN NOSAVE can kill a timedout script anyway (19 ms) +=== (scripting repl needs:debug external:skip) Starting server 127.0.0.1:21455 ok +=== () Starting server 127.0.0.1:21457 ok +[ok]: Before the replica connects we issue two EVAL commands (2 ms) +[ok]: Connect a replica to the master instance (105 ms) +[ok]: Replication of script multiple pushes to list with BLPOP (14 ms) +[ok]: Lua scripts using SELECT are replicated correctly (104 ms) +=== (scripting repl external:skip) Starting server 127.0.0.1:21459 ok +=== () Starting server 127.0.0.1:21461 ok +[ok]: Connect a replica to the master instance (107 ms) [ok]: Redis.replicate_commands() can be issued anywhere now (3 ms) -[ok]: Redis.set_repl() can be issued before replicate_commands() now (6 ms) -[ok]: Redis.set_repl() don't accept invalid values (18 ms) -[ok]: Test selective replication of certain commands from Lua (120 ms) -[ok]: PRNG is seeded randomly for command replication (2 ms) -[ok]: Using side effects is not a problem with command replication (1 ms) -=== (scripting needs:debug) Starting server 127.0.0.1:21400 ok -[ok]: test RESP2/2 big number protocol parsing (5 ms) +[ok]: Redis.set_repl() can be issued before replicate_commands() now (1 ms) +[ok]: Redis.set_repl() don't accept invalid values (2 ms) +[ok]: Test selective replication of certain commands from Lua (6 ms) +[ok]: PRNG is seeded randomly for command replication (1 ms) +[ok]: Using side effects is not a problem with command replication (2 ms) +=== (scripting needs:debug) Starting server 127.0.0.1:21463 ok +[ok]: test RESP2/2 big number protocol parsing (1 ms) [ok]: test RESP2/2 malformed big number protocol parsing (1 ms) -[ok]: test RESP2/2 map protocol parsing (1 ms) -[ok]: test RESP2/2 set protocol parsing (1 ms) +[ok]: test RESP2/2 map protocol parsing (0 ms) +[ok]: test RESP2/2 set protocol parsing (0 ms) [ok]: test RESP2/2 double protocol parsing (1 ms) [ok]: test RESP2/2 null protocol parsing (1 ms) -[ok]: test RESP2/2 verbatim protocol parsing (1 ms) +[ok]: test RESP2/2 verbatim protocol parsing (0 ms) [ok]: test RESP2/2 true protocol parsing (1 ms) -[ok]: test RESP2/2 false protocol parsing (0 ms) +[ok]: test RESP2/2 false protocol parsing (1 ms) [ok]: test RESP2/3 big number protocol parsing (1 ms) [ok]: test RESP2/3 malformed big number protocol parsing (1 ms) -[ok]: test RESP2/3 map protocol parsing (1 ms) -[ok]: test RESP2/3 set protocol parsing (2 ms) +[ok]: test RESP2/3 map protocol parsing (0 ms) +[ok]: test RESP2/3 set protocol parsing (0 ms) [ok]: test RESP2/3 double protocol parsing (1 ms) -[ok]: test RESP2/3 null protocol parsing (1 ms) +[ok]: test RESP2/3 null protocol parsing (0 ms) [ok]: test RESP2/3 verbatim protocol parsing (0 ms) [ok]: test RESP2/3 true protocol parsing (1 ms) -[ok]: test RESP2/3 false protocol parsing (1 ms) -[ok]: test RESP3/2 big number protocol parsing (2 ms) -[ok]: test RESP3/2 malformed big number protocol parsing (1 ms) +[ok]: test RESP2/3 false protocol parsing (0 ms) +[ok]: test RESP3/2 big number protocol parsing (1 ms) +[ok]: test RESP3/2 malformed big number protocol parsing (0 ms) [ok]: test RESP3/2 map protocol parsing (1 ms) [ok]: test RESP3/2 set protocol parsing (1 ms) [ok]: test RESP3/2 double protocol parsing (1 ms) -[ok]: test RESP3/2 null protocol parsing (1 ms) +[ok]: test RESP3/2 null protocol parsing (0 ms) [ok]: test RESP3/2 verbatim protocol parsing (1 ms) -[ok]: test RESP3/2 true protocol parsing (1 ms) -[ok]: test RESP3/2 false protocol parsing (1 ms) +[ok]: test RESP3/2 true protocol parsing (0 ms) +[ok]: test RESP3/2 false protocol parsing (0 ms) [ok]: test RESP3/3 big number protocol parsing (1 ms) -[ok]: test RESP3/3 malformed big number protocol parsing (1 ms) -[ok]: test RESP3/3 map protocol parsing (2 ms) -[ok]: test RESP3/3 set protocol parsing (2 ms) +[ok]: test RESP3/3 malformed big number protocol parsing (0 ms) +[ok]: test RESP3/3 map protocol parsing (1 ms) +[ok]: test RESP3/3 set protocol parsing (1 ms) [ok]: test RESP3/3 double protocol parsing (1 ms) -[ok]: test RESP3/3 null protocol parsing (1 ms) +[ok]: test RESP3/3 null protocol parsing (0 ms) [ok]: test RESP3/3 verbatim protocol parsing (1 ms) -[ok]: test RESP3/3 true protocol parsing (1 ms) -[ok]: test RESP3/3 false protocol parsing (1 ms) +[ok]: test RESP3/3 true protocol parsing (0 ms) +[ok]: test RESP3/3 false protocol parsing (0 ms) [ok]: test resp3 attribute protocol parsing (1 ms) -[ok]: Script block the time during execution (13 ms) -[ok]: Script delete the expired key (13 ms) -[ok]: TIME command using cached time (22 ms) -[ok]: Script block the time in some expiration related commands (106 ms) -[ok]: RESTORE expired keys with expiration time (31 ms) -=== (scripting) Starting server 127.0.0.1:21402 ok -[ok]: EVAL - Does Lua interpreter replies to our requests? (7 ms) +[ok]: Script block the time during execution (11 ms) +[ok]: Script delete the expired key (6 ms) +[ok]: TIME command using cached time (12 ms) +[ok]: Script block the time in some expiration related commands (86 ms) +[ok]: RESTORE expired keys with expiration time (16 ms) +=== (scripting) Starting server 127.0.0.1:21465 ok +[ok]: EVAL - Does Lua interpreter replies to our requests? (1 ms) [ok]: EVAL - Return _G (1 ms) -[ok]: EVAL - Return table with a metatable that raise error (1 ms) -[ok]: EVAL - Return table with a metatable that call server (0 ms) +[ok]: EVAL - Return table with a metatable that raise error (0 ms) +[ok]: EVAL - Return table with a metatable that call server (1 ms) [ok]: EVAL - Lua integer -> Redis protocol type conversion (0 ms) -[ok]: EVAL - Lua string -> Redis protocol type conversion (1 ms) -[ok]: EVAL - Lua true boolean -> Redis protocol type conversion (1 ms) +[ok]: EVAL - Lua string -> Redis protocol type conversion (0 ms) +[ok]: EVAL - Lua true boolean -> Redis protocol type conversion (0 ms) [ok]: EVAL - Lua false boolean -> Redis protocol type conversion (0 ms) [ok]: EVAL - Lua status code reply -> Redis protocol type conversion (0 ms) [ok]: EVAL - Lua error reply -> Redis protocol type conversion (1 ms) [ok]: EVAL - Lua table -> Redis protocol type conversion (0 ms) [ok]: EVAL - Are the KEYS and ARGV arrays populated correctly? (1 ms) -[ok]: EVAL - is Lua able to call Redis API? (1 ms) +[ok]: EVAL - is Lua able to call Redis API? (0 ms) [ok]: EVAL - Redis integer -> Lua type conversion (0 ms) -[ok]: EVAL - Lua number -> Redis integer conversion (0 ms) -[ok]: EVAL - Redis bulk -> Lua type conversion (1 ms) -[ok]: EVAL - Redis multi bulk -> Lua type conversion (1 ms) -[ok]: EVAL - Redis status reply -> Lua type conversion (1 ms) +[ok]: EVAL - Lua number -> Redis integer conversion (1 ms) +[ok]: EVAL - Redis bulk -> Lua type conversion (0 ms) +[ok]: EVAL - Redis multi bulk -> Lua type conversion (2 ms) +[ok]: EVAL - Redis status reply -> Lua type conversion (0 ms) [ok]: EVAL - Redis error reply -> Lua type conversion (1 ms) [ok]: EVAL - Redis nil bulk reply -> Lua type conversion (1 ms) [ok]: EVAL - Is the Lua client using the currently selected DB? (1 ms) [ok]: EVAL - SELECT inside Lua should not affect the caller (1 ms) [ok]: EVAL - Scripts do not block on blpop command (1 ms) [ok]: EVAL - Scripts do not block on brpop command (1 ms) -[ok]: EVAL - Scripts do not block on brpoplpush command (2 ms) -[ok]: EVAL - Scripts do not block on blmove command (2 ms) +[ok]: EVAL - Scripts do not block on brpoplpush command (1 ms) +[ok]: EVAL - Scripts do not block on blmove command (1 ms) [ok]: EVAL - Scripts do not block on bzpopmin command (1 ms) [ok]: EVAL - Scripts do not block on bzpopmax command (1 ms) [ok]: EVAL - Scripts do not block on wait (1 ms) @@ -2969,973 +3454,672 @@ [ok]: EVAL - Scripts do not block on XREAD with BLOCK option -- non empty stream (1 ms) [ok]: EVAL - Scripts do not block on XREADGROUP with BLOCK option -- non empty stream (1 ms) [ok]: EVAL - Scripts can run non-deterministic commands (1 ms) -[ok]: EVAL - No arguments to redis.call/pcall is considered an error (1 ms) -[ok]: EVAL - redis.call variant raises a Lua error on Redis cmd error (1) (1 ms) +[ok]: EVAL - No arguments to redis.call/pcall is considered an error (0 ms) [ok]: EVAL - redis.call variant raises a Lua error on Redis cmd error (1) (1 ms) +[ok]: EVAL - redis.call variant raises a Lua error on Redis cmd error (1) (0 ms) [ok]: EVAL - redis.call variant raises a Lua error on Redis cmd error (1) (1 ms) [ignore]: EVAL - JSON string encoding a string larger than 2GB: large memory flag not provided -[ok]: EVAL - JSON numeric decoding (1 ms) +[ok]: EVAL - JSON numeric decoding (0 ms) [ok]: EVAL - JSON string decoding (1 ms) [ok]: EVAL - JSON smoke test (0 ms) [ok]: EVAL - cmsgpack can pack double? (1 ms) [ok]: EVAL - cmsgpack can pack negative int64? (1 ms) -[ok]: EVAL - cmsgpack pack/unpack smoke test (1203 ms) -[ok]: EVAL - cmsgpack can pack and unpack circular references? (8 ms) -[ok]: EVAL - Numerical sanity check from bitop (6 ms) -[ok]: EVAL - Verify minimal bitop functionality (9 ms) -[ok]: EVAL - Able to parse trailing comments (8 ms) -[ok]: EVAL_RO - Successful case (12 ms) -[ok]: EVAL_RO - Cannot run write commands (10 ms) -[ok]: redis.sha1hex() implementation (18 ms) -[ok]: Measures elapsed time os.clock() (1664 ms) -[ok]: Prohibit dangerous lua methods in sandbox (8 ms) -[ok]: Verify execution of prohibit dangerous Lua methods will fail (52 ms) -[ok]: Globals protection reading an undeclared global variable (8 ms) -[ok]: Globals protection setting an undeclared global* (8 ms) -[ok]: lua bit.tohex bug (11 ms) -[ok]: Test an example script DECR_IF_GT (43 ms) -[ok]: EVAL does not leak in the Lua stack (1347 ms) -[ok]: Call Redis command with many args from Lua (issue #1764) (3 ms) +[ok]: EVAL - cmsgpack pack/unpack smoke test (457 ms) +[ok]: EVAL - cmsgpack can pack and unpack circular references? (1 ms) +[ok]: EVAL - Numerical sanity check from bitop (3 ms) +[ok]: EVAL - Verify minimal bitop functionality (0 ms) +[ok]: EVAL - Able to parse trailing comments (1 ms) +[ok]: EVAL_RO - Successful case (1 ms) +[ok]: EVAL_RO - Cannot run write commands (1 ms) +[ok]: redis.sha1hex() implementation (1 ms) +[ok]: Measures elapsed time os.clock() (1090 ms) +[ok]: Prohibit dangerous lua methods in sandbox (0 ms) +[ok]: Verify execution of prohibit dangerous Lua methods will fail (3 ms) +[ok]: Globals protection reading an undeclared global variable (1 ms) +[ok]: Globals protection setting an undeclared global* (0 ms) +[ok]: lua bit.tohex bug (1 ms) +[ok]: Test an example script DECR_IF_GT (2 ms) +[ok]: EVAL does not leak in the Lua stack (3152 ms) +[ok]: Call Redis command with many args from Lua (issue #1764) (1 ms) [ok]: Number conversion precision test (issue #1118) (1 ms) -[ok]: String containing number precision test (regression of issue #1118) (1 ms) -[ok]: Verify negative arg count is error instead of crash (issue #1842) (0 ms) +[ok]: String containing number precision test (regression of issue #1118) (0 ms) +[ok]: Verify negative arg count is error instead of crash (issue #1842) (1 ms) [ok]: Scripts can handle commands with incorrect arity (1 ms) [ok]: Correct handling of reused argv (issue #1939) (1 ms) -[ok]: Functions in the Redis namespace are able to report errors (1 ms) -[ok]: CLUSTER RESET can not be invoke from within a script (1 ms) -[ok]: Script with RESP3 map (3 ms) -[ok]: Script return recursive object (225 ms) -[ok]: Script check unpack with massive arguments (11 ms) -[ok]: Script read key with expiration set (1 ms) +[ok]: Functions in the Redis namespace are able to report errors (0 ms) +[ok]: CLUSTER RESET can not be invoke from within a script (2 ms) +[ok]: Script with RESP3 map (4 ms) +[ok]: Script return recursive object (301 ms) +[ok]: Script check unpack with massive arguments (6 ms) +[ok]: Script read key with expiration set (8 ms) [ok]: Script del key with expiration set (1 ms) -[ok]: Script ACL check (2 ms) -[ok]: Binary code loading failed (9 ms) -[ok]: Try trick global protection 1 (1 ms) +[ok]: Script ACL check (14 ms) +[ok]: Binary code loading failed (1 ms) +[ok]: Try trick global protection 1 (0 ms) [ok]: Try trick global protection 2 (1 ms) [ok]: Try trick global protection 3 (0 ms) -[ok]: Try trick global protection 4 (0 ms) +[ok]: Try trick global protection 4 (2 ms) [ok]: Try trick readonly table on valkey table (1 ms) [ok]: Try trick readonly table on json table (1 ms) -[ok]: Try trick readonly table on cmsgpack table (1 ms) +[ok]: Try trick readonly table on cmsgpack table (0 ms) [ok]: Try trick readonly table on bit table (0 ms) [ok]: Test loadfile are not available (1 ms) -[ok]: Test dofile are not available (2 ms) -[ok]: Test print are not available (1 ms) -=== (scripting) Starting server 127.0.0.1:21404 ok -[ok]: Timedout read-only scripts can be killed by SCRIPT KILL (449 ms) -[ok]: Timedout read-only scripts can be killed by SCRIPT KILL even when use pcall (139 ms) -[ok]: Timedout script does not cause a false dead client (113 ms) -[ok]: Timedout script link is still usable after Lua returns (161 ms) -[ok]: Timedout scripts and unblocked command (499 ms) -[ok]: Timedout scripts that modified data can't be killed by SCRIPT KILL (237 ms) -[ok]: SHUTDOWN NOSAVE can kill a timedout script anyway (21 ms) -=== (scripting repl needs:debug external:skip) Starting server 127.0.0.1:21406 ok -=== () Starting server 127.0.0.1:21408 ok +[ok]: Test dofile are not available (0 ms) +[ok]: Test print are not available (4 ms) +=== (scripting) Starting server 127.0.0.1:21467 ok +[ok]: Timedout read-only scripts can be killed by SCRIPT KILL (415 ms) +[ok]: Timedout read-only scripts can be killed by SCRIPT KILL even when use pcall (184 ms) +[ok]: Timedout script does not cause a false dead client (61 ms) +[ok]: Timedout script link is still usable after Lua returns (79 ms) +[ok]: Timedout scripts and unblocked command (321 ms) +[ok]: Timedout scripts that modified data can't be killed by SCRIPT KILL (215 ms) +[ok]: SHUTDOWN NOSAVE can kill a timedout script anyway (14 ms) +=== (scripting repl needs:debug external:skip) Starting server 127.0.0.1:21469 ok +=== () Starting server 127.0.0.1:21471 ok [ok]: Before the replica connects we issue two EVAL commands (1 ms) -[ok]: Connect a replica to the master instance (130 ms) -[ok]: Replication of script multiple pushes to list with BLPOP (33 ms) -[ok]: Lua scripts using SELECT are replicated correctly (1 ms) -=== (scripting repl external:skip) Starting server 127.0.0.1:21410 ok -=== () Starting server 127.0.0.1:21412 ok -[ok]: Connect a replica to the master instance (147 ms) -[ok]: Redis.replicate_commands() can be issued anywhere now (7 ms) -[ok]: Redis.set_repl() can be issued before replicate_commands() now (8 ms) -[ok]: Redis.set_repl() don't accept invalid values (8 ms) -[ok]: Test selective replication of certain commands from Lua (48 ms) -[ok]: PRNG is seeded randomly for command replication (16 ms) -[ok]: Using side effects is not a problem with command replication (31 ms) -=== (scripting needs:debug) Starting server 127.0.0.1:21414 ok -[ok]: test RESP2/2 big number protocol parsing (5 ms) -[ok]: test RESP2/2 malformed big number protocol parsing (1 ms) -[ok]: test RESP2/2 map protocol parsing (7 ms) -[ok]: test RESP2/2 set protocol parsing (5 ms) -[ok]: test RESP2/2 double protocol parsing (9 ms) -[ok]: test RESP2/2 null protocol parsing (8 ms) -[ok]: test RESP2/2 verbatim protocol parsing (1 ms) -[ok]: test RESP2/2 true protocol parsing (5 ms) -[ok]: test RESP2/2 false protocol parsing (15 ms) -[ok]: test RESP2/3 big number protocol parsing (8 ms) -[ok]: test RESP2/3 malformed big number protocol parsing (7 ms) -[ok]: test RESP2/3 map protocol parsing (16 ms) -[ok]: test RESP2/3 set protocol parsing (4 ms) -[ok]: test RESP2/3 double protocol parsing (11 ms) -[ok]: test RESP2/3 null protocol parsing (8 ms) -[ok]: test RESP2/3 verbatim protocol parsing (11 ms) -[ok]: test RESP2/3 true protocol parsing (6 ms) -[ok]: test RESP2/3 false protocol parsing (4 ms) -[ok]: test RESP3/2 big number protocol parsing (10 ms) -[ok]: test RESP3/2 malformed big number protocol parsing (2 ms) -[ok]: test RESP3/2 map protocol parsing (7 ms) -[ok]: test RESP3/2 set protocol parsing (3 ms) -[ok]: test RESP3/2 double protocol parsing (8 ms) -[ok]: test RESP3/2 null protocol parsing (3 ms) -[ok]: test RESP3/2 verbatim protocol parsing (6 ms) -[ok]: test RESP3/2 true protocol parsing (5 ms) -[ok]: test RESP3/2 false protocol parsing (6 ms) -[ok]: test RESP3/3 big number protocol parsing (8 ms) -[ok]: test RESP3/3 malformed big number protocol parsing (2 ms) -[ok]: test RESP3/3 map protocol parsing (10 ms) -[ok]: test RESP3/3 set protocol parsing (6 ms) -[ok]: test RESP3/3 double protocol parsing (5 ms) -[ok]: test RESP3/3 null protocol parsing (10 ms) -[ok]: test RESP3/3 verbatim protocol parsing (10 ms) -[ok]: test RESP3/3 true protocol parsing (8 ms) -[ok]: test RESP3/3 false protocol parsing (8 ms) -[ok]: test resp3 attribute protocol parsing (8 ms) -[ok]: Script block the time during execution (23 ms) -[ok]: Script delete the expired key (29 ms) -[ok]: TIME command using cached time (14 ms) -[ok]: Script block the time in some expiration related commands (110 ms) -[ok]: RESTORE expired keys with expiration time (15 ms) -=== (scripting) Starting server 127.0.0.1:21416 ok -[ok]: Script - disallow write on OOM (23 ms) -[ok]: EVAL - Does Lua interpreter replies to our requests? (8 ms) -[ok]: EVAL - Return _G (8 ms) -[ok]: EVAL - Return table with a metatable that raise error (7 ms) -[ok]: EVAL - Return table with a metatable that call server (13 ms) -[ok]: EVAL - Lua integer -> Redis protocol type conversion (6 ms) -[ok]: EVAL - Lua string -> Redis protocol type conversion (7 ms) -[ok]: EVAL - Lua true boolean -> Redis protocol type conversion (4 ms) -[ok]: EVAL - Lua false boolean -> Redis protocol type conversion (7 ms) -[ok]: EVAL - Lua status code reply -> Redis protocol type conversion (4 ms) -[ok]: EVAL - Lua error reply -> Redis protocol type conversion (7 ms) -[ok]: EVAL - Lua table -> Redis protocol type conversion (8 ms) -[ok]: EVAL - Are the KEYS and ARGV arrays populated correctly? (8 ms) -[ok]: EVAL - is Lua able to call Redis API? (11 ms) -[ok]: EVALSHA - Can we call a SHA1 if already defined? (7 ms) -[ok]: EVALSHA_RO - Can we call a SHA1 if already defined? (8 ms) -[ok]: EVALSHA - Can we call a SHA1 in uppercase? (8 ms) -[ok]: EVALSHA - Do we get an error on invalid SHA1? (8 ms) -[ok]: EVALSHA - Do we get an error on non defined SHA1? (4 ms) -[ok]: EVAL - Redis integer -> Lua type conversion (15 ms) -[ok]: EVAL - Lua number -> Redis integer conversion (16 ms) -[ok]: EVAL - Redis bulk -> Lua type conversion (15 ms) -[ok]: EVAL - Redis multi bulk -> Lua type conversion (37 ms) -[ok]: EVAL - Redis status reply -> Lua type conversion (7 ms) -[ok]: EVAL - Redis error reply -> Lua type conversion (5 ms) -[ok]: EVAL - Redis nil bulk reply -> Lua type conversion (19 ms) -[ok]: EVAL - Is the Lua client using the currently selected DB? (28 ms) -[ok]: EVAL - SELECT inside Lua should not affect the caller (27 ms) -[ok]: EVAL - Scripts do not block on blpop command (19 ms) -[ok]: EVAL - Scripts do not block on brpop command (13 ms) -[ok]: EVAL - Scripts do not block on brpoplpush command (17 ms) -[ok]: EVAL - Scripts do not block on blmove command (15 ms) -[ok]: EVAL - Scripts do not block on bzpopmin command (17 ms) -[ok]: EVAL - Scripts do not block on bzpopmax command (19 ms) -[ok]: EVAL - Scripts do not block on wait (4 ms) -[ok]: EVAL - Scripts do not block on waitaof (7 ms) -[ok]: EVAL - Scripts do not block on XREAD with BLOCK option (32 ms) -[ok]: EVAL - Scripts do not block on XREADGROUP with BLOCK option (17 ms) -[ok]: EVAL - Scripts do not block on XREAD with BLOCK option -- non empty stream (24 ms) -[ok]: EVAL - Scripts do not block on XREADGROUP with BLOCK option -- non empty stream (12 ms) -[ok]: EVAL - Scripts can run non-deterministic commands (8 ms) -[ok]: EVAL - No arguments to redis.call/pcall is considered an error (8 ms) -[ok]: EVAL - redis.call variant raises a Lua error on Redis cmd error (1) (8 ms) -[ok]: EVAL - redis.call variant raises a Lua error on Redis cmd error (1) (8 ms) -[ok]: EVAL - redis.call variant raises a Lua error on Redis cmd error (1) (16 ms) +[ok]: Connect a replica to the master instance (309 ms) +[ok]: Replication of script multiple pushes to list with BLPOP (16 ms) +[ok]: Lua scripts using SELECT are replicated correctly (4 ms) +=== (scripting repl external:skip) Starting server 127.0.0.1:21473 ok +=== () Starting server 127.0.0.1:21475 ok +[ok]: Connect a replica to the master instance (106 ms) +[ok]: Redis.replicate_commands() can be issued anywhere now (1 ms) +[ok]: Redis.set_repl() can be issued before replicate_commands() now (1 ms) +[ok]: Redis.set_repl() don't accept invalid values (1 ms) +[ok]: Test selective replication of certain commands from Lua (2 ms) +[ok]: PRNG is seeded randomly for command replication (1 ms) +[ok]: Using side effects is not a problem with command replication (2 ms) +=== (scripting needs:debug) Starting server 127.0.0.1:21477 ok +[ok]: test RESP2/2 big number protocol parsing (0 ms) +[ok]: test RESP2/2 malformed big number protocol parsing (0 ms) +[ok]: test RESP2/2 map protocol parsing (0 ms) +[ok]: test RESP2/2 set protocol parsing (1 ms) +[ok]: test RESP2/2 double protocol parsing (0 ms) +[ok]: test RESP2/2 null protocol parsing (0 ms) +[ok]: test RESP2/2 verbatim protocol parsing (0 ms) +[ok]: test RESP2/2 true protocol parsing (0 ms) +[ok]: test RESP2/2 false protocol parsing (1 ms) +[ok]: test RESP2/3 big number protocol parsing (1 ms) +[ok]: test RESP2/3 malformed big number protocol parsing (0 ms) +[ok]: test RESP2/3 map protocol parsing (0 ms) +[ok]: test RESP2/3 set protocol parsing (0 ms) +[ok]: test RESP2/3 double protocol parsing (0 ms) +[ok]: test RESP2/3 null protocol parsing (0 ms) +[ok]: test RESP2/3 verbatim protocol parsing (1 ms) +[ok]: test RESP2/3 true protocol parsing (0 ms) +[ok]: test RESP2/3 false protocol parsing (0 ms) +[ok]: test RESP3/2 big number protocol parsing (0 ms) +[ok]: test RESP3/2 malformed big number protocol parsing (1 ms) +[ok]: test RESP3/2 map protocol parsing (0 ms) +[ok]: test RESP3/2 set protocol parsing (1 ms) +[ok]: test RESP3/2 double protocol parsing (0 ms) +[ok]: test RESP3/2 null protocol parsing (1 ms) +[ok]: test RESP3/2 verbatim protocol parsing (0 ms) +[ok]: test RESP3/2 true protocol parsing (0 ms) +[ok]: test RESP3/2 false protocol parsing (1 ms) +[ok]: test RESP3/3 big number protocol parsing (0 ms) +[ok]: test RESP3/3 malformed big number protocol parsing (0 ms) +[ok]: test RESP3/3 map protocol parsing (1 ms) +[ok]: test RESP3/3 set protocol parsing (0 ms) +[ok]: test RESP3/3 double protocol parsing (0 ms) +[ok]: test RESP3/3 null protocol parsing (0 ms) +[ok]: test RESP3/3 verbatim protocol parsing (1 ms) +[ok]: test RESP3/3 true protocol parsing (0 ms) +[ok]: test RESP3/3 false protocol parsing (1 ms) +[ok]: test resp3 attribute protocol parsing (0 ms) +[ok]: Script block the time during execution (11 ms) +[ok]: Script delete the expired key (4 ms) +[ok]: TIME command using cached time (12 ms) +[ok]: Script block the time in some expiration related commands (85 ms) +[ok]: RESTORE expired keys with expiration time (11 ms) +=== (scripting) Starting server 127.0.0.1:21479 ok +[ok]: Script - disallow write on OOM (2 ms) +[ok]: EVAL - Does Lua interpreter replies to our requests? (2 ms) +[ok]: EVAL - Return _G (0 ms) +[ok]: EVAL - Return table with a metatable that raise error (0 ms) +[ok]: EVAL - Return table with a metatable that call server (1 ms) +[ok]: EVAL - Lua integer -> Redis protocol type conversion (1 ms) +[ok]: EVAL - Lua string -> Redis protocol type conversion (1 ms) +[ok]: EVAL - Lua true boolean -> Redis protocol type conversion (0 ms) +[ok]: EVAL - Lua false boolean -> Redis protocol type conversion (1 ms) +[ok]: EVAL - Lua status code reply -> Redis protocol type conversion (0 ms) +[ok]: EVAL - Lua error reply -> Redis protocol type conversion (1 ms) +[ok]: EVAL - Lua table -> Redis protocol type conversion (0 ms) +[ok]: EVAL - Are the KEYS and ARGV arrays populated correctly? (1 ms) +[ok]: EVAL - is Lua able to call Redis API? (0 ms) +[ok]: EVALSHA - Can we call a SHA1 if already defined? (1 ms) +[ok]: EVALSHA_RO - Can we call a SHA1 if already defined? (1 ms) +[ok]: EVALSHA - Can we call a SHA1 in uppercase? (0 ms) +[ok]: EVALSHA - Do we get an error on invalid SHA1? (1 ms) +[ok]: EVALSHA - Do we get an error on non defined SHA1? (0 ms) +[ok]: EVAL - Redis integer -> Lua type conversion (1 ms) +[ok]: EVAL - Lua number -> Redis integer conversion (0 ms) +[ok]: EVAL - Redis bulk -> Lua type conversion (0 ms) +[ok]: EVAL - Redis multi bulk -> Lua type conversion (2 ms) +[ok]: EVAL - Redis status reply -> Lua type conversion (0 ms) +[ok]: EVAL - Redis error reply -> Lua type conversion (1 ms) +[ok]: EVAL - Redis nil bulk reply -> Lua type conversion (1 ms) +[ok]: EVAL - Is the Lua client using the currently selected DB? (1 ms) +[ok]: EVAL - SELECT inside Lua should not affect the caller (1 ms) +[ok]: EVAL - Scripts do not block on blpop command (3 ms) +[ok]: EVAL - Scripts do not block on brpop command (1 ms) +[ok]: EVAL - Scripts do not block on brpoplpush command (0 ms) +[ok]: EVAL - Scripts do not block on blmove command (1 ms) +[ok]: EVAL - Scripts do not block on bzpopmin command (2 ms) +[ok]: EVAL - Scripts do not block on bzpopmax command (1 ms) +[ok]: EVAL - Scripts do not block on wait (0 ms) +[ok]: EVAL - Scripts do not block on waitaof (1 ms) +[ok]: EVAL - Scripts do not block on XREAD with BLOCK option (1 ms) +[ok]: EVAL - Scripts do not block on XREADGROUP with BLOCK option (0 ms) +[ok]: EVAL - Scripts do not block on XREAD with BLOCK option -- non empty stream (2 ms) +[ok]: EVAL - Scripts do not block on XREADGROUP with BLOCK option -- non empty stream (0 ms) +[ok]: EVAL - Scripts can run non-deterministic commands (1 ms) +[ok]: EVAL - No arguments to redis.call/pcall is considered an error (0 ms) +[ok]: EVAL - redis.call variant raises a Lua error on Redis cmd error (1) (0 ms) +[ok]: EVAL - redis.call variant raises a Lua error on Redis cmd error (1) (1 ms) +[ok]: EVAL - redis.call variant raises a Lua error on Redis cmd error (1) (1 ms) [ignore]: EVAL - JSON string encoding a string larger than 2GB: large memory flag not provided -[ok]: EVAL - JSON numeric decoding (8 ms) -[ok]: EVAL - JSON string decoding (3 ms) -[ok]: EVAL - JSON smoke test (8 ms) -[ok]: EVAL - cmsgpack can pack double? (8 ms) -[ok]: EVAL - cmsgpack can pack negative int64? (8 ms) -[ok]: EVAL - cmsgpack pack/unpack smoke test (1068 ms) -[ok]: EVAL - cmsgpack can pack and unpack circular references? (8 ms) -[ok]: EVAL - Numerical sanity check from bitop (7 ms) -[ok]: EVAL - Verify minimal bitop functionality (7 ms) -[ok]: EVAL - Able to parse trailing comments (10 ms) -[ok]: EVAL_RO - Successful case (7 ms) -[ok]: EVAL_RO - Cannot run write commands (15 ms) -[ok]: SCRIPTING FLUSH - is able to clear the scripts cache? (188 ms) -[ok]: SCRIPTING FLUSH ASYNC (742 ms) -[ok]: SCRIPT EXISTS - can detect already defined scripts? (17 ms) -[ok]: SCRIPT LOAD - is able to register scripts in the scripting cache (11 ms) -[ok]: SCRIPT SHOW - is able to dump scripts from the scripting cache (12 ms) -[ok]: SCRIPT SHOW - wrong sha1 length or invalid sha1 char return noscript error (16 ms) -[ok]: SCRIPT SHOW - script not exist return noscript error (15 ms) -[ok]: SORT is normally not alpha re-ordered for the scripting engine (23 ms) -[ok]: SORT BY output gets ordered for scripting (28 ms) -[ok]: SORT BY with GET gets ordered for scripting (20 ms) -[ok]: redis.sha1hex() implementation (16 ms) -[ok]: Measures elapsed time os.clock() (1952 ms) -[ok]: Prohibit dangerous lua methods in sandbox (6 ms) -[ok]: Verify execution of prohibit dangerous Lua methods will fail (54 ms) -[ok]: Globals protection reading an undeclared global variable (8 ms) -[ok]: Globals protection setting an undeclared global* (4 ms) -[ok]: lua bit.tohex bug (11 ms) -[ok]: Test an example script DECR_IF_GT (44 ms) -[ok]: random numbers are random now (9 ms) -[ok]: Scripting engine PRNG can be seeded correctly (31 ms) -[ok]: EVAL does not leak in the Lua stack (1131 ms) -[ok]: SPOP: We can call scripts rewriting client->argv from Lua (195 ms) -[ok]: MGET: mget shouldn't be propagated in Lua (227 ms) -[ok]: EXPIRE: We can call scripts rewriting client->argv from Lua (111 ms) -=== (scripting) Starting server 127.0.0.1:21418 ok -[ok]: INCRBYFLOAT: We can call scripts expanding client->argv from Lua (654 ms) -[ok]: Call Redis command with many args from Lua (issue #1764) (6 ms) -[ok]: Number conversion precision test (issue #1118) (3 ms) -[ok]: String containing number precision test (regression of issue #1118) (3 ms) -[ok]: Verify negative arg count is error instead of crash (issue #1842) (8 ms) -[ok]: Scripts can handle commands with incorrect arity (16 ms) -[ok]: Correct handling of reused argv (issue #1939) (4 ms) -[ok]: Functions in the Redis namespace are able to report errors (6 ms) -[ok]: CLUSTER RESET can not be invoke from within a script (4 ms) -[ok]: Script with RESP3 map (42 ms) -[ok]: Script return recursive object (203 ms) -[ok]: Script check unpack with massive arguments (11 ms) -[ok]: Script read key with expiration set (2 ms) +[ok]: EVAL - JSON numeric decoding (0 ms) +[ok]: EVAL - JSON string decoding (1 ms) +[ok]: EVAL - JSON smoke test (3 ms) +[ok]: EVAL - cmsgpack can pack double? (1 ms) +[ok]: EVAL - cmsgpack can pack negative int64? (2 ms) +[ok]: EVAL - cmsgpack pack/unpack smoke test (395 ms) +[ok]: EVAL - cmsgpack can pack and unpack circular references? (0 ms) +[ok]: EVAL - Numerical sanity check from bitop (0 ms) +[ok]: EVAL - Verify minimal bitop functionality (0 ms) +[ok]: EVAL - Able to parse trailing comments (1 ms) +[ok]: EVAL_RO - Successful case (0 ms) +[ok]: EVAL_RO - Cannot run write commands (1 ms) +[ok]: SCRIPTING FLUSH - is able to clear the scripts cache? (25 ms) +[ok]: SCRIPTING FLUSH ASYNC (29 ms) +[ok]: SCRIPT EXISTS - can detect already defined scripts? (1 ms) +[ok]: SCRIPT LOAD - is able to register scripts in the scripting cache (1 ms) +[ok]: SCRIPT SHOW - is able to dump scripts from the scripting cache (0 ms) +[ok]: SCRIPT SHOW - wrong sha1 length or invalid sha1 char return noscript error (1 ms) +[ok]: SCRIPT SHOW - script not exist return noscript error (1 ms) +[ok]: SORT is normally not alpha re-ordered for the scripting engine (1 ms) +[ok]: SORT BY output gets ordered for scripting (1 ms) +[ok]: SORT BY with GET gets ordered for scripting (1 ms) +[ok]: redis.sha1hex() implementation (1 ms) +[ok]: Measures elapsed time os.clock() (1029 ms) +[ok]: Prohibit dangerous lua methods in sandbox (0 ms) +[ok]: Verify execution of prohibit dangerous Lua methods will fail (4 ms) +[ok]: Globals protection reading an undeclared global variable (0 ms) +[ok]: Globals protection setting an undeclared global* (1 ms) +[ok]: lua bit.tohex bug (2 ms) +[ok]: Test an example script DECR_IF_GT (2 ms) +[ok]: random numbers are random now (1 ms) +[ok]: Scripting engine PRNG can be seeded correctly (2 ms) +[ok]: EVAL does not leak in the Lua stack (2172 ms) +[ok]: SPOP: We can call scripts rewriting client->argv from Lua (811 ms) +[ok]: MGET: mget shouldn't be propagated in Lua (102 ms) +[ok]: EXPIRE: We can call scripts rewriting client->argv from Lua (99 ms) +=== (scripting) Starting server 127.0.0.1:21481 ok +[ok]: INCRBYFLOAT: We can call scripts expanding client->argv from Lua (361 ms) +[ok]: Call Redis command with many args from Lua (issue #1764) (5 ms) +[ok]: Number conversion precision test (issue #1118) (0 ms) +[ok]: String containing number precision test (regression of issue #1118) (0 ms) +[ok]: Verify negative arg count is error instead of crash (issue #1842) (0 ms) +[ok]: Scripts can handle commands with incorrect arity (1 ms) +[ok]: Correct handling of reused argv (issue #1939) (1 ms) +[ok]: Functions in the Redis namespace are able to report errors (0 ms) +[ok]: CLUSTER RESET can not be invoke from within a script (0 ms) +[ok]: Script with RESP3 map (2 ms) +[ok]: Script return recursive object (145 ms) +[ok]: Script check unpack with massive arguments (4 ms) +[ok]: Script read key with expiration set (0 ms) [ok]: Script del key with expiration set (0 ms) [ok]: Script ACL check (2 ms) [ok]: Binary code loading failed (0 ms) -[ok]: Try trick global protection 1 (0 ms) -[ok]: Try trick global protection 2 (1 ms) +[ok]: Try trick global protection 1 (1 ms) +[ok]: Try trick global protection 2 (0 ms) [ok]: Try trick global protection 3 (0 ms) -[ok]: Try trick global protection 4 (1 ms) -[ok]: Try trick readonly table on valkey table (0 ms) -[ok]: Try trick readonly table on json table (1 ms) +[ok]: Try trick global protection 4 (0 ms) +[ok]: Try trick readonly table on valkey table (1 ms) +[ok]: Try trick readonly table on json table (0 ms) [ok]: Try trick readonly table on cmsgpack table (1 ms) -[ok]: Try trick readonly table on bit table (1 ms) +[ok]: Try trick readonly table on bit table (0 ms) [ok]: Test loadfile are not available (1 ms) [ok]: Test dofile are not available (0 ms) -[ok]: Test print are not available (0 ms) -=== (scripting) Starting server 127.0.0.1:21420 ok -[ok]: Timedout read-only scripts can be killed by SCRIPT KILL (460 ms) -[ok]: Timedout read-only scripts can be killed by SCRIPT KILL even when use pcall (216 ms) -[ok]: Timedout script does not cause a false dead client (207 ms) -[ok]: Timedout script link is still usable after Lua returns (195 ms) -[ok]: Timedout scripts and unblocked command (763 ms) -[ok]: Timedout scripts that modified data can't be killed by SCRIPT KILL (243 ms) -[ok]: SHUTDOWN NOSAVE can kill a timedout script anyway (13 ms) -=== (scripting repl needs:debug external:skip) Starting server 127.0.0.1:21422 ok -=== () Starting server 127.0.0.1:21424 ok -[ok]: Before the replica connects we issue two EVAL commands (15 ms) -[ok]: Connect a replica to the master instance (144 ms) -[ok]: Now use EVALSHA against the master, with both SHAs (11 ms) -[ok]: 'x' should be '4' for EVALSHA being replicated by effects (4 ms) -[ok]: Replication of script multiple pushes to list with BLPOP (19 ms) -[ok]: EVALSHA replication when first call is readonly (7 ms) -[ok]: Lua scripts using SELECT are replicated correctly (4 ms) -=== (scripting repl external:skip) Starting server 127.0.0.1:21426 ok -=== () Starting server 127.0.0.1:21428 ok -[ok]: Connect a replica to the master instance (143 ms) -[ok]: Redis.replicate_commands() can be issued anywhere now (8 ms) -[ok]: Redis.set_repl() can be issued before replicate_commands() now (4 ms) -[ok]: Redis.set_repl() don't accept invalid values (8 ms) -[ok]: Test selective replication of certain commands from Lua (48 ms) -[ok]: PRNG is seeded randomly for command replication (15 ms) -[ok]: Using side effects is not a problem with command replication (28 ms) -=== (scripting external:skip) Starting server 127.0.0.1:21430 ok -=== (scripting needs:debug external:skip) Starting server 127.0.0.1:21432 ok -[ok]: Test scripting debug protocol parsing (50 ms) -[ok]: Test scripting debug lua stack overflow (59 ms) -[ok]: Test scripting debug lua server invocations (65 ms) -=== (scripting external:skip) Starting server 127.0.0.1:21434 ok -[ok]: Lua scripts eviction does not generate many scripts (1083 ms) -[ok]: Lua scripts eviction is plain LRU (113 ms) -[ok]: Lua scripts eviction does not affect script load (268 ms) -[ok]: Lua scripts promoted from eval to script load (496 ms) -=== (scripting needs:debug) Starting server 127.0.0.1:21436 ok -[ok]: test RESP2/2 big number protocol parsing (2 ms) +[ok]: Test print are not available (1 ms) +=== (scripting) Starting server 127.0.0.1:21483 ok +[ok]: Timedout read-only scripts can be killed by SCRIPT KILL (418 ms) +[ok]: Timedout read-only scripts can be killed by SCRIPT KILL even when use pcall (135 ms) +[ok]: Timedout script does not cause a false dead client (55 ms) +[ok]: Timedout script link is still usable after Lua returns (80 ms) +[ok]: Timedout scripts and unblocked command (344 ms) +[ok]: Timedout scripts that modified data can't be killed by SCRIPT KILL (212 ms) +[ok]: SHUTDOWN NOSAVE can kill a timedout script anyway (74 ms) +=== (scripting repl needs:debug external:skip) Starting server 127.0.0.1:21485 ok +=== () Starting server 127.0.0.1:21487 ok +[ok]: Before the replica connects we issue two EVAL commands (1 ms) +[ok]: Connect a replica to the master instance (123 ms) +[ok]: Now use EVALSHA against the master, with both SHAs (1 ms) +[ok]: 'x' should be '4' for EVALSHA being replicated by effects (1 ms) +[ok]: Replication of script multiple pushes to list with BLPOP (23 ms) +[ok]: EVALSHA replication when first call is readonly (2 ms) +[ok]: Lua scripts using SELECT are replicated correctly (2 ms) +=== (scripting repl external:skip) Starting server 127.0.0.1:21489 ok +=== () Starting server 127.0.0.1:21491 ok +[ok]: Connect a replica to the master instance (103 ms) +[ok]: Redis.replicate_commands() can be issued anywhere now (0 ms) +[ok]: Redis.set_repl() can be issued before replicate_commands() now (1 ms) +[ok]: Redis.set_repl() don't accept invalid values (0 ms) +[ok]: Test selective replication of certain commands from Lua (2 ms) +[ok]: PRNG is seeded randomly for command replication (1 ms) +[ok]: Using side effects is not a problem with command replication (1 ms) +=== (scripting external:skip) Starting server 127.0.0.1:21493 ok +=== (scripting needs:debug external:skip) Starting server 127.0.0.1:21495 ok +[ok]: Test scripting debug protocol parsing (1 ms) +[ok]: Test scripting debug lua stack overflow (8 ms) +[ok]: Test scripting debug lua server invocations (8 ms) +=== (scripting external:skip) Starting server 127.0.0.1:21497 ok +[ok]: Lua scripts eviction does not generate many scripts (208 ms) +[ok]: Lua scripts eviction is plain LRU (101 ms) +[ok]: Lua scripts eviction does not affect script load (688 ms) +[ok]: Lua scripts promoted from eval to script load (192 ms) +=== (scripting needs:debug) Starting server 127.0.0.1:21499 ok +[ok]: test RESP2/2 big number protocol parsing (1 ms) [ok]: test RESP2/2 malformed big number protocol parsing (0 ms) -[ok]: test RESP2/2 map protocol parsing (0 ms) +[ok]: test RESP2/2 map protocol parsing (1 ms) [ok]: test RESP2/2 set protocol parsing (0 ms) [ok]: test RESP2/2 double protocol parsing (1 ms) [ok]: test RESP2/2 null protocol parsing (0 ms) -[ok]: test RESP2/2 verbatim protocol parsing (1 ms) -[ok]: test RESP2/2 true protocol parsing (0 ms) -[ok]: test RESP2/2 false protocol parsing (1 ms) +[ok]: test RESP2/2 verbatim protocol parsing (0 ms) +[ok]: test RESP2/2 true protocol parsing (1 ms) +[ok]: test RESP2/2 false protocol parsing (0 ms) [ok]: test RESP2/3 big number protocol parsing (0 ms) -[ok]: test RESP2/3 malformed big number protocol parsing (1 ms) -[ok]: test RESP2/3 map protocol parsing (0 ms) -[ok]: test RESP2/3 set protocol parsing (1 ms) -[ok]: test RESP2/3 double protocol parsing (1 ms) +[ok]: test RESP2/3 malformed big number protocol parsing (0 ms) +[ok]: test RESP2/3 map protocol parsing (1 ms) +[ok]: test RESP2/3 set protocol parsing (0 ms) +[ok]: test RESP2/3 double protocol parsing (0 ms) [ok]: test RESP2/3 null protocol parsing (0 ms) [ok]: test RESP2/3 verbatim protocol parsing (0 ms) [ok]: test RESP2/3 true protocol parsing (0 ms) -[ok]: test RESP2/3 false protocol parsing (0 ms) +[ok]: test RESP2/3 false protocol parsing (1 ms) [ok]: test RESP3/2 big number protocol parsing (1 ms) -[ok]: test RESP3/2 malformed big number protocol parsing (2 ms) -[ok]: test RESP3/2 map protocol parsing (1 ms) +[ok]: test RESP3/2 malformed big number protocol parsing (0 ms) +[ok]: test RESP3/2 map protocol parsing (0 ms) [ok]: test RESP3/2 set protocol parsing (0 ms) -[ok]: test RESP3/2 double protocol parsing (1 ms) -[ok]: test RESP3/2 null protocol parsing (0 ms) -[ok]: test RESP3/2 verbatim protocol parsing (1 ms) +[ok]: test RESP3/2 double protocol parsing (0 ms) +[ok]: test RESP3/2 null protocol parsing (1 ms) +[ok]: test RESP3/2 verbatim protocol parsing (0 ms) [ok]: test RESP3/2 true protocol parsing (0 ms) [ok]: test RESP3/2 false protocol parsing (1 ms) -[ok]: test RESP3/3 big number protocol parsing (0 ms) +[ok]: test RESP3/3 big number protocol parsing (1 ms) [ok]: test RESP3/3 malformed big number protocol parsing (0 ms) -[ok]: test RESP3/3 map protocol parsing (1 ms) -[ok]: test RESP3/3 set protocol parsing (0 ms) -[ok]: test RESP3/3 double protocol parsing (1 ms) +[ok]: test RESP3/3 map protocol parsing (0 ms) +[ok]: test RESP3/3 set protocol parsing (1 ms) +[ok]: test RESP3/3 double protocol parsing (0 ms) [ok]: test RESP3/3 null protocol parsing (0 ms) [ok]: test RESP3/3 verbatim protocol parsing (1 ms) -[ok]: test RESP3/3 true protocol parsing (1 ms) +[ok]: test RESP3/3 true protocol parsing (0 ms) [ok]: test RESP3/3 false protocol parsing (0 ms) [ok]: test resp3 attribute protocol parsing (0 ms) -[ok]: Script block the time during execution (12 ms) -[ok]: Script delete the expired key (4 ms) +[ok]: Script block the time during execution (11 ms) +[ok]: Script delete the expired key (3 ms) [ok]: TIME command using cached time (11 ms) -[ok]: Script block the time in some expiration related commands (100 ms) -[ok]: RESTORE expired keys with expiration time (13 ms) -=== (scripting) Starting server 127.0.0.1:21438 ok -[ok]: Shebang support for lua engine (15 ms) -[ok]: Unknown shebang option (6 ms) -[ok]: Unknown shebang flag (5 ms) -[ok]: allow-oom shebang flag (8 ms) -[ok]: no-writes shebang flag (0 ms) -=== (external:skip) Starting server 127.0.0.1:21440 ok -[ok]: no-writes shebang flag on replica (222 ms) -[ok]: not enough good replicas (51 ms) -=== (external:skip) Starting server 127.0.0.1:21442 ok -[ok]: not enough good replicas state change during long script (1399 ms) -[ok]: allow-stale shebang flag (68 ms) -[ok]: reject script do not cause a Lua stack leak (419 ms) -=== (scripting) Starting server 127.0.0.1:21444 ok -[ok]: Consistent eval error reporting (169 ms) +[ok]: Script block the time in some expiration related commands (81 ms) +[ok]: RESTORE expired keys with expiration time (10 ms) +=== (scripting) Starting server 127.0.0.1:21501 ok +[ok]: Shebang support for lua engine (1 ms) +[ok]: Unknown shebang option (0 ms) +[ok]: Unknown shebang flag (0 ms) +[ok]: allow-oom shebang flag (2 ms) +[ok]: no-writes shebang flag (1 ms) +=== (external:skip) Starting server 127.0.0.1:21503 ok +[ok]: no-writes shebang flag on replica (125 ms) +[ok]: not enough good replicas (3 ms) +=== (external:skip) Starting server 127.0.0.1:21505 ok +[ok]: not enough good replicas state change during long script (829 ms) +[ok]: allow-stale shebang flag (3 ms) +[ok]: reject script do not cause a Lua stack leak (13 ms) +=== (scripting) Starting server 127.0.0.1:21507 ok +[ok]: Consistent eval error reporting (16 ms) [ok]: LUA redis.error_reply API (1 ms) -[ok]: LUA redis.error_reply API with empty string (5 ms) -[ok]: LUA redis.status_reply API (2 ms) +[ok]: LUA redis.error_reply API with empty string (1 ms) +[ok]: LUA redis.status_reply API (0 ms) [ok]: LUA test pcall (0 ms) -[ok]: LUA test pcall with error (1 ms) -[ok]: LUA test pcall with non string/integer arg (0 ms) -[ok]: LUA test trim string as expected (3 ms) -[37/124 done]: unit/scripting (63 seconds) -Testing unit/quit -=== (quit) Starting server 127.0.0.1:21446 ok -[ok]: QUIT returns OK (35 ms) -[ok]: Pipelined commands after QUIT must not be executed (67 ms) -[ok]: Pipelined commands after QUIT that exceed read buffer size (64 ms) -[38/124 done]: unit/quit (0 seconds) -Testing unit/functions -=== (scripting) Starting server 127.0.0.1:21448 ok -[ok]: FUNCTION - Basic usage (17 ms) -[ok]: FUNCTION - Load with unknown argument (5 ms) -[ok]: FUNCTION - Create an already exiting library raise error (6 ms) -[ok]: FUNCTION - Create an already exiting library raise error (case insensitive) (8 ms) -[ok]: FUNCTION - Create a library with wrong name format (8 ms) -[ok]: FUNCTION - Create library with unexisting engine (8 ms) -[ok]: FUNCTION - Test uncompiled script (6 ms) -[ok]: FUNCTION - test replace argument (6 ms) -[ok]: FUNCTION - test function case insensitive (4 ms) -[ok]: FUNCTION - test replace argument with failure keeps old libraries (9 ms) -[ok]: FUNCTION - test function delete (3 ms) -[ok]: FUNCTION - test fcall bad arguments (5 ms) -[ok]: FUNCTION - test fcall bad number of keys arguments (4 ms) -[ok]: FUNCTION - test fcall negative number of keys (2 ms) -[ok]: FUNCTION - test delete on not exiting library (2 ms) -[ok]: FUNCTION - test function kill when function is not running (1 ms) -[ok]: FUNCTION - test wrong subcommand (1 ms) -[ok]: FUNCTION - test loading from rdb (26 ms) -[ok]: FUNCTION - test debug reload different options (73 ms) -[ok]: FUNCTION - test debug reload with nosave and noflush (79 ms) -[ok]: FUNCTION - test flushall and flushdb do not clean functions (40 ms) -[ok]: FUNCTION - test function dump and restore (56 ms) -[ok]: FUNCTION - test function dump and restore with flush argument (41 ms) -[ok]: FUNCTION - test function dump and restore with append argument (66 ms) -[ok]: FUNCTION - test function dump and restore with replace argument (71 ms) -[ok]: FUNCTION - test function restore with bad payload do not drop existing functions (32 ms) -[ok]: FUNCTION - test function restore with wrong number of arguments (8 ms) -[ok]: FUNCTION - test fcall_ro with write command (12 ms) -[ok]: FUNCTION - test fcall_ro with read only commands (20 ms) -[ok]: FUNCTION - test keys and argv (24 ms) -[ok]: FUNCTION - test command get keys on fcall (8 ms) -[ok]: FUNCTION - test command get keys on fcall_ro (7 ms) -[ok]: FUNCTION - test function kill (468 ms) -[ok]: FUNCTION - test script kill not working on function (483 ms) -[ok]: FUNCTION - test function kill not working on eval (468 ms) -[ok]: FUNCTION - test function flush (74 ms) -[ok]: FUNCTION - test function wrong argument (6 ms) -=== (scripting repl external:skip) Starting server 127.0.0.1:21450 ok -=== () Starting server 127.0.0.1:21452 ok -[ok]: Connect a replica to the master instance (157 ms) -[ok]: FUNCTION - creation is replicated to replica (16 ms) -[ok]: FUNCTION - call on replica (7 ms) -[ok]: FUNCTION - restore is replicated to replica (28 ms) -[ok]: FUNCTION - delete is replicated to replica (15 ms) -[ok]: FUNCTION - flush is replicated to replica (25 ms) -[ok]: Disconnecting the replica from master instance (142 ms) -[ok]: FUNCTION - test replication to replica on rdb phase (6 ms) -[ok]: FUNCTION - test replication to replica on rdb phase info command (4 ms) -[ok]: FUNCTION - create on read only replica (5 ms) -[ok]: FUNCTION - delete on read only replica (1 ms) -[ok]: FUNCTION - function effect is replicated to replica (27 ms) -[ok]: FUNCTION - modify key space of read only replica (6 ms) -=== () Starting server 127.0.0.1:21454 ok - -Waiting for background AOF rewrite to finish... -Waiting for background AOF rewrite to finish... [ok]: FUNCTION can processes create, delete and flush commands in AOF when doing "debug loadaof" in read-only slaves (764 ms) -=== (scripting) Starting server 127.0.0.1:21456 ok -[ok]: LIBRARIES - test shared function can access default globals (14 ms) -[ok]: LIBRARIES - usage and code sharing (29 ms) -[ok]: LIBRARIES - test registration failure revert the entire load (22 ms) -[ok]: LIBRARIES - test registration function name collision (24 ms) -[ok]: LIBRARIES - test registration function name collision on same library (11 ms) -[ok]: LIBRARIES - test registration with no argument (8 ms) -[ok]: LIBRARIES - test registration with only name (8 ms) -[ok]: LIBRARIES - test registration with to many arguments (7 ms) -[ok]: LIBRARIES - test registration with no string name (7 ms) -[ok]: LIBRARIES - test registration with wrong name format (7 ms) -[ok]: LIBRARIES - test registration with empty name (8 ms) -[ok]: LIBRARIES - math.random from function load (7 ms) -[ok]: LIBRARIES - redis.call from function load (8 ms) -[ok]: LIBRARIES - redis.setresp from function load (7 ms) -[ok]: LIBRARIES - redis.set_repl from function load (7 ms) -[ok]: LIBRARIES - redis.acl_check_cmd from function load (8 ms) -[ok]: LIBRARIES - malicious access test (40 ms) -[ok]: LIBRARIES - delete removed all functions on library (8 ms) -[ok]: LIBRARIES - register function inside a function (12 ms) -[ok]: LIBRARIES - register library with no functions (16 ms) -[ok]: LIBRARIES - load timeout (504 ms) -[ok]: LIBRARIES - verify global protection on the load run (1 ms) -[ok]: LIBRARIES - named arguments (6 ms) -[ok]: LIBRARIES - named arguments, bad function name (8 ms) -[ok]: LIBRARIES - named arguments, bad callback type (8 ms) -[ok]: LIBRARIES - named arguments, bad description (1 ms) -[ok]: LIBRARIES - named arguments, unknown argument (0 ms) -[ok]: LIBRARIES - named arguments, missing function name (1 ms) -[ok]: LIBRARIES - named arguments, missing callback (8 ms) -[ok]: FUNCTION - test function restore with function name collision (80 ms) -[ok]: FUNCTION - test function list with code (1 ms) -[ok]: FUNCTION - test function list with pattern (0 ms) -[ok]: FUNCTION - test function list wrong argument (0 ms) -[ok]: FUNCTION - test function list with bad argument to library name (0 ms) -[ok]: FUNCTION - test function list withcode multiple times (0 ms) -[ok]: FUNCTION - test function list libraryname multiple times (1 ms) -[ok]: FUNCTION - verify OOM on function load and function restore (6 ms) -[ok]: FUNCTION - verify allow-omm allows running any command (2 ms) -=== (scripting) Starting server 127.0.0.1:21458 ok -[ok]: FUNCTION - wrong flags type named arguments (13 ms) -[ok]: FUNCTION - wrong flag type (7 ms) -[ok]: FUNCTION - unknown flag (4 ms) -[ok]: FUNCTION - write script on fcall_ro (8 ms) -[ok]: FUNCTION - write script with no-writes flag (12 ms) -[ok]: FUNCTION - deny oom (18 ms) -[ok]: FUNCTION - deny oom on no-writes function (23 ms) -[ok]: FUNCTION - allow stale (58 ms) -[ok]: FUNCTION - valkey version api (18 ms) -[ok]: FUNCTION - function stats (24 ms) -[ok]: FUNCTION - function stats reloaded correctly from rdb (40 ms) -[ok]: FUNCTION - function stats delete library (16 ms) -[ok]: FUNCTION - test function stats on loading failure (20 ms) -[ok]: FUNCTION - function stats cleaned after flush (22 ms) -[ok]: FUNCTION - function test empty engine (4 ms) -[ok]: FUNCTION - function test unknown metadata value (4 ms) -[ok]: FUNCTION - function test no name (6 ms) -[ok]: FUNCTION - function test multiple names (6 ms) -[ok]: FUNCTION - function test name with quotes (3 ms) -[ok]: FUNCTION - trick global protection 1 (14 ms) -[ok]: FUNCTION - test getmetatable on script load (21 ms) -[39/124 done]: unit/functions (7 seconds) -Testing unit/pubsub -=== (pubsub network) Starting server 127.0.0.1:21460 ok -[ok]: Pub/Sub PING on RESP2 (17 ms) -[ok]: Pub/Sub PING on RESP3 (23 ms) -[ok]: PUBLISH/SUBSCRIBE basics (83 ms) -[ok]: PUBLISH/SUBSCRIBE with two clients (73 ms) -[ok]: PUBLISH/SUBSCRIBE after UNSUBSCRIBE without arguments (55 ms) -[ok]: SUBSCRIBE to one channel more than once (44 ms) -[ok]: UNSUBSCRIBE from non-subscribed channels (39 ms) -[ok]: PUBLISH/PSUBSCRIBE basics (79 ms) -[ok]: PUBLISH/PSUBSCRIBE with two clients (49 ms) -[ok]: PUBLISH/PSUBSCRIBE after PUNSUBSCRIBE without arguments (26 ms) -[ok]: PubSub messages with CLIENT REPLY OFF (116 ms) -[ok]: PUNSUBSCRIBE from non-subscribed channels (36 ms) -[ok]: NUMSUB returns numbers, not strings (#1561) (1 ms) -[ok]: NUMPATs returns the number of unique patterns (66 ms) -[ok]: Mix SUBSCRIBE and PSUBSCRIBE (23 ms) -[ok]: PUNSUBSCRIBE and UNSUBSCRIBE should always reply (1 ms) -[ok]: Keyspace notifications: we receive keyspace notifications (70 ms) -[ok]: Keyspace notifications: we receive keyevent notifications (75 ms) -[ok]: Keyspace notifications: we can receive both kind of events (27 ms) -[ok]: Keyspace notifications: we are able to mask events (23 ms) -[ok]: Keyspace notifications: general events test (20 ms) -[ok]: Keyspace notifications: list events test (23 ms) -[ok]: Keyspace notifications: set events test (24 ms) -[ok]: Keyspace notifications: zset events test (21 ms) -[ok]: Keyspace notifications: hash events test (20 ms) -[ok]: Keyspace notifications: stream events test (22 ms) -[ok]: Keyspace notifications: expired events (triggered expire) (128 ms) -[ok]: Keyspace notifications: expired events (background expire) (208 ms) -[ok]: Keyspace notifications: evicted events (65 ms) -[ok]: Keyspace notifications: test CONFIG GET/SET of event flags (61 ms) -[ok]: Keyspace notifications: new key test (81 ms) -[ok]: publish to self inside multi (55 ms) -[ok]: publish to self inside script (24 ms) -[ok]: unsubscribe inside multi, and publish to self (68 ms) -[40/124 done]: unit/pubsub (3 seconds) -Testing unit/geo -=== (geo) Starting server 127.0.0.1:21462 ok -[ok]: GEO with wrong type src key (91 ms) -[ok]: GEO with non existing src key (72 ms) -[ok]: GEO BYLONLAT with empty search (48 ms) -[ok]: GEO BYMEMBER with non existing member (47 ms) -[ok]: GEOADD create (8 ms) -[ok]: GEOADD update (7 ms) -[ok]: GEOADD update with CH option (15 ms) -[ok]: GEOADD update with NX option (16 ms) -[ok]: GEOADD update with XX option (15 ms) -[ok]: GEOADD update with CH NX option (8 ms) -[ok]: GEOADD update with CH XX option (7 ms) -[ok]: GEOADD update with XX NX option will return syntax error (8 ms) -[ok]: GEOADD update with invalid option (8 ms) -[ok]: GEOADD invalid coordinates (8 ms) -[ok]: GEOADD multi add (6 ms) -[ok]: Check geoset values (8 ms) -[ok]: GEORADIUS simple (sorted) (8 ms) -[ok]: GEORADIUS_RO simple (sorted) (1 ms) -[ok]: GEOSEARCH simple (sorted) (10 ms) -[ok]: GEOSEARCH FROMLONLAT and FROMMEMBER cannot exist at the same time (8 ms) -[ok]: GEOSEARCH FROMLONLAT and FROMMEMBER one must exist (1 ms) -[ok]: GEOSEARCH BYRADIUS and BYBOX cannot exist at the same time (11 ms) -[ok]: GEOSEARCH BYRADIUS and BYBOX one must exist (8 ms) -[ok]: GEOSEARCH with STOREDIST option (2 ms) -[ok]: GEORADIUS withdist (sorted) (6 ms) -[ok]: GEOSEARCH withdist (sorted) (8 ms) -[ok]: GEORADIUS with COUNT (8 ms) -[ok]: GEORADIUS with multiple WITH* tokens (24 ms) -[ok]: GEORADIUS with ANY not sorted by default (8 ms) -[ok]: GEORADIUS with ANY sorted by ASC (12 ms) -[ok]: GEORADIUS with ANY but no COUNT (8 ms) -[ok]: GEORADIUS with COUNT but missing integer argument (8 ms) -[ok]: GEORADIUS with COUNT DESC (8 ms) -[ok]: GEORADIUS HUGE, issue #2767 (16 ms) -[ok]: GEORADIUSBYMEMBER simple (sorted) (8 ms) -[ok]: GEORADIUSBYMEMBER_RO simple (sorted) (8 ms) -[ok]: GEORADIUSBYMEMBER search areas contain satisfied points in oblique direction (72 ms) -[ok]: GEORADIUSBYMEMBER crossing pole search (17 ms) -[ok]: GEOSEARCH FROMMEMBER simple (sorted) (3 ms) -[ok]: GEOSEARCH vs GEORADIUS (21 ms) -[ok]: GEOSEARCH non square, long and narrow (49 ms) -[ok]: GEOSEARCH corner point test (16 ms) -[ok]: GEORADIUSBYMEMBER withdist (sorted) (4 ms) -[ok]: GEOHASH is able to return geohash strings (11 ms) -[ok]: GEOHASH with only key as argument (13 ms) -[ok]: GEOPOS simple (14 ms) -[ok]: GEOPOS missing element (23 ms) -[ok]: GEOPOS with only key as argument (21 ms) -[ok]: GEODIST simple & unit (46 ms) -[ok]: GEODIST missing elements (44 ms) -[ok]: GEORADIUS STORE option: syntax error (17 ms) -[ok]: GEOSEARCHSTORE STORE option: syntax error (6 ms) -[ok]: GEORANGE STORE option: incompatible options (39 ms) -[ok]: GEORANGE STORE option: plain usage (37 ms) -[ok]: GEORADIUSBYMEMBER STORE/STOREDIST option: plain usage (27 ms) -[ok]: GEOSEARCHSTORE STORE option: plain usage (5 ms) -[ok]: GEORANGE STOREDIST option: plain usage (2 ms) -[ok]: GEOSEARCHSTORE STOREDIST option: plain usage (0 ms) -[ok]: GEORANGE STOREDIST option: COUNT ASC and DESC (2 ms) -[ok]: GEOSEARCH the box spans -180° or 180° (3 ms) -[ok]: GEOSEARCH with small distance (16 ms) -[ok]: GEOSEARCH fuzzy test - byradius (42624 ms) -[ok]: GEOSEARCH fuzzy test - bybox (40846 ms) -[ok]: GEOSEARCH box edges fuzzy test (855 ms) -[41/124 done]: unit/geo (86 seconds) -Testing unit/keyspace -=== (keyspace) Starting server 127.0.0.1:21464 ok -[ok]: DEL against a single item (13 ms) -[ok]: Vararg DEL (2 ms) -[ok]: Untagged multi-key commands (8 ms) -[ok]: KEYS with pattern (1 ms) -[ok]: KEYS to get all keys (0 ms) -[ok]: DBSIZE (0 ms) -[ok]: KEYS with hashtag (2 ms) -[ok]: DEL all keys (6 ms) -[ok]: DEL against expired key (1103 ms) -[ok]: EXISTS (5 ms) -[ok]: Zero length value in key. SET/GET/EXISTS (1 ms) -[ok]: Commands pipelining (0 ms) -[ok]: Non existing command (1 ms) -[ok]: RENAME basic usage (3 ms) -[ok]: RENAME source key should no longer exist (1 ms) -[ok]: RENAME against already existing key (1 ms) -[ok]: RENAMENX basic usage (2 ms) -[ok]: RENAMENX against already existing key (0 ms) -[ok]: RENAMENX against already existing key (2) (5 ms) -[ok]: RENAME against non existing source key (1 ms) -[ok]: RENAME where source and dest key are the same (existing) (9 ms) -[ok]: RENAMENX where source and dest key are the same (existing) (1 ms) -[ok]: RENAME where source and dest key are the same (non existing) (0 ms) -[ok]: RENAME with volatile key, should move the TTL as well (2 ms) -[ok]: RENAME with volatile key, should not inherit TTL of target key (12 ms) -[ok]: DEL all keys again (DB 0) (2 ms) -[ok]: DEL all keys again (DB 1) (0 ms) -[ok]: COPY basic usage for string (6 ms) -[ok]: COPY for string does not replace an existing key without REPLACE option (4 ms) -[ok]: COPY for string can replace an existing key with REPLACE option (5 ms) -[ok]: COPY for string ensures that copied data is independent of copying data (6 ms) -[ok]: COPY for string does not copy data to no-integer DB (1 ms) -[ok]: COPY can copy key expire metadata as well (5 ms) -[ok]: COPY does not create an expire if it does not exist (1 ms) -[ok]: COPY basic usage for list - quicklist (57 ms) -[ok]: COPY basic usage for list - listpack (3 ms) -[ok]: COPY basic usage for $type set (11 ms) -[ok]: COPY basic usage for $type set (8 ms) -[ok]: COPY basic usage for $type set (122 ms) -[ok]: COPY basic usage for listpack sorted set (17 ms) -[ok]: COPY basic usage for skiplist sorted set (97 ms) -[ok]: COPY basic usage for listpack hash (6 ms) -[ok]: COPY basic usage for hashtable hash (114 ms) -[ok]: COPY basic usage for stream (3842 ms) -[ok]: COPY basic usage for stream-cgroups (13 ms) -[ok]: MOVE basic usage (2 ms) -[ok]: MOVE against key existing in the target DB (0 ms) -[ok]: MOVE against non-integer DB (#1428) (1 ms) -[ok]: MOVE can move key expire metadata as well (2 ms) -[ok]: MOVE does not create an expire if it does not exist (2 ms) -[ok]: SET/GET keys in different DBs (5 ms) -[ok]: RANDOMKEY (24 ms) -[ok]: RANDOMKEY against empty DB (0 ms) -[ok]: RANDOMKEY regression 1 (1 ms) -[ok]: KEYS * two times with long key, Github issue #1208 (1 ms) -[ok]: Regression for pattern matching long nested loops (1 ms) -[ok]: Coverage: basic SWAPDB test and unhappy path (5 ms) -[ok]: Coverage: SWAPDB and FLUSHDB (5 ms) -[ok]: Regression for pattern matching very long nested loops (9 ms) -[42/124 done]: unit/keyspace (6 seconds) -Testing unit/introspection -=== (introspection) Starting server 127.0.0.1:21466 ok -[ok]: PING (19 ms) -[ok]: CLIENT LIST (7 ms) -[ok]: CLIENT LIST with IDs (12 ms) -[ok]: CLIENT INFO (4 ms) -[ok]: client input output and command process statistics (107 ms) -[ok]: CLIENT KILL with illegal arguments (87 ms) -CLIENT KILL maxAGE will kill old clients test attempts: 0 -[ok]: CLIENT KILL maxAGE will kill old clients (2121 ms) -[ok]: CLIENT KILL SKIPME YES/NO will kill all clients (205 ms) -[ok]: CLIENT command unhappy path coverage (118 ms) -[ok]: CLIENT KILL close the client connection during bgsave (151 ms) -[ok]: CLIENT REPLY OFF/ON: disable all commands reply (92 ms) -[ok]: CLIENT REPLY SKIP: skip the next command reply (76 ms) -[ok]: CLIENT REPLY ON: unset SKIP flag (93 ms) -[ok]: MONITOR can log executed commands (51 ms) -[ok]: MONITOR can log commands issued by the scripting engine (45 ms) -[ok]: MONITOR can log commands issued by functions (54 ms) -[ok]: MONITOR supports redacting command arguments (91 ms) -[ok]: MONITOR correctly handles multi-exec cases (91 ms) -[ok]: MONITOR log blocked command only once (164 ms) -[ok]: CLIENT GETNAME should return NIL if name is not assigned (3 ms) -[ok]: CLIENT GETNAME check if name set correctly (12 ms) -[ok]: CLIENT LIST shows empty fields for unassigned names (7 ms) -[ok]: CLIENT SETNAME does not accept spaces (4 ms) -[ok]: CLIENT SETNAME can assign a name to this connection (12 ms) -[ok]: CLIENT SETNAME can change the name of an existing connection (12 ms) -[ok]: After CLIENT SETNAME, connection can still be closed (63 ms) -[ok]: CLIENT SETINFO can set a library name to this connection (24 ms) -[ok]: CLIENT SETINFO invalid args (27 ms) -[ok]: RESET does NOT clean library name (11 ms) -[ok]: CLIENT SETINFO can clear library name (14 ms) -=== () Starting server 127.0.0.1:21468 ok -=== () Starting server 127.0.0.1:21470 ok -=== () Starting server 127.0.0.1:21472 ok -=== () Starting server 127.0.0.1:21474 ok -[ok]: CONFIG save params special case handled properly (2702 ms) -[ok]: CONFIG sanity (1232 ms) -[ok]: CONFIG REWRITE sanity (841 ms) -=== () Starting server 127.0.0.1:21476 ok -[ok]: CONFIG REWRITE handles save and shutdown properly (2596 ms) -[ok]: CONFIG SET with multiple args (50 ms) -[ok]: CONFIG SET rollback on set error (51 ms) -dummy_accept: sock55a191431120 -[ok]: CONFIG SET rollback on apply error (216 ms) -[ok]: CONFIG SET duplicate configs (4 ms) -[ok]: CONFIG SET set immutable (11 ms) -[ok]: CONFIG GET hidden configs (19 ms) -[ok]: CONFIG GET multiple args (8 ms) -[ok]: valkey-server command line arguments - error cases (1116 ms) -=== () Starting server 127.0.0.1:21479 ok -=== () Starting server 127.0.0.1:21481 ok -[ok]: tot-net-out for replica client (1382 ms) -=== () Starting server 127.0.0.1:21483 ok -[ok]: valkey-server command line arguments - allow passing option name and option value in the same arg (502 ms) -=== () Starting server 127.0.0.1:21485 ok -[ok]: valkey-server command line arguments - wrong usage that we support anyway (615 ms) -=== () Starting server 127.0.0.1:21487 ok -[ok]: valkey-server command line arguments - allow option value to use the `--` prefix (668 ms) -=== () Starting server 127.0.0.1:21489 ok -[ok]: valkey-server command line arguments - option name and option value in the same arg and `--` prefix (855 ms) -=== () Starting server 127.0.0.1:21491 ok -=== () Starting server 127.0.0.1:21493 ok -=== () Starting server 127.0.0.1:21495 ok -=== () Starting server 127.0.0.1:21497 ok -=== () Starting server 127.0.0.1:21499 ok -[ok]: valkey-server command line arguments - save with empty input (2848 ms) -=== () Starting server 127.0.0.1:21501 ok -[ok]: valkey-server command line arguments - take one bulk string with spaces for MULTI_ARG configs parsing (508 ms) -=== (introspection external:skip) Starting server 127.0.0.1:21503 ok -[ok]: cannot modify protected configuration - no (15 ms) -=== (introspection external:skip) Starting server 127.0.0.1:21505 ok -[ok]: cannot modify protected configuration - local (32 ms) -=== () Starting server 127.0.0.1:21507 ok -[ok]: config during loading (3908 ms) -=== (introspection) Starting server 127.0.0.1:21509 ok -[ok]: CONFIG REWRITE handles rename-command properly (857 ms) -=== (introspection) Starting server 127.0.0.1:21511 ok -=== (introspection) Starting server 127.0.0.1:21513 ok -[ok]: CONFIG REWRITE handles alias config properly (1930 ms) -[43/124 done]: unit/introspection (28 seconds) -Testing unit/sort -=== (sort) Starting server 127.0.0.1:21515 ok -[ok]: Listpack: SORT BY key (9 ms) -[ok]: Listpack: SORT BY key with limit (4 ms) -[ok]: Listpack: SORT BY hash field (2 ms) -[ok]: Quicklist: SORT BY key (17 ms) -[ok]: Quicklist: SORT BY key with limit (2 ms) -[ok]: Quicklist: SORT BY hash field (16 ms) -[ok]: Big Quicklist: SORT BY key (1450 ms) -[ok]: Big Quicklist: SORT BY key with limit (19 ms) -[ok]: Big Quicklist: SORT BY hash field (1499 ms) -[ok]: Intset: SORT BY key (8 ms) -[ok]: Intset: SORT BY key with limit (8 ms) -[ok]: Intset: SORT BY hash field (8 ms) -[ok]: Hash table: SORT BY key (49 ms) -[ok]: Hash table: SORT BY key with limit (4 ms) -[ok]: Hash table: SORT BY hash field (47 ms) -[ok]: Big Hash table: SORT BY key (1370 ms) -[ok]: Big Hash table: SORT BY key with limit (32 ms) -[ok]: Big Hash table: SORT BY hash field (982 ms) -[ok]: SORT GET # (1 ms) -[ok]: SORT GET (13 ms) -[ok]: SORT_RO GET (15 ms) -[ok]: SORT GET (key and hash) with sanity check (219 ms) -[ok]: SORT BY key STORE (43 ms) -[ok]: SORT BY hash field STORE (32 ms) -[ok]: SORT extracts STORE correctly (8 ms) -[ok]: SORT_RO get keys (1 ms) -[ok]: SORT extracts multiple STORE correctly (7 ms) -[ok]: SORT DESC (8 ms) -[ok]: SORT ALPHA against integer encoded strings (43 ms) -[ok]: SORT sorted set (56 ms) -[ok]: SORT sorted set BY nosort should retain ordering (69 ms) -[ok]: SORT sorted set BY nosort + LIMIT (62 ms) -[ok]: SORT sorted set BY nosort works as expected from scripts (55 ms) -[ok]: SORT sorted set: +inf and -inf handling (64 ms) -[ok]: SORT regression for issue #19, sorting floats (84 ms) -[ok]: SORT with STORE returns zero if result is empty (github issue 224) (19 ms) -[ok]: SORT with STORE does not create empty lists (github issue 224) (36 ms) -[ok]: SORT with STORE removes key if result is empty (github issue 227) (27 ms) -[ok]: SORT with BY and STORE should still order output (32 ms) -[ok]: SORT will complain with numerical sorting and bad doubles (1) (24 ms) -[ok]: SORT will complain with numerical sorting and bad doubles (2) (24 ms) -[ok]: SORT BY sub-sorts lexicographically if score is the same (23 ms) -[ok]: SORT GET with pattern ending with just -> does not get hash field (27 ms) -[ok]: SORT by nosort retains native order for lists (23 ms) -[ok]: SORT by nosort plus store retains native order for lists (28 ms) -[ok]: SORT by nosort with limit returns based on original list order (17 ms) -[ok]: SORT_RO - Successful case (22 ms) -[ok]: SORT_RO - Cannot run with STORE arg (8 ms) - - Average time to sort: 0.76000000000000001 milliseconds [ok]: SORT speed, 100 element list BY key, 100 times (77 ms) - - Average time to sort: 0.85999999999999999 milliseconds [ok]: SORT speed, 100 element list BY hash field, 100 times (87 ms) - - Average time to sort: 0.65000000000000002 milliseconds [ok]: SORT speed, 100 element list directly, 100 times (65 ms) - - Average time to sort: 0.67000000000000004 milliseconds [ok]: SORT speed, 100 element list BY , 100 times (71 ms) -[ok]: SETRANGE with huge offset (1 ms) -[ok]: SORT STORE quicklist with the right options (65 ms) -=== (external:skip cluster sort) Starting server 127.0.0.1:21517 ok -[ok]: sort by in cluster mode (24 ms) -[ok]: sort get in cluster mode (24 ms) -[ok]: sort_ro by in cluster mode (24 ms) -[ok]: sort_ro get in cluster mode (24 ms) -[44/124 done]: unit/sort (47 seconds) +[ok]: LUA test pcall with error (0 ms) +[ok]: LUA test pcall with non string/integer arg (1 ms) +[ok]: LUA test trim string as expected (2 ms) +[42/124 done]: unit/scripting (39 seconds) +Testing unit/querybuf +=== (querybuf slow) Starting server 127.0.0.1:21509 ok +[ok]: query buffer resized correctly (2715 ms) +[ok]: query buffer resized correctly when not idle (80 ms) +[ok]: query buffer resized correctly with fat argv (2929 ms) +[43/124 done]: unit/querybuf (6 seconds) +Testing unit/shutdown +=== (shutdown external:skip) Starting server 127.0.0.1:21511 ok +[ok]: Temp rdb will be deleted if we use bg_unlink when shutdown (1970 ms) +=== (shutdown external:skip) Starting server 127.0.0.1:21513 ok +[ok]: SHUTDOWN ABORT can cancel SIGTERM (32 ms) +[ok]: Temp rdb will be deleted in signal handle (136 ms) +=== (shutdown external:skip) Starting server 127.0.0.1:21515 ok +[ok]: RDB save will be failed in shutdown (65 ms) +[ok]: SHUTDOWN will abort if rdb save failed on signal (331 ms) +[ok]: SHUTDOWN will abort if rdb save failed on shutdown command (288 ms) +[ok]: SHUTDOWN can proceed if shutdown command was with nosave (57 ms) +[ok]: Clean up rdb same named folder (22 ms) +=== (shutdown external:skip) Starting server 127.0.0.1:21517 ok +[ok]: SHUTDOWN SIGTERM will abort if there's an initial AOFRW - default (162 ms) +[44/124 done]: unit/shutdown (4 seconds) +Testing unit/type/list-2 +=== (list) Starting server 127.0.0.1:21519 ok +[ok]: LTRIM stress testing - quicklist (38543 ms) +[ok]: LTRIM stress testing - listpack (45852 ms) +[45/124 done]: unit/type/list-2 (85 seconds) +Testing unit/type/hash +=== (hash) Starting server 127.0.0.1:21521 ok +[ok]: HSET/HLEN - Small hash creation (4 ms) +[ok]: Is the small hash encoded with a listpack? (0 ms) +[ok]: HRANDFIELD - listpack (27 ms) +[ok]: HRANDFIELD - hashtable (22 ms) +[ok]: HRANDFIELD with RESP3 (1 ms) +[ok]: HRANDFIELD count of 0 is handled correctly (0 ms) +[ok]: HRANDFIELD count overflow (1 ms) +[ok]: HRANDFIELD with against non existing key (0 ms) +[ok]: HRANDFIELD count of 0 is handled correctly - emptyarray (1 ms) +[ok]: HRANDFIELD with against non existing key - emptyarray (0 ms) +[ok]: HRANDFIELD with - hashtable (510 ms) +[ok]: HRANDFIELD with - listpack (978 ms) +[ok]: HSET/HLEN - Big hash creation (881 ms) +[ok]: Is the big hash encoded with an hash table? (7 ms) +[ok]: HGET against the small hash (9 ms) +[ok]: HGET against the big hash (1033 ms) +[ok]: HGET against non existing key (4 ms) +[ok]: HSET in update and insert mode (15 ms) +[ok]: HSETNX target key missing - small hash (1 ms) +[ok]: HSETNX target key exists - small hash (1 ms) +[ok]: HSETNX target key missing - big hash (1 ms) +[ok]: HSETNX target key exists - big hash (1 ms) +[ok]: HSET/HMSET wrong number of args (7 ms) +[ok]: HMSET - small hash (0 ms) +[ok]: HMSET - big hash (115 ms) +[ok]: HMGET against non existing key and fields (8 ms) +[ok]: Hash commands against wrong type (16 ms) +[ok]: HMGET - small hash (1 ms) +[ok]: HMGET - big hash (95 ms) +[ok]: HKEYS - small hash (1 ms) +[ok]: HKEYS - big hash (56 ms) +[ok]: HVALS - small hash (0 ms) +[ok]: HVALS - big hash (13 ms) +[ok]: HGETALL - small hash (0 ms) +[ok]: HGETALL - big hash (80 ms) +[ok]: HGETALL against non-existing key (1 ms) +[ok]: HDEL and return value (4 ms) +[ok]: HDEL - more than a single value (4 ms) +[ok]: HDEL - hash becomes empty before deleting all specified fields (1 ms) +[ok]: HEXISTS (3 ms) +[ok]: Is a ziplist encoded Hash promoted on big payload? (3 ms) +[ok]: HINCRBY against non existing database key (0 ms) +[ok]: HINCRBY HINCRBYFLOAT against non-integer increment value (1 ms) +[ok]: HINCRBY against non existing hash key (4 ms) +[ok]: HINCRBY against hash key created by hincrby itself (4 ms) +[ok]: HINCRBY against hash key originally set with HSET (1 ms) +[ok]: HINCRBY over 32bit value (0 ms) +[ok]: HINCRBY over 32bit value with over 32bit increment (1 ms) +[ok]: HINCRBY fails against hash value with spaces (left) (1 ms) +[ok]: HINCRBY fails against hash value with spaces (right) (4 ms) +[ok]: HINCRBY can detect overflows (1 ms) +[ok]: HINCRBYFLOAT against non existing database key (0 ms) +[ok]: HINCRBYFLOAT against non existing hash key (7 ms) +[ok]: HINCRBYFLOAT against hash key created by hincrby itself (1 ms) +[ok]: HINCRBYFLOAT against hash key originally set with HSET (1 ms) +[ok]: HINCRBYFLOAT over 32bit value (8 ms) +[ok]: HINCRBYFLOAT over 32bit value with over 32bit increment (1 ms) +[ok]: HINCRBYFLOAT fails against hash value with spaces (left) (6 ms) +[ok]: HINCRBYFLOAT fails against hash value with spaces (right) (1 ms) +[ok]: HINCRBYFLOAT fails against hash value that contains a null-terminator in the middle (0 ms) +[ok]: HSTRLEN against the small hash (2 ms) +[ok]: HSTRLEN against the big hash (725 ms) +[ok]: HSTRLEN against non existing field (0 ms) +[ok]: HSTRLEN corner cases (26 ms) +[ok]: HINCRBYFLOAT over hash-max-listpack-value encoded with a listpack (3 ms) +[ok]: Hash ziplist regression test for large keys (16 ms) +[ok]: Hash fuzzing #1 - 10 fields (352 ms) +[ok]: Hash fuzzing #2 - 10 fields (192 ms) +[ok]: Hash fuzzing #1 - 512 fields (7465 ms) +[ok]: Hash fuzzing #2 - 512 fields (2173 ms) +[ok]: Stress test the hash ziplist -> hashtable encoding conversion (2617 ms) +[ok]: Test HINCRBYFLOAT for correct float representation (issue #2846) (9 ms) +[ok]: Hash ziplist of various encodings (106 ms) +[ok]: Hash ziplist of various encodings - sanitize dump (9 ms) +[ok]: HINCRBYFLOAT does not allow NaN or Infinity (1 ms) +[46/124 done]: unit/type/hash (20 seconds) +Testing unit/type/list-3 +=== (list ziplist) Starting server 127.0.0.1:21523 ok +[ok]: Explicit regression for a list bug (16 ms) +[ok]: Regression for quicklist #3343 bug (144 ms) +[ok]: Check compression with recompress (121 ms) +[ok]: Crash due to wrongly recompress after lrem (238 ms) +[ok]: LINSERT correctly recompress full quicklistNode after inserting a element before it (61 ms) +[ok]: LINSERT correctly recompress full quicklistNode after inserting a element after it (48 ms) +[ok]: Stress tester for #3343-alike bugs comp: 2 (13738 ms) +[ok]: Stress tester for #3343-alike bugs comp: 1 (11923 ms) +[ok]: Stress tester for #3343-alike bugs comp: 0 (14542 ms) +[ok]: ziplist implementation: value encoding and backlink (20205 ms) +[ok]: ziplist implementation: encoding stress testing (67065 ms) +[47/124 done]: unit/type/list-3 (128 seconds) Testing unit/type/string -=== (string) Starting server 127.0.0.1:21519 ok -[ok]: SET and GET an item (12 ms) -[ok]: SET and GET an empty item (14 ms) -[ok]: Very big payload in GET/SET (264 ms) -[ok]: Very big payload random access (11117 ms) -[ok]: SET 10000 numeric keys and access all them in reverse order (9505 ms) -[ok]: DBSIZE should be 10000 now (1 ms) +=== (string) Starting server 127.0.0.1:21525 ok +[ok]: SET and GET an item (0 ms) +[ok]: SET and GET an empty item (0 ms) +[ok]: Very big payload in GET/SET (93 ms) +[ok]: Very big payload random access (7231 ms) +[ok]: SET 10000 numeric keys and access all them in reverse order (7701 ms) +[ok]: DBSIZE should be 10000 now (0 ms) [ok]: SETNX target key missing (0 ms) [ok]: SETNX target key exists (1 ms) [ok]: SETNX against not-expired volatile key (1 ms) -[ok]: SETNX against expired volatile key (8087 ms) -[ok]: GETEX EX option (33 ms) -[ok]: GETEX PX option (32 ms) -[ok]: GETEX EXAT option (28 ms) -[ok]: GETEX PXAT option (32 ms) -[ok]: GETEX PERSIST option (40 ms) -[ok]: GETEX no option (26 ms) -[ok]: GETEX syntax errors (6 ms) -[ok]: GETEX and GET expired key or not exist (36 ms) -[ok]: GETEX no arguments (8 ms) -[ok]: GETDEL command (28 ms) -[ok]: GETDEL propagate as DEL command to replica (144 ms) -[ok]: GETEX without argument does not propagate to replica (308 ms) -[ok]: MGET (46 ms) -[ok]: MGET against non existing key (8 ms) -[ok]: MGET against non-string key (20 ms) -[ok]: GETSET (set new value) (23 ms) -[ok]: GETSET (replace old value) (23 ms) -[ok]: MSET base case (12 ms) -[ok]: MSET/MSETNX wrong number of args (12 ms) -[ok]: MSET with already existing - same key twice (20 ms) -[ok]: MSETNX with already existent key (21 ms) -[ok]: MSETNX with not existing keys (27 ms) -[ok]: MSETNX with not existing keys - same key twice (19 ms) -[ok]: MSETNX with already existing keys - same key twice (12 ms) -[ok]: STRLEN against non-existing key (8 ms) -[ok]: STRLEN against integer-encoded value (16 ms) -[ok]: STRLEN against plain string (14 ms) -[ok]: SETBIT against non-existing key (22 ms) -[ok]: SETBIT against string-encoded key (32 ms) -[ok]: SETBIT against integer-encoded key (40 ms) -[ok]: SETBIT against key with wrong type (24 ms) -[ok]: SETBIT with out of range bit offset (24 ms) -[ok]: SETBIT with non-bit argument (32 ms) -[ok]: SETBIT fuzzing (15019 ms) -[ok]: GETBIT against non-existing key (15 ms) -[ok]: GETBIT against string-encoded key (55 ms) -[ok]: GETBIT against integer-encoded key (68 ms) -[ok]: SETRANGE against non-existing key (69 ms) -[ok]: SETRANGE against string-encoded key (94 ms) -[ok]: SETRANGE against integer-encoded key (160 ms) -[ok]: SETRANGE against key with wrong type (23 ms) -[ok]: SETRANGE with out of range offset (36 ms) -[ok]: GETRANGE against non-existing key (15 ms) -[ok]: GETRANGE against wrong key type (13 ms) -[ok]: GETRANGE against string value (47 ms) -[ok]: GETRANGE against integer-encoded value (55 ms) -[ok]: GETRANGE fuzzing (6212 ms) -[ok]: Coverage: SUBSTR (10 ms) -[ok]: trim on SET with big value (54 ms) -[ok]: Extended SET can detect syntax errors (0 ms) -[ok]: Extended SET NX option (7 ms) -[ok]: Extended SET XX option (1 ms) +[ok]: SETNX against expired volatile key (4987 ms) +[ok]: GETEX EX option (4 ms) +[ok]: GETEX PX option (2 ms) +[ok]: GETEX EXAT option (2 ms) +[ok]: GETEX PXAT option (1 ms) +[ok]: GETEX PERSIST option (4 ms) +[ok]: GETEX no option (1 ms) +[ok]: GETEX syntax errors (1 ms) +[ok]: GETEX and GET expired key or not exist (4 ms) +[ok]: GETEX no arguments (3 ms) +[ok]: GETDEL command (2 ms) +[ok]: GETDEL propagate as DEL command to replica (146 ms) +[ok]: GETEX without argument does not propagate to replica (108 ms) +[ok]: MGET (2 ms) +[ok]: MGET against non existing key (0 ms) +[ok]: MGET against non-string key (2 ms) +[ok]: GETSET (set new value) (1 ms) +[ok]: GETSET (replace old value) (1 ms) +[ok]: MSET base case (2 ms) +[ok]: MSET/MSETNX wrong number of args (2 ms) +[ok]: MSET with already existing - same key twice (1 ms) +[ok]: MSETNX with already existent key (2 ms) +[ok]: MSETNX with not existing keys (2 ms) +[ok]: MSETNX with not existing keys - same key twice (1 ms) +[ok]: MSETNX with already existing keys - same key twice (1 ms) +[ok]: STRLEN against non-existing key (1 ms) +[ok]: STRLEN against integer-encoded value (10 ms) +[ok]: STRLEN against plain string (1 ms) +[ok]: SETBIT against non-existing key (2 ms) +[ok]: SETBIT against string-encoded key (3 ms) +[ok]: SETBIT against integer-encoded key (2 ms) +[ok]: SETBIT against key with wrong type (1 ms) +[ok]: SETBIT with out of range bit offset (2 ms) +[ok]: SETBIT with non-bit argument (2 ms) +[ok]: SETBIT fuzzing (4509 ms) +[ok]: GETBIT against non-existing key (1 ms) +[ok]: GETBIT against string-encoded key (2 ms) +[ok]: GETBIT against integer-encoded key (2 ms) +[ok]: SETRANGE against non-existing key (1 ms) +[ok]: SETRANGE against string-encoded key (3 ms) +[ok]: SETRANGE against integer-encoded key (11 ms) +[ok]: SETRANGE against key with wrong type (1 ms) +[ok]: SETRANGE with out of range offset (1 ms) +[ok]: GETRANGE against non-existing key (0 ms) +[ok]: GETRANGE against wrong key type (1 ms) +[ok]: GETRANGE against string value (2 ms) +[ok]: GETRANGE against integer-encoded value (1 ms) +[ok]: GETRANGE fuzzing (5992 ms) +[ok]: Coverage: SUBSTR (2 ms) +[ok]: trim on SET with big value (100 ms) +[ok]: Extended SET can detect syntax errors (1 ms) +[ok]: Extended SET NX option (1 ms) +[ok]: Extended SET XX option (2 ms) [ok]: Extended SET GET option (1 ms) -[ok]: Extended SET GET option with no previous value (5 ms) +[ok]: Extended SET GET option with no previous value (0 ms) [ok]: Extended SET GET option with XX (1 ms) [ok]: Extended SET GET option with XX and no previous value (1 ms) [ok]: Extended SET GET option with NX (0 ms) [ok]: Extended SET GET option with NX and previous value (1 ms) [ok]: Extended SET GET with incorrect type should result in wrong type error (1 ms) -[ok]: Extended SET EX option (0 ms) +[ok]: Extended SET EX option (1 ms) [ok]: Extended SET PX option (1 ms) [ok]: Extended SET EXAT option (0 ms) -[ok]: Extended SET PXAT option (7 ms) -[ok]: SET EXAT / PXAT Expiration time is expired (132 ms) -[ok]: Extended SET using multiple options at once (0 ms) -[ok]: GETRANGE with huge ranges, Github issue #1844 (1 ms) -[ok]: LCS basic (3 ms) -[ok]: LCS len (4 ms) -[ok]: LCS indexes (1 ms) -[ok]: LCS indexes with match len (6 ms) -[ok]: LCS indexes with match len and minimum match len (0 ms) +[ok]: Extended SET PXAT option (0 ms) +[ok]: SET EXAT / PXAT Expiration time is expired (89 ms) +[ok]: Extended SET using multiple options at once (1 ms) +[ok]: GETRANGE with huge ranges, Github issue #1844 (0 ms) +[ok]: LCS basic (1 ms) +[ok]: LCS len (1 ms) +[ok]: LCS indexes (0 ms) +[ok]: LCS indexes with match len (1 ms) +[ok]: LCS indexes with match len and minimum match len (1 ms) [ok]: SETRANGE with huge offset (0 ms) -[ok]: APPEND modifies the encoding from int to raw (7 ms) -[45/124 done]: unit/type/string (53 seconds) -Testing unit/type/hash -=== (hash) Starting server 127.0.0.1:21521 ok -[ok]: HSET/HLEN - Small hash creation (67 ms) -[ok]: Is the small hash encoded with a listpack? (4 ms) -[ok]: HRANDFIELD - listpack (754 ms) -[ok]: HRANDFIELD - hashtable (738 ms) -[ok]: HRANDFIELD with RESP3 (28 ms) -[ok]: HRANDFIELD count of 0 is handled correctly (8 ms) -[ok]: HRANDFIELD count overflow (32 ms) -[ok]: HRANDFIELD with against non existing key (8 ms) -[ok]: HRANDFIELD count of 0 is handled correctly - emptyarray (8 ms) -[ok]: HRANDFIELD with against non existing key - emptyarray (8 ms) -[ok]: HRANDFIELD with - hashtable (5252 ms) -[ok]: HRANDFIELD with - listpack (4586 ms) -[ok]: HSET/HLEN - Big hash creation (489 ms) -[ok]: Is the big hash encoded with an hash table? (1 ms) -[ok]: HGET against the small hash (4 ms) -[ok]: HGET against the big hash (636 ms) -[ok]: HGET against non existing key (1 ms) -[ok]: HSET in update and insert mode (8 ms) -[ok]: HSETNX target key missing - small hash (6 ms) -[ok]: HSETNX target key exists - small hash (8 ms) -[ok]: HSETNX target key missing - big hash (1 ms) -[ok]: HSETNX target key exists - big hash (1 ms) -[ok]: HSET/HMSET wrong number of args (10 ms) -[ok]: HMSET - small hash (1 ms) -[ok]: HMSET - big hash (119 ms) -[ok]: HMGET against non existing key and fields (7 ms) -[ok]: Hash commands against wrong type (25 ms) -[ok]: HMGET - small hash (2 ms) -[ok]: HMGET - big hash (136 ms) -[ok]: HKEYS - small hash (10 ms) -[ok]: HKEYS - big hash (88 ms) -[ok]: HVALS - small hash (1 ms) -[ok]: HVALS - big hash (18 ms) -[ok]: HGETALL - small hash (6 ms) -[ok]: HGETALL - big hash (193 ms) -[ok]: HGETALL against non-existing key (13 ms) -[ok]: HDEL and return value (12 ms) -[ok]: HDEL - more than a single value (7 ms) -[ok]: HDEL - hash becomes empty before deleting all specified fields (2 ms) -[ok]: HEXISTS (2 ms) -[ok]: Is a ziplist encoded Hash promoted on big payload? (10 ms) -[ok]: HINCRBY against non existing database key (2 ms) -[ok]: HINCRBY HINCRBYFLOAT against non-integer increment value (1 ms) -[ok]: HINCRBY against non existing hash key (7 ms) -[ok]: HINCRBY against hash key created by hincrby itself (1 ms) -[ok]: HINCRBY against hash key originally set with HSET (0 ms) -[ok]: HINCRBY over 32bit value (0 ms) -[ok]: HINCRBY over 32bit value with over 32bit increment (5 ms) -[ok]: HINCRBY fails against hash value with spaces (left) (8 ms) -[ok]: HINCRBY fails against hash value with spaces (right) (1 ms) -[ok]: HINCRBY can detect overflows (1 ms) -[ok]: HINCRBYFLOAT against non existing database key (1 ms) -[ok]: HINCRBYFLOAT against non existing hash key (1 ms) -[ok]: HINCRBYFLOAT against hash key created by hincrby itself (3 ms) -[ok]: HINCRBYFLOAT against hash key originally set with HSET (1 ms) -[ok]: HINCRBYFLOAT over 32bit value (10 ms) -[ok]: HINCRBYFLOAT over 32bit value with over 32bit increment (1 ms) -[ok]: HINCRBYFLOAT fails against hash value with spaces (left) (1 ms) -[ok]: HINCRBYFLOAT fails against hash value with spaces (right) (1 ms) -[ok]: HINCRBYFLOAT fails against hash value that contains a null-terminator in the middle (1 ms) -[ok]: HSTRLEN against the small hash (3 ms) -[ok]: HSTRLEN against the big hash (424 ms) -[ok]: HSTRLEN against non existing field (0 ms) -[ok]: HSTRLEN corner cases (6 ms) -[ok]: HINCRBYFLOAT over hash-max-listpack-value encoded with a listpack (13 ms) -[ok]: Hash ziplist regression test for large keys (2 ms) -[ok]: Hash fuzzing #1 - 10 fields (142 ms) -[ok]: Hash fuzzing #2 - 10 fields (108 ms) -[ok]: Hash fuzzing #1 - 512 fields (17000 ms) -[ok]: Hash fuzzing #2 - 512 fields (8635 ms) -[ok]: Stress test the hash ziplist -> hashtable encoding conversion (22010 ms) -[ok]: Test HINCRBYFLOAT for correct float representation (issue #2846) (1 ms) -[ok]: Hash ziplist of various encodings (112 ms) -[ok]: Hash ziplist of various encodings - sanitize dump (5 ms) -[ok]: HINCRBYFLOAT does not allow NaN or Infinity (15 ms) -[46/124 done]: unit/type/hash (63 seconds) +[ok]: APPEND modifies the encoding from int to raw (2 ms) +[48/124 done]: unit/type/string (33 seconds) +Testing unit/type/stream +=== (stream) Starting server 127.0.0.1:21527 ok +[ok]: XADD wrong number of args (1 ms) +[ok]: XADD can add entries into a stream that XRANGE can fetch (3 ms) +[ok]: XADD IDs are incremental (1 ms) +[ok]: XADD IDs are incremental when ms is the same as well (9 ms) +[ok]: XADD IDs correctly report an error when overflowing (3 ms) +[ok]: XADD auto-generated sequence is incremented for last ID (1 ms) +[ok]: XADD auto-generated sequence is zero for future timestamp ID (4 ms) +[ok]: XADD auto-generated sequence can't be smaller than last ID (1 ms) +[ok]: XADD auto-generated sequence can't overflow (2 ms) +[ok]: XADD 0-* should succeed (3 ms) +[ok]: XADD with MAXLEN option (730 ms) +[ok]: XADD with MAXLEN option and the '=' argument (902 ms) +[ok]: XADD with MAXLEN option and the '~' argument (773 ms) +[ok]: XADD with NOMKSTREAM option (9 ms) +[ok]: XADD with MINID option (1599 ms) +[ok]: XTRIM with MINID option (27 ms) +[ok]: XTRIM with MINID option, big delta from master record (20 ms) +[ok]: XADD mass insertion and XLEN (14768 ms) +[ok]: XADD with ID 0-0 (1 ms) +[ok]: XADD with LIMIT delete entries no more than limit (2 ms) +[ok]: XRANGE COUNT works as expected (1 ms) +[ok]: XREVRANGE COUNT works as expected (2 ms) +[ok]: XRANGE can be used to iterate the whole stream (924 ms) +[ok]: XREVRANGE returns the reverse of XRANGE (4121 ms) +[ok]: XRANGE exclusive ranges (9 ms) +[ok]: XREAD with non empty stream (1 ms) +[ok]: Non blocking XREAD with empty streams (2 ms) +[ok]: XREAD with non empty second stream (7324 ms) +[ok]: Blocking XREAD waiting new data (17 ms) +[ok]: Blocking XREAD waiting old data (23 ms) +[ok]: Blocking XREAD will not reply with an empty array (90 ms) +[ok]: Blocking XREAD for stream that ran dry (issue #5299) (105 ms) +[ok]: XREAD last element from non-empty stream (2 ms) +[ok]: XREAD last element from empty stream (1 ms) +[ok]: XREAD last element blocking from empty stream (28 ms) +[ok]: XREAD last element blocking from non-empty stream (2 ms) +[ok]: XREAD last element from multiple streams (2 ms) +[ok]: XREAD last element with count > 1 (1 ms) +[ok]: XREAD: XADD + DEL should not awake client (16 ms) +[ok]: XREAD: XADD + DEL + LPUSH should not awake client (20 ms) +[ok]: XREAD with same stream name multiple times should work (15 ms) +[ok]: XREAD + multiple XADD inside transaction (17 ms) +[ok]: XDEL basic test (2 ms) +[ok]: XDEL multiply id test (1 ms) +[ok]: XDEL fuzz test (9046 ms) +[ok]: XRANGE fuzzing (17453 ms) +[ok]: XREVRANGE regression test for issue #5006 (3 ms) +[ok]: XREAD streamID edge (no-blocking) (6 ms) +[ok]: XREAD streamID edge (blocking) (51 ms) +[ok]: XADD streamID edge (8 ms) +[ok]: XTRIM with MAXLEN option basic test (848 ms) +[ok]: XADD with LIMIT consecutive calls (69 ms) +[ok]: XTRIM with ~ is limited (109 ms) +[ok]: XTRIM without ~ is not limited (103 ms) +[ok]: XTRIM without ~ and with LIMIT (73 ms) +[ok]: XTRIM with LIMIT delete entries no more than limit (1 ms) +=== (stream needs:debug) Starting server 127.0.0.1:21529 ok +[ok]: XADD with MAXLEN > xlen can propagate correctly (26 ms) +=== (stream needs:debug) Starting server 127.0.0.1:21531 ok +[ok]: XADD with MINID > lastid can propagate correctly (120 ms) +=== (stream needs:debug) Starting server 127.0.0.1:21533 ok +[ok]: XADD with ~ MAXLEN can propagate correctly (109 ms) +=== (stream needs:debug) Starting server 127.0.0.1:21535 ok +[ok]: XADD with ~ MAXLEN and LIMIT can propagate correctly (112 ms) +=== (stream needs:debug) Starting server 127.0.0.1:21537 ok +[ok]: XADD with ~ MINID can propagate correctly (108 ms) +=== (stream needs:debug) Starting server 127.0.0.1:21539 ok +[ok]: XADD with ~ MINID and LIMIT can propagate correctly (185 ms) +=== (stream needs:debug) Starting server 127.0.0.1:21541 ok +[ok]: XTRIM with ~ MAXLEN can propagate correctly (41 ms) +=== (stream) Starting server 127.0.0.1:21543 ok +[ok]: XADD can CREATE an empty stream (0 ms) +[ok]: XSETID can set a specific ID (1 ms) +[ok]: XSETID cannot SETID with smaller ID (2 ms) +[ok]: XSETID cannot SETID on non-existent key (1 ms) +[ok]: XSETID cannot run with an offset but without a maximal tombstone (1 ms) +[ok]: XSETID cannot run with a maximal tombstone but without an offset (1 ms) +[ok]: XSETID errors on negstive offset (0 ms) +[ok]: XSETID cannot set the maximal tombstone with larger ID (2 ms) +[ok]: XSETID cannot set the offset to less than the length (2 ms) +[ok]: XSETID cannot set smaller ID than current MAXDELETEDID (7 ms) +=== (stream) Starting server 127.0.0.1:21545 ok +[ok]: XADD advances the entries-added counter and sets the recorded-first-entry-id (4 ms) +[ok]: XDEL/TRIM are reflected by recorded first entry (11 ms) +[ok]: Maximum XDEL ID behaves correctly (7 ms) +[ok]: XADD with artial ID with maximal seq (1 ms) +=== (stream needs:debug) Starting server 127.0.0.1:21547 ok + +Waiting for background AOF rewrite to finish... +Waiting for background AOF rewrite to finish... [ok]: Empty stream can be rewrite into AOF correctly (135 ms) + +Waiting for background AOF rewrite to finish... +Waiting for background AOF rewrite to finish... [ok]: Stream can be rewrite into AOF correctly after XDEL lastid (124 ms) +=== (stream) Starting server 127.0.0.1:21549 ok +[ok]: XGROUP HELP should not have unexpected options (8 ms) +[ok]: XINFO HELP should not have unexpected options (0 ms) +[49/124 done]: unit/type/stream (67 seconds) Testing unit/type/list -=== () Starting server 127.0.0.1:21523 ok -[ok]: packed node check compression with insert and pop (119 ms) -[ok]: packed node check compression combined with trim (43 ms) -[ok]: packed node check compression with lset (36 ms) -[ok]: plain node check compression with insert and pop (478 ms) -[ok]: plain node check compression combined with trim (186 ms) -[ok]: plain node check compression with lset (228 ms) -=== () Starting server 127.0.0.1:21525 ok -[ok]: Test LPUSH and LPOP on plain nodes (159 ms) -[ok]: Test LINDEX and LINSERT on plain nodes (79 ms) -[ok]: Test LTRIM on plain nodes (67 ms) -[ok]: Test LREM on plain nodes (76 ms) -[ok]: Test LPOS on plain nodes (67 ms) -[ok]: Test LMOVE on plain nodes (112 ms) -[ok]: Test LSET with packed / plain combinations (119 ms) -[ok]: Test LSET with packed is split in the middle (131 ms) -[ok]: Test LSET with packed consist only one item (131 ms) -[ok]: Crash due to delete entry from a compress quicklist node (79 ms) -[ok]: Crash due to split quicklist node wrongly (59 ms) -[ok]: Test LPUSH and LPOP on plain nodes (139 ms) -[ok]: Test LINDEX and LINSERT on plain nodes (107 ms) -[ok]: Test LTRIM on plain nodes (51 ms) -[ok]: Test LREM on plain nodes (68 ms) -[ok]: Test LPOS on plain nodes (48 ms) -[ok]: Test LMOVE on plain nodes (110 ms) -[ok]: Test LSET with packed / plain combinations (103 ms) -[ok]: Test LSET with packed is split in the middle (128 ms) -[ok]: Test LSET with packed consist only one item (131 ms) -[ok]: Crash due to delete entry from a compress quicklist node (75 ms) -[ok]: Crash due to split quicklist node wrongly (64 ms) -=== () Starting server 127.0.0.1:21527 ok +=== () Starting server 127.0.0.1:21551 ok +[ok]: packed node check compression with insert and pop (5 ms) +[ok]: packed node check compression combined with trim (2 ms) +[ok]: packed node check compression with lset (6 ms) +[ok]: plain node check compression with insert and pop (366 ms) +[ok]: plain node check compression combined with trim (138 ms) +[ok]: plain node check compression with lset (184 ms) +=== () Starting server 127.0.0.1:21553 ok +[ok]: Test LPUSH and LPOP on plain nodes (27 ms) +[ok]: Test LINDEX and LINSERT on plain nodes (2 ms) +[ok]: Test LTRIM on plain nodes (3 ms) +[ok]: Test LREM on plain nodes (4 ms) +[ok]: Test LPOS on plain nodes (8 ms) +[ok]: Test LMOVE on plain nodes (8 ms) +[ok]: Test LSET with packed / plain combinations (189 ms) +[ok]: Test LSET with packed is split in the middle (4 ms) +[ok]: Test LSET with packed consist only one item (205 ms) +[ok]: Crash due to delete entry from a compress quicklist node (5 ms) +[ok]: Crash due to split quicklist node wrongly (4 ms) +[ok]: Test LPUSH and LPOP on plain nodes (15 ms) +[ok]: Test LINDEX and LINSERT on plain nodes (3 ms) +[ok]: Test LTRIM on plain nodes (1 ms) +[ok]: Test LREM on plain nodes (5 ms) +[ok]: Test LPOS on plain nodes (4 ms) +[ok]: Test LMOVE on plain nodes (4 ms) +[ok]: Test LSET with packed / plain combinations (189 ms) +[ok]: Test LSET with packed is split in the middle (9 ms) +[ok]: Test LSET with packed consist only one item (181 ms) +[ok]: Crash due to delete entry from a compress quicklist node (2 ms) +[ok]: Crash due to split quicklist node wrongly (2 ms) +=== () Starting server 127.0.0.1:21555 ok [ignore]: Test LPUSH and LPOP on plain nodes over 4GB: large memory flag not provided [ignore]: Test LINDEX and LINSERT on plain nodes over 4GB: large memory flag not provided [ignore]: Test LTRIM on plain nodes over 4GB: large memory flag not provided @@ -3945,1121 +4129,970 @@ [ignore]: Test LSET splits a quicklist node, and then merge: large memory flag not provided [ignore]: Test LSET splits a LZF compressed quicklist node, and then merge: large memory flag not provided [ignore]: Test LMOVE on plain nodes over 4GB: large memory flag not provided -=== (list) Starting server 127.0.0.1:21529 ok -[ok]: LPOS basic usage - quicklist (71 ms) -[ok]: LPOS RANK (positive, negative and zero rank) option (55 ms) -[ok]: LPOS COUNT option (32 ms) -[ok]: LPOS COUNT + RANK option (12 ms) -[ok]: LPOS non existing key (4 ms) -[ok]: LPOS no match (18 ms) -[ok]: LPOS MAXLEN (36 ms) -[ok]: LPOS when RANK is greater than matches (23 ms) -[ok]: LPUSH, RPUSH, LLENGTH, LINDEX, LPOP - quicklist (252 ms) -[ok]: LPOP/RPOP with wrong number of arguments (16 ms) -[ok]: RPOP/LPOP with the optional count argument - quicklist (92 ms) -[ok]: LPOS basic usage - listpack (31 ms) -[ok]: LPOS RANK (positive, negative and zero rank) option (48 ms) -[ok]: LPOS COUNT option (31 ms) -[ok]: LPOS COUNT + RANK option (15 ms) -[ok]: LPOS non existing key (8 ms) -[ok]: LPOS no match (11 ms) -[ok]: LPOS MAXLEN (31 ms) -[ok]: LPOS when RANK is greater than matches (16 ms) -[ok]: LPUSH, RPUSH, LLENGTH, LINDEX, LPOP - listpack (163 ms) -[ok]: LPOP/RPOP with wrong number of arguments (16 ms) -[ok]: RPOP/LPOP with the optional count argument - listpack (56 ms) -[ok]: LPOP/RPOP with the count 0 returns an empty array in RESP3 (20 ms) -[ok]: LPOP/RPOP against non existing key in RESP3 (23 ms) -[ok]: LPOP/RPOP with against non existing key in RESP3 (31 ms) -[ok]: LPOP/RPOP with the count 0 returns an empty array in RESP2 (23 ms) -[ok]: LPOP/RPOP against non existing key in RESP2 (23 ms) -[ok]: LPOP/RPOP with against non existing key in RESP2 (35 ms) -[ok]: Variadic RPUSH/LPUSH (28 ms) -[ok]: DEL a list (19 ms) -[ok]: BLPOP: single existing list - quicklist (176 ms) -[ok]: BLPOP: multiple existing lists - quicklist (252 ms) -[ok]: BLPOP: second list has an entry - quicklist (164 ms) -[ok]: BLMPOP_LEFT: single existing list - quicklist (172 ms) -[ok]: BLMPOP_LEFT: multiple existing lists - quicklist (244 ms) -[ok]: BLMPOP_LEFT: second list has an entry - quicklist (159 ms) -[ok]: BRPOPLPUSH - quicklist (164 ms) -[ok]: BLMOVE left left - quicklist (149 ms) -[ok]: BLMOVE left right - quicklist (157 ms) -[ok]: BLMOVE right left - quicklist (156 ms) -[ok]: BLMOVE right right - quicklist (153 ms) -[ok]: BLPOP: single existing list - listpack (115 ms) -[ok]: BLPOP: multiple existing lists - listpack (163 ms) -[ok]: BLPOP: second list has an entry - listpack (92 ms) -[ok]: BLMPOP_LEFT: single existing list - listpack (112 ms) -[ok]: BLMPOP_LEFT: multiple existing lists - listpack (176 ms) -[ok]: BLMPOP_LEFT: second list has an entry - listpack (116 ms) -[ok]: BRPOPLPUSH - listpack (112 ms) -[ok]: BLMOVE left left - listpack (127 ms) -[ok]: BLMOVE left right - listpack (104 ms) -[ok]: BLMOVE right left - listpack (111 ms) -[ok]: BLMOVE right right - listpack (120 ms) -[ok]: BLPOP, LPUSH + DEL should not awake blocked client (84 ms) -[ok]: BLPOP, LPUSH + DEL + SET should not awake blocked client (92 ms) -[ok]: BLMPOP_LEFT, LPUSH + DEL should not awake blocked client (73 ms) -[ok]: BLMPOP_LEFT, LPUSH + DEL + SET should not awake blocked client (108 ms) -[ok]: BLPOP with same key multiple times should work (issue #801) (99 ms) -[ok]: MULTI/EXEC is isolated from the point of view of BLPOP (96 ms) -[ok]: BLPOP with variadic LPUSH (64 ms) -[ok]: MULTI/EXEC is isolated from the point of view of BLMPOP_LEFT (69 ms) -[ok]: BLMPOP_LEFT with variadic LPUSH (55 ms) -[ok]: BRPOPLPUSH with zero timeout should block indefinitely (67 ms) -[ok]: BLMOVE left left with zero timeout should block indefinitely (68 ms) -[ok]: BLMOVE left right with zero timeout should block indefinitely (68 ms) -[ok]: BLMOVE right left with zero timeout should block indefinitely (69 ms) -[ok]: BLMOVE right right with zero timeout should block indefinitely (67 ms) -[ok]: BLMOVE (left, left) with a client BLPOPing the target list (104 ms) -[ok]: BLMOVE (left, right) with a client BLPOPing the target list (95 ms) -[ok]: BLMOVE (right, left) with a client BLPOPing the target list (116 ms) -[ok]: BLMOVE (right, right) with a client BLPOPing the target list (120 ms) -[ok]: BRPOPLPUSH with wrong source type (51 ms) -[ok]: BRPOPLPUSH with wrong destination type (140 ms) -[ok]: BRPOPLPUSH maintains order of elements after failure (63 ms) -[ok]: BRPOPLPUSH with multiple blocked clients (120 ms) -[ok]: BLMPOP with multiple blocked clients (204 ms) -[ok]: Linked LMOVEs (123 ms) -[ok]: Circular BRPOPLPUSH (95 ms) -[ok]: Self-referential BRPOPLPUSH (59 ms) -[ok]: BRPOPLPUSH inside a transaction (58 ms) -[ok]: PUSH resulting from BRPOPLPUSH affect WATCH (129 ms) -[ok]: BRPOPLPUSH does not affect WATCH while still blocked (159 ms) -[ok]: BRPOPLPUSH timeout (1068 ms) -[ok]: SWAPDB awakes blocked client (52 ms) -[ok]: SWAPDB wants to wake blocked client, but the key already expired (317 ms) -[ok]: MULTI + LPUSH + EXPIRE + DEBUG SLEEP on blocked client, key already expired (411 ms) -[ok]: BLPOP unblock but the key is expired and then block again - reprocessing command (1086 ms) -[ok]: BLPOP when new key is moved into place (29 ms) -[ok]: BLPOP when result key is created by SORT..STORE (39 ms) -[ok]: BLMPOP_LEFT when new key is moved into place (47 ms) -[ok]: BLMPOP_LEFT when result key is created by SORT..STORE (44 ms) -[ok]: BLPOP: timeout value out of range (2 ms) -[ok]: BLPOP: with single empty list argument (35 ms) -[ok]: BLPOP: with negative timeout (51 ms) -[ok]: BLPOP: with non-integer timeout (41 ms) -[ok]: BLPOP: with zero timeout should block indefinitely (54 ms) -[ok]: BLPOP: with 0.001 timeout should not block indefinitely (52 ms) -[ok]: BLPOP: second argument is not a list (43 ms) -[ok]: BLPOP: timeout (1096 ms) -[ok]: BLPOP: arguments are empty (44 ms) -[ok]: BRPOP: with single empty list argument (41 ms) -[ok]: BRPOP: with negative timeout (24 ms) -[ok]: BRPOP: with non-integer timeout (19 ms) -[ok]: BRPOP: with zero timeout should block indefinitely (26 ms) -[ok]: BRPOP: with 0.001 timeout should not block indefinitely (50 ms) -[ok]: BRPOP: second argument is not a list (28 ms) -[ok]: BRPOP: timeout (1114 ms) -[ok]: BRPOP: arguments are empty (34 ms) -[ok]: BLMPOP_LEFT: with single empty list argument (20 ms) -[ok]: BLMPOP_LEFT: with negative timeout (26 ms) -[ok]: BLMPOP_LEFT: with non-integer timeout (22 ms) -[ok]: BLMPOP_LEFT: with zero timeout should block indefinitely (23 ms) -[ok]: BLMPOP_LEFT: with 0.001 timeout should not block indefinitely (52 ms) -[ok]: BLMPOP_LEFT: second argument is not a list (32 ms) -[ok]: BLMPOP_LEFT: timeout (1130 ms) -[ok]: BLMPOP_LEFT: arguments are empty (42 ms) -[ok]: BLMPOP_RIGHT: with single empty list argument (33 ms) -[ok]: BLMPOP_RIGHT: with negative timeout (23 ms) -[ok]: BLMPOP_RIGHT: with non-integer timeout (28 ms) -[ok]: BLMPOP_RIGHT: with zero timeout should block indefinitely (33 ms) -[ok]: BLMPOP_RIGHT: with 0.001 timeout should not block indefinitely (62 ms) -[ok]: BLMPOP_RIGHT: second argument is not a list (40 ms) -[ok]: BLMPOP_RIGHT: timeout (1092 ms) -[ok]: BLMPOP_RIGHT: arguments are empty (34 ms) -[ok]: BLPOP inside a transaction (9 ms) +=== (list) Starting server 127.0.0.1:21557 ok +[ok]: LPOS basic usage - quicklist (48 ms) +[ok]: LPOS RANK (positive, negative and zero rank) option (8 ms) +[ok]: LPOS COUNT option (1 ms) +[ok]: LPOS COUNT + RANK option (1 ms) +[ok]: LPOS non existing key (0 ms) +[ok]: LPOS no match (1 ms) +[ok]: LPOS MAXLEN (4 ms) +[ok]: LPOS when RANK is greater than matches (2 ms) +[ok]: LPUSH, RPUSH, LLENGTH, LINDEX, LPOP - quicklist (100 ms) +[ok]: LPOP/RPOP with wrong number of arguments (0 ms) +[ok]: RPOP/LPOP with the optional count argument - quicklist (49 ms) +[ok]: LPOS basic usage - listpack (1 ms) +[ok]: LPOS RANK (positive, negative and zero rank) option (1 ms) +[ok]: LPOS COUNT option (1 ms) +[ok]: LPOS COUNT + RANK option (1 ms) +[ok]: LPOS non existing key (0 ms) +[ok]: LPOS no match (0 ms) +[ok]: LPOS MAXLEN (0 ms) +[ok]: LPOS when RANK is greater than matches (0 ms) +[ok]: LPUSH, RPUSH, LLENGTH, LINDEX, LPOP - listpack (5 ms) +[ok]: LPOP/RPOP with wrong number of arguments (0 ms) +[ok]: RPOP/LPOP with the optional count argument - listpack (15 ms) +[ok]: LPOP/RPOP with the count 0 returns an empty array in RESP3 (2 ms) +[ok]: LPOP/RPOP against non existing key in RESP3 (6 ms) +[ok]: LPOP/RPOP with against non existing key in RESP3 (3 ms) +[ok]: LPOP/RPOP with the count 0 returns an empty array in RESP2 (1 ms) +[ok]: LPOP/RPOP against non existing key in RESP2 (0 ms) +[ok]: LPOP/RPOP with against non existing key in RESP2 (1 ms) +[ok]: Variadic RPUSH/LPUSH (0 ms) +[ok]: DEL a list (0 ms) +[ok]: BLPOP: single existing list - quicklist (71 ms) +[ok]: BLPOP: multiple existing lists - quicklist (118 ms) +[ok]: BLPOP: second list has an entry - quicklist (79 ms) +[ok]: BLMPOP_LEFT: single existing list - quicklist (82 ms) +[ok]: BLMPOP_LEFT: multiple existing lists - quicklist (113 ms) +[ok]: BLMPOP_LEFT: second list has an entry - quicklist (58 ms) +[ok]: BRPOPLPUSH - quicklist (71 ms) +[ok]: BLMOVE left left - quicklist (67 ms) +[ok]: BLMOVE left right - quicklist (73 ms) +[ok]: BLMOVE right left - quicklist (79 ms) +[ok]: BLMOVE right right - quicklist (69 ms) +[ok]: BLPOP: single existing list - listpack (14 ms) +[ok]: BLPOP: multiple existing lists - listpack (17 ms) +[ok]: BLPOP: second list has an entry - listpack (14 ms) +[ok]: BLMPOP_LEFT: single existing list - listpack (14 ms) +[ok]: BLMPOP_LEFT: multiple existing lists - listpack (17 ms) +[ok]: BLMPOP_LEFT: second list has an entry - listpack (12 ms) +[ok]: BRPOPLPUSH - listpack (12 ms) +[ok]: BLMOVE left left - listpack (12 ms) +[ok]: BLMOVE left right - listpack (11 ms) +[ok]: BLMOVE right left - listpack (13 ms) +[ok]: BLMOVE right right - listpack (13 ms) +[ok]: BLPOP, LPUSH + DEL should not awake blocked client (114 ms) +[ok]: BLPOP, LPUSH + DEL + SET should not awake blocked client (119 ms) +[ok]: BLMPOP_LEFT, LPUSH + DEL should not awake blocked client (12 ms) +[ok]: BLMPOP_LEFT, LPUSH + DEL + SET should not awake blocked client (13 ms) +[ok]: BLPOP with same key multiple times should work (issue #801) (16 ms) +[ok]: MULTI/EXEC is isolated from the point of view of BLPOP (12 ms) +[ok]: BLPOP with variadic LPUSH (113 ms) +[ok]: MULTI/EXEC is isolated from the point of view of BLMPOP_LEFT (139 ms) +[ok]: BLMPOP_LEFT with variadic LPUSH (14 ms) +[ok]: BRPOPLPUSH with zero timeout should block indefinitely (36 ms) +[ok]: BLMOVE left left with zero timeout should block indefinitely (17 ms) +[ok]: BLMOVE left right with zero timeout should block indefinitely (24 ms) +[ok]: BLMOVE right left with zero timeout should block indefinitely (29 ms) +[ok]: BLMOVE right right with zero timeout should block indefinitely (10 ms) +[ok]: BLMOVE (left, left) with a client BLPOPing the target list (24 ms) +[ok]: BLMOVE (left, right) with a client BLPOPing the target list (32 ms) +[ok]: BLMOVE (right, left) with a client BLPOPing the target list (50 ms) +[ok]: BLMOVE (right, right) with a client BLPOPing the target list (74 ms) +[ok]: BRPOPLPUSH with wrong source type (14 ms) +[ok]: BRPOPLPUSH with wrong destination type (44 ms) +[ok]: BRPOPLPUSH maintains order of elements after failure (120 ms) +[ok]: BRPOPLPUSH with multiple blocked clients (35 ms) +[ok]: BLMPOP with multiple blocked clients (112 ms) +[ok]: Linked LMOVEs (46 ms) +[ok]: Circular BRPOPLPUSH (85 ms) +[ok]: Self-referential BRPOPLPUSH (41 ms) +[ok]: BRPOPLPUSH inside a transaction (7 ms) +[ok]: PUSH resulting from BRPOPLPUSH affect WATCH (2161 ms) +[ok]: BRPOPLPUSH does not affect WATCH while still blocked (82 ms) +[ok]: BRPOPLPUSH timeout (1049 ms) +[ok]: SWAPDB awakes blocked client (35 ms) +[ok]: SWAPDB wants to wake blocked client, but the key already expired (276 ms) +[ok]: MULTI + LPUSH + EXPIRE + DEBUG SLEEP on blocked client, key already expired (402 ms) +[ok]: BLPOP unblock but the key is expired and then block again - reprocessing command (1081 ms) +[ok]: BLPOP when new key is moved into place (63 ms) +[ok]: BLPOP when result key is created by SORT..STORE (46 ms) +[ok]: BLMPOP_LEFT when new key is moved into place (57 ms) +[ok]: BLMPOP_LEFT when result key is created by SORT..STORE (62 ms) +[ok]: BLPOP: timeout value out of range (1 ms) +[ok]: BLPOP: with single empty list argument (42 ms) +[ok]: BLPOP: with negative timeout (37 ms) +[ok]: BLPOP: with non-integer timeout (39 ms) +[ok]: BLPOP: with zero timeout should block indefinitely (42 ms) +[ok]: BLPOP: with 0.001 timeout should not block indefinitely (65 ms) +[ok]: BLPOP: second argument is not a list (36 ms) +[ok]: BLPOP: timeout (1043 ms) +[ok]: BLPOP: arguments are empty (61 ms) +[ok]: BRPOP: with single empty list argument (40 ms) +[ok]: BRPOP: with negative timeout (33 ms) +[ok]: BRPOP: with non-integer timeout (26 ms) +[ok]: BRPOP: with zero timeout should block indefinitely (22 ms) +[ok]: BRPOP: with 0.001 timeout should not block indefinitely (68 ms) +[ok]: BRPOP: second argument is not a list (51 ms) +[ok]: BRPOP: timeout (1127 ms) +[ok]: BRPOP: arguments are empty (28 ms) +[ok]: BLMPOP_LEFT: with single empty list argument (30 ms) +[ok]: BLMPOP_LEFT: with negative timeout (30 ms) +[ok]: BLMPOP_LEFT: with non-integer timeout (32 ms) +[ok]: BLMPOP_LEFT: with zero timeout should block indefinitely (20 ms) +[ok]: BLMPOP_LEFT: with 0.001 timeout should not block indefinitely (35 ms) +[ok]: BLMPOP_LEFT: second argument is not a list (20 ms) +[ok]: BLMPOP_LEFT: timeout (1069 ms) +[ok]: BLMPOP_LEFT: arguments are empty (20 ms) +[ok]: BLMPOP_RIGHT: with single empty list argument (13 ms) +[ok]: BLMPOP_RIGHT: with negative timeout (15 ms) +[ok]: BLMPOP_RIGHT: with non-integer timeout (18 ms) +[ok]: BLMPOP_RIGHT: with zero timeout should block indefinitely (17 ms) +[ok]: BLMPOP_RIGHT: with 0.001 timeout should not block indefinitely (43 ms) +[ok]: BLMPOP_RIGHT: second argument is not a list (36 ms) +[ok]: BLMPOP_RIGHT: timeout (1112 ms) +[ok]: BLMPOP_RIGHT: arguments are empty (22 ms) +[ok]: BLPOP inside a transaction (2 ms) [ok]: BLMPOP_LEFT inside a transaction (2 ms) -[ok]: BLMPOP propagate as pop with count command to replica (224 ms) -[ok]: LPUSHX, RPUSHX - generic (39 ms) -[ok]: LPUSHX, RPUSHX - quicklist (108 ms) -[ok]: LINSERT - quicklist (188 ms) -[ok]: LPUSHX, RPUSHX - listpack (67 ms) -[ok]: LINSERT - listpack (141 ms) -[ok]: LINSERT raise error on bad syntax (7 ms) -[ok]: LINSERT against non-list value error (16 ms) -[ok]: LINSERT against non existing key (2 ms) -[ok]: LINDEX consistency test - listpack (3280 ms) -[ok]: LINDEX random access - listpack (281 ms) -[ok]: Check if list is still ok after a DEBUG RELOAD - listpack (7227 ms) -[ok]: LINDEX consistency test - listpack (7152 ms) -[ok]: LINDEX random access - listpack (576 ms) -[ok]: Check if list is still ok after a DEBUG RELOAD - listpack (1059 ms) -[ok]: LINDEX consistency test - quicklist (3137 ms) -[ok]: LINDEX random access - quicklist (3632 ms) -[ok]: Check if list is still ok after a DEBUG RELOAD - quicklist (6932 ms) -[ok]: LINDEX consistency test - quicklist (3417 ms) -[ok]: LINDEX random access - quicklist (404 ms) -[ok]: Check if list is still ok after a DEBUG RELOAD - quicklist (10348 ms) -[ok]: LLEN against non-list value error (16 ms) -[ok]: LLEN against non existing key (4 ms) +[ok]: BLMPOP propagate as pop with count command to replica (432 ms) +[ok]: LPUSHX, RPUSHX - generic (20 ms) +[ok]: LPUSHX, RPUSHX - quicklist (52 ms) +[ok]: LINSERT - quicklist (65 ms) +[ok]: LPUSHX, RPUSHX - listpack (2 ms) +[ok]: LINSERT - listpack (5 ms) +[ok]: LINSERT raise error on bad syntax (0 ms) +[ok]: LINSERT against non-list value error (3 ms) +[ok]: LINSERT against non existing key (1 ms) +[ok]: LINDEX consistency test - listpack (771 ms) +[ok]: LINDEX random access - listpack (993 ms) +[ok]: Check if list is still ok after a DEBUG RELOAD - listpack (1266 ms) +[ok]: LINDEX consistency test - listpack (756 ms) +[ok]: LINDEX random access - listpack (235 ms) +[ok]: Check if list is still ok after a DEBUG RELOAD - listpack (1276 ms) +[ok]: LINDEX consistency test - quicklist (430 ms) +[ok]: LINDEX random access - quicklist (274 ms) +[ok]: Check if list is still ok after a DEBUG RELOAD - quicklist (489 ms) +[ok]: LINDEX consistency test - quicklist (1107 ms) +[ok]: LINDEX random access - quicklist (725 ms) +[ok]: Check if list is still ok after a DEBUG RELOAD - quicklist (1386 ms) +[ok]: LLEN against non-list value error (2 ms) +[ok]: LLEN against non existing key (1 ms) [ok]: LINDEX against non-list value error (4 ms) -[ok]: LINDEX against non existing key (8 ms) +[ok]: LINDEX against non existing key (0 ms) [ok]: LPUSH against non-list value error (4 ms) -[ok]: RPUSH against non-list value error (4 ms) -[ok]: RPOPLPUSH base case - quicklist (151 ms) -[ok]: LMOVE left left base case - quicklist (124 ms) -[ok]: LMOVE left right base case - quicklist (132 ms) -[ok]: LMOVE right left base case - quicklist (132 ms) -[ok]: LMOVE right right base case - quicklist (132 ms) -[ok]: RPOPLPUSH with the same list as src and dst - quicklist (100 ms) -[ok]: LMOVE left left with the same list as src and dst - quicklist (95 ms) -[ok]: LMOVE left right with the same list as src and dst - quicklist (96 ms) -[ok]: LMOVE right left with the same list as src and dst - quicklist (104 ms) -[ok]: LMOVE right right with the same list as src and dst - quicklist (104 ms) -[ok]: RPOPLPUSH with quicklist source and existing target quicklist (188 ms) -[ok]: LMOVE left left with quicklist source and existing target quicklist (196 ms) -[ok]: LMOVE left right with quicklist source and existing target quicklist (200 ms) -[ok]: LMOVE right left with quicklist source and existing target quicklist (190 ms) -[ok]: LMOVE right right with quicklist source and existing target quicklist (176 ms) -[ok]: RPOPLPUSH with quicklist source and existing target listpack (143 ms) -[ok]: LMOVE left left with quicklist source and existing target listpack (156 ms) -[ok]: LMOVE left right with quicklist source and existing target listpack (136 ms) -[ok]: LMOVE right left with quicklist source and existing target listpack (120 ms) -[ok]: LMOVE right right with quicklist source and existing target listpack (130 ms) -[ok]: RPOPLPUSH base case - listpack (96 ms) -[ok]: LMOVE left left base case - listpack (88 ms) -[ok]: LMOVE left right base case - listpack (88 ms) -[ok]: LMOVE right left base case - listpack (80 ms) -[ok]: LMOVE right right base case - listpack (84 ms) -[ok]: RPOPLPUSH with the same list as src and dst - listpack (51 ms) -[ok]: LMOVE left left with the same list as src and dst - listpack (52 ms) -[ok]: LMOVE left right with the same list as src and dst - listpack (60 ms) -[ok]: LMOVE right left with the same list as src and dst - listpack (64 ms) -[ok]: LMOVE right right with the same list as src and dst - listpack (57 ms) -[ok]: RPOPLPUSH with listpack source and existing target quicklist (140 ms) -[ok]: LMOVE left left with listpack source and existing target quicklist (85 ms) +[ok]: RPUSH against non-list value error (1 ms) +[ok]: RPOPLPUSH base case - quicklist (52 ms) +[ok]: LMOVE left left base case - quicklist (62 ms) +[ok]: LMOVE left right base case - quicklist (46 ms) +[ok]: LMOVE right left base case - quicklist (44 ms) +[ok]: LMOVE right right base case - quicklist (57 ms) +[ok]: RPOPLPUSH with the same list as src and dst - quicklist (52 ms) +[ok]: LMOVE left left with the same list as src and dst - quicklist (53 ms) +[ok]: LMOVE left right with the same list as src and dst - quicklist (48 ms) +[ok]: LMOVE right left with the same list as src and dst - quicklist (54 ms) +[ok]: LMOVE right right with the same list as src and dst - quicklist (46 ms) +[ok]: RPOPLPUSH with quicklist source and existing target quicklist (100 ms) +[ok]: LMOVE left left with quicklist source and existing target quicklist (121 ms) +[ok]: LMOVE left right with quicklist source and existing target quicklist (109 ms) +[ok]: LMOVE right left with quicklist source and existing target quicklist (97 ms) +[ok]: LMOVE right right with quicklist source and existing target quicklist (122 ms) +[ok]: RPOPLPUSH with quicklist source and existing target listpack (58 ms) +[ok]: LMOVE left left with quicklist source and existing target listpack (59 ms) +[ok]: LMOVE left right with quicklist source and existing target listpack (58 ms) +[ok]: LMOVE right left with quicklist source and existing target listpack (58 ms) +[ok]: LMOVE right right with quicklist source and existing target listpack (79 ms) +[ok]: RPOPLPUSH base case - listpack (5 ms) +[ok]: LMOVE left left base case - listpack (4 ms) +[ok]: LMOVE left right base case - listpack (9 ms) +[ok]: LMOVE right left base case - listpack (7 ms) +[ok]: LMOVE right right base case - listpack (5 ms) +[ok]: RPOPLPUSH with the same list as src and dst - listpack (2 ms) +[ok]: LMOVE left left with the same list as src and dst - listpack (2 ms) +[ok]: LMOVE left right with the same list as src and dst - listpack (4 ms) +[ok]: LMOVE right left with the same list as src and dst - listpack (3 ms) +[ok]: LMOVE right right with the same list as src and dst - listpack (2 ms) +[ok]: RPOPLPUSH with listpack source and existing target quicklist (57 ms) +[ok]: LMOVE left left with listpack source and existing target quicklist (69 ms) [ok]: LMOVE left right with listpack source and existing target quicklist (58 ms) -[ok]: LMOVE right left with listpack source and existing target quicklist (49 ms) -[ok]: LMOVE right right with listpack source and existing target quicklist (48 ms) -[ok]: RPOPLPUSH with listpack source and existing target listpack (4 ms) -[ok]: LMOVE left left with listpack source and existing target listpack (17 ms) -[ok]: LMOVE left right with listpack source and existing target listpack (21 ms) -[ok]: LMOVE right left with listpack source and existing target listpack (6 ms) -[ok]: LMOVE right right with listpack source and existing target listpack (4 ms) -[ok]: RPOPLPUSH against non existing key (6 ms) -[ok]: RPOPLPUSH against non list src key (8 ms) -[ok]: RPOPLPUSH against non list dst key - quicklist (50 ms) -[ok]: RPOPLPUSH against non list dst key - listpack (11 ms) -[ok]: RPOPLPUSH against non existing src key (0 ms) -[ok]: Basic LPOP/RPOP/LMPOP - quicklist (94 ms) -[ok]: Basic LPOP/RPOP/LMPOP - listpack (11 ms) -[ok]: LPOP/RPOP/LMPOP against empty list (9 ms) -[ok]: LPOP/RPOP/LMPOP NON-BLOCK or BLOCK against non list value (5 ms) -[ok]: Mass RPOP/LPOP - listpack (257 ms) -[ok]: Mass RPOP/LPOP - listpack (489 ms) -[ok]: LMPOP with illegal argument (18 ms) -[ok]: LMPOP single existing list - quicklist (165 ms) -[ok]: LMPOP multiple existing lists - quicklist (205 ms) -[ok]: LMPOP single existing list - listpack (14 ms) -[ok]: LMPOP multiple existing lists - listpack (27 ms) -[ok]: LMPOP propagate as pop with count command to replica (151 ms) -[ok]: LRANGE basics - quicklist (55 ms) -[ok]: LRANGE inverted indexes - quicklist (52 ms) -[ok]: LRANGE out of range indexes including the full list - quicklist (44 ms) -[ok]: LRANGE out of range negative end index - quicklist (52 ms) -[ok]: LRANGE basics - listpack (16 ms) -[ok]: LRANGE inverted indexes - listpack (32 ms) -[ok]: LRANGE out of range indexes including the full list - listpack (3 ms) -[ok]: LRANGE out of range negative end index - listpack (15 ms) -[ok]: LRANGE against non existing key (7 ms) -[ok]: LRANGE with start > end yields an empty array for backward compatibility (2 ms) -[ok]: LTRIM basics - quicklist (646 ms) -[ok]: LTRIM out of range negative end index - quicklist (115 ms) -[ok]: LSET - quicklist (44 ms) -[ok]: LSET out of range index - quicklist (6 ms) -[ok]: LTRIM basics - listpack (80 ms) -[ok]: LTRIM out of range negative end index - listpack (26 ms) +[ok]: LMOVE right left with listpack source and existing target quicklist (77 ms) +[ok]: LMOVE right right with listpack source and existing target quicklist (63 ms) +[ok]: RPOPLPUSH with listpack source and existing target listpack (11 ms) +[ok]: LMOVE left left with listpack source and existing target listpack (3 ms) +[ok]: LMOVE left right with listpack source and existing target listpack (3 ms) +[ok]: LMOVE right left with listpack source and existing target listpack (2 ms) +[ok]: LMOVE right right with listpack source and existing target listpack (3 ms) +[ok]: RPOPLPUSH against non existing key (5 ms) +[ok]: RPOPLPUSH against non list src key (3 ms) +[ok]: RPOPLPUSH against non list dst key - quicklist (58 ms) +[ok]: RPOPLPUSH against non list dst key - listpack (13 ms) +[ok]: RPOPLPUSH against non existing src key (2 ms) +[ok]: Basic LPOP/RPOP/LMPOP - quicklist (106 ms) +[ok]: Basic LPOP/RPOP/LMPOP - listpack (18 ms) +[ok]: LPOP/RPOP/LMPOP against empty list (4 ms) +[ok]: LPOP/RPOP/LMPOP NON-BLOCK or BLOCK against non list value (23 ms) +[ok]: Mass RPOP/LPOP - listpack (282 ms) +[ok]: Mass RPOP/LPOP - listpack (520 ms) +[ok]: LMPOP with illegal argument (14 ms) +[ok]: LMPOP single existing list - quicklist (186 ms) +[ok]: LMPOP multiple existing lists - quicklist (209 ms) +[ok]: LMPOP single existing list - listpack (9 ms) +[ok]: LMPOP multiple existing lists - listpack (10 ms) +[ok]: LMPOP propagate as pop with count command to replica (79 ms) +[ok]: LRANGE basics - quicklist (58 ms) +[ok]: LRANGE inverted indexes - quicklist (49 ms) +[ok]: LRANGE out of range indexes including the full list - quicklist (45 ms) +[ok]: LRANGE out of range negative end index - quicklist (53 ms) +[ok]: LRANGE basics - listpack (4 ms) +[ok]: LRANGE inverted indexes - listpack (3 ms) +[ok]: LRANGE out of range indexes including the full list - listpack (2 ms) +[ok]: LRANGE out of range negative end index - listpack (1 ms) +[ok]: LRANGE against non existing key (0 ms) +[ok]: LRANGE with start > end yields an empty array for backward compatibility (4 ms) +[ok]: LTRIM basics - quicklist (601 ms) +[ok]: LTRIM out of range negative end index - quicklist (97 ms) +[ok]: LSET - quicklist (57 ms) +[ok]: LSET out of range index - quicklist (0 ms) +[ok]: LTRIM basics - listpack (56 ms) +[ok]: LTRIM out of range negative end index - listpack (3 ms) [ok]: LSET - listpack (3 ms) [ok]: LSET out of range index - listpack (1 ms) -[ok]: LSET against non existing key (2 ms) -[ok]: LSET against non list value (5 ms) -[ok]: LREM remove all the occurrences - quicklist (50 ms) -[ok]: LREM remove the first occurrence - quicklist (9 ms) -[ok]: LREM remove non existing element - quicklist (1 ms) -[ok]: LREM starting from tail with negative count - quicklist (49 ms) -[ok]: LREM starting from tail with negative count (2) - quicklist (2 ms) -[ok]: LREM deleting objects that may be int encoded - quicklist (54 ms) -[ok]: LREM remove all the occurrences - listpack (3 ms) -[ok]: LREM remove the first occurrence - listpack (1 ms) -[ok]: LREM remove non existing element - listpack (14 ms) -[ok]: LREM starting from tail with negative count - listpack (12 ms) -[ok]: LREM starting from tail with negative count (2) - listpack (1 ms) -[ok]: LREM deleting objects that may be int encoded - listpack (1 ms) -[ok]: Regression for bug 593 - chaining BRPOPLPUSH with other blocking cmds (85 ms) -[ok]: BLPOP/BLMOVE should increase dirty (54 ms) -[ok]: client unblock tests (130 ms) -[ok]: client unblock tests (139 ms) -[ok]: List listpack -> quicklist encoding conversion (258 ms) -[ok]: List listpack -> quicklist encoding conversion (431 ms) -[ok]: List quicklist -> listpack encoding conversion (659 ms) -[ok]: List encoding conversion when RDB loading (203 ms) -[ok]: List invalid list-max-listpack-size config (136 ms) -[ok]: List of various encodings (126 ms) -[ok]: List of various encodings - sanitize dump (24 ms) -[ok]: Unblock fairness is kept while pipelining (134 ms) -[ok]: Unblock fairness is kept during nested unblock (189 ms) -[ok]: Blocking command accounted only once in commandstats (88 ms) -[ok]: Blocking command accounted only once in commandstats after timeout (97 ms) -[ok]: Command being unblocked cause another command to get unblocked execution order test (400 ms) -[47/124 done]: unit/type/list (80 seconds) +[ok]: LSET against non existing key (1 ms) +[ok]: LSET against non list value (1 ms) +[ok]: LREM remove all the occurrences - quicklist (51 ms) +[ok]: LREM remove the first occurrence - quicklist (2 ms) +[ok]: LREM remove non existing element - quicklist (3 ms) +[ok]: LREM starting from tail with negative count - quicklist (64 ms) +[ok]: LREM starting from tail with negative count (2) - quicklist (1 ms) +[ok]: LREM deleting objects that may be int encoded - quicklist (51 ms) +[ok]: LREM remove all the occurrences - listpack (11 ms) +[ok]: LREM remove the first occurrence - listpack (0 ms) +[ok]: LREM remove non existing element - listpack (2 ms) +[ok]: LREM starting from tail with negative count - listpack (9 ms) +[ok]: LREM starting from tail with negative count (2) - listpack (2 ms) +[ok]: LREM deleting objects that may be int encoded - listpack (9 ms) +[ok]: Regression for bug 593 - chaining BRPOPLPUSH with other blocking cmds (31 ms) +[ok]: BLPOP/BLMOVE should increase dirty (17 ms) +[ok]: client unblock tests (15 ms) +[ok]: client unblock tests (20 ms) +[ok]: List listpack -> quicklist encoding conversion (16 ms) +[ok]: List listpack -> quicklist encoding conversion (189 ms) +[ok]: List quicklist -> listpack encoding conversion (244 ms) +[ok]: List encoding conversion when RDB loading (47 ms) +[ok]: List invalid list-max-listpack-size config (50 ms) +[ok]: List of various encodings (98 ms) +[ok]: List of various encodings - sanitize dump (4 ms) +[ok]: Unblock fairness is kept while pipelining (33 ms) +[ok]: Unblock fairness is kept during nested unblock (51 ms) +[ok]: Blocking command accounted only once in commandstats (20 ms) +[ok]: Blocking command accounted only once in commandstats after timeout (15 ms) +[ok]: Command being unblocked cause another command to get unblocked execution order test (170 ms) +[50/124 done]: unit/type/list (33 seconds) Testing unit/type/set -=== (set) Starting server 127.0.0.1:21531 ok -[ok]: SADD, SCARD, SISMEMBER, SMISMEMBER, SMEMBERS basics - listpack (110 ms) -[ok]: SADD, SCARD, SISMEMBER, SMISMEMBER, SMEMBERS basics - hashtable (1110 ms) -[ok]: SADD, SCARD, SISMEMBER, SMISMEMBER, SMEMBERS basics - intset (98 ms) -[ok]: SMISMEMBER SMEMBERS SCARD against non set (24 ms) -[ok]: SMISMEMBER SMEMBERS SCARD against non existing key (31 ms) -[ok]: SMISMEMBER requires one or more members (28 ms) -[ok]: SADD against non set (16 ms) -[ok]: SADD a non-integer against a small intset (55 ms) -[ok]: SADD a non-integer against a large intset (919 ms) -[ok]: SADD an integer larger than 64 bits (36 ms) -[ok]: SADD an integer larger than 64 bits to a large intset (980 ms) -[ok]: SADD overflows the maximum allowed integers in an intset - single (3664 ms) -[ok]: SADD overflows the maximum allowed elements in a listpack - single (1036 ms) -[ok]: SADD overflows the maximum allowed integers in an intset - multiple (48 ms) -[ok]: SADD overflows the maximum allowed elements in a listpack - multiple (35 ms) -[ok]: SADD overflows the maximum allowed integers in an intset - single_multiple (100 ms) -[ok]: SADD overflows the maximum allowed elements in a listpack - single_multiple (48 ms) -[ok]: Variadic SADD (24 ms) -[ok]: Set encoding after DEBUG RELOAD (6886 ms) -[ok]: SREM basics - $type (8 ms) -[ok]: SREM basics - $type (99 ms) -[ok]: SREM basics - intset (10 ms) -[ok]: SREM with multiple arguments (2 ms) -[ok]: SREM variadic version with more args needed to destroy the key (2 ms) -[ok]: SINTERCARD with illegal arguments (18 ms) -[ok]: SINTERCARD against non-set should throw error (2 ms) -[ok]: SINTERCARD against non-existing key (6 ms) -[ok]: Generated sets must be encoded correctly - regular (17 ms) +=== (set) Starting server 127.0.0.1:21559 ok +[ok]: SADD, SCARD, SISMEMBER, SMISMEMBER, SMEMBERS basics - listpack (4 ms) +[ok]: SADD, SCARD, SISMEMBER, SMISMEMBER, SMEMBERS basics - hashtable (30 ms) +[ok]: SADD, SCARD, SISMEMBER, SMISMEMBER, SMEMBERS basics - intset (4 ms) +[ok]: SMISMEMBER SMEMBERS SCARD against non set (1 ms) +[ok]: SMISMEMBER SMEMBERS SCARD against non existing key (1 ms) +[ok]: SMISMEMBER requires one or more members (1 ms) +[ok]: SADD against non set (0 ms) +[ok]: SADD a non-integer against a small intset (2 ms) +[ok]: SADD a non-integer against a large intset (25 ms) +[ok]: SADD an integer larger than 64 bits (2 ms) +[ok]: SADD an integer larger than 64 bits to a large intset (25 ms) +[ok]: SADD overflows the maximum allowed integers in an intset - single (104 ms) +[ok]: SADD overflows the maximum allowed elements in a listpack - single (23 ms) +[ok]: SADD overflows the maximum allowed integers in an intset - multiple (3 ms) +[ok]: SADD overflows the maximum allowed elements in a listpack - multiple (2 ms) +[ok]: SADD overflows the maximum allowed integers in an intset - single_multiple (48 ms) +[ok]: SADD overflows the maximum allowed elements in a listpack - single_multiple (3 ms) +[ok]: Variadic SADD (2 ms) +[ok]: Set encoding after DEBUG RELOAD (336 ms) +[ok]: SREM basics - $type (1 ms) +[ok]: SREM basics - $type (22 ms) +[ok]: SREM basics - intset (2 ms) +[ok]: SREM with multiple arguments (1 ms) +[ok]: SREM variadic version with more args needed to destroy the key (0 ms) +[ok]: SINTERCARD with illegal arguments (3 ms) +[ok]: SINTERCARD against non-set should throw error (1 ms) +[ok]: SINTERCARD against non-existing key (0 ms) +[ok]: Generated sets must be encoded correctly - regular (1 ms) [ok]: SINTER with two sets - regular (1 ms) -[ok]: SINTERCARD with two sets - regular (2 ms) -[ok]: SINTERSTORE with two sets - regular (2 ms) -[ok]: SINTERSTORE with two sets, after a DEBUG RELOAD - regular (682 ms) -[ok]: SUNION with two sets - regular (52 ms) -[ok]: SUNIONSTORE with two sets - regular (47 ms) +[ok]: SINTERCARD with two sets - regular (1 ms) +[ok]: SINTERSTORE with two sets - regular (1 ms) +[ok]: SINTERSTORE with two sets, after a DEBUG RELOAD - regular (11 ms) +[ok]: SUNION with two sets - regular (7 ms) +[ok]: SUNIONSTORE with two sets - regular (7 ms) [ok]: SINTER against three sets - regular (1 ms) -[ok]: SINTERCARD against three sets - regular (2 ms) -[ok]: SINTERSTORE with three sets - regular (2 ms) -[ok]: SUNION with non existing keys - regular (42 ms) -[ok]: SDIFF with two sets - regular (17 ms) -[ok]: SDIFF with three sets - regular (1 ms) -[ok]: SDIFFSTORE with three sets - regular (10 ms) -[ok]: SINTER/SUNION/SDIFF with three same sets - regular (40 ms) -[ok]: Generated sets must be encoded correctly - intset (40 ms) -[ok]: SINTER with two sets - intset (8 ms) -[ok]: SINTERCARD with two sets - intset (23 ms) -[ok]: SINTERSTORE with two sets - intset (24 ms) -[ok]: SINTERSTORE with two sets, after a DEBUG RELOAD - intset (84 ms) -[ok]: SUNION with two sets - intset (63 ms) -[ok]: SUNIONSTORE with two sets - intset (79 ms) -[ok]: SINTER against three sets - intset (8 ms) -[ok]: SINTERCARD against three sets - intset (23 ms) -[ok]: SINTERSTORE with three sets - intset (16 ms) -[ok]: SUNION with non existing keys - intset (64 ms) -[ok]: SDIFF with two sets - intset (12 ms) -[ok]: SDIFF with three sets - intset (8 ms) -[ok]: SDIFFSTORE with three sets - intset (24 ms) -[ok]: SINTER/SUNION/SDIFF with three same sets - intset (56 ms) -[ok]: SINTERSTORE with two listpack sets where result is intset (60 ms) -[ok]: SINTERSTORE with two hashtable sets where result is intset (1812 ms) -[ok]: SUNION hashtable and listpack (116 ms) -[ok]: SDIFF with first set empty (32 ms) -[ok]: SDIFF with same set two times (20 ms) -[ok]: SDIFF fuzzing (16005 ms) -[ok]: SDIFF against non-set should throw error (7 ms) -[ok]: SDIFF should handle non existing key as empty (2 ms) +[ok]: SINTERCARD against three sets - regular (1 ms) +[ok]: SINTERSTORE with three sets - regular (0 ms) +[ok]: SUNION with non existing keys - regular (6 ms) +[ok]: SDIFF with two sets - regular (1 ms) +[ok]: SDIFF with three sets - regular (0 ms) +[ok]: SDIFFSTORE with three sets - regular (1 ms) +[ok]: SINTER/SUNION/SDIFF with three same sets - regular (6 ms) +[ok]: Generated sets must be encoded correctly - intset (1 ms) +[ok]: SINTER with two sets - intset (0 ms) +[ok]: SINTERCARD with two sets - intset (1 ms) +[ok]: SINTERSTORE with two sets - intset (1 ms) +[ok]: SINTERSTORE with two sets, after a DEBUG RELOAD - intset (12 ms) +[ok]: SUNION with two sets - intset (6 ms) +[ok]: SUNIONSTORE with two sets - intset (6 ms) +[ok]: SINTER against three sets - intset (0 ms) +[ok]: SINTERCARD against three sets - intset (1 ms) +[ok]: SINTERSTORE with three sets - intset (1 ms) +[ok]: SUNION with non existing keys - intset (5 ms) +[ok]: SDIFF with two sets - intset (1 ms) +[ok]: SDIFF with three sets - intset (0 ms) +[ok]: SDIFFSTORE with three sets - intset (1 ms) +[ok]: SINTER/SUNION/SDIFF with three same sets - intset (4 ms) +[ok]: SINTERSTORE with two listpack sets where result is intset (2 ms) +[ok]: SINTERSTORE with two hashtable sets where result is intset (45 ms) +[ok]: SUNION hashtable and listpack (3 ms) +[ok]: SDIFF with first set empty (1 ms) +[ok]: SDIFF with same set two times (1 ms) +[ok]: SDIFF fuzzing (6911 ms) +[ok]: SDIFF against non-set should throw error (2 ms) +[ok]: SDIFF should handle non existing key as empty (1 ms) [ok]: SDIFFSTORE against non-set should throw error (3 ms) -[ok]: SDIFFSTORE should handle non existing key as empty (3 ms) -[ok]: SINTER against non-set should throw error (5 ms) +[ok]: SDIFFSTORE should handle non existing key as empty (2 ms) +[ok]: SINTER against non-set should throw error (1 ms) [ok]: SINTER should handle non existing key as empty (1 ms) -[ok]: SINTER with same integer elements but different encoding (2 ms) -[ok]: SINTERSTORE against non-set should throw error (9 ms) -[ok]: SINTERSTORE against non existing keys should delete dstkey (2 ms) -[ok]: SUNION against non-set should throw error (2 ms) -[ok]: SUNION should handle non existing key as empty (1 ms) +[ok]: SINTER with same integer elements but different encoding (1 ms) +[ok]: SINTERSTORE against non-set should throw error (6 ms) +[ok]: SINTERSTORE against non existing keys should delete dstkey (3 ms) +[ok]: SUNION against non-set should throw error (1 ms) +[ok]: SUNION should handle non existing key as empty (0 ms) [ok]: SUNIONSTORE against non-set should throw error (3 ms) [ok]: SUNIONSTORE should handle non existing key as empty (3 ms) -[ok]: SUNIONSTORE against non existing keys should delete dstkey (1 ms) -[ok]: SPOP basics - listpack (2 ms) -[ok]: SPOP with =1 - listpack (9 ms) -[ok]: SRANDMEMBER - listpack (22 ms) -[ok]: SPOP basics - intset (3 ms) +[ok]: SUNIONSTORE against non existing keys should delete dstkey (0 ms) +[ok]: SPOP basics - listpack (1 ms) +[ok]: SPOP with =1 - listpack (2 ms) +[ok]: SRANDMEMBER - listpack (18 ms) +[ok]: SPOP basics - intset (2 ms) [ok]: SPOP with =1 - intset (1 ms) -[ok]: SRANDMEMBER - intset (19 ms) -[ok]: SPOP integer from listpack set (9 ms) +[ok]: SRANDMEMBER - intset (17 ms) +[ok]: SPOP integer from listpack set (3 ms) [ok]: SPOP with - listpack (6 ms) -[ok]: SPOP with - intset (11 ms) -[ok]: SPOP with - hashtable (10 ms) +[ok]: SPOP with - intset (6 ms) +[ok]: SPOP with - hashtable (5 ms) [ok]: SPOP using integers, testing Knuth's and Floyd's algorithm (5 ms) -[ok]: SPOP using integers with Knuth's algorithm (1 ms) -[ok]: SPOP new implementation: code path #1 intset (8 ms) -[ok]: SPOP new implementation: code path #2 intset (10 ms) -[ok]: SPOP new implementation: code path #3 intset (4 ms) -[ok]: SPOP new implementation: code path #1 listpack (9 ms) -[ok]: SPOP new implementation: code path #2 listpack (6 ms) -[ok]: SPOP new implementation: code path #3 listpack (4 ms) -[ok]: SPOP new implementation: code path #1 propagate as DEL or UNLINK (100 ms) -[ok]: SRANDMEMBER count of 0 is handled correctly (1 ms) -[ok]: SRANDMEMBER with against non existing key (0 ms) -[ok]: SRANDMEMBER count overflow (1 ms) -[ok]: SRANDMEMBER count of 0 is handled correctly - emptyarray (0 ms) +[ok]: SPOP using integers with Knuth's algorithm (0 ms) +[ok]: SPOP new implementation: code path #1 intset (4 ms) +[ok]: SPOP new implementation: code path #2 intset (4 ms) +[ok]: SPOP new implementation: code path #3 intset (5 ms) +[ok]: SPOP new implementation: code path #1 listpack (4 ms) +[ok]: SPOP new implementation: code path #2 listpack (4 ms) +[ok]: SPOP new implementation: code path #3 listpack (5 ms) +[ok]: SPOP new implementation: code path #1 propagate as DEL or UNLINK (121 ms) +[ok]: SRANDMEMBER count of 0 is handled correctly (0 ms) +[ok]: SRANDMEMBER with against non existing key (1 ms) +[ok]: SRANDMEMBER count overflow (0 ms) +[ok]: SRANDMEMBER count of 0 is handled correctly - emptyarray (1 ms) [ok]: SRANDMEMBER with against non existing key - emptyarray (0 ms) -[ok]: SRANDMEMBER with - listpack (77 ms) -[ok]: SRANDMEMBER with - intset (55 ms) -[ok]: SRANDMEMBER with - hashtable (99 ms) -[ok]: SRANDMEMBER histogram distribution - listpack (227 ms) -[ok]: SRANDMEMBER histogram distribution - intset (226 ms) -[ok]: SRANDMEMBER histogram distribution - hashtable (248 ms) +[ok]: SRANDMEMBER with - listpack (33 ms) +[ok]: SRANDMEMBER with - intset (32 ms) +[ok]: SRANDMEMBER with - hashtable (39 ms) +[ok]: SRANDMEMBER histogram distribution - listpack (152 ms) +[ok]: SRANDMEMBER histogram distribution - intset (135 ms) +[ok]: SRANDMEMBER histogram distribution - hashtable (153 ms) Waiting for background save to finish... Hash table 0 stats (main hash table): table size: 4096 number of elements: 499 - different slots: 21 - max chain length: 32 - avg chain length (counted): 23.76 - avg chain length (computed): 23.76 + different slots: 20 + max chain length: 34 + avg chain length (counted): 24.95 + avg chain length (computed): 24.95 Chain length distribution: - 0: 4075 (99.49%) - 9: 1 (0.02%) - 17: 1 (0.02%) - 19: 1 (0.02%) - 21: 4 (0.10%) - 22: 1 (0.02%) - 23: 2 (0.05%) + 0: 4076 (99.51%) + 10: 1 (0.02%) + 18: 1 (0.02%) + 20: 3 (0.07%) + 21: 1 (0.02%) + 22: 2 (0.05%) 24: 1 (0.02%) - 25: 1 (0.02%) - 26: 2 (0.05%) - 27: 2 (0.05%) + 26: 4 (0.10%) 28: 1 (0.02%) - 29: 3 (0.07%) - 32: 1 (0.02%) + 30: 1 (0.02%) + 31: 3 (0.07%) + 33: 1 (0.02%) + 34: 1 (0.02%) Waiting for background save to finish... -Waiting for background save to finish... [ok]: SRANDMEMBER with a dict containing long chain (104241 ms) -[ok]: SMOVE basics - from regular set to intset (90 ms) -[ok]: SMOVE basics - from intset to regular set (4 ms) -[ok]: SMOVE non existing key (4 ms) -[ok]: SMOVE non existing src set (12 ms) -[ok]: SMOVE from regular set to non existing destination set (6 ms) -[ok]: SMOVE from intset to non existing destination set (9 ms) +Waiting for background save to finish... [ok]: SRANDMEMBER with a dict containing long chain (29990 ms) +[ok]: SMOVE basics - from regular set to intset (6 ms) +[ok]: SMOVE basics - from intset to regular set (2 ms) +[ok]: SMOVE non existing key (3 ms) +[ok]: SMOVE non existing src set (4 ms) +[ok]: SMOVE from regular set to non existing destination set (4 ms) +[ok]: SMOVE from intset to non existing destination set (4 ms) [ok]: SMOVE wrong src key type (0 ms) -[ok]: SMOVE wrong dst key type (1 ms) -[ok]: SMOVE with identical source and destination (6 ms) -[ok]: SMOVE only notify dstset when the addition is successful (37 ms) -[ok]: intsets implementation stress testing (28804 ms) -=== () Starting server 127.0.0.1:21533 ok +[ok]: SMOVE wrong dst key type (0 ms) +[ok]: SMOVE with identical source and destination (9 ms) +[ok]: SMOVE only notify dstset when the addition is successful (16 ms) +[ok]: intsets implementation stress testing (5907 ms) +=== () Starting server 127.0.0.1:21561 ok [ignore]: SADD, SCARD, SISMEMBER - large data: large memory flag not provided -[48/124 done]: unit/type/set (174 seconds) -Testing unit/type/incr -=== (incr) Starting server 127.0.0.1:21535 ok -[ok]: INCR against non existing key (14 ms) -[ok]: INCR against key created by incr itself (5 ms) -[ok]: DECR against key created by incr (3 ms) -[ok]: DECR against key is not exist and incr (19 ms) -[ok]: INCR against key originally set with SET (16 ms) -[ok]: INCR over 32bit value (12 ms) -[ok]: INCRBY over 32bit value with over 32bit increment (10 ms) -[ok]: INCR fails against key with spaces (left) (12 ms) -[ok]: INCR fails against key with spaces (right) (16 ms) -[ok]: INCR fails against key with spaces (both) (11 ms) -[ok]: DECRBY negation overflow (16 ms) -[ok]: INCR fails against a key holding a list (20 ms) -[ok]: DECRBY over 32bit value with over 32bit increment, negative res (15 ms) -[ok]: DECRBY against key is not exist (11 ms) -[ok]: INCR uses shared objects in the 0-9999 range (52 ms) -[ok]: INCR can modify objects in-place (35 ms) -[ok]: INCRBYFLOAT against non existing key (37 ms) -[ok]: INCRBYFLOAT against key originally set with SET (18 ms) -[ok]: INCRBYFLOAT over 32bit value (8 ms) -[ok]: INCRBYFLOAT over 32bit value with over 32bit increment (16 ms) -[ok]: INCRBYFLOAT fails against key with spaces (left) (16 ms) -[ok]: INCRBYFLOAT fails against key with spaces (right) (16 ms) -[ok]: INCRBYFLOAT fails against key with spaces (both) (12 ms) -[ok]: INCRBYFLOAT fails against a key holding a list (27 ms) -[ok]: INCRBYFLOAT does not allow NaN or Infinity (11 ms) -[ok]: INCRBYFLOAT decrement (16 ms) -[ok]: string to double with null terminator (16 ms) -[ok]: No negative zero (23 ms) -[ok]: INCRBY INCRBYFLOAT DECRBY against unhappy path (51 ms) -[ok]: incr operation should update encoding from raw to int (60 ms) -[ok]: decr operation should update encoding from raw to int (52 ms) -[ok]: incrby operation should update encoding from raw to int (56 ms) -[ok]: decrby operation should update encoding from raw to int (68 ms) -[49/124 done]: unit/type/incr (1 seconds) -Testing unit/type/list-2 -=== (list) Starting server 127.0.0.1:21537 ok -[ok]: LTRIM stress testing - quicklist (70722 ms) -[ok]: LTRIM stress testing - listpack (62381 ms) -[50/124 done]: unit/type/list-2 (134 seconds) +[51/124 done]: unit/type/set (45 seconds) +Testing unit/type/stream-cgroups +=== (stream) Starting server 127.0.0.1:21563 ok +[ok]: XGROUP CREATE: creation and duplicate group name detection (1 ms) +[ok]: XGROUP CREATE: with ENTRIESREAD parameter (3 ms) +[ok]: XGROUP CREATE: automatic stream creation fails without MKSTREAM (0 ms) +[ok]: XGROUP CREATE: automatic stream creation works with MKSTREAM (1 ms) +[ok]: XREADGROUP will return only new elements (1 ms) +[ok]: XREADGROUP can read the history of the elements we own (1 ms) +[ok]: XPENDING is able to return pending items (1 ms) +[ok]: XPENDING can return single consumer items (0 ms) +[ok]: XPENDING only group (0 ms) +[ok]: XPENDING with IDLE (22 ms) +[ok]: XPENDING with exclusive range intervals works as expected (1 ms) +[ok]: XACK is able to remove items from the consumer/group PEL (1 ms) +[ok]: XACK can't remove the same item multiple times (1 ms) +[ok]: XACK is able to accept multiple arguments (0 ms) +[ok]: XACK should fail if got at least one invalid ID (2 ms) +[ok]: PEL NACK reassignment after XGROUP SETID event (2 ms) +[ok]: XREADGROUP will not report data on empty history. Bug #5577 (3 ms) +[ok]: XREADGROUP history reporting of deleted entries. Bug #5570 (1 ms) +[ok]: Blocking XREADGROUP will not reply with an empty array (36 ms) +[ok]: Blocking XREADGROUP: key deleted (16 ms) +[ok]: Blocking XREADGROUP: key type changed with SET (17 ms) +[ok]: Blocking XREADGROUP: key type changed with transaction (16 ms) +[ok]: Blocking XREADGROUP: flushed DB (11 ms) +[ok]: Blocking XREADGROUP: swapped DB, key doesn't exist (13 ms) +[ok]: Blocking XREADGROUP: swapped DB, key is not a stream (10 ms) +[ok]: XREAD and XREADGROUP against wrong parameter (2 ms) +[ok]: Blocking XREAD: key deleted (9 ms) +[ok]: Blocking XREAD: key type changed with SET (14 ms) +[ok]: Blocking XREADGROUP for stream that ran dry (issue #5299) (102 ms) +[ok]: Blocking XREADGROUP will ignore BLOCK if ID is not > (13 ms) +[ok]: Blocking XREADGROUP for stream key that has clients blocked on list (22 ms) +[ok]: Blocking XREADGROUP for stream key that has clients blocked on stream - avoid endless loop (28 ms) +[ok]: Blocking XREADGROUP for stream key that has clients blocked on stream - reprocessing command (1050 ms) +[ok]: XGROUP DESTROY should unblock XREADGROUP with -NOGROUP (15 ms) +[ok]: RENAME can unblock XREADGROUP with data (13 ms) +[ok]: RENAME can unblock XREADGROUP with -NOGROUP (11 ms) +[ok]: XCLAIM can claim PEL items from another consumer (618 ms) +[ok]: XCLAIM without JUSTID increments delivery count (412 ms) +[ok]: XCLAIM same consumer (213 ms) +[ok]: XAUTOCLAIM can claim PEL items from another consumer (1678 ms) +[ok]: XAUTOCLAIM as an iterator (207 ms) +[ok]: XAUTOCLAIM COUNT must be > 0 (9 ms) +[ok]: XCLAIM with XDEL (3 ms) +[ok]: XCLAIM with trimming (9 ms) +[ok]: XAUTOCLAIM with XDEL (19 ms) +[ok]: XAUTOCLAIM with XDEL and count (3 ms) +[ok]: XAUTOCLAIM with out of range count (0 ms) +[ok]: XCLAIM with trimming (17 ms) +[ok]: XINFO FULL output (14 ms) +[ok]: Consumer seen-time and active-time (228 ms) +[ok]: XGROUP CREATECONSUMER: create consumer if does not exist (10 ms) +[ok]: XGROUP CREATECONSUMER: group must exist (1 ms) +=== (stream needs:debug) Starting server 127.0.0.1:21565 ok +[ok]: XREADGROUP with NOACK creates consumer (364 ms) + +Waiting for background AOF rewrite to finish... +Waiting for background AOF rewrite to finish... +Waiting for background AOF rewrite to finish... +Waiting for background AOF rewrite to finish... +Waiting for background AOF rewrite to finish... +Waiting for background AOF rewrite to finish... [ok]: Consumer without PEL is present in AOF after AOFRW (3190 ms) +[ok]: Consumer group read counter and lag in empty streams (7 ms) +[ok]: Consumer group read counter and lag sanity (20 ms) +[ok]: Consumer group lag with XDELs (30 ms) +[ok]: Loading from legacy (Redis <= v6.2.x, rdb_ver < 10) persistence (3 ms) +[ok]: Loading from legacy (Redis <= v7.0.x, rdb_ver < 11) persistence (2 ms) +=== (external:skip) Starting server 127.0.0.1:21567 ok +[ok]: Consumer group last ID propagation to slave (NOACK=0) (545 ms) +[ok]: Consumer group last ID propagation to slave (NOACK=1) (279 ms) +=== (external:skip) Starting server 127.0.0.1:21569 ok +[ok]: Replication tests of XCLAIM with deleted entries (autoclaim=0) (554 ms) +[ok]: Replication tests of XCLAIM with deleted entries (autoclaim=1) (19 ms) +[ok]: XREADGROUP ACK would propagate entries-read (11 ms) +=== (stream needs:debug) Starting server 127.0.0.1:21571 ok + +Waiting for background AOF rewrite to finish... +Waiting for background AOF rewrite to finish... +Waiting for background AOF rewrite to finish... +Waiting for background AOF rewrite to finish... +Waiting for background AOF rewrite to finish... +Waiting for background AOF rewrite to finish... [ok]: Empty stream with no lastid can be rewrite into AOF correctly (1153 ms) +[52/124 done]: unit/type/stream-cgroups (15 seconds) Testing unit/type/zset -=== (zset) Starting server 127.0.0.1:21539 ok -[ok]: Check encoding - listpack (16 ms) +=== (zset) Starting server 127.0.0.1:21573 ok +[ok]: Check encoding - listpack (1 ms) [ok]: ZSET basic ZADD and score update - listpack (2 ms) -[ok]: ZSET element can't be set to NaN with ZADD - listpack (5 ms) -[ok]: ZSET element can't be set to NaN with ZINCRBY - listpack (1 ms) +[ok]: ZSET element can't be set to NaN with ZADD - listpack (0 ms) +[ok]: ZSET element can't be set to NaN with ZINCRBY - listpack (2 ms) [ok]: ZADD with options syntax error with incomplete pair - listpack (1 ms) -[ok]: ZADD XX option without key - listpack (7 ms) -[ok]: ZADD XX existing key - listpack (6 ms) -[ok]: ZADD XX returns the number of elements actually added - listpack (6 ms) -[ok]: ZADD XX updates existing elements score - listpack (3 ms) -[ok]: ZADD GT updates existing elements when new scores are greater - listpack (5 ms) -[ok]: ZADD LT updates existing elements when new scores are lower - listpack (13 ms) +[ok]: ZADD XX option without key - listpack (1 ms) +[ok]: ZADD XX existing key - listpack (1 ms) +[ok]: ZADD XX returns the number of elements actually added - listpack (1 ms) +[ok]: ZADD XX updates existing elements score - listpack (2 ms) +[ok]: ZADD GT updates existing elements when new scores are greater - listpack (4 ms) +[ok]: ZADD LT updates existing elements when new scores are lower - listpack (2 ms) [ok]: ZADD GT XX updates existing elements when new scores are greater and skips new elements - listpack (2 ms) -[ok]: ZADD LT XX updates existing elements when new scores are lower and skips new elements - listpack (9 ms) +[ok]: ZADD LT XX updates existing elements when new scores are lower and skips new elements - listpack (2 ms) [ok]: ZADD XX and NX are not compatible - listpack (1 ms) -[ok]: ZADD NX with non existing key - listpack (5 ms) -[ok]: ZADD NX only add new elements without updating old ones - listpack (8 ms) +[ok]: ZADD NX with non existing key - listpack (1 ms) +[ok]: ZADD NX only add new elements without updating old ones - listpack (2 ms) [ok]: ZADD GT and NX are not compatible - listpack (2 ms) -[ok]: ZADD LT and NX are not compatible - listpack (1 ms) +[ok]: ZADD LT and NX are not compatible - listpack (0 ms) [ok]: ZADD LT and GT are not compatible - listpack (0 ms) -[ok]: ZADD INCR LT/GT replies with nill if score not updated - listpack (8 ms) -[ok]: ZADD INCR LT/GT with inf - listpack (12 ms) -[ok]: ZADD INCR works like ZINCRBY - listpack (3 ms) -[ok]: ZADD INCR works with a single score-element pair - listpack (2 ms) -[ok]: ZADD CH option changes return value to all changed elements - listpack (9 ms) +[ok]: ZADD INCR LT/GT replies with nill if score not updated - listpack (2 ms) +[ok]: ZADD INCR LT/GT with inf - listpack (4 ms) +[ok]: ZADD INCR works like ZINCRBY - listpack (2 ms) +[ok]: ZADD INCR works with a single score-element pair - listpack (1 ms) +[ok]: ZADD CH option changes return value to all changed elements - listpack (1 ms) [ok]: ZINCRBY calls leading to NaN result in error - listpack (1 ms) [ok]: ZINCRBY against invalid incr value - listpack (1 ms) [ok]: ZADD - Variadic version base case - listpack (1 ms) [ok]: ZADD - Return value is the number of actually added items - listpack (1 ms) -[ok]: ZADD - Variadic version does not add nothing on single parsing err - listpack (10 ms) -[ok]: ZADD - Variadic version will raise error on missing arg - listpack (1 ms) +[ok]: ZADD - Variadic version does not add nothing on single parsing err - listpack (4 ms) +[ok]: ZADD - Variadic version will raise error on missing arg - listpack (0 ms) [ok]: ZINCRBY does not work variadic even if shares ZADD implementation - listpack (1 ms) [ok]: ZCARD basics - listpack (2 ms) [ok]: ZREM removes key after last element is removed - listpack (8 ms) -[ok]: ZREM variadic version - listpack (2 ms) -[ok]: ZREM variadic version -- remove elements after key deletion - listpack (8 ms) -[ok]: ZRANGE basics - listpack (22 ms) -[ok]: ZREVRANGE basics - listpack (7 ms) -[ok]: ZRANK/ZREVRANK basics - listpack (21 ms) -[ok]: ZRANK - after deletion - listpack (8 ms) -[ok]: ZINCRBY - can create a new sorted set - listpack (7 ms) +[ok]: ZREM variadic version - listpack (1 ms) +[ok]: ZREM variadic version -- remove elements after key deletion - listpack (1 ms) +[ok]: ZRANGE basics - listpack (12 ms) +[ok]: ZREVRANGE basics - listpack (16 ms) +[ok]: ZRANK/ZREVRANK basics - listpack (7 ms) +[ok]: ZRANK - after deletion - listpack (1 ms) +[ok]: ZINCRBY - can create a new sorted set - listpack (2 ms) [ok]: ZINCRBY - increment and decrement - listpack (2 ms) -[ok]: ZINCRBY return value - listpack (1 ms) -[ok]: ZRANGEBYSCORE/ZREVRANGEBYSCORE/ZCOUNT basics - listpack (28 ms) -[ok]: ZRANGEBYSCORE with WITHSCORES - listpack (4 ms) -[ok]: ZRANGEBYSCORE with LIMIT - listpack (24 ms) -[ok]: ZRANGEBYSCORE with LIMIT and WITHSCORES - listpack (3 ms) -[ok]: ZRANGEBYSCORE with non-value min or max - listpack (11 ms) -[ok]: ZRANGEBYLEX/ZREVRANGEBYLEX/ZLEXCOUNT basics - listpack (7 ms) -[ok]: ZLEXCOUNT advanced - listpack (5 ms) -[ok]: ZRANGEBYLEX with LIMIT - listpack (21 ms) -[ok]: ZRANGEBYLEX with invalid lex range specifiers - listpack (1 ms) -[ok]: ZREMRANGEBYSCORE basics - listpack (42 ms) -[ok]: ZREMRANGEBYSCORE with non-value min or max - listpack (2 ms) -[ok]: ZREMRANGEBYRANK basics - listpack (18 ms) -[ok]: ZREMRANGEBYLEX basics - listpack (61 ms) +[ok]: ZINCRBY return value - listpack (0 ms) +[ok]: ZRANGEBYSCORE/ZREVRANGEBYSCORE/ZCOUNT basics - listpack (16 ms) +[ok]: ZRANGEBYSCORE with WITHSCORES - listpack (5 ms) +[ok]: ZRANGEBYSCORE with LIMIT - listpack (26 ms) +[ok]: ZRANGEBYSCORE with LIMIT and WITHSCORES - listpack (4 ms) +[ok]: ZRANGEBYSCORE with non-value min or max - listpack (3 ms) +[ok]: ZRANGEBYLEX/ZREVRANGEBYLEX/ZLEXCOUNT basics - listpack (13 ms) +[ok]: ZLEXCOUNT advanced - listpack (3 ms) +[ok]: ZRANGEBYLEX with LIMIT - listpack (23 ms) +[ok]: ZRANGEBYLEX with invalid lex range specifiers - listpack (7 ms) +[ok]: ZREMRANGEBYSCORE basics - listpack (36 ms) +[ok]: ZREMRANGEBYSCORE with non-value min or max - listpack (0 ms) +[ok]: ZREMRANGEBYRANK basics - listpack (10 ms) +[ok]: ZREMRANGEBYLEX basics - listpack (32 ms) [ok]: ZUNIONSTORE against non-existing key doesn't set destination - listpack (1 ms) [ok]: ZUNION/ZINTER/ZINTERCARD/ZDIFF against non-existing key - listpack (2 ms) [ok]: ZUNIONSTORE with empty set - listpack (1 ms) -[ok]: ZUNION/ZINTER/ZINTERCARD/ZDIFF with empty set - listpack (8 ms) -[ok]: ZUNIONSTORE basics - listpack (4 ms) -[ok]: ZUNION/ZINTER/ZINTERCARD/ZDIFF with integer members - listpack (3 ms) -[ok]: ZUNIONSTORE with weights - listpack (2 ms) -[ok]: ZUNION with weights - listpack (5 ms) -[ok]: ZUNIONSTORE with a regular set and weights - listpack (6 ms) -[ok]: ZUNIONSTORE with AGGREGATE MIN - listpack (1 ms) +[ok]: ZUNION/ZINTER/ZINTERCARD/ZDIFF with empty set - listpack (7 ms) +[ok]: ZUNIONSTORE basics - listpack (1 ms) +[ok]: ZUNION/ZINTER/ZINTERCARD/ZDIFF with integer members - listpack (6 ms) +[ok]: ZUNIONSTORE with weights - listpack (0 ms) +[ok]: ZUNION with weights - listpack (0 ms) +[ok]: ZUNIONSTORE with a regular set and weights - listpack (2 ms) +[ok]: ZUNIONSTORE with AGGREGATE MIN - listpack (0 ms) [ok]: ZUNION/ZINTER with AGGREGATE MIN - listpack (1 ms) -[ok]: ZUNIONSTORE with AGGREGATE MAX - listpack (6 ms) -[ok]: ZUNION/ZINTER with AGGREGATE MAX - listpack (3 ms) +[ok]: ZUNIONSTORE with AGGREGATE MAX - listpack (0 ms) +[ok]: ZUNION/ZINTER with AGGREGATE MAX - listpack (0 ms) [ok]: ZINTERSTORE basics - listpack (1 ms) -[ok]: ZINTER basics - listpack (1 ms) -[ok]: ZINTERCARD with illegal arguments (1 ms) -[ok]: ZINTERCARD basics - listpack (2 ms) -[ok]: ZINTER RESP3 - listpack (1 ms) +[ok]: ZINTER basics - listpack (0 ms) +[ok]: ZINTERCARD with illegal arguments (3 ms) +[ok]: ZINTERCARD basics - listpack (1 ms) +[ok]: ZINTER RESP3 - listpack (2 ms) [ok]: ZINTERSTORE with weights - listpack (5 ms) -[ok]: ZINTER with weights - listpack (2 ms) +[ok]: ZINTER with weights - listpack (0 ms) [ok]: ZINTERSTORE with a regular set and weights - listpack (2 ms) [ok]: ZINTERSTORE with AGGREGATE MIN - listpack (1 ms) -[ok]: ZINTERSTORE with AGGREGATE MAX - listpack (1 ms) -[ok]: ZUNIONSTORE with +inf/-inf scores - listpack (10 ms) -[ok]: ZUNIONSTORE with NaN weights - listpack (3 ms) -[ok]: ZINTERSTORE with +inf/-inf scores - listpack (12 ms) -[ok]: ZINTERSTORE with NaN weights - listpack (1 ms) -[ok]: ZDIFFSTORE basics - listpack (1 ms) +[ok]: ZINTERSTORE with AGGREGATE MAX - listpack (0 ms) +[ok]: ZUNIONSTORE with +inf/-inf scores - listpack (7 ms) +[ok]: ZUNIONSTORE with NaN weights - listpack (1 ms) +[ok]: ZINTERSTORE with +inf/-inf scores - listpack (10 ms) +[ok]: ZINTERSTORE with NaN weights - listpack (11 ms) +[ok]: ZDIFFSTORE basics - listpack (0 ms) [ok]: ZDIFF basics - listpack (0 ms) -[ok]: ZDIFFSTORE with a regular set - listpack (8 ms) -[ok]: ZDIFF subtracting set from itself - listpack (1 ms) -[ok]: ZDIFF algorithm 1 - listpack (2 ms) -[ok]: ZDIFF algorithm 2 - listpack (5 ms) -[ok]: ZDIFF fuzzing - listpack (32426 ms) -[ok]: ZPOPMIN with the count 0 returns an empty array (12 ms) +[ok]: ZDIFFSTORE with a regular set - listpack (2 ms) +[ok]: ZDIFF subtracting set from itself - listpack (0 ms) +[ok]: ZDIFF algorithm 1 - listpack (9 ms) +[ok]: ZDIFF algorithm 2 - listpack (3 ms) +[ok]: ZDIFF fuzzing - listpack (12565 ms) +[ok]: ZPOPMIN with the count 0 returns an empty array (1 ms) [ok]: ZPOPMIN with negative count (1 ms) [ok]: ZPOPMAX with the count 0 returns an empty array (1 ms) -[ok]: ZPOPMAX with negative count (1 ms) -[ok]: Basic ZPOPMIN/ZPOPMAX with a single key - listpack (9 ms) -[ok]: ZPOPMIN/ZPOPMAX with count - listpack (6 ms) +[ok]: ZPOPMAX with negative count (2 ms) +[ok]: Basic ZPOPMIN/ZPOPMAX with a single key - listpack (3 ms) +[ok]: ZPOPMIN/ZPOPMAX with count - listpack (2 ms) [ok]: Basic ZMPOP_MIN/ZMPOP_MAX with a single key - listpack (3 ms) -[ok]: ZMPOP_MIN/ZMPOP_MAX with count - listpack (12 ms) -[ok]: BZPOPMIN/BZPOPMAX with a single existing sorted set - listpack (43 ms) -[ok]: BZPOPMIN/BZPOPMAX with multiple existing sorted sets - listpack (52 ms) -[ok]: BZPOPMIN/BZPOPMAX second sorted set has members - listpack (91 ms) -[ok]: BZMPOP_MIN/BZMPOP_MAX with a single existing sorted set - listpack (100 ms) -[ok]: BZMPOP_MIN/BZMPOP_MAX with multiple existing sorted sets - listpack (123 ms) -[ok]: BZMPOP_MIN/BZMPOP_MAX second sorted set has members - listpack (91 ms) -[ok]: Basic ZPOPMIN/ZPOPMAX - listpack RESP3 (64 ms) -[ok]: ZPOPMIN/ZPOPMAX with count - listpack RESP3 (60 ms) -[ok]: Basic ZMPOP_MIN/ZMPOP_MAX - listpack RESP3 (64 ms) -[ok]: ZMPOP_MIN/ZMPOP_MAX with count - listpack RESP3 (52 ms) -[ok]: BZPOPMIN/BZPOPMAX - listpack RESP3 (120 ms) -[ok]: BZMPOP_MIN/BZMPOP_MAX - listpack RESP3 (123 ms) -[ok]: Check encoding - skiplist (27 ms) -[ok]: ZSET basic ZADD and score update - skiplist (48 ms) -[ok]: ZSET element can't be set to NaN with ZADD - skiplist (12 ms) -[ok]: ZSET element can't be set to NaN with ZINCRBY - skiplist (5 ms) -[ok]: ZADD with options syntax error with incomplete pair - skiplist (18 ms) -[ok]: ZADD XX option without key - skiplist (23 ms) -[ok]: ZADD XX existing key - skiplist (32 ms) -[ok]: ZADD XX returns the number of elements actually added - skiplist (24 ms) -[ok]: ZADD XX updates existing elements score - skiplist (24 ms) -[ok]: ZADD GT updates existing elements when new scores are greater - skiplist (39 ms) -[ok]: ZADD LT updates existing elements when new scores are lower - skiplist (56 ms) -[ok]: ZADD GT XX updates existing elements when new scores are greater and skips new elements - skiplist (51 ms) -[ok]: ZADD LT XX updates existing elements when new scores are lower and skips new elements - skiplist (56 ms) -[ok]: ZADD XX and NX are not compatible - skiplist (16 ms) -[ok]: ZADD NX with non existing key - skiplist (23 ms) -[ok]: ZADD NX only add new elements without updating old ones - skiplist (56 ms) -[ok]: ZADD GT and NX are not compatible - skiplist (12 ms) -[ok]: ZADD LT and NX are not compatible - skiplist (16 ms) -[ok]: ZADD LT and GT are not compatible - skiplist (16 ms) -[ok]: ZADD INCR LT/GT replies with nill if score not updated - skiplist (47 ms) -[ok]: ZADD INCR LT/GT with inf - skiplist (140 ms) -[ok]: ZADD INCR works like ZINCRBY - skiplist (23 ms) -[ok]: ZADD INCR works with a single score-element pair - skiplist (20 ms) -[ok]: ZADD CH option changes return value to all changed elements - skiplist (31 ms) -[ok]: ZINCRBY calls leading to NaN result in error - skiplist (16 ms) -[ok]: ZINCRBY against invalid incr value - skiplist (24 ms) -[ok]: ZADD - Variadic version base case - skiplist (24 ms) -[ok]: ZADD - Return value is the number of actually added items - skiplist (16 ms) -[ok]: ZADD - Variadic version does not add nothing on single parsing err - skiplist (27 ms) -[ok]: ZADD - Variadic version will raise error on missing arg - skiplist (16 ms) -[ok]: ZINCRBY does not work variadic even if shares ZADD implementation - skiplist (14 ms) -[ok]: ZCARD basics - skiplist (21 ms) -[ok]: ZREM removes key after last element is removed - skiplist (64 ms) -[ok]: ZREM variadic version - skiplist (35 ms) -[ok]: ZREM variadic version -- remove elements after key deletion - skiplist (19 ms) -[ok]: ZRANGE basics - skiplist (148 ms) -[ok]: ZREVRANGE basics - skiplist (150 ms) -[ok]: ZRANK/ZREVRANK basics - skiplist (148 ms) -[ok]: ZRANK - after deletion - skiplist (40 ms) -[ok]: ZINCRBY - can create a new sorted set - skiplist (32 ms) -[ok]: ZINCRBY - increment and decrement - skiplist (71 ms) -[ok]: ZINCRBY return value - skiplist (13 ms) -[ok]: ZRANGEBYSCORE/ZREVRANGEBYSCORE/ZCOUNT basics - skiplist (331 ms) -[ok]: ZRANGEBYSCORE with WITHSCORES - skiplist (80 ms) -[ok]: ZRANGEBYSCORE with LIMIT - skiplist (408 ms) -[ok]: ZRANGEBYSCORE with LIMIT and WITHSCORES - skiplist (80 ms) -[ok]: ZRANGEBYSCORE with non-value min or max - skiplist (24 ms) -[ok]: ZRANGEBYLEX/ZREVRANGEBYLEX/ZLEXCOUNT basics - skiplist (207 ms) -[ok]: ZLEXCOUNT advanced - skiplist (167 ms) -[ok]: ZRANGEBYLEX with LIMIT - skiplist (408 ms) -[ok]: ZRANGEBYLEX with invalid lex range specifiers - skiplist (36 ms) -[ok]: ZREMRANGEBYSCORE basics - skiplist (784 ms) -[ok]: ZREMRANGEBYSCORE with non-value min or max - skiplist (18 ms) -[ok]: ZREMRANGEBYRANK basics - skiplist (396 ms) -[ok]: ZREMRANGEBYLEX basics - skiplist (999 ms) -[ok]: ZUNIONSTORE against non-existing key doesn't set destination - skiplist (15 ms) -[ok]: ZUNION/ZINTER/ZINTERCARD/ZDIFF against non-existing key - skiplist (39 ms) -[ok]: ZUNIONSTORE with empty set - skiplist (39 ms) -[ok]: ZUNION/ZINTER/ZINTERCARD/ZDIFF with empty set - skiplist (49 ms) -[ok]: ZUNIONSTORE basics - skiplist (75 ms) -[ok]: ZUNION/ZINTER/ZINTERCARD/ZDIFF with integer members - skiplist (83 ms) -[ok]: ZUNIONSTORE with weights - skiplist (11 ms) -[ok]: ZUNION with weights - skiplist (15 ms) -[ok]: ZUNIONSTORE with a regular set and weights - skiplist (47 ms) -[ok]: ZUNIONSTORE with AGGREGATE MIN - skiplist (15 ms) -[ok]: ZUNION/ZINTER with AGGREGATE MIN - skiplist (15 ms) -[ok]: ZUNIONSTORE with AGGREGATE MAX - skiplist (16 ms) -[ok]: ZUNION/ZINTER with AGGREGATE MAX - skiplist (15 ms) -[ok]: ZINTERSTORE basics - skiplist (15 ms) -[ok]: ZINTER basics - skiplist (7 ms) -[ok]: ZINTERCARD with illegal arguments (40 ms) -[ok]: ZINTERCARD basics - skiplist (27 ms) -[ok]: ZINTER RESP3 - skiplist (23 ms) -[ok]: ZINTERSTORE with weights - skiplist (19 ms) -[ok]: ZINTER with weights - skiplist (8 ms) -[ok]: ZINTERSTORE with a regular set and weights - skiplist (41 ms) -[ok]: ZINTERSTORE with AGGREGATE MIN - skiplist (10 ms) -[ok]: ZINTERSTORE with AGGREGATE MAX - skiplist (15 ms) -[ok]: ZUNIONSTORE with +inf/-inf scores - skiplist (123 ms) -[ok]: ZUNIONSTORE with NaN weights - skiplist (23 ms) -[ok]: ZINTERSTORE with +inf/-inf scores - skiplist (123 ms) -[ok]: ZINTERSTORE with NaN weights - skiplist (27 ms) -[ok]: ZDIFFSTORE basics - skiplist (19 ms) -[ok]: ZDIFF basics - skiplist (10 ms) -[ok]: ZDIFFSTORE with a regular set - skiplist (56 ms) -[ok]: ZDIFF subtracting set from itself - skiplist (16 ms) -[ok]: ZDIFF algorithm 1 - skiplist (60 ms) -[ok]: ZDIFF algorithm 2 - skiplist (72 ms) -[ok]: ZDIFF fuzzing - skiplist (38178 ms) -[ok]: ZPOPMIN with the count 0 returns an empty array (0 ms) +[ok]: ZMPOP_MIN/ZMPOP_MAX with count - listpack (1 ms) +[ok]: BZPOPMIN/BZPOPMAX with a single existing sorted set - listpack (14 ms) +[ok]: BZPOPMIN/BZPOPMAX with multiple existing sorted sets - listpack (18 ms) +[ok]: BZPOPMIN/BZPOPMAX second sorted set has members - listpack (13 ms) +[ok]: BZMPOP_MIN/BZMPOP_MAX with a single existing sorted set - listpack (15 ms) +[ok]: BZMPOP_MIN/BZMPOP_MAX with multiple existing sorted sets - listpack (14 ms) +[ok]: BZMPOP_MIN/BZMPOP_MAX second sorted set has members - listpack (33 ms) +[ok]: Basic ZPOPMIN/ZPOPMAX - listpack RESP3 (11 ms) +[ok]: ZPOPMIN/ZPOPMAX with count - listpack RESP3 (8 ms) +[ok]: Basic ZMPOP_MIN/ZMPOP_MAX - listpack RESP3 (10 ms) +[ok]: ZMPOP_MIN/ZMPOP_MAX with count - listpack RESP3 (2 ms) +[ok]: BZPOPMIN/BZPOPMAX - listpack RESP3 (31 ms) +[ok]: BZMPOP_MIN/BZMPOP_MAX - listpack RESP3 (47 ms) +[ok]: Check encoding - skiplist (2 ms) +[ok]: ZSET basic ZADD and score update - skiplist (4 ms) +[ok]: ZSET element can't be set to NaN with ZADD - skiplist (1 ms) +[ok]: ZSET element can't be set to NaN with ZINCRBY - skiplist (0 ms) +[ok]: ZADD with options syntax error with incomplete pair - skiplist (1 ms) +[ok]: ZADD XX option without key - skiplist (1 ms) +[ok]: ZADD XX existing key - skiplist (1 ms) +[ok]: ZADD XX returns the number of elements actually added - skiplist (0 ms) +[ok]: ZADD XX updates existing elements score - skiplist (7 ms) +[ok]: ZADD GT updates existing elements when new scores are greater - skiplist (13 ms) +[ok]: ZADD LT updates existing elements when new scores are lower - skiplist (33 ms) +[ok]: ZADD GT XX updates existing elements when new scores are greater and skips new elements - skiplist (11 ms) +[ok]: ZADD LT XX updates existing elements when new scores are lower and skips new elements - skiplist (2 ms) +[ok]: ZADD XX and NX are not compatible - skiplist (0 ms) +[ok]: ZADD NX with non existing key - skiplist (10 ms) +[ok]: ZADD NX only add new elements without updating old ones - skiplist (5 ms) +[ok]: ZADD GT and NX are not compatible - skiplist (6 ms) +[ok]: ZADD LT and NX are not compatible - skiplist (4 ms) +[ok]: ZADD LT and GT are not compatible - skiplist (0 ms) +[ok]: ZADD INCR LT/GT replies with nill if score not updated - skiplist (14 ms) +[ok]: ZADD INCR LT/GT with inf - skiplist (30 ms) +[ok]: ZADD INCR works like ZINCRBY - skiplist (11 ms) +[ok]: ZADD INCR works with a single score-element pair - skiplist (1 ms) +[ok]: ZADD CH option changes return value to all changed elements - skiplist (1 ms) +[ok]: ZINCRBY calls leading to NaN result in error - skiplist (0 ms) +[ok]: ZINCRBY against invalid incr value - skiplist (6 ms) +[ok]: ZADD - Variadic version base case - skiplist (8 ms) +[ok]: ZADD - Return value is the number of actually added items - skiplist (1 ms) +[ok]: ZADD - Variadic version does not add nothing on single parsing err - skiplist (7 ms) +[ok]: ZADD - Variadic version will raise error on missing arg - skiplist (1 ms) +[ok]: ZINCRBY does not work variadic even if shares ZADD implementation - skiplist (0 ms) +[ok]: ZCARD basics - skiplist (5 ms) +[ok]: ZREM removes key after last element is removed - skiplist (5 ms) +[ok]: ZREM variadic version - skiplist (2 ms) +[ok]: ZREM variadic version -- remove elements after key deletion - skiplist (22 ms) +[ok]: ZRANGE basics - skiplist (21 ms) +[ok]: ZREVRANGE basics - skiplist (38 ms) +[ok]: ZRANK/ZREVRANK basics - skiplist (22 ms) +[ok]: ZRANK - after deletion - skiplist (1 ms) +[ok]: ZINCRBY - can create a new sorted set - skiplist (9 ms) +[ok]: ZINCRBY - increment and decrement - skiplist (13 ms) +[ok]: ZINCRBY return value - skiplist (1 ms) +[ok]: ZRANGEBYSCORE/ZREVRANGEBYSCORE/ZCOUNT basics - skiplist (35 ms) +[ok]: ZRANGEBYSCORE with WITHSCORES - skiplist (11 ms) +[ok]: ZRANGEBYSCORE with LIMIT - skiplist (37 ms) +[ok]: ZRANGEBYSCORE with LIMIT and WITHSCORES - skiplist (2 ms) +[ok]: ZRANGEBYSCORE with non-value min or max - skiplist (5 ms) +[ok]: ZRANGEBYLEX/ZREVRANGEBYLEX/ZLEXCOUNT basics - skiplist (10 ms) +[ok]: ZLEXCOUNT advanced - skiplist (10 ms) +[ok]: ZRANGEBYLEX with LIMIT - skiplist (55 ms) +[ok]: ZRANGEBYLEX with invalid lex range specifiers - skiplist (20 ms) +[ok]: ZREMRANGEBYSCORE basics - skiplist (203 ms) +[ok]: ZREMRANGEBYSCORE with non-value min or max - skiplist (7 ms) +[ok]: ZREMRANGEBYRANK basics - skiplist (30 ms) +[ok]: ZREMRANGEBYLEX basics - skiplist (62 ms) +[ok]: ZUNIONSTORE against non-existing key doesn't set destination - skiplist (1 ms) +[ok]: ZUNION/ZINTER/ZINTERCARD/ZDIFF against non-existing key - skiplist (1 ms) +[ok]: ZUNIONSTORE with empty set - skiplist (5 ms) +[ok]: ZUNION/ZINTER/ZINTERCARD/ZDIFF with empty set - skiplist (3 ms) +[ok]: ZUNIONSTORE basics - skiplist (4 ms) +[ok]: ZUNION/ZINTER/ZINTERCARD/ZDIFF with integer members - skiplist (5 ms) +[ok]: ZUNIONSTORE with weights - skiplist (2 ms) +[ok]: ZUNION with weights - skiplist (1 ms) +[ok]: ZUNIONSTORE with a regular set and weights - skiplist (6 ms) +[ok]: ZUNIONSTORE with AGGREGATE MIN - skiplist (0 ms) +[ok]: ZUNION/ZINTER with AGGREGATE MIN - skiplist (5 ms) +[ok]: ZUNIONSTORE with AGGREGATE MAX - skiplist (0 ms) +[ok]: ZUNION/ZINTER with AGGREGATE MAX - skiplist (13 ms) +[ok]: ZINTERSTORE basics - skiplist (0 ms) +[ok]: ZINTER basics - skiplist (0 ms) +[ok]: ZINTERCARD with illegal arguments (1 ms) +[ok]: ZINTERCARD basics - skiplist (1 ms) +[ok]: ZINTER RESP3 - skiplist (8 ms) +[ok]: ZINTERSTORE with weights - skiplist (1 ms) +[ok]: ZINTER with weights - skiplist (1 ms) +[ok]: ZINTERSTORE with a regular set and weights - skiplist (7 ms) +[ok]: ZINTERSTORE with AGGREGATE MIN - skiplist (1 ms) +[ok]: ZINTERSTORE with AGGREGATE MAX - skiplist (0 ms) +[ok]: ZUNIONSTORE with +inf/-inf scores - skiplist (35 ms) +[ok]: ZUNIONSTORE with NaN weights - skiplist (7 ms) +[ok]: ZINTERSTORE with +inf/-inf scores - skiplist (16 ms) +[ok]: ZINTERSTORE with NaN weights - skiplist (2 ms) +[ok]: ZDIFFSTORE basics - skiplist (1 ms) +[ok]: ZDIFF basics - skiplist (6 ms) +[ok]: ZDIFFSTORE with a regular set - skiplist (2 ms) +[ok]: ZDIFF subtracting set from itself - skiplist (0 ms) +[ok]: ZDIFF algorithm 1 - skiplist (17 ms) +[ok]: ZDIFF algorithm 2 - skiplist (10 ms) +[ok]: ZDIFF fuzzing - skiplist (12880 ms) +[ok]: ZPOPMIN with the count 0 returns an empty array (2 ms) [ok]: ZPOPMIN with negative count (1 ms) [ok]: ZPOPMAX with the count 0 returns an empty array (1 ms) [ok]: ZPOPMAX with negative count (1 ms) -[ok]: Basic ZPOPMIN/ZPOPMAX with a single key - skiplist (2 ms) -[ok]: ZPOPMIN/ZPOPMAX with count - skiplist (4 ms) -[ok]: Basic ZMPOP_MIN/ZMPOP_MAX with a single key - skiplist (6 ms) -[ok]: ZMPOP_MIN/ZMPOP_MAX with count - skiplist (6 ms) -[ok]: BZPOPMIN/BZPOPMAX with a single existing sorted set - skiplist (36 ms) -[ok]: BZPOPMIN/BZPOPMAX with multiple existing sorted sets - skiplist (36 ms) -[ok]: BZPOPMIN/BZPOPMAX second sorted set has members - skiplist (45 ms) -[ok]: BZMPOP_MIN/BZMPOP_MAX with a single existing sorted set - skiplist (40 ms) -[ok]: BZMPOP_MIN/BZMPOP_MAX with multiple existing sorted sets - skiplist (34 ms) -[ok]: BZMPOP_MIN/BZMPOP_MAX second sorted set has members - skiplist (38 ms) -[ok]: Basic ZPOPMIN/ZPOPMAX - skiplist RESP3 (2 ms) -[ok]: ZPOPMIN/ZPOPMAX with count - skiplist RESP3 (8 ms) -[ok]: Basic ZMPOP_MIN/ZMPOP_MAX - skiplist RESP3 (3 ms) -[ok]: ZMPOP_MIN/ZMPOP_MAX with count - skiplist RESP3 (8 ms) -[ok]: BZPOPMIN/BZPOPMAX - skiplist RESP3 (36 ms) -[ok]: BZMPOP_MIN/BZMPOP_MAX - skiplist RESP3 (43 ms) -[ok]: ZPOP/ZMPOP against wrong type (2 ms) -[ok]: ZMPOP with illegal argument (9 ms) -[ok]: ZMPOP propagate as pop with count command to replica (180 ms) -[ok]: ZPOPMIN/ZPOPMAX readraw in RESP3 (16 ms) -[ok]: BZPOPMIN/BZPOPMAX readraw in RESP3 (143 ms) +[ok]: Basic ZPOPMIN/ZPOPMAX with a single key - skiplist (3 ms) +[ok]: ZPOPMIN/ZPOPMAX with count - skiplist (2 ms) +[ok]: Basic ZMPOP_MIN/ZMPOP_MAX with a single key - skiplist (3 ms) +[ok]: ZMPOP_MIN/ZMPOP_MAX with count - skiplist (2 ms) +[ok]: BZPOPMIN/BZPOPMAX with a single existing sorted set - skiplist (9 ms) +[ok]: BZPOPMIN/BZPOPMAX with multiple existing sorted sets - skiplist (11 ms) +[ok]: BZPOPMIN/BZPOPMAX second sorted set has members - skiplist (10 ms) +[ok]: BZMPOP_MIN/BZMPOP_MAX with a single existing sorted set - skiplist (10 ms) +[ok]: BZMPOP_MIN/BZMPOP_MAX with multiple existing sorted sets - skiplist (11 ms) +[ok]: BZMPOP_MIN/BZMPOP_MAX second sorted set has members - skiplist (9 ms) +[ok]: Basic ZPOPMIN/ZPOPMAX - skiplist RESP3 (3 ms) +[ok]: ZPOPMIN/ZPOPMAX with count - skiplist RESP3 (2 ms) +[ok]: Basic ZMPOP_MIN/ZMPOP_MAX - skiplist RESP3 (2 ms) +[ok]: ZMPOP_MIN/ZMPOP_MAX with count - skiplist RESP3 (2 ms) +[ok]: BZPOPMIN/BZPOPMAX - skiplist RESP3 (10 ms) +[ok]: BZMPOP_MIN/BZMPOP_MAX - skiplist RESP3 (11 ms) +[ok]: ZPOP/ZMPOP against wrong type (4 ms) +[ok]: ZMPOP with illegal argument (4 ms) +[ok]: ZMPOP propagate as pop with count command to replica (145 ms) +[ok]: ZPOPMIN/ZPOPMAX readraw in RESP3 (3 ms) +[ok]: BZPOPMIN/BZPOPMAX readraw in RESP3 (1183 ms) [ok]: ZMPOP readraw in RESP3 (3 ms) -[ok]: BZMPOP readraw in RESP3 (203 ms) -[ok]: ZPOPMIN/ZPOPMAX readraw in RESP2 (16 ms) -[ok]: BZPOPMIN/BZPOPMAX readraw in RESP2 (151 ms) -[ok]: ZMPOP readraw in RESP2 (12 ms) -[ok]: BZMPOP readraw in RESP2 (187 ms) -[ok]: ZINTERSTORE regression with two sets, intset+hashtable (11 ms) -[ok]: ZUNIONSTORE regression, should not create NaN in scores (2 ms) -[ok]: ZINTERSTORE #516 regression, mixed sets and ziplist zsets (1 ms) -[ok]: ZUNIONSTORE result is sorted (305 ms) -[ok]: ZUNIONSTORE/ZINTERSTORE/ZDIFFSTORE error if using WITHSCORES (14 ms) +[ok]: BZMPOP readraw in RESP3 (205 ms) +[ok]: ZPOPMIN/ZPOPMAX readraw in RESP2 (2 ms) +[ok]: BZPOPMIN/BZPOPMAX readraw in RESP2 (187 ms) +[ok]: ZMPOP readraw in RESP2 (10 ms) +[ok]: BZMPOP readraw in RESP2 (194 ms) +[ok]: ZINTERSTORE regression with two sets, intset+hashtable (9 ms) +[ok]: ZUNIONSTORE regression, should not create NaN in scores (3 ms) +[ok]: ZINTERSTORE #516 regression, mixed sets and ziplist zsets (6 ms) +[ok]: ZUNIONSTORE result is sorted (148 ms) +[ok]: ZUNIONSTORE/ZINTERSTORE/ZDIFFSTORE error if using WITHSCORES (1 ms) [ok]: ZMSCORE retrieve (1 ms) -[ok]: ZMSCORE retrieve from empty set (1 ms) -[ok]: ZMSCORE retrieve with missing member (1 ms) -[ok]: ZMSCORE retrieve single member (6 ms) -[ok]: ZMSCORE retrieve requires one or more members (7 ms) +[ok]: ZMSCORE retrieve from empty set (0 ms) +[ok]: ZMSCORE retrieve with missing member (0 ms) +[ok]: ZMSCORE retrieve single member (2 ms) +[ok]: ZMSCORE retrieve requires one or more members (1 ms) [ok]: ZSET commands don't accept the empty strings as valid score (1 ms) -[ok]: zunionInterDiffGenericCommand at least 1 input key (8 ms) -[ok]: ZSCORE - listpack (114 ms) -[ok]: ZMSCORE - listpack (71 ms) -[ok]: ZSCORE after a DEBUG RELOAD - listpack (113 ms) -[ok]: ZSET sorting stresser - listpack (89 ms) -[ok]: ZRANGEBYSCORE fuzzy test, 100 ranges in 128 element sorted set - listpack (19200 ms) -[ok]: ZRANGEBYLEX fuzzy test, 100 ranges in 128 element sorted set - listpack (581 ms) -[ok]: ZREMRANGEBYLEX fuzzy test, 100 ranges in 128 element sorted set - listpack (1026 ms) -[ok]: ZSETs skiplist implementation backlink consistency test - listpack (142 ms) -[ok]: ZSETs ZRANK augmented skip list stress testing - listpack (7363 ms) -[ok]: BZPOPMIN, ZADD + DEL should not awake blocked client (83 ms) -[ok]: BZPOPMIN, ZADD + DEL + SET should not awake blocked client (112 ms) -[ok]: BZMPOP_MIN, ZADD + DEL should not awake blocked client (100 ms) -[ok]: BZMPOP_MIN, ZADD + DEL + SET should not awake blocked client (104 ms) -Time: 1063 -[ok]: BZPOPMIN unblock but the key is expired and then block again - reprocessing command (1120 ms) -[ok]: BZPOPMIN with same key multiple times should work (106 ms) -[ok]: MULTI/EXEC is isolated from the point of view of BZPOPMIN (61 ms) -[ok]: BZPOPMIN with variadic ZADD (51 ms) -[ok]: BZPOPMIN with zero timeout should block indefinitely (1047 ms) -[ok]: MULTI/EXEC is isolated from the point of view of BZMPOP_MIN (101 ms) -[ok]: BZMPOP_MIN with variadic ZADD (63 ms) -[ok]: BZMPOP_MIN with zero timeout should block indefinitely (1070 ms) -[ok]: ZSCORE - skiplist (1412 ms) -[ok]: ZMSCORE - skiplist (1444 ms) -[ok]: ZSCORE after a DEBUG RELOAD - skiplist (1619 ms) -[ok]: ZSET sorting stresser - skiplist (1761 ms) -[ok]: ZRANGEBYSCORE fuzzy test, 100 ranges in 100 element sorted set - skiplist (19577 ms) -[ok]: ZRANGEBYLEX fuzzy test, 100 ranges in 100 element sorted set - skiplist (964 ms) -[ok]: ZREMRANGEBYLEX fuzzy test, 100 ranges in 100 element sorted set - skiplist (1414 ms) -[ok]: ZSETs skiplist implementation backlink consistency test - skiplist (137 ms) -[ok]: ZSETs ZRANK augmented skip list stress testing - skiplist (5465 ms) -[ok]: BZPOPMIN, ZADD + DEL should not awake blocked client (48 ms) -[ok]: BZPOPMIN, ZADD + DEL + SET should not awake blocked client (53 ms) -[ok]: BZMPOP_MIN, ZADD + DEL should not awake blocked client (52 ms) -[ok]: BZMPOP_MIN, ZADD + DEL + SET should not awake blocked client (44 ms) -Time: 1037 -[ok]: BZPOPMIN unblock but the key is expired and then block again - reprocessing command (1092 ms) -[ok]: BZPOPMIN with same key multiple times should work (49 ms) -[ok]: MULTI/EXEC is isolated from the point of view of BZPOPMIN (46 ms) -[ok]: BZPOPMIN with variadic ZADD (53 ms) -[ok]: BZPOPMIN with zero timeout should block indefinitely (1050 ms) -[ok]: MULTI/EXEC is isolated from the point of view of BZMPOP_MIN (73 ms) -[ok]: BZMPOP_MIN with variadic ZADD (69 ms) -[ok]: BZMPOP_MIN with zero timeout should block indefinitely (1071 ms) -[ok]: BZPOP/BZMPOP against wrong type (22 ms) -[ok]: BZMPOP with illegal argument (10 ms) -[ok]: BZMPOP with multiple blocked clients (315 ms) -[ok]: BZMPOP propagate as pop with count command to replica (358 ms) -[ok]: BZMPOP should not blocks on non key arguments - #10762 (157 ms) -[ok]: ZSET skiplist order consistency when elements are moved (9754 ms) -[ok]: ZRANGESTORE basic (2 ms) +[ok]: zunionInterDiffGenericCommand at least 1 input key (5 ms) +[ok]: ZSCORE - listpack (52 ms) +[ok]: ZMSCORE - listpack (201 ms) +[ok]: ZSCORE after a DEBUG RELOAD - listpack (141 ms) +[ok]: ZSET sorting stresser - listpack (93 ms) +[ok]: ZRANGEBYSCORE fuzzy test, 100 ranges in 128 element sorted set - listpack (10203 ms) +[ok]: ZRANGEBYLEX fuzzy test, 100 ranges in 128 element sorted set - listpack (271 ms) +[ok]: ZREMRANGEBYLEX fuzzy test, 100 ranges in 128 element sorted set - listpack (663 ms) +[ok]: ZSETs skiplist implementation backlink consistency test - listpack (73 ms) +[ok]: ZSETs ZRANK augmented skip list stress testing - listpack (1969 ms) +[ok]: BZPOPMIN, ZADD + DEL should not awake blocked client (12 ms) +[ok]: BZPOPMIN, ZADD + DEL + SET should not awake blocked client (121 ms) +[ok]: BZMPOP_MIN, ZADD + DEL should not awake blocked client (133 ms) +[ok]: BZMPOP_MIN, ZADD + DEL + SET should not awake blocked client (29 ms) +Time: 1013 +[ok]: BZPOPMIN unblock but the key is expired and then block again - reprocessing command (1040 ms) +[ok]: BZPOPMIN with same key multiple times should work (13 ms) +[ok]: MULTI/EXEC is isolated from the point of view of BZPOPMIN (18 ms) +[ok]: BZPOPMIN with variadic ZADD (9 ms) +[ok]: BZPOPMIN with zero timeout should block indefinitely (1754 ms) +[ok]: MULTI/EXEC is isolated from the point of view of BZMPOP_MIN (120 ms) +[ok]: BZMPOP_MIN with variadic ZADD (30 ms) +[ok]: BZMPOP_MIN with zero timeout should block indefinitely (1032 ms) +[ok]: ZSCORE - skiplist (110 ms) +[ok]: ZMSCORE - skiplist (115 ms) +[ok]: ZSCORE after a DEBUG RELOAD - skiplist (147 ms) +[ok]: ZSET sorting stresser - skiplist (75 ms) +[ok]: ZRANGEBYSCORE fuzzy test, 100 ranges in 100 element sorted set - skiplist (8740 ms) +[ok]: ZRANGEBYLEX fuzzy test, 100 ranges in 100 element sorted set - skiplist (391 ms) +[ok]: ZREMRANGEBYLEX fuzzy test, 100 ranges in 100 element sorted set - skiplist (559 ms) +[ok]: ZSETs skiplist implementation backlink consistency test - skiplist (53 ms) +[ok]: ZSETs ZRANK augmented skip list stress testing - skiplist (2403 ms) +[ok]: BZPOPMIN, ZADD + DEL should not awake blocked client (123 ms) +[ok]: BZPOPMIN, ZADD + DEL + SET should not awake blocked client (14 ms) +[ok]: BZMPOP_MIN, ZADD + DEL should not awake blocked client (13 ms) +[ok]: BZMPOP_MIN, ZADD + DEL + SET should not awake blocked client (14 ms) +Time: 1015 +[ok]: BZPOPMIN unblock but the key is expired and then block again - reprocessing command (1033 ms) +[ok]: BZPOPMIN with same key multiple times should work (26 ms) +[ok]: MULTI/EXEC is isolated from the point of view of BZPOPMIN (17 ms) +[ok]: BZPOPMIN with variadic ZADD (14 ms) +[ok]: BZPOPMIN with zero timeout should block indefinitely (1017 ms) +[ok]: MULTI/EXEC is isolated from the point of view of BZMPOP_MIN (14 ms) +[ok]: BZMPOP_MIN with variadic ZADD (25 ms) +[ok]: BZMPOP_MIN with zero timeout should block indefinitely (1016 ms) +[ok]: BZPOP/BZMPOP against wrong type (8 ms) +[ok]: BZMPOP with illegal argument (25 ms) +[ok]: BZMPOP with multiple blocked clients (1226 ms) +[ok]: BZMPOP propagate as pop with count command to replica (373 ms) +[ok]: BZMPOP should not blocks on non key arguments - #10762 (25 ms) +[ok]: ZSET skiplist order consistency when elements are moved (2546 ms) +[ok]: ZRANGESTORE basic (1 ms) [ok]: ZRANGESTORE RESP3 (1 ms) [ok]: ZRANGESTORE range (0 ms) [ok]: ZRANGESTORE BYLEX (1 ms) -[ok]: ZRANGESTORE BYSCORE (6 ms) +[ok]: ZRANGESTORE BYSCORE (2 ms) [ok]: ZRANGESTORE BYSCORE LIMIT (1 ms) -[ok]: ZRANGESTORE BYSCORE REV LIMIT (1 ms) +[ok]: ZRANGESTORE BYSCORE REV LIMIT (0 ms) [ok]: ZRANGE BYSCORE REV LIMIT (0 ms) -[ok]: ZRANGESTORE - src key missing (1 ms) -[ok]: ZRANGESTORE - src key wrong type (1 ms) -[ok]: ZRANGESTORE - empty range (1 ms) -[ok]: ZRANGESTORE BYLEX - empty range (1 ms) -[ok]: ZRANGESTORE BYSCORE - empty range (1 ms) +[ok]: ZRANGESTORE - src key missing (0 ms) +[ok]: ZRANGESTORE - src key wrong type (0 ms) +[ok]: ZRANGESTORE - empty range (0 ms) +[ok]: ZRANGESTORE BYLEX - empty range (0 ms) +[ok]: ZRANGESTORE BYSCORE - empty range (0 ms) [ok]: ZRANGE BYLEX (0 ms) [ok]: ZRANGESTORE invalid syntax (1 ms) -[ok]: ZRANGESTORE with zset-max-listpack-entries 0 #10767 case (8 ms) +[ok]: ZRANGESTORE with zset-max-listpack-entries 0 #10767 case (2 ms) [ok]: ZRANGESTORE with zset-max-listpack-entries 1 dst key should use skiplist encoding (2 ms) [ok]: ZRANGE invalid syntax (1 ms) -[ok]: ZRANDMEMBER - listpack (22 ms) -[ok]: ZRANDMEMBER - skiplist (29 ms) +[ok]: ZRANDMEMBER - listpack (20 ms) +[ok]: ZRANDMEMBER - skiplist (33 ms) [ok]: ZRANDMEMBER with RESP3 (1 ms) -[ok]: ZRANDMEMBER count of 0 is handled correctly (7 ms) -[ok]: ZRANDMEMBER with against non existing key (1 ms) +[ok]: ZRANDMEMBER count of 0 is handled correctly (1 ms) +[ok]: ZRANDMEMBER with against non existing key (0 ms) [ok]: ZRANDMEMBER count overflow (1 ms) [ok]: ZRANDMEMBER count of 0 is handled correctly - emptyarray (0 ms) -[ok]: ZRANDMEMBER with against non existing key - emptyarray (1 ms) -[ok]: ZRANDMEMBER with - skiplist (672 ms) -[ok]: ZRANDMEMBER with - listpack (858 ms) +[ok]: ZRANDMEMBER with against non existing key - emptyarray (0 ms) +[ok]: ZRANDMEMBER with - skiplist (596 ms) +[ok]: ZRANDMEMBER with - listpack (794 ms) [ok]: zset score double range (1 ms) -[ok]: zunionInterDiffGenericCommand acts on SET and ZSET (54 ms) -[ok]: ZADD overflows the maximum allowed elements in a listpack - single (38 ms) -[ok]: ZADD overflows the maximum allowed elements in a listpack - multiple (11 ms) -[ok]: ZADD overflows the maximum allowed elements in a listpack - single_multiple (4 ms) -[51/124 done]: unit/type/zset (164 seconds) -Testing unit/type/stream-cgroups -=== (stream) Starting server 127.0.0.1:21541 ok -[ok]: XGROUP CREATE: creation and duplicate group name detection (26 ms) -[ok]: XGROUP CREATE: with ENTRIESREAD parameter (64 ms) -[ok]: XGROUP CREATE: automatic stream creation fails without MKSTREAM (19 ms) -[ok]: XGROUP CREATE: automatic stream creation works with MKSTREAM (11 ms) -[ok]: XREADGROUP will return only new elements (20 ms) -[ok]: XREADGROUP can read the history of the elements we own (28 ms) -[ok]: XPENDING is able to return pending items (8 ms) -[ok]: XPENDING can return single consumer items (8 ms) -[ok]: XPENDING only group (8 ms) -[ok]: XPENDING with IDLE (52 ms) -[ok]: XPENDING with exclusive range intervals works as expected (16 ms) -[ok]: XACK is able to remove items from the consumer/group PEL (32 ms) -[ok]: XACK can't remove the same item multiple times (8 ms) -[ok]: XACK is able to accept multiple arguments (8 ms) -[ok]: XACK should fail if got at least one invalid ID (43 ms) -[ok]: PEL NACK reassignment after XGROUP SETID event (62 ms) -[ok]: XREADGROUP will not report data on empty history. Bug #5577 (55 ms) -[ok]: XREADGROUP history reporting of deleted entries. Bug #5570 (43 ms) -[ok]: Blocking XREADGROUP will not reply with an empty array (116 ms) -[ok]: Blocking XREADGROUP: key deleted (72 ms) -[ok]: Blocking XREADGROUP: key type changed with SET (60 ms) -[ok]: Blocking XREADGROUP: key type changed with transaction (87 ms) -[ok]: Blocking XREADGROUP: flushed DB (76 ms) -[ok]: Blocking XREADGROUP: swapped DB, key doesn't exist (115 ms) -[ok]: Blocking XREADGROUP: swapped DB, key is not a stream (111 ms) -[ok]: XREAD and XREADGROUP against wrong parameter (35 ms) -[ok]: Blocking XREAD: key deleted (89 ms) -[ok]: Blocking XREAD: key type changed with SET (79 ms) -[ok]: Blocking XREADGROUP for stream that ran dry (issue #5299) (184 ms) -[ok]: Blocking XREADGROUP will ignore BLOCK if ID is not > (231 ms) -[ok]: Blocking XREADGROUP for stream key that has clients blocked on list (180 ms) -[ok]: Blocking XREADGROUP for stream key that has clients blocked on stream - avoid endless loop (142 ms) -[ok]: Blocking XREADGROUP for stream key that has clients blocked on stream - reprocessing command (1195 ms) -[ok]: XGROUP DESTROY should unblock XREADGROUP with -NOGROUP (85 ms) -[ok]: RENAME can unblock XREADGROUP with data (64 ms) -[ok]: RENAME can unblock XREADGROUP with -NOGROUP (52 ms) -[ok]: XCLAIM can claim PEL items from another consumer (710 ms) -[ok]: XCLAIM without JUSTID increments delivery count (484 ms) -[ok]: XCLAIM same consumer (260 ms) -[ok]: XAUTOCLAIM can claim PEL items from another consumer (696 ms) -[ok]: XAUTOCLAIM as an iterator (272 ms) -[ok]: XAUTOCLAIM COUNT must be > 0 (8 ms) -[ok]: XCLAIM with XDEL (53 ms) -[ok]: XCLAIM with trimming (62 ms) -[ok]: XAUTOCLAIM with XDEL (68 ms) -[ok]: XAUTOCLAIM with XDEL and count (84 ms) -[ok]: XAUTOCLAIM with out of range count (7 ms) -[ok]: XCLAIM with trimming (76 ms) -[ok]: XINFO FULL output (109 ms) -[ok]: Consumer seen-time and active-time (304 ms) -[ok]: XGROUP CREATECONSUMER: create consumer if does not exist (78 ms) -[ok]: XGROUP CREATECONSUMER: group must exist (19 ms) -=== (stream needs:debug) Starting server 127.0.0.1:21543 ok -[ok]: XREADGROUP with NOACK creates consumer (244 ms) - -Waiting for background AOF rewrite to finish... -Waiting for background AOF rewrite to finish... [ok]: Consumer without PEL is present in AOF after AOFRW (367 ms) -[ok]: Consumer group read counter and lag in empty streams (40 ms) -[ok]: Consumer group read counter and lag sanity (89 ms) -[ok]: Consumer group lag with XDELs (159 ms) -[ok]: Loading from legacy (Redis <= v6.2.x, rdb_ver < 10) persistence (12 ms) -[ok]: Loading from legacy (Redis <= v7.0.x, rdb_ver < 11) persistence (19 ms) -=== (external:skip) Starting server 127.0.0.1:21545 ok -[ok]: Consumer group last ID propagation to slave (NOACK=0) (208 ms) -[ok]: Consumer group last ID propagation to slave (NOACK=1) (200 ms) -=== (external:skip) Starting server 127.0.0.1:21547 ok -[ok]: Replication tests of XCLAIM with deleted entries (autoclaim=0) (918 ms) -[ok]: Replication tests of XCLAIM with deleted entries (autoclaim=1) (135 ms) -[ok]: XREADGROUP ACK would propagate entries-read (88 ms) -=== (stream needs:debug) Starting server 127.0.0.1:21549 ok - -Waiting for background AOF rewrite to finish... [ok]: Empty stream with no lastid can be rewrite into AOF correctly (177 ms) -[52/124 done]: unit/type/stream-cgroups (12 seconds) -Testing unit/type/list-3 -=== (list ziplist) Starting server 127.0.0.1:21551 ok -[ok]: Explicit regression for a list bug (39 ms) -[ok]: Regression for quicklist #3343 bug (198 ms) -[ok]: Check compression with recompress (124 ms) -[ok]: Crash due to wrongly recompress after lrem (296 ms) -[ok]: LINSERT correctly recompress full quicklistNode after inserting a element before it (110 ms) -[ok]: LINSERT correctly recompress full quicklistNode after inserting a element after it (124 ms) -[ok]: Stress tester for #3343-alike bugs comp: 2 (10947 ms) -[ok]: Stress tester for #3343-alike bugs comp: 1 (9347 ms) -[ok]: Stress tester for #3343-alike bugs comp: 0 (13517 ms) -[ok]: ziplist implementation: value encoding and backlink (17636 ms) -[ok]: ziplist implementation: encoding stress testing (91679 ms) -[53/124 done]: unit/type/list-3 (145 seconds) -Testing unit/type/stream -=== (stream) Starting server 127.0.0.1:21553 ok -[ok]: XADD wrong number of args (23 ms) -[ok]: XADD can add entries into a stream that XRANGE can fetch (24 ms) -[ok]: XADD IDs are incremental (23 ms) -[ok]: XADD IDs are incremental when ms is the same as well (29 ms) -[ok]: XADD IDs correctly report an error when overflowing (19 ms) -[ok]: XADD auto-generated sequence is incremented for last ID (17 ms) -[ok]: XADD auto-generated sequence is zero for future timestamp ID (19 ms) -[ok]: XADD auto-generated sequence can't be smaller than last ID (24 ms) -[ok]: XADD auto-generated sequence can't overflow (24 ms) -[ok]: XADD 0-* should succeed (12 ms) -[ok]: XADD with MAXLEN option (1529 ms) -[ok]: XADD with MAXLEN option and the '=' argument (365 ms) -[ok]: XADD with MAXLEN option and the '~' argument (355 ms) -[ok]: XADD with NOMKSTREAM option (7 ms) -[ok]: XADD with MINID option (355 ms) -[ok]: XTRIM with MINID option (3 ms) -[ok]: XTRIM with MINID option, big delta from master record (1 ms) -[ok]: XADD mass insertion and XLEN (8691 ms) -[ok]: XADD with ID 0-0 (7 ms) -[ok]: XADD with LIMIT delete entries no more than limit (1 ms) -[ok]: XRANGE COUNT works as expected (1 ms) -[ok]: XREVRANGE COUNT works as expected (9 ms) -[ok]: XRANGE can be used to iterate the whole stream (2501 ms) -[ok]: XREVRANGE returns the reverse of XRANGE (5964 ms) -[ok]: XRANGE exclusive ranges (149 ms) -[ok]: XREAD with non empty stream (7 ms) -[ok]: Non blocking XREAD with empty streams (7 ms) -[ok]: XREAD with non empty second stream (12656 ms) -[ok]: Blocking XREAD waiting new data (56 ms) -[ok]: Blocking XREAD waiting old data (44 ms) -[ok]: Blocking XREAD will not reply with an empty array (167 ms) -[ok]: Blocking XREAD for stream that ran dry (issue #5299) (128 ms) -[ok]: XREAD last element from non-empty stream (63 ms) -[ok]: XREAD last element from empty stream (35 ms) -[ok]: XREAD last element blocking from empty stream (73 ms) -[ok]: XREAD last element blocking from non-empty stream (35 ms) -[ok]: XREAD last element from multiple streams (68 ms) -[ok]: XREAD last element with count > 1 (31 ms) -[ok]: XREAD: XADD + DEL should not awake client (97 ms) -[ok]: XREAD: XADD + DEL + LPUSH should not awake client (115 ms) -[ok]: XREAD with same stream name multiple times should work (63 ms) -[ok]: XREAD + multiple XADD inside transaction (104 ms) -[ok]: XDEL basic test (41 ms) -[ok]: XDEL multiply id test (62 ms) -[ok]: XDEL fuzz test (25941 ms) -[ok]: XRANGE fuzzing (38292 ms) -[ok]: XREVRANGE regression test for issue #5006 (7 ms) -[ok]: XREAD streamID edge (no-blocking) (8 ms) -[ok]: XREAD streamID edge (blocking) (43 ms) -[ok]: XADD streamID edge (2 ms) -[ok]: XTRIM with MAXLEN option basic test (601 ms) -[ok]: XADD with LIMIT consecutive calls (75 ms) -[ok]: XTRIM with ~ is limited (71 ms) -[ok]: XTRIM without ~ is not limited (75 ms) -[ok]: XTRIM without ~ and with LIMIT (77 ms) -[ok]: XTRIM with LIMIT delete entries no more than limit (3 ms) -=== (stream needs:debug) Starting server 127.0.0.1:21555 ok -[ok]: XADD with MAXLEN > xlen can propagate correctly (773 ms) -=== (stream needs:debug) Starting server 127.0.0.1:21557 ok -[ok]: XADD with MINID > lastid can propagate correctly (859 ms) -=== (stream needs:debug) Starting server 127.0.0.1:21559 ok -[ok]: XADD with ~ MAXLEN can propagate correctly (804 ms) -=== (stream needs:debug) Starting server 127.0.0.1:21561 ok -[ok]: XADD with ~ MAXLEN and LIMIT can propagate correctly (855 ms) -=== (stream needs:debug) Starting server 127.0.0.1:21563 ok -[ok]: XADD with ~ MINID can propagate correctly (875 ms) -=== (stream needs:debug) Starting server 127.0.0.1:21565 ok -[ok]: XADD with ~ MINID and LIMIT can propagate correctly (816 ms) -=== (stream needs:debug) Starting server 127.0.0.1:21567 ok -[ok]: XTRIM with ~ MAXLEN can propagate correctly (748 ms) -=== (stream) Starting server 127.0.0.1:21569 ok -[ok]: XADD can CREATE an empty stream (12 ms) -[ok]: XSETID can set a specific ID (18 ms) -[ok]: XSETID cannot SETID with smaller ID (24 ms) -[ok]: XSETID cannot SETID on non-existent key (1 ms) -[ok]: XSETID cannot run with an offset but without a maximal tombstone (11 ms) -[ok]: XSETID cannot run with a maximal tombstone but without an offset (7 ms) -[ok]: XSETID errors on negstive offset (8 ms) -[ok]: XSETID cannot set the maximal tombstone with larger ID (27 ms) -[ok]: XSETID cannot set the offset to less than the length (19 ms) -[ok]: XSETID cannot set smaller ID than current MAXDELETEDID (52 ms) -=== (stream) Starting server 127.0.0.1:21571 ok -[ok]: XADD advances the entries-added counter and sets the recorded-first-entry-id (40 ms) -[ok]: XDEL/TRIM are reflected by recorded first entry (100 ms) -[ok]: Maximum XDEL ID behaves correctly (68 ms) -[ok]: XADD with artial ID with maximal seq (24 ms) -=== (stream needs:debug) Starting server 127.0.0.1:21573 ok - -Waiting for background AOF rewrite to finish... -Waiting for background AOF rewrite to finish... [ok]: Empty stream can be rewrite into AOF correctly (196 ms) - -Waiting for background AOF rewrite to finish... -Waiting for background AOF rewrite to finish... [ok]: Stream can be rewrite into AOF correctly after XDEL lastid (256 ms) -=== (stream) Starting server 127.0.0.1:21575 ok -[ok]: XGROUP HELP should not have unexpected options (11 ms) -[ok]: XINFO HELP should not have unexpected options (7 ms) -[54/124 done]: unit/type/stream (113 seconds) -Testing unit/cluster/slot-stats +[ok]: zunionInterDiffGenericCommand acts on SET and ZSET (29 ms) +[ok]: ZADD overflows the maximum allowed elements in a listpack - single (16 ms) +[ok]: ZADD overflows the maximum allowed elements in a listpack - multiple (3 ms) +[ok]: ZADD overflows the maximum allowed elements in a listpack - single_multiple (3 ms) +[53/124 done]: unit/type/zset (69 seconds) +Testing unit/type/incr +=== (incr) Starting server 127.0.0.1:21575 ok +[ok]: INCR against non existing key (1 ms) +[ok]: INCR against key created by incr itself (0 ms) +[ok]: DECR against key created by incr (0 ms) +[ok]: DECR against key is not exist and incr (0 ms) +[ok]: INCR against key originally set with SET (1 ms) +[ok]: INCR over 32bit value (0 ms) +[ok]: INCRBY over 32bit value with over 32bit increment (1 ms) +[ok]: INCR fails against key with spaces (left) (0 ms) +[ok]: INCR fails against key with spaces (right) (1 ms) +[ok]: INCR fails against key with spaces (both) (0 ms) +[ok]: DECRBY negation overflow (0 ms) +[ok]: INCR fails against a key holding a list (1 ms) +[ok]: DECRBY over 32bit value with over 32bit increment, negative res (0 ms) +[ok]: DECRBY against key is not exist (0 ms) +[ok]: INCR uses shared objects in the 0-9999 range (2 ms) +[ok]: INCR can modify objects in-place (1 ms) +[ok]: INCRBYFLOAT against non existing key (1 ms) +[ok]: INCRBYFLOAT against key originally set with SET (0 ms) +[ok]: INCRBYFLOAT over 32bit value (1 ms) +[ok]: INCRBYFLOAT over 32bit value with over 32bit increment (0 ms) +[ok]: INCRBYFLOAT fails against key with spaces (left) (1 ms) +[ok]: INCRBYFLOAT fails against key with spaces (right) (0 ms) +[ok]: INCRBYFLOAT fails against key with spaces (both) (1 ms) +[ok]: INCRBYFLOAT fails against a key holding a list (1 ms) +[ok]: INCRBYFLOAT does not allow NaN or Infinity (1 ms) +[ok]: INCRBYFLOAT decrement (0 ms) +[ok]: string to double with null terminator (1 ms) +[ok]: No negative zero (1 ms) +[ok]: INCRBY INCRBYFLOAT DECRBY against unhappy path (2 ms) +[ok]: incr operation should update encoding from raw to int (2 ms) +[ok]: decr operation should update encoding from raw to int (2 ms) +[ok]: incrby operation should update encoding from raw to int (1 ms) +[ok]: decrby operation should update encoding from raw to int (1 ms) +[54/124 done]: unit/type/incr (0 seconds) +Testing unit/cluster/slave-stop-cond === (external:skip cluster) Starting server 127.0.0.1:21577 ok -[ok]: CLUSTER SLOT-STATS cpu-usec reset upon CONFIG RESETSTAT. (6300 ms) -[ok]: CLUSTER SLOT-STATS cpu-usec reset upon slot migration. (22476 ms) -[ok]: CLUSTER SLOT-STATS cpu-usec for non-slot specific commands. (8585 ms) -[ok]: CLUSTER SLOT-STATS cpu-usec for slot specific commands. (8652 ms) -[ok]: CLUSTER SLOT-STATS cpu-usec for blocking commands, unblocked on keyspace update. (21698 ms) -[ok]: CLUSTER SLOT-STATS cpu-usec for blocking commands, unblocked on timeout. (14464 ms) -[ok]: CLUSTER SLOT-STATS cpu-usec for transactions. (35941 ms) -[ok]: CLUSTER SLOT-STATS cpu-usec for lua-scripts, without cross-slot keys. (17075 ms) -[ok]: CLUSTER SLOT-STATS cpu-usec for lua-scripts, with cross-slot keys. (22153 ms) -[ok]: CLUSTER SLOT-STATS cpu-usec for functions, without cross-slot keys. (15974 ms) -[ok]: CLUSTER SLOT-STATS cpu-usec for functions, with cross-slot keys. (14875 ms) === (external:skip cluster) Starting server 127.0.0.1:21579 ok -[ok]: CLUSTER SLOT-STATS network-bytes-in, multi bulk buffer processing. (6702 ms) -[ok]: CLUSTER SLOT-STATS network-bytes-in, in-line buffer processing. (14563 ms) -[ok]: CLUSTER SLOT-STATS network-bytes-in, blocking command. (22803 ms) -[ok]: CLUSTER SLOT-STATS network-bytes-in, multi-exec transaction. (11126 ms) -[ok]: CLUSTER SLOT-STATS network-bytes-in, non slot specific command. (10368 ms) -[ok]: CLUSTER SLOT-STATS network-bytes-in, pub/sub. (11409 ms) === (external:skip cluster) Starting server 127.0.0.1:21581 ok === (external:skip cluster) Starting server 127.0.0.1:21583 ok -[ok]: CLUSTER SLOT-STATS network-bytes-in, sharded pub/sub. (24289 ms) === (external:skip cluster) Starting server 127.0.0.1:21585 ok -[ok]: CLUSTER SLOT-STATS network-bytes-out, for non-slot specific commands. (11003 ms) -[ok]: CLUSTER SLOT-STATS network-bytes-out, for slot specific commands. (11438 ms) -[ok]: CLUSTER SLOT-STATS network-bytes-out, blocking commands. (22085 ms) === (external:skip cluster) Starting server 127.0.0.1:21587 ok === (external:skip cluster) Starting server 127.0.0.1:21589 ok -[ok]: CLUSTER SLOT-STATS network-bytes-out, replication stream egress. (9672 ms) === (external:skip cluster) Starting server 127.0.0.1:21591 ok === (external:skip cluster) Starting server 127.0.0.1:21593 ok -[ok]: CLUSTER SLOT-STATS network-bytes-out, sharded pub/sub, single channel. (22126 ms) -[ok]: CLUSTER SLOT-STATS network-bytes-out, sharded pub/sub, cross-slot channels. (21888 ms) === (external:skip cluster) Starting server 127.0.0.1:21595 ok -[ok]: CLUSTER SLOT-STATS contains default value upon valkey-server startup (7745 ms) -[ok]: CLUSTER SLOT-STATS contains correct metrics upon key introduction (14065 ms) -[ok]: CLUSTER SLOT-STATS contains correct metrics upon key mutation (12376 ms) -[ok]: CLUSTER SLOT-STATS contains correct metrics upon key deletion (19081 ms) -[ok]: CLUSTER SLOT-STATS slot visibility based on slot ownership changes (33695 ms) +[ok]: Cluster is up (3 ms) +[ok]: The first master has actually one slave (0 ms) +[ok]: Slaves of #0 is instance #5 as expected (0 ms) +[ok]: Instance #5 synced with the master (0 ms) +[ok]: Lower the slave validity factor of #5 to the value of 2 (1 ms) +[ok]: Break master-slave link and prevent further reconnections (20318 ms) +[ok]: Slave #5 is reachable and alive (5 ms) +[ok]: Slave #5 should not be able to failover (10003 ms) +[ok]: Cluster should be down (715 ms) +[55/124 done]: unit/cluster/slave-stop-cond (45 seconds) +Testing unit/cluster/update-msg === (external:skip cluster) Starting server 127.0.0.1:21597 ok -[ok]: CLUSTER SLOT-STATS SLOTSRANGE all slots present (6 ms) -[ok]: CLUSTER SLOT-STATS SLOTSRANGE some slots missing (20 ms) === (external:skip cluster) Starting server 127.0.0.1:21599 ok -[ok]: CLUSTER SLOT-STATS ORDERBY DESC correct ordering (60206 ms) -[ok]: CLUSTER SLOT-STATS ORDERBY ASC correct ordering (48968 ms) -[ok]: CLUSTER SLOT-STATS ORDERBY LIMIT correct response pagination, where limit is less than number of assigned slots (100 ms) -[ok]: CLUSTER SLOT-STATS ORDERBY LIMIT correct response pagination, where limit is greater than number of assigned slots (92 ms) -[ok]: CLUSTER SLOT-STATS ORDERBY arg sanity check. (56 ms) === (external:skip cluster) Starting server 127.0.0.1:21601 ok === (external:skip cluster) Starting server 127.0.0.1:21603 ok -[ok]: CLUSTER SLOT-STATS metrics replication for new keys (19200 ms) -[ok]: CLUSTER SLOT-STATS metrics replication for existing keys (18938 ms) -[ok]: CLUSTER SLOT-STATS metrics replication for deleting keys (24194 ms) -[55/124 done]: unit/cluster/slot-stats (720 seconds) -Testing unit/cluster/misc === (external:skip cluster) Starting server 127.0.0.1:21605 ok === (external:skip cluster) Starting server 127.0.0.1:21607 ok === (external:skip cluster) Starting server 127.0.0.1:21609 ok === (external:skip cluster) Starting server 127.0.0.1:21611 ok -[ok]: Key lazy expires during key migration (90 ms) -[ok]: Coverage: Basic cluster commands (91 ms) -[56/124 done]: unit/cluster/misc (13 seconds) -Testing unit/cluster/cluster-response-tls -=== (external:skip cluster tls) Starting server 127.0.0.1:21613 ok -=== (external:skip cluster tls) Starting server 127.0.0.1:21615 ok -=== (external:skip cluster tls) Starting server 127.0.0.1:21617 ok -=== (external:skip cluster tls) Starting server 127.0.0.1:21619 ok -=== (external:skip cluster tls) Starting server 127.0.0.1:21621 ok -=== (external:skip cluster tls) Starting server 127.0.0.1:21623 ok -[ok]: CLUSTER SLOTS cached using EVAL over TLS -- tls-cluster yes (43 ms) -[ok]: CLUSTER SLOTS with different connection type -- tls-cluster yes (158 ms) -[ok]: CLUSTER NODES return port according to connection type -- tls-cluster yes (157 ms) -[ok]: Set many keys in the cluster -- tls-cluster yes (59820 ms) -[ok]: Test cluster responses during migration of slot x -- tls-cluster yes (63 ms) -=== (external:skip cluster tls) Starting server 127.0.0.1:21625 ok -=== (external:skip cluster tls) Starting server 127.0.0.1:21627 ok -=== (external:skip cluster tls) Starting server 127.0.0.1:21629 ok -=== (external:skip cluster tls) Starting server 127.0.0.1:21631 ok -=== (external:skip cluster tls) Starting server 127.0.0.1:21633 ok -=== (external:skip cluster tls) Starting server 127.0.0.1:21635 ok -[ok]: CLUSTER SLOTS cached using EVAL over TLS -- tls-cluster no (44 ms) -[ok]: CLUSTER SLOTS with different connection type -- tls-cluster no (261 ms) -[ok]: CLUSTER NODES return port according to connection type -- tls-cluster no (226 ms) -[ok]: Set many keys in the cluster -- tls-cluster no (54616 ms) -[ok]: Test cluster responses during migration of slot x -- tls-cluster no (43 ms) -[57/124 done]: unit/cluster/cluster-response-tls (146 seconds) -Testing unit/cluster/sharded-pubsub +=== (external:skip cluster) Starting server 127.0.0.1:21613 ok +=== (external:skip cluster) Starting server 127.0.0.1:21615 ok +[ok]: Cluster is up (3 ms) +[ok]: Cluster is writable (157 ms) +[ok]: Instance #5 is a slave (1 ms) +[ok]: Instance #5 synced with the master (1 ms) +[ok]: Killing one master node (53 ms) +[ok]: Wait for failover (4272 ms) +[ok]: Cluster should eventually be up again (439 ms) +[ok]: Cluster is writable (132 ms) +[ok]: Instance #5 is now a master (1 ms) +[ok]: Killing the new master #5 (39 ms) +[ok]: Cluster should be down now (3509 ms) +[ok]: Restarting the old master node (1 ms) +[ok]: Instance #0 gets converted into a slave (130 ms) +[ok]: Restarting the new master node (1 ms) +[ok]: Cluster is up again (1603 ms) +[56/124 done]: unit/cluster/update-msg (31 seconds) +Testing unit/cluster/cluster-reliable-meet +[ignore]: Not supported in tls mode +[57/124 done]: unit/cluster/cluster-reliable-meet (0 seconds) +Testing unit/cluster/info +=== (external:skip cluster) Starting server 127.0.0.1:21617 ok +=== (external:skip cluster) Starting server 127.0.0.1:21619 ok +[ok]: Cluster should start ok (1 ms) +[ok]: errorstats: rejected call due to MOVED Redirection (3 ms) +[58/124 done]: unit/cluster/info (3 seconds) +Testing unit/cluster/consistency-check +=== (external:skip cluster) Starting server 127.0.0.1:21621 ok +=== (external:skip cluster) Starting server 127.0.0.1:21623 ok +=== (external:skip cluster) Starting server 127.0.0.1:21625 ok +=== (external:skip cluster) Starting server 127.0.0.1:21627 ok +=== (external:skip cluster) Starting server 127.0.0.1:21629 ok +=== (external:skip cluster) Starting server 127.0.0.1:21631 ok +=== (external:skip cluster) Starting server 127.0.0.1:21633 ok +=== (external:skip cluster) Starting server 127.0.0.1:21635 ok === (external:skip cluster) Starting server 127.0.0.1:21637 ok === (external:skip cluster) Starting server 127.0.0.1:21639 ok -[ok]: Sharded pubsub publish behavior within multi/exec (1 ms) -[ok]: Sharded pubsub within multi/exec with cross slot operation (1 ms) -[ok]: Sharded pubsub publish behavior within multi/exec with read operation on primary (1 ms) -[ok]: Sharded pubsub publish behavior within multi/exec with read operation on replica (7 ms) -[ok]: Sharded pubsub publish behavior within multi/exec with write operation on primary (2 ms) -[ok]: Sharded pubsub publish behavior within multi/exec with write operation on replica (1 ms) -[58/124 done]: unit/cluster/sharded-pubsub (12 seconds) -Testing unit/cluster/slot-migration-response +[ok]: Cluster should start ok (2 ms) +[ok]: Cluster is writable (133 ms) +[ok]: Slave expired keys is loaded when restarted: appendonly=no (3287 ms) +[ok]: Slave expired keys is loaded when restarted: appendonly=yes (3238 ms) +[59/124 done]: unit/cluster/consistency-check (19 seconds) +Testing unit/cluster/manual-takeover === (external:skip cluster) Starting server 127.0.0.1:21641 ok === (external:skip cluster) Starting server 127.0.0.1:21643 ok -[ok]: Cluster is up (15 ms) -[ok]: Set many keys in the cluster (22479 ms) -[ok]: Test cluster responses during migration of slot x (9 ms) -[59/124 done]: unit/cluster/slot-migration-response (26 seconds) -Testing unit/cluster/half-migrated-slot === (external:skip cluster) Starting server 127.0.0.1:21645 ok === (external:skip cluster) Starting server 127.0.0.1:21647 ok -[ok]: Cluster is up (15 ms) -[ok]: Half init migration in 'migrating' is fixable (435 ms) -[ok]: Half init migration in 'importing' is fixable (356 ms) -[ok]: Init migration and move key (1051 ms) -[ok]: Move key again (164 ms) -[ok]: Half-finish migration (550 ms) -[ok]: Move key back (92 ms) -[ok]: Half-finish importing (779 ms) -[60/124 done]: unit/cluster/half-migrated-slot (7 seconds) -Testing unit/cluster/diskless-load-swapdb === (external:skip cluster) Starting server 127.0.0.1:21649 ok === (external:skip cluster) Starting server 127.0.0.1:21651 ok -[ok]: Cluster should start ok (6 ms) -[ok]: Cluster is writable (1777 ms) -[ok]: Main db not affected when fail to diskless load (114420 ms) -Waiting for process 2367310 to exit... -Waiting for process 2367310 to exit... -Waiting for process 2367310 to exit... -[61/124 done]: unit/cluster/diskless-load-swapdb (139 seconds) -Testing unit/cluster/human-announced-nodename === (external:skip cluster) Starting server 127.0.0.1:21653 ok === (external:skip cluster) Starting server 127.0.0.1:21655 ok === (external:skip cluster) Starting server 127.0.0.1:21657 ok -[ok]: Set cluster human announced nodename and let it propagate (183 ms) -[ok]: Human nodenames are visible in log messages (3898 ms) -[62/124 done]: unit/cluster/human-announced-nodename (8 seconds) -Testing unit/cluster/no-failover-option === (external:skip cluster) Starting server 127.0.0.1:21659 ok +[ok]: Cluster is up (4 ms) +[ok]: Cluster is writable (148 ms) +[ok]: Killing majority of master nodes (122 ms) +[ok]: Cluster should eventually be down (3339 ms) +[ok]: Use takeover to bring slaves back (18 ms) +[ok]: Cluster should eventually be up again (3205 ms) +[ok]: Cluster is writable (122 ms) +[ok]: Instance #5, #6, #7 are now masters (1 ms) +[ok]: Restarting the previously killed master nodes (3 ms) +[ok]: Instance #0, #1, #2 gets converted into a slaves (61 ms) +[60/124 done]: unit/cluster/manual-takeover (20 seconds) +Testing unit/cluster/sharded-pubsub === (external:skip cluster) Starting server 127.0.0.1:21661 ok === (external:skip cluster) Starting server 127.0.0.1:21663 ok +[ok]: Sharded pubsub publish behavior within multi/exec (1 ms) +[ok]: Sharded pubsub within multi/exec with cross slot operation (1 ms) +[ok]: Sharded pubsub publish behavior within multi/exec with read operation on primary (1 ms) +[ok]: Sharded pubsub publish behavior within multi/exec with read operation on replica (1 ms) +[ok]: Sharded pubsub publish behavior within multi/exec with write operation on primary (1 ms) +[ok]: Sharded pubsub publish behavior within multi/exec with write operation on replica (1 ms) +[61/124 done]: unit/cluster/sharded-pubsub (12 seconds) +Testing unit/cluster/many-slot-migration === (external:skip cluster) Starting server 127.0.0.1:21665 ok === (external:skip cluster) Starting server 127.0.0.1:21667 ok === (external:skip cluster) Starting server 127.0.0.1:21669 ok -[ok]: Cluster is up (48 ms) -[ok]: Instance #3 is a replica (168 ms) -[ok]: Instance #3 synced with the master (5 ms) -[ok]: The nofailover flag is propagated (59 ms) -[ok]: Killing one master node (180 ms) -[ok]: Cluster should be still down after some time (3082 ms) -[ok]: Instance #3 is still a replica (9 ms) -[ok]: Restarting the previously killed master node (20 ms) -[63/124 done]: unit/cluster/no-failover-option (17 seconds) -Testing unit/cluster/failover === (external:skip cluster) Starting server 127.0.0.1:21671 ok === (external:skip cluster) Starting server 127.0.0.1:21673 ok === (external:skip cluster) Starting server 127.0.0.1:21675 ok @@ -5067,93 +5100,2256 @@ === (external:skip cluster) Starting server 127.0.0.1:21679 ok === (external:skip cluster) Starting server 127.0.0.1:21681 ok === (external:skip cluster) Starting server 127.0.0.1:21683 ok +[ok]: Cluster is up (3 ms) +[ok]: Set many keys (6608 ms) +[ok]: Keys are accessible (6075 ms) +[ok]: Init migration of many slots (960 ms) +[ok]: Fix cluster (23419 ms) +[ok]: Keys are accessible (8754 ms) +[62/124 done]: unit/cluster/many-slot-migration (53 seconds) +Testing unit/cluster/replica-in-sync === (external:skip cluster) Starting server 127.0.0.1:21685 ok === (external:skip cluster) Starting server 127.0.0.1:21687 ok +[ok]: Cluster is up (1 ms) +[ok]: Cluster is writable (134 ms) +[ok]: Fill up primary with data (3 ms) +[ok]: Add new node as replica (1 ms) +[ok]: Check digest and replica state (2 ms) +[ok]: Replica in loading state is hidden (6479 ms) +[ok]: Check disconnected replica not hidden from slots (117 ms) +[63/124 done]: unit/cluster/replica-in-sync (18 seconds) +Testing unit/cluster/slot-migration === (external:skip cluster) Starting server 127.0.0.1:21689 ok -[ok]: Cluster is up (63 ms) -[ok]: Cluster is writable (1692 ms) -[ok]: Instance #5 is a slave (10 ms) -[ok]: Instance #5 synced with the master (7 ms) -[ok]: Killing one master node (290 ms) -[ok]: Wait for failover (4116 ms) -[ok]: Cluster should eventually be up again (2134 ms) -[ok]: Cluster is writable (1994 ms) -[ok]: Instance #5 is now a master (15 ms) -[ok]: Restarting the previously killed master node (22 ms) -[ok]: Instance #0 gets converted into a slave (645 ms) === (external:skip cluster) Starting server 127.0.0.1:21691 ok === (external:skip cluster) Starting server 127.0.0.1:21693 ok === (external:skip cluster) Starting server 127.0.0.1:21695 ok === (external:skip cluster) Starting server 127.0.0.1:21697 ok === (external:skip cluster) Starting server 127.0.0.1:21699 ok +[ok]: Slot migration states are replicated (23 ms) +[ok]: Migration target is auto-updated after failover in target shard (3780 ms) +R 0 ROLE: master 226 {{127.0.0.1 21693 226}} +[err]: Migration source is auto-updated after failover in source shard in tests/unit/cluster/slot-migration.tcl +R 0 didn't assume the replication slave in time +[ok]: Replica redirects key access in migrating slots (2 ms) +[ok]: Replica of migrating node returns ASK redirect after READONLY (3 ms) +[ok]: Replica of migrating node returns TRYAGAIN after READONLY (3 ms) +[ok]: Replica of importing node returns TRYAGAIN after READONLY and ASKING (3 ms) +[ok]: New replica inherits migrating slot (12568 ms) + +===== Start of server log (pid 2067216) ===== + +[err]: New replica inherits importing slot in tests/unit/cluster/slot-migration.tcl +cluster config did not reach a consistent state +### Starting server for test +2067216:C 12 Apr 2026 23:36:26.948 # WARNING: Changing databases number from 16 to 1 since we are in cluster mode +2067216:C 12 Apr 2026 23:36:26.948 # WARNING Memory overcommit must be enabled! Without it, a background save or replication may fail under low memory condition. Being disabled, it can also cause failures without low memory condition, see https://github.com/jemalloc/jemalloc/issues/1328. To fix this issue add 'vm.overcommit_memory = 1' to /etc/sysctl.conf and then reboot or run the command 'sysctl vm.overcommit_memory=1' for this to take effect. +2067216:C 12 Apr 2026 23:36:26.948 * oO0OoO0OoO0Oo Valkey is starting oO0OoO0OoO0Oo +2067216:C 12 Apr 2026 23:36:26.948 * Valkey version=8.0.2, bits=64, commit=00000000, modified=0, pid=2067216, just started +2067216:C 12 Apr 2026 23:36:26.948 * Configuration loaded +2067216:M 12 Apr 2026 23:36:26.949 * Increased maximum number of open files to 10032 (it was originally set to 1024). +2067216:M 12 Apr 2026 23:36:26.949 * monotonic clock: POSIX clock_gettime +2067216:M 12 Apr 2026 23:36:26.952 # Failed to write PID file: Permission denied + .+^+. + .+#########+. + .+########+########+. Valkey 8.0.2 (00000000/0) 64 bit + .+########+' '+########+. + .########+' .+. '+########. Running in cluster mode + |####+' .+#######+. '+####| Port: 21700 + |###| .+###############+. |###| PID: 2067216 + |###| |#####*'' ''*#####| |###| + |###| |####' .-. '####| |###| + |###| |###( (@@@) )###| |###| https://valkey.io + |###| |####. '-' .####| |###| + |###| |#####*. .*#####| |###| + |###| '+#####| |#####+' |###| + |####+. +##| |#+' .+####| + '#######+ |##| .+########' + '+###| |##| .+########+' + '| |####+########+' + +#########+' + '+v+' + +2067216:M 12 Apr 2026 23:36:26.953 * No cluster configuration found, I'm 27c281c71471f64d57c2f90a6fce7ca7d11f3a9e +2067216:M 12 Apr 2026 23:36:27.051 * Server initialized +2067216:M 12 Apr 2026 23:36:27.051 * Ready to accept connections tcp +2067216:M 12 Apr 2026 23:36:27.051 * Ready to accept connections unix +2067216:M 12 Apr 2026 23:36:27.051 * Ready to accept connections tls +2067216:M 12 Apr 2026 23:36:27.184 - Accepted 127.0.0.1:40163 +2067216:M 12 Apr 2026 23:36:27.224 - Reading from client: error:0A000126:SSL routines::unexpected eof while reading +2067216:M 12 Apr 2026 23:36:27.248 - Accepted 127.0.0.1:37223 +2067216:M 12 Apr 2026 23:36:27.420 - Accepting cluster node connection from 127.0.0.1:37906 +2067216:M 12 Apr 2026 23:36:27.454 * IP address for this node updated to 127.0.0.1 +2067216:M 12 Apr 2026 23:36:27.454 * Successfully completed handshake with ceb41590bdb53727bffcb314c80a81989180d13a () +2067216:M 12 Apr 2026 23:36:27.455 * Address updated for node ceb41590bdb53727bffcb314c80a81989180d13a (), now 127.0.0.1:21693 +2067216:M 12 Apr 2026 23:36:27.462 * configEpoch collision with node ceb41590bdb53727bffcb314c80a81989180d13a (). configEpoch set to 1 +2067216:M 12 Apr 2026 23:36:27.477 - Accepting cluster node connection from 127.0.0.1:37908 +2067216:M 12 Apr 2026 23:36:27.484 - Accepting cluster node connection from 127.0.0.1:37920 +2067216:M 12 Apr 2026 23:36:27.490 * Successfully completed handshake with 7cbfdaa3b3c0ebf9bf9116dc86ed8bceffd31db0 () +2067216:M 12 Apr 2026 23:36:27.490 * Address updated for node 7cbfdaa3b3c0ebf9bf9116dc86ed8bceffd31db0 (), now 127.0.0.1:21691 +2067216:M 12 Apr 2026 23:36:27.506 - Accepting cluster node connection from 127.0.0.1:37932 +2067216:M 12 Apr 2026 23:36:27.508 - Accepting cluster node connection from 127.0.0.1:37938 +2067216:M 12 Apr 2026 23:36:27.521 * Successfully completed handshake with ad56f48bae1934f4d1c3141ffee95fa8bf2fa37c () +2067216:M 12 Apr 2026 23:36:27.521 * Address updated for node ad56f48bae1934f4d1c3141ffee95fa8bf2fa37c (), now 127.0.0.1:21697 +2067216:M 12 Apr 2026 23:36:27.532 * Successfully completed handshake with 71cce21972c3b30e1e72d50b69266c87048e6a42 () +2067216:M 12 Apr 2026 23:36:27.532 * Address updated for node 71cce21972c3b30e1e72d50b69266c87048e6a42 (), now 127.0.0.1:21689 +2067216:M 12 Apr 2026 23:36:27.534 * Successfully completed handshake with d24296bca8e44a56521bbf138bec235238274ad7 () +2067216:M 12 Apr 2026 23:36:27.534 * Address updated for node d24296bca8e44a56521bbf138bec235238274ad7 (), now 127.0.0.1:21695 +2067216:M 12 Apr 2026 23:36:27.619 - Accepting cluster node connection from 127.0.0.1:37940 +2067216:M 12 Apr 2026 23:36:27.645 * configEpoch collision with node 71cce21972c3b30e1e72d50b69266c87048e6a42 (). configEpoch set to 2 +2067216:M 12 Apr 2026 23:36:27.699 * configEpoch collision with node ad56f48bae1934f4d1c3141ffee95fa8bf2fa37c (). configEpoch set to 3 +2067216:M 12 Apr 2026 23:36:27.835 * configEpoch collision with node ceb41590bdb53727bffcb314c80a81989180d13a (). configEpoch set to 4 +2067216:M 12 Apr 2026 23:36:27.913 * configEpoch collision with node 7cbfdaa3b3c0ebf9bf9116dc86ed8bceffd31db0 (). configEpoch set to 5 +2067216:M 12 Apr 2026 23:36:27.953 - Accepted 127.0.0.1:43358 +2067216:M 12 Apr 2026 23:36:27.957 * Node ceb41590bdb53727bffcb314c80a81989180d13a () is no longer primary of shard 7be85fc7a35a355988332b0ade6bb85642f90b76; removed all 0 slot(s) it used to own +2067216:M 12 Apr 2026 23:36:27.957 * Node ceb41590bdb53727bffcb314c80a81989180d13a () is now part of shard 0ccab0ca3bdf2df4fb1d88a6388f0d2e2e9492d9 +2067216:M 12 Apr 2026 23:36:27.957 * Node ceb41590bdb53727bffcb314c80a81989180d13a () is now a replica of node 27c281c71471f64d57c2f90a6fce7ca7d11f3a9e () in shard 0ccab0ca3bdf2df4fb1d88a6388f0d2e2e9492d9 +2067216:M 12 Apr 2026 23:36:27.973 * Node 7cbfdaa3b3c0ebf9bf9116dc86ed8bceffd31db0 () is no longer primary of shard 6d5a08da00fb925618c9429f0a0abf50cb184dee; removed all 0 slot(s) it used to own +2067216:M 12 Apr 2026 23:36:27.973 * Node 7cbfdaa3b3c0ebf9bf9116dc86ed8bceffd31db0 () is now part of shard fb1332a38fbff59dddf699f62b6c363413af8689 +2067216:M 12 Apr 2026 23:36:27.973 * Node 7cbfdaa3b3c0ebf9bf9116dc86ed8bceffd31db0 () is now a replica of node ad56f48bae1934f4d1c3141ffee95fa8bf2fa37c () in shard fb1332a38fbff59dddf699f62b6c363413af8689 +2067216:M 12 Apr 2026 23:36:27.975 * Node 71cce21972c3b30e1e72d50b69266c87048e6a42 () is no longer primary of shard 797d84842d0558c353ef2fd200fec2d129d7c96f; removed all 0 slot(s) it used to own +2067216:M 12 Apr 2026 23:36:27.975 * Node 71cce21972c3b30e1e72d50b69266c87048e6a42 () is now part of shard b553502a57202365d35122f1f841b02047c3c23d +2067216:M 12 Apr 2026 23:36:27.975 * Node 71cce21972c3b30e1e72d50b69266c87048e6a42 () is now a replica of node d24296bca8e44a56521bbf138bec235238274ad7 () in shard b553502a57202365d35122f1f841b02047c3c23d +2067216:M 12 Apr 2026 23:36:27.981 * Replica 127.0.0.1:21693 asks for synchronization +2067216:M 12 Apr 2026 23:36:27.981 * Full resync requested by replica 127.0.0.1:21693 +2067216:M 12 Apr 2026 23:36:27.981 * Replication backlog created, my new replication IDs are '3434ac96bb7f1503c85d89d9ee94a4c782639d94' and '0000000000000000000000000000000000000000' +2067216:M 12 Apr 2026 23:36:27.981 * Starting BGSAVE for SYNC with target: replicas sockets using: normal sync +2067216:M 12 Apr 2026 23:36:27.982 * Background RDB transfer started by pid 2067749 to pipe through parent process +2067216:M 12 Apr 2026 23:36:27.985 # DEBUG LOG: ========== I am primary 0 ========== +2067749:C 12 Apr 2026 23:36:27.987 * Fork CoW for RDB: current 0 MB, peak 0 MB, average 0 MB +2067216:M 12 Apr 2026 23:36:27.989 * Diskless rdb transfer, done reading from pipe, 1 replicas still up. +2067216:M 12 Apr 2026 23:36:28.004 * Background RDB transfer terminated with success +2067216:M 12 Apr 2026 23:36:28.004 * Streamed RDB transfer with replica 127.0.0.1:21693 succeeded (socket). Waiting for REPLCONF ACK from replica to enable streaming +2067216:M 12 Apr 2026 23:36:28.004 * Synchronization with replica 127.0.0.1:21693 succeeded +2067216:M 12 Apr 2026 23:36:29.087 * Cluster state changed: ok +2067216:M 12 Apr 2026 23:36:35.592 * NODE 71cce21972c3b30e1e72d50b69266c87048e6a42 () possibly failing. +### Starting test Slot migration states are replicated in tests/unit/cluster/slot-migration.tcl +2067216:M 12 Apr 2026 23:36:37.789 * Migrating slot 609 to node ad56f48bae1934f4d1c3141ffee95fa8bf2fa37c () +### Starting test Migration target is auto-updated after failover in target shard in tests/unit/cluster/slot-migration.tcl +2067216:M 12 Apr 2026 23:36:39.000 * NODE ad56f48bae1934f4d1c3141ffee95fa8bf2fa37c () possibly failing. +2067216:M 12 Apr 2026 23:36:39.104 * Node d24296bca8e44a56521bbf138bec235238274ad7 () reported node ad56f48bae1934f4d1c3141ffee95fa8bf2fa37c () as not reachable. +2067216:M 12 Apr 2026 23:36:39.105 * Marking node ad56f48bae1934f4d1c3141ffee95fa8bf2fa37c () as failing (quorum reached). +2067216:M 12 Apr 2026 23:36:39.105 # Cluster state changed: fail +2067216:M 12 Apr 2026 23:36:39.928 * Failover auth granted to 7cbfdaa3b3c0ebf9bf9116dc86ed8bceffd31db0 () for epoch 6 +2067216:M 12 Apr 2026 23:36:39.982 * Failover occurred in migration target. Slot 609 is now being migrated to node 7cbfdaa3b3c0ebf9bf9116dc86ed8bceffd31db0 () in shard fb1332a38fbff59dddf699f62b6c363413af8689. +2067216:M 12 Apr 2026 23:36:39.984 * Node 7cbfdaa3b3c0ebf9bf9116dc86ed8bceffd31db0 () reported node ad56f48bae1934f4d1c3141ffee95fa8bf2fa37c () as not reachable. +2067216:M 12 Apr 2026 23:36:39.985 * Cluster state changed: ok +2067216:M 12 Apr 2026 23:36:40.925 - Node ad56f48bae1934f4d1c3141ffee95fa8bf2fa37c has old slots configuration, sending an UPDATE message about 7cbfdaa3b3c0ebf9bf9116dc86ed8bceffd31db0 +2067216:M 12 Apr 2026 23:36:41.178 * A failover occurred in shard fb1332a38fbff59dddf699f62b6c363413af8689; node ad56f48bae1934f4d1c3141ffee95fa8bf2fa37c () failed over to node 7cbfdaa3b3c0ebf9bf9116dc86ed8bceffd31db0 () with a config epoch of 6 +2067216:M 12 Apr 2026 23:36:41.178 * Node ad56f48bae1934f4d1c3141ffee95fa8bf2fa37c () is now a replica of node 7cbfdaa3b3c0ebf9bf9116dc86ed8bceffd31db0 () in shard fb1332a38fbff59dddf699f62b6c363413af8689 +2067216:M 12 Apr 2026 23:36:41.182 * Clear FAIL state for node ad56f48bae1934f4d1c3141ffee95fa8bf2fa37c (): replica is reachable again. +2067216:M 12 Apr 2026 23:36:41.291 * Node d24296bca8e44a56521bbf138bec235238274ad7 () reported node ad56f48bae1934f4d1c3141ffee95fa8bf2fa37c () is back online. +2067216:M 12 Apr 2026 23:36:41.293 * Failover auth granted to ad56f48bae1934f4d1c3141ffee95fa8bf2fa37c () for epoch 7 +2067216:M 12 Apr 2026 23:36:41.321 * Node 7cbfdaa3b3c0ebf9bf9116dc86ed8bceffd31db0 () reported node ad56f48bae1934f4d1c3141ffee95fa8bf2fa37c () is back online. +2067216:M 12 Apr 2026 23:36:41.337 * Failover occurred in migration target. Slot 609 is now being migrated to node ad56f48bae1934f4d1c3141ffee95fa8bf2fa37c () in shard fb1332a38fbff59dddf699f62b6c363413af8689. +2067216:M 12 Apr 2026 23:36:41.463 * A failover occurred in shard fb1332a38fbff59dddf699f62b6c363413af8689; node 7cbfdaa3b3c0ebf9bf9116dc86ed8bceffd31db0 () failed over to node ad56f48bae1934f4d1c3141ffee95fa8bf2fa37c () with a config epoch of 7 +2067216:M 12 Apr 2026 23:36:41.463 * Node 7cbfdaa3b3c0ebf9bf9116dc86ed8bceffd31db0 () is now a replica of node ad56f48bae1934f4d1c3141ffee95fa8bf2fa37c () in shard fb1332a38fbff59dddf699f62b6c363413af8689 +### Starting test Migration source is auto-updated after failover in source shard in tests/unit/cluster/slot-migration.tcl +2067216:M 12 Apr 2026 23:36:45.705 - Accepting cluster node connection from 127.0.0.1:41370 +2067216:M 12 Apr 2026 23:36:45.708 - Error accepting cluster node connection: error:0A000126:SSL routines::unexpected eof while reading +2067216:M 12 Apr 2026 23:36:45.709 - Accepting cluster node connection from 127.0.0.1:41372 +2067216:M 12 Apr 2026 23:36:45.711 - Error accepting cluster node connection: error:0A000126:SSL routines::unexpected eof while reading +2067216:M 12 Apr 2026 23:36:45.711 - Accepting cluster node connection from 127.0.0.1:41386 +2067216:M 12 Apr 2026 23:36:45.714 - Error accepting cluster node connection: error:0A000126:SSL routines::unexpected eof while reading +2067216:M 12 Apr 2026 23:36:45.714 - Accepting cluster node connection from 127.0.0.1:41388 +2067216:M 12 Apr 2026 23:36:45.717 - Error accepting cluster node connection: error:0A000126:SSL routines::unexpected eof while reading +2067216:M 12 Apr 2026 23:36:45.717 - Accepting cluster node connection from 127.0.0.1:41390 +2067216:M 12 Apr 2026 23:36:45.720 - Error accepting cluster node connection: error:0A000126:SSL routines::unexpected eof while reading +2067216:M 12 Apr 2026 23:36:45.720 - Accepting cluster node connection from 127.0.0.1:41398 +2067216:M 12 Apr 2026 23:36:45.728 - Accepting cluster node connection from 127.0.0.1:41402 +2067216:M 12 Apr 2026 23:36:45.737 - Accepting cluster node connection from 127.0.0.1:41410 +2067216:M 12 Apr 2026 23:36:45.760 - Accepting cluster node connection from 127.0.0.1:41412 +2067216:M 12 Apr 2026 23:36:46.727 * NODE ceb41590bdb53727bffcb314c80a81989180d13a () possibly failing. +2067216:M 12 Apr 2026 23:36:47.009 * Node d24296bca8e44a56521bbf138bec235238274ad7 () reported node ceb41590bdb53727bffcb314c80a81989180d13a () as not reachable. +2067216:M 12 Apr 2026 23:36:47.009 * Marking node ceb41590bdb53727bffcb314c80a81989180d13a () as failing (quorum reached). +2067216:M 12 Apr 2026 23:36:47.117 * Node ad56f48bae1934f4d1c3141ffee95fa8bf2fa37c () reported node ceb41590bdb53727bffcb314c80a81989180d13a () as not reachable. +2067216:M 12 Apr 2026 23:36:47.179 - DB 0: 1 keys (0 volatile) in 4 slots HT. +2067216:M 12 Apr 2026 23:36:47.255 - Accepting cluster node connection from 127.0.0.1:41414 +2067216:M 12 Apr 2026 23:36:47.282 * Clear FAIL state for node ceb41590bdb53727bffcb314c80a81989180d13a (): replica is reachable again. +2067216:M 12 Apr 2026 23:36:47.397 * Node ad56f48bae1934f4d1c3141ffee95fa8bf2fa37c () reported node ceb41590bdb53727bffcb314c80a81989180d13a () is back online. +2067216:M 12 Apr 2026 23:36:47.397 * Node d24296bca8e44a56521bbf138bec235238274ad7 () reported node ceb41590bdb53727bffcb314c80a81989180d13a () is back online. +2067216:M 12 Apr 2026 23:36:52.293 - DB 0: 1 keys (0 volatile) in 4 slots HT. +2067216:M 12 Apr 2026 23:36:53.741 * NODE 7cbfdaa3b3c0ebf9bf9116dc86ed8bceffd31db0 () possibly failing. +2067216:M 12 Apr 2026 23:36:53.862 * Node ad56f48bae1934f4d1c3141ffee95fa8bf2fa37c () reported node 7cbfdaa3b3c0ebf9bf9116dc86ed8bceffd31db0 () as not reachable. +2067216:M 12 Apr 2026 23:36:53.862 * Marking node 7cbfdaa3b3c0ebf9bf9116dc86ed8bceffd31db0 () as failing (quorum reached). +2067216:M 12 Apr 2026 23:36:54.041 * Node d24296bca8e44a56521bbf138bec235238274ad7 () reported node 7cbfdaa3b3c0ebf9bf9116dc86ed8bceffd31db0 () as not reachable. +2067216:M 12 Apr 2026 23:36:54.056 * Clear FAIL state for node 7cbfdaa3b3c0ebf9bf9116dc86ed8bceffd31db0 (): replica is reachable again. +2067216:M 12 Apr 2026 23:36:54.249 * Node d24296bca8e44a56521bbf138bec235238274ad7 () reported node 7cbfdaa3b3c0ebf9bf9116dc86ed8bceffd31db0 () is back online. +2067216:M 12 Apr 2026 23:36:54.327 * Node ad56f48bae1934f4d1c3141ffee95fa8bf2fa37c () reported node 7cbfdaa3b3c0ebf9bf9116dc86ed8bceffd31db0 () is back online. +2067216:M 12 Apr 2026 23:36:55.251 - Accepting cluster node connection from 127.0.0.1:51402 +### Starting test Replica redirects key access in migrating slots in tests/unit/cluster/slot-migration.tcl +### Starting test Replica of migrating node returns ASK redirect after READONLY in tests/unit/cluster/slot-migration.tcl +### Starting test Replica of migrating node returns TRYAGAIN after READONLY in tests/unit/cluster/slot-migration.tcl +### Starting test Replica of importing node returns TRYAGAIN after READONLY and ASKING in tests/unit/cluster/slot-migration.tcl +### Starting test New replica inherits migrating slot in tests/unit/cluster/slot-migration.tcl +2067216:M 12 Apr 2026 23:36:57.967 - Client closed connection id=11 addr=127.0.0.1:43358 laddr=127.0.0.1:21699 fd=24 name= age=30 idle=0 flags=S db=0 sub=0 psub=0 ssub=0 multi=-1 watch=0 qbuf=0 qbuf-free=20474 argv-mem=0 multi-mem=0 rbs=1024 rbp=0 obl=0 oll=1 omem=20504 tot-mem=42904 events=r cmd=replconf user=default redir=-1 resp=2 lib-name= lib-ver= tot-net-in=1111 tot-net-out=248 tot-cmds=31 +2067216:M 12 Apr 2026 23:36:57.967 * Connection with replica 127.0.0.1:21693 lost. +2067216:M 12 Apr 2026 23:36:58.742 - DB 0: 1 keys (0 volatile) in 4 slots HT. +2067216:M 12 Apr 2026 23:36:59.047 * NODE ceb41590bdb53727bffcb314c80a81989180d13a () possibly failing. +2067216:M 12 Apr 2026 23:36:59.633 * FAIL message received from ad56f48bae1934f4d1c3141ffee95fa8bf2fa37c () about ceb41590bdb53727bffcb314c80a81989180d13a () +2067216:M 12 Apr 2026 23:36:59.636 * Node ad56f48bae1934f4d1c3141ffee95fa8bf2fa37c () reported node ceb41590bdb53727bffcb314c80a81989180d13a () as not reachable. +2067216:M 12 Apr 2026 23:36:59.637 * Node d24296bca8e44a56521bbf138bec235238274ad7 () reported node ceb41590bdb53727bffcb314c80a81989180d13a () as not reachable. +2067216:M 12 Apr 2026 23:37:00.749 - Accepting cluster node connection from 127.0.0.1:51410 +2067216:M 12 Apr 2026 23:37:00.943 * Clear FAIL state for node ceb41590bdb53727bffcb314c80a81989180d13a (): replica is reachable again. +2067216:M 12 Apr 2026 23:37:00.956 * Node ad56f48bae1934f4d1c3141ffee95fa8bf2fa37c () reported node ceb41590bdb53727bffcb314c80a81989180d13a () is back online. +2067216:M 12 Apr 2026 23:37:00.956 * Node d24296bca8e44a56521bbf138bec235238274ad7 () reported node ceb41590bdb53727bffcb314c80a81989180d13a () is back online. +2067216:M 12 Apr 2026 23:37:01.648 - Accepting cluster node connection from 127.0.0.1:51416 +2067216:M 12 Apr 2026 23:37:02.077 * NODE ceb41590bdb53727bffcb314c80a81989180d13a () possibly failing. +2067216:M 12 Apr 2026 23:37:02.271 * Node d24296bca8e44a56521bbf138bec235238274ad7 () reported node ceb41590bdb53727bffcb314c80a81989180d13a () as not reachable. +2067216:M 12 Apr 2026 23:37:02.271 * Marking node ceb41590bdb53727bffcb314c80a81989180d13a () as failing (quorum reached). +2067216:M 12 Apr 2026 23:37:04.040 * Node d24296bca8e44a56521bbf138bec235238274ad7 () reported node ad56f48bae1934f4d1c3141ffee95fa8bf2fa37c () as not reachable. +2067216:M 12 Apr 2026 23:37:04.043 - Accepting cluster node connection from 127.0.0.1:44446 +2067216:M 12 Apr 2026 23:37:04.046 - Accepting cluster node connection from 127.0.0.1:44448 +2067216:M 12 Apr 2026 23:37:04.076 * Node d24296bca8e44a56521bbf138bec235238274ad7 () reported node 7cbfdaa3b3c0ebf9bf9116dc86ed8bceffd31db0 () as not reachable. +2067216:M 12 Apr 2026 23:37:04.266 - Error accepting cluster node connection: Connection reset by peer +2067216:M 12 Apr 2026 23:37:04.348 * Clear FAIL state for node ceb41590bdb53727bffcb314c80a81989180d13a (): primary without slots is reachable again. +2067216:M 12 Apr 2026 23:37:04.429 * Node ad56f48bae1934f4d1c3141ffee95fa8bf2fa37c () reported node ceb41590bdb53727bffcb314c80a81989180d13a () as not reachable. +2067216:M 12 Apr 2026 23:37:04.437 - Accepting cluster node connection from 127.0.0.1:44456 +2067216:M 12 Apr 2026 23:37:04.618 * Node d24296bca8e44a56521bbf138bec235238274ad7 () reported node 7cbfdaa3b3c0ebf9bf9116dc86ed8bceffd31db0 () is back online. +2067216:M 12 Apr 2026 23:37:05.086 * Node d24296bca8e44a56521bbf138bec235238274ad7 () reported node ad56f48bae1934f4d1c3141ffee95fa8bf2fa37c () is back online. +2067216:M 12 Apr 2026 23:37:05.207 * Node ad56f48bae1934f4d1c3141ffee95fa8bf2fa37c () reported node ceb41590bdb53727bffcb314c80a81989180d13a () is back online. +2067216:M 12 Apr 2026 23:37:05.207 * Node ad56f48bae1934f4d1c3141ffee95fa8bf2fa37c () reported node 71cce21972c3b30e1e72d50b69266c87048e6a42 () as not reachable. +2067216:M 12 Apr 2026 23:37:05.289 * Node d24296bca8e44a56521bbf138bec235238274ad7 () reported node ceb41590bdb53727bffcb314c80a81989180d13a () as not reachable. +2067216:M 12 Apr 2026 23:37:05.391 * Node ad56f48bae1934f4d1c3141ffee95fa8bf2fa37c () reported node 71cce21972c3b30e1e72d50b69266c87048e6a42 () is back online. +2067216:M 12 Apr 2026 23:37:06.002 - Accepting cluster node connection from 127.0.0.1:44464 +2067216:M 12 Apr 2026 23:37:06.021 - Accepting cluster node connection from 127.0.0.1:44474 +2067216:M 12 Apr 2026 23:37:06.251 * NODE ceb41590bdb53727bffcb314c80a81989180d13a () possibly failing. +2067216:M 12 Apr 2026 23:37:06.436 * Marking node ceb41590bdb53727bffcb314c80a81989180d13a () as failing (quorum reached). +2067216:M 12 Apr 2026 23:37:06.907 * Node ad56f48bae1934f4d1c3141ffee95fa8bf2fa37c () reported node ceb41590bdb53727bffcb314c80a81989180d13a () as not reachable. +2067216:M 12 Apr 2026 23:37:06.963 - DB 0: 1 keys (0 volatile) in 4 slots HT. +2067216:M 12 Apr 2026 23:37:08.577 * Clear FAIL state for node ceb41590bdb53727bffcb314c80a81989180d13a (): primary without slots is reachable again. +2067216:M 12 Apr 2026 23:37:08.790 * Node d24296bca8e44a56521bbf138bec235238274ad7 () reported node ceb41590bdb53727bffcb314c80a81989180d13a () is back online. +2067216:M 12 Apr 2026 23:37:08.856 - Accepted 127.0.0.1:40500 +2067216:M 12 Apr 2026 23:37:08.985 * Node ad56f48bae1934f4d1c3141ffee95fa8bf2fa37c () reported node ceb41590bdb53727bffcb314c80a81989180d13a () is back online. +2067216:M 12 Apr 2026 23:37:09.065 * A failover occurred in shard 0ccab0ca3bdf2df4fb1d88a6388f0d2e2e9492d9; node ceb41590bdb53727bffcb314c80a81989180d13a () failed over to node 27c281c71471f64d57c2f90a6fce7ca7d11f3a9e () with a config epoch of 5 +2067216:M 12 Apr 2026 23:37:09.065 * Node ceb41590bdb53727bffcb314c80a81989180d13a () is now a replica of node 27c281c71471f64d57c2f90a6fce7ca7d11f3a9e () in shard 0ccab0ca3bdf2df4fb1d88a6388f0d2e2e9492d9 +2067216:M 12 Apr 2026 23:37:09.274 * Replica 127.0.0.1:21693 asks for synchronization +2067216:M 12 Apr 2026 23:37:09.274 * Full resync requested by replica 127.0.0.1:21693 +2067216:M 12 Apr 2026 23:37:09.274 * Starting BGSAVE for SYNC with target: replicas sockets using: normal sync +2067216:M 12 Apr 2026 23:37:09.275 * Background RDB transfer started by pid 2082883 to pipe through parent process +2082883:C 12 Apr 2026 23:37:09.277 - Fork CoW for RDB: current 0 MB, peak 0 MB, average 0 MB +2082883:C 12 Apr 2026 23:37:09.293 * Fork CoW for RDB: current 1 MB, peak 1 MB, average 1 MB +2067216:M 12 Apr 2026 23:37:09.293 * Diskless rdb transfer, done reading from pipe, 1 replicas still up. +2067216:M 12 Apr 2026 23:37:09.388 * Background RDB transfer terminated with success +2067216:M 12 Apr 2026 23:37:09.388 * Streamed RDB transfer with replica 127.0.0.1:21693 succeeded (socket). Waiting for REPLCONF ACK from replica to enable streaming +2067216:M 12 Apr 2026 23:37:09.388 * Synchronization with replica 127.0.0.1:21693 succeeded +2067216:M 12 Apr 2026 23:37:10.420 * NODE ceb41590bdb53727bffcb314c80a81989180d13a () possibly failing. +### Starting test New replica inherits importing slot in tests/unit/cluster/slot-migration.tcl +2067216:M 12 Apr 2026 23:37:12.524 * NODE 71cce21972c3b30e1e72d50b69266c87048e6a42 () possibly failing. +2067216:M 12 Apr 2026 23:37:12.525 * NODE 7cbfdaa3b3c0ebf9bf9116dc86ed8bceffd31db0 () possibly failing. +2067216:M 12 Apr 2026 23:37:12.762 * Node d24296bca8e44a56521bbf138bec235238274ad7 () reported node 71cce21972c3b30e1e72d50b69266c87048e6a42 () as not reachable. +2067216:M 12 Apr 2026 23:37:12.762 * Node d24296bca8e44a56521bbf138bec235238274ad7 () reported node 7cbfdaa3b3c0ebf9bf9116dc86ed8bceffd31db0 () as not reachable. +2067216:M 12 Apr 2026 23:37:12.762 * Marking node 7cbfdaa3b3c0ebf9bf9116dc86ed8bceffd31db0 () as failing (quorum reached). +2067216:M 12 Apr 2026 23:37:12.762 * FAIL message received from ad56f48bae1934f4d1c3141ffee95fa8bf2fa37c () about 71cce21972c3b30e1e72d50b69266c87048e6a42 () +2067216:M 12 Apr 2026 23:37:12.837 * Clear FAIL state for node 7cbfdaa3b3c0ebf9bf9116dc86ed8bceffd31db0 (): replica is reachable again. +2067216:M 12 Apr 2026 23:37:12.838 * FAIL message received from ad56f48bae1934f4d1c3141ffee95fa8bf2fa37c () about 7cbfdaa3b3c0ebf9bf9116dc86ed8bceffd31db0 () +2067216:M 12 Apr 2026 23:37:13.004 * Clear FAIL state for node 7cbfdaa3b3c0ebf9bf9116dc86ed8bceffd31db0 (): primary without slots is reachable again. +2067216:M 12 Apr 2026 23:37:13.019 * Node ad56f48bae1934f4d1c3141ffee95fa8bf2fa37c () reported node 71cce21972c3b30e1e72d50b69266c87048e6a42 () as not reachable. +2067216:M 12 Apr 2026 23:37:13.020 * Node d24296bca8e44a56521bbf138bec235238274ad7 () reported node 71cce21972c3b30e1e72d50b69266c87048e6a42 () is back online. +2067216:M 12 Apr 2026 23:37:13.020 * Clear FAIL state for node 71cce21972c3b30e1e72d50b69266c87048e6a42 (): replica is reachable again. +2067216:M 12 Apr 2026 23:37:13.193 * Node d24296bca8e44a56521bbf138bec235238274ad7 () reported node 7cbfdaa3b3c0ebf9bf9116dc86ed8bceffd31db0 () is back online. +2067216:M 12 Apr 2026 23:37:13.195 * Node ad56f48bae1934f4d1c3141ffee95fa8bf2fa37c () reported node 71cce21972c3b30e1e72d50b69266c87048e6a42 () is back online. +2067216:M 12 Apr 2026 23:37:14.209 - DB 0: 1 keys (0 volatile) in 4 slots HT. +2067216:M 12 Apr 2026 23:37:19.294 - DB 0: 1 keys (0 volatile) in 4 slots HT. +2067216:M 12 Apr 2026 23:37:22.226 * NODE 71cce21972c3b30e1e72d50b69266c87048e6a42 () possibly failing. +2067216:M 12 Apr 2026 23:37:22.445 * NODE ceb41590bdb53727bffcb314c80a81989180d13a () possibly failing. +2067216:M 12 Apr 2026 23:37:22.445 * NODE d24296bca8e44a56521bbf138bec235238274ad7 () possibly failing. +2067216:M 12 Apr 2026 23:37:22.445 * NODE 7cbfdaa3b3c0ebf9bf9116dc86ed8bceffd31db0 () possibly failing. +2067216:M 12 Apr 2026 23:37:22.445 * NODE ad56f48bae1934f4d1c3141ffee95fa8bf2fa37c () possibly failing. +2067216:M 12 Apr 2026 23:37:22.445 # Cluster state changed: fail +2067216:M 12 Apr 2026 23:37:23.683 * Cluster state changed: ok +2067216:M 12 Apr 2026 23:37:27.134 - DB 0: 1 keys (0 volatile) in 4 slots HT. +2067216:M 12 Apr 2026 23:37:29.979 * NODE 7cbfdaa3b3c0ebf9bf9116dc86ed8bceffd31db0 () possibly failing. +2067216:M 12 Apr 2026 23:37:30.089 * FAIL message received from ad56f48bae1934f4d1c3141ffee95fa8bf2fa37c () about 7cbfdaa3b3c0ebf9bf9116dc86ed8bceffd31db0 () +2067216:M 12 Apr 2026 23:37:30.192 * Node d24296bca8e44a56521bbf138bec235238274ad7 () reported node 7cbfdaa3b3c0ebf9bf9116dc86ed8bceffd31db0 () as not reachable. +2067216:M 12 Apr 2026 23:37:30.204 * Node ad56f48bae1934f4d1c3141ffee95fa8bf2fa37c () reported node 7cbfdaa3b3c0ebf9bf9116dc86ed8bceffd31db0 () as not reachable. +2067216:M 12 Apr 2026 23:37:30.236 * Clear FAIL state for node 7cbfdaa3b3c0ebf9bf9116dc86ed8bceffd31db0 (): primary without slots is reachable again. +2067216:M 12 Apr 2026 23:37:30.379 * Node d24296bca8e44a56521bbf138bec235238274ad7 () reported node 7cbfdaa3b3c0ebf9bf9116dc86ed8bceffd31db0 () is back online. +2067216:M 12 Apr 2026 23:37:30.400 * Node ad56f48bae1934f4d1c3141ffee95fa8bf2fa37c () reported node 7cbfdaa3b3c0ebf9bf9116dc86ed8bceffd31db0 () is back online. +2067216:M 12 Apr 2026 23:37:32.341 - DB 0: 1 keys (0 volatile) in 4 slots HT. +2067216:M 12 Apr 2026 23:37:37.487 - DB 0: 1 keys (0 volatile) in 4 slots HT. +2067216:M 12 Apr 2026 23:37:39.665 * NODE ceb41590bdb53727bffcb314c80a81989180d13a () possibly failing. +2067216:M 12 Apr 2026 23:37:39.665 * NODE 7cbfdaa3b3c0ebf9bf9116dc86ed8bceffd31db0 () possibly failing. +2067216:M 12 Apr 2026 23:37:39.764 - Accepting cluster node connection from 127.0.0.1:50630 +2067216:M 12 Apr 2026 23:37:39.816 - Accepting cluster node connection from 127.0.0.1:50640 +2067216:M 12 Apr 2026 23:37:39.841 * Node d24296bca8e44a56521bbf138bec235238274ad7 () reported node 7cbfdaa3b3c0ebf9bf9116dc86ed8bceffd31db0 () as not reachable. +2067216:M 12 Apr 2026 23:37:39.842 * Marking node 7cbfdaa3b3c0ebf9bf9116dc86ed8bceffd31db0 () as failing (quorum reached). +2067216:M 12 Apr 2026 23:37:39.842 * Node d24296bca8e44a56521bbf138bec235238274ad7 () reported node ceb41590bdb53727bffcb314c80a81989180d13a () as not reachable. +2067216:M 12 Apr 2026 23:37:39.842 * Marking node ceb41590bdb53727bffcb314c80a81989180d13a () as failing (quorum reached). +2067216:M 12 Apr 2026 23:37:39.842 * Node ad56f48bae1934f4d1c3141ffee95fa8bf2fa37c () reported node d24296bca8e44a56521bbf138bec235238274ad7 () as not reachable. +2067216:M 12 Apr 2026 23:37:39.842 * Node ad56f48bae1934f4d1c3141ffee95fa8bf2fa37c () reported node ceb41590bdb53727bffcb314c80a81989180d13a () as not reachable. +2067216:M 12 Apr 2026 23:37:39.885 * Clear FAIL state for node ceb41590bdb53727bffcb314c80a81989180d13a (): replica is reachable again. +2067216:M 12 Apr 2026 23:37:39.885 * FAIL message received from d24296bca8e44a56521bbf138bec235238274ad7 () about ceb41590bdb53727bffcb314c80a81989180d13a () +2067216:M 12 Apr 2026 23:37:39.885 * Node d24296bca8e44a56521bbf138bec235238274ad7 () reported node 7cbfdaa3b3c0ebf9bf9116dc86ed8bceffd31db0 () is back online. +2067216:M 12 Apr 2026 23:37:39.904 * Clear FAIL state for node 7cbfdaa3b3c0ebf9bf9116dc86ed8bceffd31db0 (): primary without slots is reachable again. +2067216:M 12 Apr 2026 23:37:39.925 * FAIL message received from 71cce21972c3b30e1e72d50b69266c87048e6a42 () about 7cbfdaa3b3c0ebf9bf9116dc86ed8bceffd31db0 () +2067216:M 12 Apr 2026 23:37:39.975 * Node ad56f48bae1934f4d1c3141ffee95fa8bf2fa37c () reported node d24296bca8e44a56521bbf138bec235238274ad7 () is back online. +2067216:M 12 Apr 2026 23:37:39.975 * Node ad56f48bae1934f4d1c3141ffee95fa8bf2fa37c () reported node ceb41590bdb53727bffcb314c80a81989180d13a () is back online. +2067216:M 12 Apr 2026 23:37:40.009 * Node d24296bca8e44a56521bbf138bec235238274ad7 () reported node ceb41590bdb53727bffcb314c80a81989180d13a () is back online. +2067216:M 12 Apr 2026 23:37:40.057 * Clear FAIL state for node ceb41590bdb53727bffcb314c80a81989180d13a (): replica is reachable again. +2067216:M 12 Apr 2026 23:37:40.060 * Clear FAIL state for node 7cbfdaa3b3c0ebf9bf9116dc86ed8bceffd31db0 (): primary without slots is reachable again. +2067216:M 12 Apr 2026 23:37:43.940 - DB 0: 1 keys (0 volatile) in 4 slots HT. +2067216:M 12 Apr 2026 23:37:49.345 * NODE 7cbfdaa3b3c0ebf9bf9116dc86ed8bceffd31db0 () possibly failing. +2067216:M 12 Apr 2026 23:37:50.577 - DB 0: 1 keys (0 volatile) in 4 slots HT. +2067216:M 12 Apr 2026 23:37:55.790 - DB 0: 1 keys (0 volatile) in 4 slots HT. +2067216:M 12 Apr 2026 23:37:58.586 - Accepting cluster node connection from 127.0.0.1:39674 +2067216:M 12 Apr 2026 23:37:58.588 - Error accepting cluster node connection: error:0A000126:SSL routines::unexpected eof while reading +2067216:M 12 Apr 2026 23:37:58.588 - Accepting cluster node connection from 127.0.0.1:39682 +2067216:M 12 Apr 2026 23:37:58.625 - Error accepting cluster node connection: error:0A000126:SSL routines::unexpected eof while reading +2067216:M 12 Apr 2026 23:37:58.625 - Accepting cluster node connection from 127.0.0.1:39688 +2067216:M 12 Apr 2026 23:37:58.669 - Error accepting cluster node connection: error:0A000126:SSL routines::unexpected eof while reading +2067216:M 12 Apr 2026 23:37:58.669 * Node ad56f48bae1934f4d1c3141ffee95fa8bf2fa37c () reported node 71cce21972c3b30e1e72d50b69266c87048e6a42 () as not reachable. +2067216:M 12 Apr 2026 23:37:58.669 - Accepting cluster node connection from 127.0.0.1:39704 +2067216:M 12 Apr 2026 23:37:58.672 - Error accepting cluster node connection: error:0A000126:SSL routines::unexpected eof while reading +2067216:M 12 Apr 2026 23:37:58.672 - Accepting cluster node connection from 127.0.0.1:39708 +2067216:M 12 Apr 2026 23:37:58.719 - Accepting cluster node connection from 127.0.0.1:39720 +2067216:M 12 Apr 2026 23:37:58.753 - Accepting cluster node connection from 127.0.0.1:39728 +2067216:M 12 Apr 2026 23:37:58.756 - Accepting cluster node connection from 127.0.0.1:39732 +2067216:M 12 Apr 2026 23:37:58.851 * Node ad56f48bae1934f4d1c3141ffee95fa8bf2fa37c () reported node 71cce21972c3b30e1e72d50b69266c87048e6a42 () is back online. +2067216:M 12 Apr 2026 23:37:59.569 # Failover auth denied to ceb41590bdb53727bffcb314c80a81989180d13a () for epoch 9: its primary is up +2067216:M 12 Apr 2026 23:37:59.677 - Client closed connection id=33 addr=127.0.0.1:40500 laddr=127.0.0.1:21699 fd=24 name= age=51 idle=1 flags=S db=0 sub=0 psub=0 ssub=0 multi=-1 watch=0 qbuf=0 qbuf-free=20474 argv-mem=0 multi-mem=0 rbs=1024 rbp=0 obl=0 oll=1 omem=20504 tot-mem=42904 events=r cmd=replconf user=default redir=-1 resp=2 lib-name= lib-ver= tot-net-in=1635 tot-net-out=78 tot-cmds=45 +2067216:M 12 Apr 2026 23:37:59.677 * Connection with replica 127.0.0.1:21693 lost. +2067216:M 12 Apr 2026 23:37:59.721 * Configuration change detected. Reconfiguring myself as a replica of node ceb41590bdb53727bffcb314c80a81989180d13a () in shard 0ccab0ca3bdf2df4fb1d88a6388f0d2e2e9492d9 +2067216:S 12 Apr 2026 23:37:59.721 * Before turning into a replica, using my own primary parameters to synthesize a cached primary: I may be able to synchronize with the new primary with just a partial transfer. +2067216:S 12 Apr 2026 23:37:59.721 * Connecting to PRIMARY 127.0.0.1:21693 +2067216:S 12 Apr 2026 23:37:59.721 * PRIMARY <-> REPLICA sync started +2067216:S 12 Apr 2026 23:37:59.787 * Non blocking connect for SYNC fired the event. +2067216:S 12 Apr 2026 23:37:59.801 * Primary replied to PING, replication can continue... +2067216:S 12 Apr 2026 23:37:59.814 * Trying a partial resynchronization (request 3434ac96bb7f1503c85d89d9ee94a4c782639d94:283). +2067216:S 12 Apr 2026 23:37:59.821 * Successful partial resynchronization with primary. +2067216:S 12 Apr 2026 23:37:59.821 * Primary replication ID changed to 0b3f1854e94de8dad9f58c42bfd4fbae2be7ceef +2067216:S 12 Apr 2026 23:37:59.821 * PRIMARY <-> REPLICA sync: Primary accepted a Partial Resynchronization. +2067216:S 12 Apr 2026 23:38:03.023 - DB 0: 1 keys (0 volatile) in 4 slots HT. +2067216:S 12 Apr 2026 23:38:07.610 * NODE 71cce21972c3b30e1e72d50b69266c87048e6a42 () possibly failing. +2067216:S 12 Apr 2026 23:38:07.855 * Node d24296bca8e44a56521bbf138bec235238274ad7 () reported node 71cce21972c3b30e1e72d50b69266c87048e6a42 () as not reachable. +2067216:S 12 Apr 2026 23:38:07.873 * Node ad56f48bae1934f4d1c3141ffee95fa8bf2fa37c () reported node 71cce21972c3b30e1e72d50b69266c87048e6a42 () as not reachable. +2067216:S 12 Apr 2026 23:38:07.873 * Marking node 71cce21972c3b30e1e72d50b69266c87048e6a42 () as failing (quorum reached). +2067216:S 12 Apr 2026 23:38:07.915 * Node ceb41590bdb53727bffcb314c80a81989180d13a () reported node 71cce21972c3b30e1e72d50b69266c87048e6a42 () as not reachable. +2067216:S 12 Apr 2026 23:38:07.938 * Clear FAIL state for node 71cce21972c3b30e1e72d50b69266c87048e6a42 (): replica is reachable again. +2067216:S 12 Apr 2026 23:38:08.024 * Node ad56f48bae1934f4d1c3141ffee95fa8bf2fa37c () reported node 71cce21972c3b30e1e72d50b69266c87048e6a42 () is back online. +2067216:S 12 Apr 2026 23:38:08.049 - Accepting cluster node connection from 127.0.0.1:41496 +2067216:S 12 Apr 2026 23:38:08.064 * Node d24296bca8e44a56521bbf138bec235238274ad7 () reported node 71cce21972c3b30e1e72d50b69266c87048e6a42 () is back online. +2067216:S 12 Apr 2026 23:38:08.229 - DB 0: 1 keys (0 volatile) in 4 slots HT. +2067216:S 12 Apr 2026 23:38:08.234 * Node ceb41590bdb53727bffcb314c80a81989180d13a () reported node 71cce21972c3b30e1e72d50b69266c87048e6a42 () is back online. +2067216:S 12 Apr 2026 23:38:13.370 - DB 0: 1 keys (0 volatile) in 4 slots HT. +2067216:S 12 Apr 2026 23:38:16.489 * NODE 71cce21972c3b30e1e72d50b69266c87048e6a42 () possibly failing. +2067216:S 12 Apr 2026 23:38:16.561 * Node d24296bca8e44a56521bbf138bec235238274ad7 () reported node 71cce21972c3b30e1e72d50b69266c87048e6a42 () as not reachable. +2067216:S 12 Apr 2026 23:38:16.561 * Node d24296bca8e44a56521bbf138bec235238274ad7 () reported node ceb41590bdb53727bffcb314c80a81989180d13a () as not reachable. +2067216:S 12 Apr 2026 23:38:16.593 * NODE ceb41590bdb53727bffcb314c80a81989180d13a () possibly failing. +2067216:S 12 Apr 2026 23:38:18.053 * NODE ad56f48bae1934f4d1c3141ffee95fa8bf2fa37c () possibly failing. +2067216:S 12 Apr 2026 23:38:18.053 # Cluster state changed: fail +2067216:S 12 Apr 2026 23:38:18.054 - Accepting cluster node connection from 127.0.0.1:52098 +2067216:S 12 Apr 2026 23:38:18.103 * Node d24296bca8e44a56521bbf138bec235238274ad7 () reported node 71cce21972c3b30e1e72d50b69266c87048e6a42 () is back online. +2067216:S 12 Apr 2026 23:38:18.157 * Node ad56f48bae1934f4d1c3141ffee95fa8bf2fa37c () reported node d24296bca8e44a56521bbf138bec235238274ad7 () as not reachable. +2067216:S 12 Apr 2026 23:38:18.157 * Node ad56f48bae1934f4d1c3141ffee95fa8bf2fa37c () reported node 7cbfdaa3b3c0ebf9bf9116dc86ed8bceffd31db0 () as not reachable. +2067216:S 12 Apr 2026 23:38:18.251 - Accepting cluster node connection from 127.0.0.1:52110 +2067216:S 12 Apr 2026 23:38:18.290 * Node ad56f48bae1934f4d1c3141ffee95fa8bf2fa37c () reported node d24296bca8e44a56521bbf138bec235238274ad7 () is back online. +2067216:S 12 Apr 2026 23:38:18.290 * Node ad56f48bae1934f4d1c3141ffee95fa8bf2fa37c () reported node 7cbfdaa3b3c0ebf9bf9116dc86ed8bceffd31db0 () is back online. +2067216:S 12 Apr 2026 23:38:18.296 * Cluster state changed: ok +2067216:S 12 Apr 2026 23:38:18.394 * Node ad56f48bae1934f4d1c3141ffee95fa8bf2fa37c () reported node ceb41590bdb53727bffcb314c80a81989180d13a () as not reachable. +2067216:S 12 Apr 2026 23:38:18.495 * Node d24296bca8e44a56521bbf138bec235238274ad7 () reported node ceb41590bdb53727bffcb314c80a81989180d13a () is back online. +2067216:S 12 Apr 2026 23:38:19.537 - DB 0: 1 keys (0 volatile) in 4 slots HT. +2067216:S 12 Apr 2026 23:38:20.176 * Node ad56f48bae1934f4d1c3141ffee95fa8bf2fa37c () reported node ceb41590bdb53727bffcb314c80a81989180d13a () is back online. +2067216:S 12 Apr 2026 23:38:24.730 - DB 0: 1 keys (0 volatile) in 4 slots HT. +2067216:S 12 Apr 2026 23:38:26.705 * FAIL message received from ceb41590bdb53727bffcb314c80a81989180d13a () about 7cbfdaa3b3c0ebf9bf9116dc86ed8bceffd31db0 () +2067216:S 12 Apr 2026 23:38:26.752 * Clear FAIL state for node 7cbfdaa3b3c0ebf9bf9116dc86ed8bceffd31db0 (): primary without slots is reachable again. +2067216:S 12 Apr 2026 23:38:26.759 * Node ad56f48bae1934f4d1c3141ffee95fa8bf2fa37c () reported node 7cbfdaa3b3c0ebf9bf9116dc86ed8bceffd31db0 () as not reachable. +2067216:S 12 Apr 2026 23:38:26.774 * Node d24296bca8e44a56521bbf138bec235238274ad7 () reported node 7cbfdaa3b3c0ebf9bf9116dc86ed8bceffd31db0 () as not reachable. +2067216:S 12 Apr 2026 23:38:26.958 * Node ad56f48bae1934f4d1c3141ffee95fa8bf2fa37c () reported node 7cbfdaa3b3c0ebf9bf9116dc86ed8bceffd31db0 () is back online. +2067216:S 12 Apr 2026 23:38:26.961 * Node d24296bca8e44a56521bbf138bec235238274ad7 () reported node 7cbfdaa3b3c0ebf9bf9116dc86ed8bceffd31db0 () is back online. +2067216:S 12 Apr 2026 23:38:31.498 - DB 0: 1 keys (0 volatile) in 4 slots HT. +2067216:S 12 Apr 2026 23:38:35.499 * NODE 71cce21972c3b30e1e72d50b69266c87048e6a42 () possibly failing. +2067216:S 12 Apr 2026 23:38:35.499 * NODE ceb41590bdb53727bffcb314c80a81989180d13a () possibly failing. +2067216:S 12 Apr 2026 23:38:35.499 * NODE d24296bca8e44a56521bbf138bec235238274ad7 () possibly failing. +2067216:S 12 Apr 2026 23:38:35.499 * NODE 7cbfdaa3b3c0ebf9bf9116dc86ed8bceffd31db0 () possibly failing. +2067216:S 12 Apr 2026 23:38:35.499 * NODE ad56f48bae1934f4d1c3141ffee95fa8bf2fa37c () possibly failing. +2067216:S 12 Apr 2026 23:38:35.499 # Cluster state changed: fail +2067216:S 12 Apr 2026 23:38:35.571 * Cluster state changed: ok +2067216:S 12 Apr 2026 23:38:37.343 - DB 0: 1 keys (0 volatile) in 4 slots HT. +2067216:S 12 Apr 2026 23:38:42.549 - DB 0: 1 keys (0 volatile) in 4 slots HT. +2067216:S 12 Apr 2026 23:38:44.683 * NODE d24296bca8e44a56521bbf138bec235238274ad7 () possibly failing. +2067216:S 12 Apr 2026 23:38:44.683 * NODE 7cbfdaa3b3c0ebf9bf9116dc86ed8bceffd31db0 () possibly failing. +2067216:S 12 Apr 2026 23:38:44.801 - Accepting cluster node connection from 127.0.0.1:60498 +2067216:S 12 Apr 2026 23:38:44.804 * Node ad56f48bae1934f4d1c3141ffee95fa8bf2fa37c () reported node 7cbfdaa3b3c0ebf9bf9116dc86ed8bceffd31db0 () as not reachable. +2067216:S 12 Apr 2026 23:38:44.835 * Node ceb41590bdb53727bffcb314c80a81989180d13a () reported node 7cbfdaa3b3c0ebf9bf9116dc86ed8bceffd31db0 () as not reachable. +2067216:S 12 Apr 2026 23:38:44.835 * Marking node 7cbfdaa3b3c0ebf9bf9116dc86ed8bceffd31db0 () as failing (quorum reached). +2067216:S 12 Apr 2026 23:38:44.835 * Node ceb41590bdb53727bffcb314c80a81989180d13a () reported node d24296bca8e44a56521bbf138bec235238274ad7 () as not reachable. +2067216:S 12 Apr 2026 23:38:44.838 * Node d24296bca8e44a56521bbf138bec235238274ad7 () reported node ad56f48bae1934f4d1c3141ffee95fa8bf2fa37c () as not reachable. +2067216:S 12 Apr 2026 23:38:44.838 * Node d24296bca8e44a56521bbf138bec235238274ad7 () reported node 7cbfdaa3b3c0ebf9bf9116dc86ed8bceffd31db0 () as not reachable. +2067216:S 12 Apr 2026 23:38:44.838 * Node d24296bca8e44a56521bbf138bec235238274ad7 () reported node ceb41590bdb53727bffcb314c80a81989180d13a () as not reachable. +2067216:S 12 Apr 2026 23:38:44.893 * Node d24296bca8e44a56521bbf138bec235238274ad7 () reported node ad56f48bae1934f4d1c3141ffee95fa8bf2fa37c () is back online. +2067216:S 12 Apr 2026 23:38:44.894 * Node d24296bca8e44a56521bbf138bec235238274ad7 () reported node ceb41590bdb53727bffcb314c80a81989180d13a () is back online. +2067216:S 12 Apr 2026 23:38:44.895 * Clear FAIL state for node 7cbfdaa3b3c0ebf9bf9116dc86ed8bceffd31db0 (): primary without slots is reachable again. +2067216:S 12 Apr 2026 23:38:44.905 * Node ceb41590bdb53727bffcb314c80a81989180d13a () reported node d24296bca8e44a56521bbf138bec235238274ad7 () is back online. +2067216:S 12 Apr 2026 23:38:44.988 * Node ceb41590bdb53727bffcb314c80a81989180d13a () reported node 7cbfdaa3b3c0ebf9bf9116dc86ed8bceffd31db0 () is back online. +2067216:S 12 Apr 2026 23:38:44.992 * Node ad56f48bae1934f4d1c3141ffee95fa8bf2fa37c () reported node 7cbfdaa3b3c0ebf9bf9116dc86ed8bceffd31db0 () is back online. +2067216:S 12 Apr 2026 23:38:45.005 * Node d24296bca8e44a56521bbf138bec235238274ad7 () reported node 7cbfdaa3b3c0ebf9bf9116dc86ed8bceffd31db0 () is back online. +2067216:S 12 Apr 2026 23:38:48.673 - DB 0: 1 keys (0 volatile) in 4 slots HT. +2067216:S 12 Apr 2026 23:38:54.077 * NODE 71cce21972c3b30e1e72d50b69266c87048e6a42 () possibly failing. +2067216:S 12 Apr 2026 23:38:54.077 * NODE 7cbfdaa3b3c0ebf9bf9116dc86ed8bceffd31db0 () possibly failing. +2067216:S 12 Apr 2026 23:38:54.077 * NODE ad56f48bae1934f4d1c3141ffee95fa8bf2fa37c () possibly failing. +2067216:S 12 Apr 2026 23:38:54.222 - Accepting cluster node connection from 127.0.0.1:60794 +2067216:S 12 Apr 2026 23:38:55.368 - DB 0: 1 keys (0 volatile) in 4 slots HT. +2067216:S 12 Apr 2026 23:39:00.456 - DB 0: 1 keys (0 volatile) in 4 slots HT. +2067216:S 12 Apr 2026 23:39:01.989 * NODE 7cbfdaa3b3c0ebf9bf9116dc86ed8bceffd31db0 () possibly failing. +2067216:S 12 Apr 2026 23:39:02.039 * Node ceb41590bdb53727bffcb314c80a81989180d13a () reported node 7cbfdaa3b3c0ebf9bf9116dc86ed8bceffd31db0 () as not reachable. +2067216:S 12 Apr 2026 23:39:03.152 * NODE d24296bca8e44a56521bbf138bec235238274ad7 () possibly failing. +2067216:S 12 Apr 2026 23:39:03.152 * NODE ad56f48bae1934f4d1c3141ffee95fa8bf2fa37c () possibly failing. +2067216:S 12 Apr 2026 23:39:03.152 # Cluster state changed: fail +2067216:S 12 Apr 2026 23:39:03.184 * FAIL message received from d24296bca8e44a56521bbf138bec235238274ad7 () about 7cbfdaa3b3c0ebf9bf9116dc86ed8bceffd31db0 () +2067216:S 12 Apr 2026 23:39:03.283 * Node ad56f48bae1934f4d1c3141ffee95fa8bf2fa37c () reported node 7cbfdaa3b3c0ebf9bf9116dc86ed8bceffd31db0 () as not reachable. +2067216:S 12 Apr 2026 23:39:03.310 * Clear FAIL state for node 7cbfdaa3b3c0ebf9bf9116dc86ed8bceffd31db0 (): primary without slots is reachable again. +2067216:S 12 Apr 2026 23:39:03.311 * Node ad56f48bae1934f4d1c3141ffee95fa8bf2fa37c () reported node 7cbfdaa3b3c0ebf9bf9116dc86ed8bceffd31db0 () is back online. +2067216:S 12 Apr 2026 23:39:03.311 * Cluster state changed: ok +2067216:S 12 Apr 2026 23:39:03.666 * Node ceb41590bdb53727bffcb314c80a81989180d13a () reported node 7cbfdaa3b3c0ebf9bf9116dc86ed8bceffd31db0 () is back online. +2067216:S 12 Apr 2026 23:39:06.510 - DB 0: 1 keys (0 volatile) in 4 slots HT. +2067216:S 12 Apr 2026 23:39:13.487 - DB 0: 1 keys (0 volatile) in 4 slots HT. +2067216:S 12 Apr 2026 23:39:18.549 - DB 0: 1 keys (0 volatile) in 4 slots HT. +2067216:S 12 Apr 2026 23:39:21.128 - Accepting cluster node connection from 127.0.0.1:42674 +2067216:S 12 Apr 2026 23:39:21.131 - Accepting cluster node connection from 127.0.0.1:42686 +2067216:S 12 Apr 2026 23:39:21.138 - Accepting cluster node connection from 127.0.0.1:42688 +2067216:S 12 Apr 2026 23:39:21.152 * Node ceb41590bdb53727bffcb314c80a81989180d13a () reported node 71cce21972c3b30e1e72d50b69266c87048e6a42 () as not reachable. +2067216:S 12 Apr 2026 23:39:21.152 * Node ceb41590bdb53727bffcb314c80a81989180d13a () reported node 7cbfdaa3b3c0ebf9bf9116dc86ed8bceffd31db0 () as not reachable. +2067216:S 12 Apr 2026 23:39:21.152 * Node ceb41590bdb53727bffcb314c80a81989180d13a () reported node ad56f48bae1934f4d1c3141ffee95fa8bf2fa37c () as not reachable. +2067216:S 12 Apr 2026 23:39:21.152 * Node ceb41590bdb53727bffcb314c80a81989180d13a () reported node d24296bca8e44a56521bbf138bec235238274ad7 () as not reachable. +2067216:S 12 Apr 2026 23:39:21.153 * Node ad56f48bae1934f4d1c3141ffee95fa8bf2fa37c () reported node 71cce21972c3b30e1e72d50b69266c87048e6a42 () as not reachable. +2067216:S 12 Apr 2026 23:39:21.153 * Node ad56f48bae1934f4d1c3141ffee95fa8bf2fa37c () reported node d24296bca8e44a56521bbf138bec235238274ad7 () as not reachable. +2067216:S 12 Apr 2026 23:39:21.153 * Node ad56f48bae1934f4d1c3141ffee95fa8bf2fa37c () reported node ceb41590bdb53727bffcb314c80a81989180d13a () as not reachable. +2067216:S 12 Apr 2026 23:39:21.153 * Node ad56f48bae1934f4d1c3141ffee95fa8bf2fa37c () reported node 7cbfdaa3b3c0ebf9bf9116dc86ed8bceffd31db0 () as not reachable. +2067216:S 12 Apr 2026 23:39:21.163 * FAIL message received from ceb41590bdb53727bffcb314c80a81989180d13a () about 71cce21972c3b30e1e72d50b69266c87048e6a42 () +2067216:S 12 Apr 2026 23:39:21.228 * Node ceb41590bdb53727bffcb314c80a81989180d13a () reported node ad56f48bae1934f4d1c3141ffee95fa8bf2fa37c () is back online. +2067216:S 12 Apr 2026 23:39:21.228 * Node ceb41590bdb53727bffcb314c80a81989180d13a () reported node d24296bca8e44a56521bbf138bec235238274ad7 () is back online. +2067216:S 12 Apr 2026 23:39:21.228 * Node ceb41590bdb53727bffcb314c80a81989180d13a () reported node 71cce21972c3b30e1e72d50b69266c87048e6a42 () is back online. +2067216:S 12 Apr 2026 23:39:21.228 * Node ad56f48bae1934f4d1c3141ffee95fa8bf2fa37c () reported node d24296bca8e44a56521bbf138bec235238274ad7 () is back online. +2067216:S 12 Apr 2026 23:39:21.228 * Node ad56f48bae1934f4d1c3141ffee95fa8bf2fa37c () reported node 7cbfdaa3b3c0ebf9bf9116dc86ed8bceffd31db0 () is back online. +2067216:S 12 Apr 2026 23:39:21.228 * Node ad56f48bae1934f4d1c3141ffee95fa8bf2fa37c () reported node ceb41590bdb53727bffcb314c80a81989180d13a () is back online. +2067216:S 12 Apr 2026 23:39:21.228 * Node d24296bca8e44a56521bbf138bec235238274ad7 () reported node 71cce21972c3b30e1e72d50b69266c87048e6a42 () as not reachable. +2067216:S 12 Apr 2026 23:39:21.228 * Clear FAIL state for node 71cce21972c3b30e1e72d50b69266c87048e6a42 (): replica is reachable again. +2067216:S 12 Apr 2026 23:39:21.430 * Node ceb41590bdb53727bffcb314c80a81989180d13a () reported node 7cbfdaa3b3c0ebf9bf9116dc86ed8bceffd31db0 () is back online. +2067216:S 12 Apr 2026 23:39:21.431 * Node ad56f48bae1934f4d1c3141ffee95fa8bf2fa37c () reported node 71cce21972c3b30e1e72d50b69266c87048e6a42 () is back online. +2067216:S 12 Apr 2026 23:39:21.538 * Node d24296bca8e44a56521bbf138bec235238274ad7 () reported node 71cce21972c3b30e1e72d50b69266c87048e6a42 () is back online. +===== End of server log (pid 2067216) ===== + + +===== Start of server stderr log (pid 2067216) ===== + + +===== End of server stderr log (pid 2067216) ===== + + +===== Start of server log (pid 2067052) ===== + +### Starting server for test +2067052:C 12 Apr 2026 23:36:26.721 # WARNING: Changing databases number from 16 to 1 since we are in cluster mode +2067052:C 12 Apr 2026 23:36:26.721 # WARNING Memory overcommit must be enabled! Without it, a background save or replication may fail under low memory condition. Being disabled, it can also cause failures without low memory condition, see https://github.com/jemalloc/jemalloc/issues/1328. To fix this issue add 'vm.overcommit_memory = 1' to /etc/sysctl.conf and then reboot or run the command 'sysctl vm.overcommit_memory=1' for this to take effect. +2067052:C 12 Apr 2026 23:36:26.721 * oO0OoO0OoO0Oo Valkey is starting oO0OoO0OoO0Oo +2067052:C 12 Apr 2026 23:36:26.721 * Valkey version=8.0.2, bits=64, commit=00000000, modified=0, pid=2067052, just started +2067052:C 12 Apr 2026 23:36:26.721 * Configuration loaded +2067052:M 12 Apr 2026 23:36:26.722 * Increased maximum number of open files to 10032 (it was originally set to 1024). +2067052:M 12 Apr 2026 23:36:26.722 * monotonic clock: POSIX clock_gettime +2067052:M 12 Apr 2026 23:36:26.723 # Failed to write PID file: Permission denied + .+^+. + .+#########+. + .+########+########+. Valkey 8.0.2 (00000000/0) 64 bit + .+########+' '+########+. + .########+' .+. '+########. Running in cluster mode + |####+' .+#######+. '+####| Port: 21698 + |###| .+###############+. |###| PID: 2067052 + |###| |#####*'' ''*#####| |###| + |###| |####' .-. '####| |###| + |###| |###( (@@@) )###| |###| https://valkey.io + |###| |####. '-' .####| |###| + |###| |#####*. .*#####| |###| + |###| '+#####| |#####+' |###| + |####+. +##| |#+' .+####| + '#######+ |##| .+########' + '+###| |##| .+########+' + '| |####+########+' + +#########+' + '+v+' + +2067052:M 12 Apr 2026 23:36:26.724 * No cluster configuration found, I'm ad56f48bae1934f4d1c3141ffee95fa8bf2fa37c +2067052:M 12 Apr 2026 23:36:26.755 * Server initialized +2067052:M 12 Apr 2026 23:36:26.755 * Ready to accept connections tcp +2067052:M 12 Apr 2026 23:36:26.755 * Ready to accept connections unix +2067052:M 12 Apr 2026 23:36:26.755 * Ready to accept connections tls +2067052:M 12 Apr 2026 23:36:26.856 - Accepted 127.0.0.1:35299 +2067052:M 12 Apr 2026 23:36:26.886 - Reading from client: error:0A000126:SSL routines::unexpected eof while reading +2067052:M 12 Apr 2026 23:36:26.908 - Accepted 127.0.0.1:39123 +2067052:M 12 Apr 2026 23:36:27.365 - Accepting cluster node connection from 127.0.0.1:51840 +2067052:M 12 Apr 2026 23:36:27.411 * IP address for this node updated to 127.0.0.1 +2067052:M 12 Apr 2026 23:36:27.518 - Accepting cluster node connection from 127.0.0.1:51842 +2067052:M 12 Apr 2026 23:36:27.569 - Accepting cluster node connection from 127.0.0.1:51858 +2067052:M 12 Apr 2026 23:36:27.581 - Accepting cluster node connection from 127.0.0.1:51866 +2067052:M 12 Apr 2026 23:36:27.616 - Handshake: we already know node ceb41590bdb53727bffcb314c80a81989180d13a (), updating the address if needed. +2067052:M 12 Apr 2026 23:36:27.617 - Accepting cluster node connection from 127.0.0.1:51872 +2067052:M 12 Apr 2026 23:36:27.617 - Error accepting cluster node connection: error:0A000126:SSL routines::unexpected eof while reading +2067052:M 12 Apr 2026 23:36:27.622 - Accepting cluster node connection from 127.0.0.1:51874 +2067052:M 12 Apr 2026 23:36:27.639 * configEpoch collision with node ceb41590bdb53727bffcb314c80a81989180d13a (). configEpoch set to 2 +2067052:M 12 Apr 2026 23:36:27.640 * Successfully completed handshake with 71cce21972c3b30e1e72d50b69266c87048e6a42 () +2067052:M 12 Apr 2026 23:36:27.646 * Successfully completed handshake with 7cbfdaa3b3c0ebf9bf9116dc86ed8bceffd31db0 () +2067052:M 12 Apr 2026 23:36:27.646 * Address updated for node 7cbfdaa3b3c0ebf9bf9116dc86ed8bceffd31db0 (), now 127.0.0.1:21691 +2067052:M 12 Apr 2026 23:36:27.723 - Accepting cluster node connection from 127.0.0.1:51878 +2067052:M 12 Apr 2026 23:36:27.726 - Accepting cluster node connection from 127.0.0.1:51894 +2067052:M 12 Apr 2026 23:36:27.753 * Address updated for node 71cce21972c3b30e1e72d50b69266c87048e6a42 (), now 127.0.0.1:21689 +2067052:M 12 Apr 2026 23:36:27.786 - Accepting cluster node connection from 127.0.0.1:51898 +2067052:M 12 Apr 2026 23:36:27.956 * Node ceb41590bdb53727bffcb314c80a81989180d13a () is no longer primary of shard 7be85fc7a35a355988332b0ade6bb85642f90b76; removed all 0 slot(s) it used to own +2067052:M 12 Apr 2026 23:36:27.956 * Node ceb41590bdb53727bffcb314c80a81989180d13a () is now part of shard 0ccab0ca3bdf2df4fb1d88a6388f0d2e2e9492d9 +2067052:M 12 Apr 2026 23:36:27.956 * Node ceb41590bdb53727bffcb314c80a81989180d13a () is now a replica of node 27c281c71471f64d57c2f90a6fce7ca7d11f3a9e () in shard 0ccab0ca3bdf2df4fb1d88a6388f0d2e2e9492d9 +2067052:M 12 Apr 2026 23:36:27.965 - Accepted 127.0.0.1:47628 +2067052:M 12 Apr 2026 23:36:27.968 * Node 7cbfdaa3b3c0ebf9bf9116dc86ed8bceffd31db0 () is no longer primary of shard 6d5a08da00fb925618c9429f0a0abf50cb184dee; removed all 0 slot(s) it used to own +2067052:M 12 Apr 2026 23:36:27.968 * Node 7cbfdaa3b3c0ebf9bf9116dc86ed8bceffd31db0 () is now part of shard fb1332a38fbff59dddf699f62b6c363413af8689 +2067052:M 12 Apr 2026 23:36:27.968 * Node 7cbfdaa3b3c0ebf9bf9116dc86ed8bceffd31db0 () is now a replica of node ad56f48bae1934f4d1c3141ffee95fa8bf2fa37c () in shard fb1332a38fbff59dddf699f62b6c363413af8689 +2067052:M 12 Apr 2026 23:36:27.988 * Node 71cce21972c3b30e1e72d50b69266c87048e6a42 () is no longer primary of shard 797d84842d0558c353ef2fd200fec2d129d7c96f; removed all 0 slot(s) it used to own +2067052:M 12 Apr 2026 23:36:27.989 * Node 71cce21972c3b30e1e72d50b69266c87048e6a42 () is now part of shard b553502a57202365d35122f1f841b02047c3c23d +2067052:M 12 Apr 2026 23:36:27.989 * Node 71cce21972c3b30e1e72d50b69266c87048e6a42 () is now a replica of node d24296bca8e44a56521bbf138bec235238274ad7 () in shard b553502a57202365d35122f1f841b02047c3c23d +2067052:M 12 Apr 2026 23:36:27.990 # DEBUG LOG: ========== I am primary 1 ========== +2067052:M 12 Apr 2026 23:36:28.008 * Replica 127.0.0.1:21691 asks for synchronization +2067052:M 12 Apr 2026 23:36:28.008 * Full resync requested by replica 127.0.0.1:21691 +2067052:M 12 Apr 2026 23:36:28.008 * Replication backlog created, my new replication IDs are '950dbe04066d7028f8350da2375aa7b0709610d8' and '0000000000000000000000000000000000000000' +2067052:M 12 Apr 2026 23:36:28.008 * Starting BGSAVE for SYNC with target: replicas sockets using: normal sync +2067052:M 12 Apr 2026 23:36:28.009 * Background RDB transfer started by pid 2067759 to pipe through parent process +2067759:C 12 Apr 2026 23:36:28.018 * Fork CoW for RDB: current 0 MB, peak 0 MB, average 0 MB +2067052:M 12 Apr 2026 23:36:28.019 * Diskless rdb transfer, done reading from pipe, 1 replicas still up. +2067052:M 12 Apr 2026 23:36:28.035 * Background RDB transfer terminated with success +2067052:M 12 Apr 2026 23:36:28.035 * Streamed RDB transfer with replica 127.0.0.1:21691 succeeded (socket). Waiting for REPLCONF ACK from replica to enable streaming +2067052:M 12 Apr 2026 23:36:28.035 * Synchronization with replica 127.0.0.1:21691 succeeded +2067052:M 12 Apr 2026 23:36:28.804 * Cluster state changed: ok +2067052:M 12 Apr 2026 23:36:35.639 * NODE 71cce21972c3b30e1e72d50b69266c87048e6a42 () possibly failing. +### Starting test Slot migration states are replicated in tests/unit/cluster/slot-migration.tcl +2067052:M 12 Apr 2026 23:36:37.793 * Importing slot 609 from node 27c281c71471f64d57c2f90a6fce7ca7d11f3a9e () +### Starting test Migration target is auto-updated after failover in target shard in tests/unit/cluster/slot-migration.tcl +2067052:M 12 Apr 2026 23:36:40.916 - Accepting cluster node connection from 127.0.0.1:60168 +2067052:M 12 Apr 2026 23:36:40.918 - Error accepting cluster node connection: error:0A000126:SSL routines::unexpected eof while reading +2067052:M 12 Apr 2026 23:36:40.919 - Accepting cluster node connection from 127.0.0.1:60172 +2067052:M 12 Apr 2026 23:36:40.955 - Error accepting cluster node connection: error:0A000126:SSL routines::unexpected eof while reading +2067052:M 12 Apr 2026 23:36:40.969 - Accepting cluster node connection from 127.0.0.1:60174 +2067052:M 12 Apr 2026 23:36:40.972 - Error accepting cluster node connection: error:0A000126:SSL routines::unexpected eof while reading +2067052:M 12 Apr 2026 23:36:40.972 - Client closed connection id=11 addr=127.0.0.1:47628 laddr=127.0.0.1:21697 fd=24 name= age=13 idle=0 flags=S db=0 sub=0 psub=0 ssub=0 multi=-1 watch=0 qbuf=0 qbuf-free=20474 argv-mem=0 multi-mem=0 rbs=1024 rbp=0 obl=0 oll=1 omem=20504 tot-mem=42904 events=r cmd=replconf user=default redir=-1 resp=2 lib-name= lib-ver= tot-net-in=678 tot-net-out=197 tot-cmds=19 +2067052:M 12 Apr 2026 23:36:40.972 * Connection with replica 127.0.0.1:21691 lost. +2067052:M 12 Apr 2026 23:36:40.981 * Configuration change detected. Reconfiguring myself as a replica of node 7cbfdaa3b3c0ebf9bf9116dc86ed8bceffd31db0 () in shard fb1332a38fbff59dddf699f62b6c363413af8689 +2067052:S 12 Apr 2026 23:36:40.981 * Before turning into a replica, using my own primary parameters to synthesize a cached primary: I may be able to synchronize with the new primary with just a partial transfer. +2067052:S 12 Apr 2026 23:36:40.981 * Connecting to PRIMARY 127.0.0.1:21691 +2067052:S 12 Apr 2026 23:36:40.981 * PRIMARY <-> REPLICA sync started +2067052:S 12 Apr 2026 23:36:41.010 - Accepting cluster node connection from 127.0.0.1:60184 +2067052:S 12 Apr 2026 23:36:41.013 - Error accepting cluster node connection: error:0A000126:SSL routines::unexpected eof while reading +2067052:S 12 Apr 2026 23:36:41.021 - Accepting cluster node connection from 127.0.0.1:60190 +2067052:S 12 Apr 2026 23:36:41.033 - Error accepting cluster node connection: error:0A000126:SSL routines::unexpected eof while reading +2067052:S 12 Apr 2026 23:36:41.033 - Accepting cluster node connection from 127.0.0.1:60194 +2067052:S 12 Apr 2026 23:36:41.036 - Error accepting cluster node connection: error:0A000126:SSL routines::unexpected eof while reading +2067052:S 12 Apr 2026 23:36:41.056 * Non blocking connect for SYNC fired the event. +2067052:S 12 Apr 2026 23:36:41.059 - Accepting cluster node connection from 127.0.0.1:60210 +2067052:S 12 Apr 2026 23:36:41.061 - Error accepting cluster node connection: error:0A000126:SSL routines::unexpected eof while reading +2067052:S 12 Apr 2026 23:36:41.069 * Primary replied to PING, replication can continue... +2067052:S 12 Apr 2026 23:36:41.069 - Accepting cluster node connection from 127.0.0.1:60212 +2067052:S 12 Apr 2026 23:36:41.072 - Error accepting cluster node connection: error:0A000126:SSL routines::unexpected eof while reading +2067052:S 12 Apr 2026 23:36:41.077 - Accepting cluster node connection from 127.0.0.1:60220 +2067052:S 12 Apr 2026 23:36:41.083 - Error accepting cluster node connection: error:0A000126:SSL routines::unexpected eof while reading +2067052:S 12 Apr 2026 23:36:41.083 - Accepting cluster node connection from 127.0.0.1:60234 +2067052:S 12 Apr 2026 23:36:41.090 - Error accepting cluster node connection: error:0A000126:SSL routines::unexpected eof while reading +2067052:S 12 Apr 2026 23:36:41.091 * Trying a partial resynchronization (request 950dbe04066d7028f8350da2375aa7b0709610d8:176). +2067052:S 12 Apr 2026 23:36:41.094 - Accepting cluster node connection from 127.0.0.1:60236 +2067052:S 12 Apr 2026 23:36:41.109 - Accepting cluster node connection from 127.0.0.1:60248 +2067052:S 12 Apr 2026 23:36:41.122 - Accepting cluster node connection from 127.0.0.1:60250 +2067052:S 12 Apr 2026 23:36:41.137 - Accepting cluster node connection from 127.0.0.1:60266 +2067052:S 12 Apr 2026 23:36:41.146 - Accepting cluster node connection from 127.0.0.1:60280 +2067052:S 12 Apr 2026 23:36:41.161 * Successful partial resynchronization with primary. +2067052:S 12 Apr 2026 23:36:41.162 * Primary replication ID changed to 04bc904ae99c9565f5e08319325b2bf2948666fc +2067052:S 12 Apr 2026 23:36:41.162 * PRIMARY <-> REPLICA sync: Primary accepted a Partial Resynchronization. +2067052:S 12 Apr 2026 23:36:41.292 * Manual failover user request accepted (user request from 'id=4 addr=127.0.0.1:39123 laddr=127.0.0.1:21697 fd=13 name= age=15 idle=0 flags=N db=0 sub=0 psub=0 ssub=0 multi=-1 watch=0 qbuf=0 qbuf-free=0 argv-mem=15 multi-mem=0 rbs=4096 rbp=2048 obl=0 oll=0 omem=0 tot-mem=5023 events=r cmd=cluster|failover user=default redir=-1 resp=2 lib-name= lib-ver= tot-net-in=1189 tot-net-out=35719 tot-cmds=39'). +2067052:S 12 Apr 2026 23:36:41.292 * Received replication offset for paused primary manual failover: 175 +2067052:S 12 Apr 2026 23:36:41.292 * All primary replication stream processed, manual failover can start. +2067052:S 12 Apr 2026 23:36:41.292 * Start of election delayed for 0 milliseconds (rank #0, offset 175). +2067052:S 12 Apr 2026 23:36:41.292 * Starting a failover election for epoch 7. +2067052:S 12 Apr 2026 23:36:41.309 * Failover election won: I'm the new primary. +2067052:S 12 Apr 2026 23:36:41.309 * configEpoch set to 7 after successful failover +2067052:M 12 Apr 2026 23:36:41.309 * Connection with primary lost. +2067052:M 12 Apr 2026 23:36:41.309 * Caching the disconnected primary state. +2067052:M 12 Apr 2026 23:36:41.310 * Discarding previously cached primary state. +2067052:M 12 Apr 2026 23:36:41.310 * Setting secondary replication ID to 04bc904ae99c9565f5e08319325b2bf2948666fc, valid up to offset: 176. New replication ID is 823749a7ecd27acd70ba1ca1af68e1d71a98c40e +2067052:M 12 Apr 2026 23:36:41.335 - Node 7cbfdaa3b3c0ebf9bf9116dc86ed8bceffd31db0 has old slots configuration, sending an UPDATE message about ad56f48bae1934f4d1c3141ffee95fa8bf2fa37c +2067052:M 12 Apr 2026 23:36:41.345 - Accepted 127.0.0.1:40762 +2067052:M 12 Apr 2026 23:36:41.386 * A failover occurred in shard fb1332a38fbff59dddf699f62b6c363413af8689; node 7cbfdaa3b3c0ebf9bf9116dc86ed8bceffd31db0 () failed over to node ad56f48bae1934f4d1c3141ffee95fa8bf2fa37c () with a config epoch of 7 +2067052:M 12 Apr 2026 23:36:41.386 * Node 7cbfdaa3b3c0ebf9bf9116dc86ed8bceffd31db0 () is now a replica of node ad56f48bae1934f4d1c3141ffee95fa8bf2fa37c () in shard fb1332a38fbff59dddf699f62b6c363413af8689 +2067052:M 12 Apr 2026 23:36:41.397 * Replica 127.0.0.1:21691 asks for synchronization +2067052:M 12 Apr 2026 23:36:41.397 * Partial resynchronization request from 127.0.0.1:21691 accepted. Sending 0 bytes of backlog starting from offset 176. +### Starting test Migration source is auto-updated after failover in source shard in tests/unit/cluster/slot-migration.tcl +2067052:M 12 Apr 2026 23:36:42.711 * NODE 27c281c71471f64d57c2f90a6fce7ca7d11f3a9e () possibly failing. +2067052:M 12 Apr 2026 23:36:42.925 * Node d24296bca8e44a56521bbf138bec235238274ad7 () reported node 27c281c71471f64d57c2f90a6fce7ca7d11f3a9e () as not reachable. +2067052:M 12 Apr 2026 23:36:42.925 * Marking node 27c281c71471f64d57c2f90a6fce7ca7d11f3a9e () as failing (quorum reached). +2067052:M 12 Apr 2026 23:36:42.925 # Cluster state changed: fail +2067052:M 12 Apr 2026 23:36:45.746 * Clear FAIL state for node 27c281c71471f64d57c2f90a6fce7ca7d11f3a9e (): is reachable again and nobody is serving its slots after some time. +2067052:M 12 Apr 2026 23:36:45.747 * Cluster state changed: ok +2067052:M 12 Apr 2026 23:36:46.111 * Node d24296bca8e44a56521bbf138bec235238274ad7 () reported node 27c281c71471f64d57c2f90a6fce7ca7d11f3a9e () is back online. +2067052:M 12 Apr 2026 23:36:46.727 * NODE ceb41590bdb53727bffcb314c80a81989180d13a () possibly failing. +2067052:M 12 Apr 2026 23:36:46.905 * Node d24296bca8e44a56521bbf138bec235238274ad7 () reported node ceb41590bdb53727bffcb314c80a81989180d13a () as not reachable. +2067052:M 12 Apr 2026 23:36:46.905 * Marking node ceb41590bdb53727bffcb314c80a81989180d13a () as failing (quorum reached). +2067052:M 12 Apr 2026 23:36:47.121 * Node 27c281c71471f64d57c2f90a6fce7ca7d11f3a9e () reported node ceb41590bdb53727bffcb314c80a81989180d13a () as not reachable. +2067052:M 12 Apr 2026 23:36:47.227 # Failover auth denied to ceb41590bdb53727bffcb314c80a81989180d13a () for epoch 8: its primary is up +2067052:M 12 Apr 2026 23:36:47.272 * Clear FAIL state for node ceb41590bdb53727bffcb314c80a81989180d13a (): replica is reachable again. +2067052:M 12 Apr 2026 23:36:47.398 * Node d24296bca8e44a56521bbf138bec235238274ad7 () reported node ceb41590bdb53727bffcb314c80a81989180d13a () is back online. +2067052:M 12 Apr 2026 23:36:47.398 * Node 27c281c71471f64d57c2f90a6fce7ca7d11f3a9e () reported node ceb41590bdb53727bffcb314c80a81989180d13a () is back online. +2067052:M 12 Apr 2026 23:36:53.760 * NODE 7cbfdaa3b3c0ebf9bf9116dc86ed8bceffd31db0 () possibly failing. +2067052:M 12 Apr 2026 23:36:53.999 * FAIL message received from d24296bca8e44a56521bbf138bec235238274ad7 () about 7cbfdaa3b3c0ebf9bf9116dc86ed8bceffd31db0 () +2067052:M 12 Apr 2026 23:36:54.216 * Node d24296bca8e44a56521bbf138bec235238274ad7 () reported node 7cbfdaa3b3c0ebf9bf9116dc86ed8bceffd31db0 () as not reachable. +2067052:M 12 Apr 2026 23:36:54.216 * Node 27c281c71471f64d57c2f90a6fce7ca7d11f3a9e () reported node 7cbfdaa3b3c0ebf9bf9116dc86ed8bceffd31db0 () as not reachable. +2067052:M 12 Apr 2026 23:36:54.227 * Clear FAIL state for node 7cbfdaa3b3c0ebf9bf9116dc86ed8bceffd31db0 (): replica is reachable again. +2067052:M 12 Apr 2026 23:36:54.328 * Node d24296bca8e44a56521bbf138bec235238274ad7 () reported node 7cbfdaa3b3c0ebf9bf9116dc86ed8bceffd31db0 () is back online. +2067052:M 12 Apr 2026 23:36:54.361 * Node 27c281c71471f64d57c2f90a6fce7ca7d11f3a9e () reported node 7cbfdaa3b3c0ebf9bf9116dc86ed8bceffd31db0 () is back online. +### Starting test Replica redirects key access in migrating slots in tests/unit/cluster/slot-migration.tcl +### Starting test Replica of migrating node returns ASK redirect after READONLY in tests/unit/cluster/slot-migration.tcl +### Starting test Replica of migrating node returns TRYAGAIN after READONLY in tests/unit/cluster/slot-migration.tcl +### Starting test Replica of importing node returns TRYAGAIN after READONLY and ASKING in tests/unit/cluster/slot-migration.tcl +### Starting test New replica inherits migrating slot in tests/unit/cluster/slot-migration.tcl +2067052:M 12 Apr 2026 23:36:59.045 * NODE ceb41590bdb53727bffcb314c80a81989180d13a () possibly failing. +2067052:M 12 Apr 2026 23:36:59.407 * Node d24296bca8e44a56521bbf138bec235238274ad7 () reported node ceb41590bdb53727bffcb314c80a81989180d13a () as not reachable. +2067052:M 12 Apr 2026 23:36:59.407 * Marking node ceb41590bdb53727bffcb314c80a81989180d13a () as failing (quorum reached). +2067052:M 12 Apr 2026 23:36:59.633 * Node 27c281c71471f64d57c2f90a6fce7ca7d11f3a9e () reported node ceb41590bdb53727bffcb314c80a81989180d13a () as not reachable. +2067052:M 12 Apr 2026 23:37:00.763 * Clear FAIL state for node ceb41590bdb53727bffcb314c80a81989180d13a (): replica is reachable again. +2067052:M 12 Apr 2026 23:37:00.956 * Node 27c281c71471f64d57c2f90a6fce7ca7d11f3a9e () reported node ceb41590bdb53727bffcb314c80a81989180d13a () is back online. +2067052:M 12 Apr 2026 23:37:01.058 * Node d24296bca8e44a56521bbf138bec235238274ad7 () reported node ceb41590bdb53727bffcb314c80a81989180d13a () is back online. +2067052:M 12 Apr 2026 23:37:01.647 - Accepting cluster node connection from 127.0.0.1:46712 +2067052:M 12 Apr 2026 23:37:02.179 * NODE ceb41590bdb53727bffcb314c80a81989180d13a () possibly failing. +2067052:M 12 Apr 2026 23:37:02.179 * Node 27c281c71471f64d57c2f90a6fce7ca7d11f3a9e () reported node ceb41590bdb53727bffcb314c80a81989180d13a () as not reachable. +2067052:M 12 Apr 2026 23:37:02.180 * Marking node ceb41590bdb53727bffcb314c80a81989180d13a () as failing (quorum reached). +2067052:M 12 Apr 2026 23:37:03.737 * Node d24296bca8e44a56521bbf138bec235238274ad7 () reported node ceb41590bdb53727bffcb314c80a81989180d13a () as not reachable. +2067052:M 12 Apr 2026 23:37:03.839 - Accepting cluster node connection from 127.0.0.1:42430 +2067052:M 12 Apr 2026 23:37:04.263 - Accepting cluster node connection from 127.0.0.1:42442 +2067052:M 12 Apr 2026 23:37:04.290 - Accepting cluster node connection from 127.0.0.1:42450 +2067052:M 12 Apr 2026 23:37:04.320 - Error accepting cluster node connection: error:0A000126:SSL routines::unexpected eof while reading +2067052:M 12 Apr 2026 23:37:04.430 * Node 27c281c71471f64d57c2f90a6fce7ca7d11f3a9e () reported node ceb41590bdb53727bffcb314c80a81989180d13a () is back online. +2067052:M 12 Apr 2026 23:37:04.437 - Accepting cluster node connection from 127.0.0.1:42458 +2067052:M 12 Apr 2026 23:37:04.439 * Clear FAIL state for node ceb41590bdb53727bffcb314c80a81989180d13a (): primary without slots is reachable again. +2067052:M 12 Apr 2026 23:37:05.034 * NODE 71cce21972c3b30e1e72d50b69266c87048e6a42 () possibly failing. +2067052:M 12 Apr 2026 23:37:05.034 * NODE d24296bca8e44a56521bbf138bec235238274ad7 () possibly failing. +2067052:M 12 Apr 2026 23:37:06.433 * NODE ceb41590bdb53727bffcb314c80a81989180d13a () possibly failing. +2067052:M 12 Apr 2026 23:37:06.434 * Marking node ceb41590bdb53727bffcb314c80a81989180d13a () as failing (quorum reached). +2067052:M 12 Apr 2026 23:37:06.901 * Node 27c281c71471f64d57c2f90a6fce7ca7d11f3a9e () reported node ceb41590bdb53727bffcb314c80a81989180d13a () as not reachable. +2067052:M 12 Apr 2026 23:37:08.575 * Clear FAIL state for node ceb41590bdb53727bffcb314c80a81989180d13a (): primary without slots is reachable again. +2067052:M 12 Apr 2026 23:37:08.789 * Node 27c281c71471f64d57c2f90a6fce7ca7d11f3a9e () reported node ceb41590bdb53727bffcb314c80a81989180d13a () is back online. +2067052:M 12 Apr 2026 23:37:08.789 * Node d24296bca8e44a56521bbf138bec235238274ad7 () reported node ceb41590bdb53727bffcb314c80a81989180d13a () is back online. +2067052:M 12 Apr 2026 23:37:09.064 * A failover occurred in shard 0ccab0ca3bdf2df4fb1d88a6388f0d2e2e9492d9; node ceb41590bdb53727bffcb314c80a81989180d13a () failed over to node 27c281c71471f64d57c2f90a6fce7ca7d11f3a9e () with a config epoch of 5 +2067052:M 12 Apr 2026 23:37:09.064 * Node ceb41590bdb53727bffcb314c80a81989180d13a () is now a replica of node 27c281c71471f64d57c2f90a6fce7ca7d11f3a9e () in shard 0ccab0ca3bdf2df4fb1d88a6388f0d2e2e9492d9 +### Starting test New replica inherits importing slot in tests/unit/cluster/slot-migration.tcl +2067052:M 12 Apr 2026 23:37:10.538 - Client closed connection id=15 addr=127.0.0.1:40762 laddr=127.0.0.1:21697 fd=14 name= age=29 idle=1 flags=S db=0 sub=0 psub=0 ssub=0 multi=-1 watch=0 qbuf=0 qbuf-free=20474 argv-mem=0 multi-mem=0 rbs=1024 rbp=0 obl=0 oll=1 omem=20504 tot-mem=42904 events=r cmd=replconf user=default redir=-1 resp=2 lib-name= lib-ver= tot-net-in=992 tot-net-out=50 tot-cmds=26 +2067052:M 12 Apr 2026 23:37:10.538 * Connection with replica 127.0.0.1:21691 lost. +2067052:M 12 Apr 2026 23:37:12.544 - Accepting cluster node connection from 127.0.0.1:42474 +2067052:M 12 Apr 2026 23:37:12.594 - Error accepting cluster node connection: error:0A000126:SSL routines::unexpected eof while reading +2067052:M 12 Apr 2026 23:37:12.645 * NODE 71cce21972c3b30e1e72d50b69266c87048e6a42 () possibly failing. +2067052:M 12 Apr 2026 23:37:12.645 * NODE 7cbfdaa3b3c0ebf9bf9116dc86ed8bceffd31db0 () possibly failing. +2067052:M 12 Apr 2026 23:37:12.758 * Node d24296bca8e44a56521bbf138bec235238274ad7 () reported node 71cce21972c3b30e1e72d50b69266c87048e6a42 () as not reachable. +2067052:M 12 Apr 2026 23:37:12.758 * Marking node 71cce21972c3b30e1e72d50b69266c87048e6a42 () as failing (quorum reached). +2067052:M 12 Apr 2026 23:37:12.758 * Node d24296bca8e44a56521bbf138bec235238274ad7 () reported node 7cbfdaa3b3c0ebf9bf9116dc86ed8bceffd31db0 () as not reachable. +2067052:M 12 Apr 2026 23:37:12.758 * Marking node 7cbfdaa3b3c0ebf9bf9116dc86ed8bceffd31db0 () as failing (quorum reached). +2067052:M 12 Apr 2026 23:37:12.796 * Clear FAIL state for node 7cbfdaa3b3c0ebf9bf9116dc86ed8bceffd31db0 (): replica is reachable again. +2067052:M 12 Apr 2026 23:37:12.857 * FAIL message received from 27c281c71471f64d57c2f90a6fce7ca7d11f3a9e () about 7cbfdaa3b3c0ebf9bf9116dc86ed8bceffd31db0 () +2067052:M 12 Apr 2026 23:37:12.858 * Node 27c281c71471f64d57c2f90a6fce7ca7d11f3a9e () reported node 7cbfdaa3b3c0ebf9bf9116dc86ed8bceffd31db0 () as not reachable. +2067052:M 12 Apr 2026 23:37:13.019 * Node 27c281c71471f64d57c2f90a6fce7ca7d11f3a9e () reported node 71cce21972c3b30e1e72d50b69266c87048e6a42 () as not reachable. +2067052:M 12 Apr 2026 23:37:13.021 * Node d24296bca8e44a56521bbf138bec235238274ad7 () reported node 71cce21972c3b30e1e72d50b69266c87048e6a42 () is back online. +2067052:M 12 Apr 2026 23:37:13.021 * Clear FAIL state for node 71cce21972c3b30e1e72d50b69266c87048e6a42 (): replica is reachable again. +2067052:M 12 Apr 2026 23:37:13.021 * Clear FAIL state for node 7cbfdaa3b3c0ebf9bf9116dc86ed8bceffd31db0 (): primary without slots is reachable again. +2067052:M 12 Apr 2026 23:37:13.195 * Node 27c281c71471f64d57c2f90a6fce7ca7d11f3a9e () reported node 7cbfdaa3b3c0ebf9bf9116dc86ed8bceffd31db0 () is back online. +2067052:M 12 Apr 2026 23:37:13.195 * Node 27c281c71471f64d57c2f90a6fce7ca7d11f3a9e () reported node 71cce21972c3b30e1e72d50b69266c87048e6a42 () is back online. +2067052:M 12 Apr 2026 23:37:13.196 * Node d24296bca8e44a56521bbf138bec235238274ad7 () reported node 7cbfdaa3b3c0ebf9bf9116dc86ed8bceffd31db0 () is back online. +2067052:M 12 Apr 2026 23:37:22.549 - Accepting cluster node connection from 127.0.0.1:58872 +2067052:M 12 Apr 2026 23:37:22.655 * Node 27c281c71471f64d57c2f90a6fce7ca7d11f3a9e () reported node 71cce21972c3b30e1e72d50b69266c87048e6a42 () as not reachable. +2067052:M 12 Apr 2026 23:37:22.655 * Node 27c281c71471f64d57c2f90a6fce7ca7d11f3a9e () reported node ceb41590bdb53727bffcb314c80a81989180d13a () as not reachable. +2067052:M 12 Apr 2026 23:37:22.655 * Node 27c281c71471f64d57c2f90a6fce7ca7d11f3a9e () reported node d24296bca8e44a56521bbf138bec235238274ad7 () as not reachable. +2067052:M 12 Apr 2026 23:37:22.655 * Node 27c281c71471f64d57c2f90a6fce7ca7d11f3a9e () reported node 7cbfdaa3b3c0ebf9bf9116dc86ed8bceffd31db0 () as not reachable. +2067052:M 12 Apr 2026 23:37:22.666 * Node 27c281c71471f64d57c2f90a6fce7ca7d11f3a9e () reported node ceb41590bdb53727bffcb314c80a81989180d13a () is back online. +2067052:M 12 Apr 2026 23:37:22.839 * Node 27c281c71471f64d57c2f90a6fce7ca7d11f3a9e () reported node 71cce21972c3b30e1e72d50b69266c87048e6a42 () is back online. +2067052:M 12 Apr 2026 23:37:22.839 * Node 27c281c71471f64d57c2f90a6fce7ca7d11f3a9e () reported node d24296bca8e44a56521bbf138bec235238274ad7 () is back online. +2067052:M 12 Apr 2026 23:37:22.855 * Node 27c281c71471f64d57c2f90a6fce7ca7d11f3a9e () reported node 7cbfdaa3b3c0ebf9bf9116dc86ed8bceffd31db0 () is back online. +2067052:M 12 Apr 2026 23:37:29.978 * NODE 7cbfdaa3b3c0ebf9bf9116dc86ed8bceffd31db0 () possibly failing. +2067052:M 12 Apr 2026 23:37:30.085 * Node 27c281c71471f64d57c2f90a6fce7ca7d11f3a9e () reported node 7cbfdaa3b3c0ebf9bf9116dc86ed8bceffd31db0 () as not reachable. +2067052:M 12 Apr 2026 23:37:30.085 * Marking node 7cbfdaa3b3c0ebf9bf9116dc86ed8bceffd31db0 () as failing (quorum reached). +2067052:M 12 Apr 2026 23:37:30.205 * Node d24296bca8e44a56521bbf138bec235238274ad7 () reported node 7cbfdaa3b3c0ebf9bf9116dc86ed8bceffd31db0 () as not reachable. +2067052:M 12 Apr 2026 23:37:30.248 * Clear FAIL state for node 7cbfdaa3b3c0ebf9bf9116dc86ed8bceffd31db0 (): primary without slots is reachable again. +2067052:M 12 Apr 2026 23:37:30.301 * Node 27c281c71471f64d57c2f90a6fce7ca7d11f3a9e () reported node 7cbfdaa3b3c0ebf9bf9116dc86ed8bceffd31db0 () is back online. +2067052:M 12 Apr 2026 23:37:30.380 * Node d24296bca8e44a56521bbf138bec235238274ad7 () reported node 7cbfdaa3b3c0ebf9bf9116dc86ed8bceffd31db0 () is back online. +2067052:M 12 Apr 2026 23:37:39.656 * NODE d24296bca8e44a56521bbf138bec235238274ad7 () possibly failing. +2067052:M 12 Apr 2026 23:37:39.656 * NODE 27c281c71471f64d57c2f90a6fce7ca7d11f3a9e () possibly failing. +2067052:M 12 Apr 2026 23:37:39.656 * NODE ceb41590bdb53727bffcb314c80a81989180d13a () possibly failing. +2067052:M 12 Apr 2026 23:37:39.656 * NODE 7cbfdaa3b3c0ebf9bf9116dc86ed8bceffd31db0 () possibly failing. +2067052:M 12 Apr 2026 23:37:39.656 # Cluster state changed: fail +2067052:M 12 Apr 2026 23:37:39.814 * Node d24296bca8e44a56521bbf138bec235238274ad7 () reported node 7cbfdaa3b3c0ebf9bf9116dc86ed8bceffd31db0 () as not reachable. +2067052:M 12 Apr 2026 23:37:39.814 * Node d24296bca8e44a56521bbf138bec235238274ad7 () reported node ceb41590bdb53727bffcb314c80a81989180d13a () as not reachable. +2067052:M 12 Apr 2026 23:37:39.814 * Marking node ceb41590bdb53727bffcb314c80a81989180d13a () as failing (quorum reached). +2067052:M 12 Apr 2026 23:37:39.867 * Node 27c281c71471f64d57c2f90a6fce7ca7d11f3a9e () reported node ceb41590bdb53727bffcb314c80a81989180d13a () as not reachable. +2067052:M 12 Apr 2026 23:37:39.867 * Node 27c281c71471f64d57c2f90a6fce7ca7d11f3a9e () reported node 7cbfdaa3b3c0ebf9bf9116dc86ed8bceffd31db0 () as not reachable. +2067052:M 12 Apr 2026 23:37:39.868 * FAIL message received from 27c281c71471f64d57c2f90a6fce7ca7d11f3a9e () about 7cbfdaa3b3c0ebf9bf9116dc86ed8bceffd31db0 () +2067052:M 12 Apr 2026 23:37:39.885 * Clear FAIL state for node ceb41590bdb53727bffcb314c80a81989180d13a (): replica is reachable again. +2067052:M 12 Apr 2026 23:37:39.903 * Clear FAIL state for node 7cbfdaa3b3c0ebf9bf9116dc86ed8bceffd31db0 (): primary without slots is reachable again. +2067052:M 12 Apr 2026 23:37:39.929 * FAIL message received from 71cce21972c3b30e1e72d50b69266c87048e6a42 () about 7cbfdaa3b3c0ebf9bf9116dc86ed8bceffd31db0 () +2067052:M 12 Apr 2026 23:37:40.076 * Clear FAIL state for node 7cbfdaa3b3c0ebf9bf9116dc86ed8bceffd31db0 (): primary without slots is reachable again. +2067052:M 12 Apr 2026 23:37:40.076 * Node d24296bca8e44a56521bbf138bec235238274ad7 () reported node 7cbfdaa3b3c0ebf9bf9116dc86ed8bceffd31db0 () is back online. +2067052:M 12 Apr 2026 23:37:40.213 * Node d24296bca8e44a56521bbf138bec235238274ad7 () reported node ceb41590bdb53727bffcb314c80a81989180d13a () is back online. +2067052:M 12 Apr 2026 23:37:40.255 * Node 27c281c71471f64d57c2f90a6fce7ca7d11f3a9e () reported node ceb41590bdb53727bffcb314c80a81989180d13a () is back online. +2067052:M 12 Apr 2026 23:37:40.255 * Node 27c281c71471f64d57c2f90a6fce7ca7d11f3a9e () reported node 7cbfdaa3b3c0ebf9bf9116dc86ed8bceffd31db0 () is back online. +2067052:M 12 Apr 2026 23:37:40.899 * Cluster state changed: ok +2067052:M 12 Apr 2026 23:37:49.977 - Accepting cluster node connection from 127.0.0.1:53068 +2067052:M 12 Apr 2026 23:37:49.980 - Accepting cluster node connection from 127.0.0.1:53070 +2067052:M 12 Apr 2026 23:37:49.991 - Accepting cluster node connection from 127.0.0.1:53076 +2067052:M 12 Apr 2026 23:37:50.013 - Accepting cluster node connection from 127.0.0.1:53084 +2067052:M 12 Apr 2026 23:37:50.025 * NODE 7cbfdaa3b3c0ebf9bf9116dc86ed8bceffd31db0 () possibly failing. +2067052:M 12 Apr 2026 23:37:57.845 * NODE 27c281c71471f64d57c2f90a6fce7ca7d11f3a9e () possibly failing. +2067052:M 12 Apr 2026 23:37:57.942 * Node d24296bca8e44a56521bbf138bec235238274ad7 () reported node 27c281c71471f64d57c2f90a6fce7ca7d11f3a9e () as not reachable. +2067052:M 12 Apr 2026 23:37:57.942 * Marking node 27c281c71471f64d57c2f90a6fce7ca7d11f3a9e () as failing (quorum reached). +2067052:M 12 Apr 2026 23:37:57.942 # Cluster state changed: fail +2067052:M 12 Apr 2026 23:37:58.361 * NODE 71cce21972c3b30e1e72d50b69266c87048e6a42 () possibly failing. +2067052:M 12 Apr 2026 23:37:58.473 * Node d24296bca8e44a56521bbf138bec235238274ad7 () reported node 71cce21972c3b30e1e72d50b69266c87048e6a42 () as not reachable. +2067052:M 12 Apr 2026 23:37:58.473 * Marking node 71cce21972c3b30e1e72d50b69266c87048e6a42 () as failing (quorum reached). +2067052:M 12 Apr 2026 23:37:58.575 * NODE ceb41590bdb53727bffcb314c80a81989180d13a () possibly failing. +2067052:M 12 Apr 2026 23:37:58.575 * NODE 7cbfdaa3b3c0ebf9bf9116dc86ed8bceffd31db0 () possibly failing. +2067052:M 12 Apr 2026 23:37:58.714 - Accepting cluster node connection from 127.0.0.1:37760 +2067052:M 12 Apr 2026 23:37:58.716 * Node d24296bca8e44a56521bbf138bec235238274ad7 () reported node 7cbfdaa3b3c0ebf9bf9116dc86ed8bceffd31db0 () as not reachable. +2067052:M 12 Apr 2026 23:37:58.753 * Node d24296bca8e44a56521bbf138bec235238274ad7 () reported node 7cbfdaa3b3c0ebf9bf9116dc86ed8bceffd31db0 () is back online. +2067052:M 12 Apr 2026 23:37:58.753 * Node d24296bca8e44a56521bbf138bec235238274ad7 () reported node 71cce21972c3b30e1e72d50b69266c87048e6a42 () is back online. +2067052:M 12 Apr 2026 23:37:58.764 * Clear FAIL state for node 71cce21972c3b30e1e72d50b69266c87048e6a42 (): replica is reachable again. +2067052:M 12 Apr 2026 23:37:59.569 * Failover auth granted to ceb41590bdb53727bffcb314c80a81989180d13a () for epoch 9 +2067052:M 12 Apr 2026 23:37:59.720 * Failover occurred in migration source. Update importing source for slot 609 to node ceb41590bdb53727bffcb314c80a81989180d13a () in shard 0ccab0ca3bdf2df4fb1d88a6388f0d2e2e9492d9. +2067052:M 12 Apr 2026 23:37:59.720 * Node ceb41590bdb53727bffcb314c80a81989180d13a () reported node 27c281c71471f64d57c2f90a6fce7ca7d11f3a9e () as not reachable. +2067052:M 12 Apr 2026 23:37:59.720 * Cluster state changed: ok +2067052:M 12 Apr 2026 23:37:59.801 * Clear FAIL state for node 27c281c71471f64d57c2f90a6fce7ca7d11f3a9e (): primary without slots is reachable again. +2067052:M 12 Apr 2026 23:37:59.801 * A failover occurred in shard 0ccab0ca3bdf2df4fb1d88a6388f0d2e2e9492d9; node 27c281c71471f64d57c2f90a6fce7ca7d11f3a9e () failed over to node ceb41590bdb53727bffcb314c80a81989180d13a () with a config epoch of 9 +2067052:M 12 Apr 2026 23:37:59.801 * Node 27c281c71471f64d57c2f90a6fce7ca7d11f3a9e () is now a replica of node ceb41590bdb53727bffcb314c80a81989180d13a () in shard 0ccab0ca3bdf2df4fb1d88a6388f0d2e2e9492d9 +2067052:M 12 Apr 2026 23:37:59.884 * Node ceb41590bdb53727bffcb314c80a81989180d13a () reported node 27c281c71471f64d57c2f90a6fce7ca7d11f3a9e () is back online. +2067052:M 12 Apr 2026 23:37:59.985 * Node d24296bca8e44a56521bbf138bec235238274ad7 () reported node 27c281c71471f64d57c2f90a6fce7ca7d11f3a9e () is back online. +2067052:M 12 Apr 2026 23:38:07.763 * NODE 71cce21972c3b30e1e72d50b69266c87048e6a42 () possibly failing. +2067052:M 12 Apr 2026 23:38:07.885 * FAIL message received from ceb41590bdb53727bffcb314c80a81989180d13a () about 71cce21972c3b30e1e72d50b69266c87048e6a42 () +2067052:M 12 Apr 2026 23:38:07.887 * Node d24296bca8e44a56521bbf138bec235238274ad7 () reported node 71cce21972c3b30e1e72d50b69266c87048e6a42 () as not reachable. +2067052:M 12 Apr 2026 23:38:07.940 * Clear FAIL state for node 71cce21972c3b30e1e72d50b69266c87048e6a42 (): replica is reachable again. +2067052:M 12 Apr 2026 23:38:07.954 * Node d24296bca8e44a56521bbf138bec235238274ad7 () reported node 71cce21972c3b30e1e72d50b69266c87048e6a42 () is back online. +2067052:M 12 Apr 2026 23:38:16.489 * NODE 71cce21972c3b30e1e72d50b69266c87048e6a42 () possibly failing. +2067052:M 12 Apr 2026 23:38:16.489 * NODE ceb41590bdb53727bffcb314c80a81989180d13a () possibly failing. +2067052:M 12 Apr 2026 23:38:17.881 * NODE d24296bca8e44a56521bbf138bec235238274ad7 () possibly failing. +2067052:M 12 Apr 2026 23:38:17.882 * NODE 27c281c71471f64d57c2f90a6fce7ca7d11f3a9e () possibly failing. +2067052:M 12 Apr 2026 23:38:17.882 * NODE 7cbfdaa3b3c0ebf9bf9116dc86ed8bceffd31db0 () possibly failing. +2067052:M 12 Apr 2026 23:38:17.882 # Cluster state changed: fail +2067052:M 12 Apr 2026 23:38:17.882 * Node d24296bca8e44a56521bbf138bec235238274ad7 () reported node 71cce21972c3b30e1e72d50b69266c87048e6a42 () as not reachable. +2067052:M 12 Apr 2026 23:38:17.882 * Marking node 71cce21972c3b30e1e72d50b69266c87048e6a42 () as failing (quorum reached). +2067052:M 12 Apr 2026 23:38:17.882 * Node d24296bca8e44a56521bbf138bec235238274ad7 () reported node ceb41590bdb53727bffcb314c80a81989180d13a () as not reachable. +2067052:M 12 Apr 2026 23:38:17.882 * Marking node ceb41590bdb53727bffcb314c80a81989180d13a () as failing (quorum reached). +2067052:M 12 Apr 2026 23:38:17.882 - Accepting cluster node connection from 127.0.0.1:40044 +2067052:M 12 Apr 2026 23:38:17.933 * Node d24296bca8e44a56521bbf138bec235238274ad7 () reported node 71cce21972c3b30e1e72d50b69266c87048e6a42 () is back online. +2067052:M 12 Apr 2026 23:38:18.101 * Clear FAIL state for node 71cce21972c3b30e1e72d50b69266c87048e6a42 (): replica is reachable again. +2067052:M 12 Apr 2026 23:38:18.150 * Node d24296bca8e44a56521bbf138bec235238274ad7 () reported node ceb41590bdb53727bffcb314c80a81989180d13a () is back online. +2067052:M 12 Apr 2026 23:38:18.164 - Accepting cluster node connection from 127.0.0.1:40052 +2067052:M 12 Apr 2026 23:38:19.970 * Clear FAIL state for node ceb41590bdb53727bffcb314c80a81989180d13a (): is reachable again and nobody is serving its slots after some time. +2067052:M 12 Apr 2026 23:38:19.970 * Cluster state changed: ok +2067052:M 12 Apr 2026 23:38:26.453 * NODE 7cbfdaa3b3c0ebf9bf9116dc86ed8bceffd31db0 () possibly failing. +2067052:M 12 Apr 2026 23:38:26.706 * FAIL message received from ceb41590bdb53727bffcb314c80a81989180d13a () about 7cbfdaa3b3c0ebf9bf9116dc86ed8bceffd31db0 () +2067052:M 12 Apr 2026 23:38:26.706 * Node ceb41590bdb53727bffcb314c80a81989180d13a () reported node 7cbfdaa3b3c0ebf9bf9116dc86ed8bceffd31db0 () as not reachable. +2067052:M 12 Apr 2026 23:38:26.759 * Node d24296bca8e44a56521bbf138bec235238274ad7 () reported node 7cbfdaa3b3c0ebf9bf9116dc86ed8bceffd31db0 () as not reachable. +2067052:M 12 Apr 2026 23:38:26.855 * Node ceb41590bdb53727bffcb314c80a81989180d13a () reported node 7cbfdaa3b3c0ebf9bf9116dc86ed8bceffd31db0 () is back online. +2067052:M 12 Apr 2026 23:38:26.864 * Clear FAIL state for node 7cbfdaa3b3c0ebf9bf9116dc86ed8bceffd31db0 (): primary without slots is reachable again. +2067052:M 12 Apr 2026 23:38:27.079 * Node d24296bca8e44a56521bbf138bec235238274ad7 () reported node 7cbfdaa3b3c0ebf9bf9116dc86ed8bceffd31db0 () is back online. +2067052:M 12 Apr 2026 23:38:35.505 * NODE ceb41590bdb53727bffcb314c80a81989180d13a () possibly failing. +2067052:M 12 Apr 2026 23:38:35.505 - Accepting cluster node connection from 127.0.0.1:34846 +2067052:M 12 Apr 2026 23:38:44.685 * NODE d24296bca8e44a56521bbf138bec235238274ad7 () possibly failing. +2067052:M 12 Apr 2026 23:38:44.685 * NODE 7cbfdaa3b3c0ebf9bf9116dc86ed8bceffd31db0 () possibly failing. +2067052:M 12 Apr 2026 23:38:44.789 - Accepting cluster node connection from 127.0.0.1:54596 +2067052:M 12 Apr 2026 23:38:44.845 * Node ceb41590bdb53727bffcb314c80a81989180d13a () reported node 7cbfdaa3b3c0ebf9bf9116dc86ed8bceffd31db0 () as not reachable. +2067052:M 12 Apr 2026 23:38:44.845 * Marking node 7cbfdaa3b3c0ebf9bf9116dc86ed8bceffd31db0 () as failing (quorum reached). +2067052:M 12 Apr 2026 23:38:44.845 * Node ceb41590bdb53727bffcb314c80a81989180d13a () reported node d24296bca8e44a56521bbf138bec235238274ad7 () as not reachable. +2067052:M 12 Apr 2026 23:38:44.845 * Node d24296bca8e44a56521bbf138bec235238274ad7 () reported node 7cbfdaa3b3c0ebf9bf9116dc86ed8bceffd31db0 () as not reachable. +2067052:M 12 Apr 2026 23:38:44.845 * Node d24296bca8e44a56521bbf138bec235238274ad7 () reported node ceb41590bdb53727bffcb314c80a81989180d13a () as not reachable. +2067052:M 12 Apr 2026 23:38:44.845 * Node d24296bca8e44a56521bbf138bec235238274ad7 () reported node 27c281c71471f64d57c2f90a6fce7ca7d11f3a9e () as not reachable. +2067052:M 12 Apr 2026 23:38:44.893 * Node d24296bca8e44a56521bbf138bec235238274ad7 () reported node 27c281c71471f64d57c2f90a6fce7ca7d11f3a9e () is back online. +2067052:M 12 Apr 2026 23:38:44.896 * Clear FAIL state for node 7cbfdaa3b3c0ebf9bf9116dc86ed8bceffd31db0 (): primary without slots is reachable again. +2067052:M 12 Apr 2026 23:38:44.897 * Node ceb41590bdb53727bffcb314c80a81989180d13a () reported node d24296bca8e44a56521bbf138bec235238274ad7 () is back online. +2067052:M 12 Apr 2026 23:38:44.997 * Node ceb41590bdb53727bffcb314c80a81989180d13a () reported node 7cbfdaa3b3c0ebf9bf9116dc86ed8bceffd31db0 () is back online. +2067052:M 12 Apr 2026 23:38:45.005 * Node d24296bca8e44a56521bbf138bec235238274ad7 () reported node 7cbfdaa3b3c0ebf9bf9116dc86ed8bceffd31db0 () is back online. +2067052:M 12 Apr 2026 23:38:45.202 * Node d24296bca8e44a56521bbf138bec235238274ad7 () reported node ceb41590bdb53727bffcb314c80a81989180d13a () is back online. +2067052:M 12 Apr 2026 23:38:54.107 - Accepting cluster node connection from 127.0.0.1:50724 +2067052:M 12 Apr 2026 23:38:54.110 - Error accepting cluster node connection: error:0A000126:SSL routines::unexpected eof while reading +2067052:M 12 Apr 2026 23:38:54.121 - Accepting cluster node connection from 127.0.0.1:50740 +2067052:M 12 Apr 2026 23:38:54.213 - Accepting cluster node connection from 127.0.0.1:50744 +2067052:M 12 Apr 2026 23:38:54.228 * Node d24296bca8e44a56521bbf138bec235238274ad7 () reported node 71cce21972c3b30e1e72d50b69266c87048e6a42 () as not reachable. +2067052:M 12 Apr 2026 23:38:54.228 * Node d24296bca8e44a56521bbf138bec235238274ad7 () reported node 7cbfdaa3b3c0ebf9bf9116dc86ed8bceffd31db0 () as not reachable. +2067052:M 12 Apr 2026 23:38:54.228 * Node d24296bca8e44a56521bbf138bec235238274ad7 () reported node ceb41590bdb53727bffcb314c80a81989180d13a () as not reachable. +2067052:M 12 Apr 2026 23:38:54.228 * Node d24296bca8e44a56521bbf138bec235238274ad7 () reported node 27c281c71471f64d57c2f90a6fce7ca7d11f3a9e () as not reachable. +2067052:M 12 Apr 2026 23:38:54.261 - Accepting cluster node connection from 127.0.0.1:50756 +2067052:M 12 Apr 2026 23:38:54.409 * Node d24296bca8e44a56521bbf138bec235238274ad7 () reported node 71cce21972c3b30e1e72d50b69266c87048e6a42 () is back online. +2067052:M 12 Apr 2026 23:38:54.409 * Node d24296bca8e44a56521bbf138bec235238274ad7 () reported node ceb41590bdb53727bffcb314c80a81989180d13a () is back online. +2067052:M 12 Apr 2026 23:38:54.409 * Node d24296bca8e44a56521bbf138bec235238274ad7 () reported node 27c281c71471f64d57c2f90a6fce7ca7d11f3a9e () is back online. +2067052:M 12 Apr 2026 23:38:54.427 * Node d24296bca8e44a56521bbf138bec235238274ad7 () reported node 7cbfdaa3b3c0ebf9bf9116dc86ed8bceffd31db0 () is back online. +2067052:M 12 Apr 2026 23:39:03.159 * NODE 7cbfdaa3b3c0ebf9bf9116dc86ed8bceffd31db0 () possibly failing. +2067052:M 12 Apr 2026 23:39:03.165 - Accepting cluster node connection from 127.0.0.1:50770 +2067052:M 12 Apr 2026 23:39:03.167 - Error accepting cluster node connection: error:0A000126:SSL routines::unexpected eof while reading +2067052:M 12 Apr 2026 23:39:03.167 - Accepting cluster node connection from 127.0.0.1:47956 +2067052:M 12 Apr 2026 23:39:03.173 - Accepting cluster node connection from 127.0.0.1:47968 +2067052:M 12 Apr 2026 23:39:03.175 - Error accepting cluster node connection: error:0A000126:SSL routines::unexpected eof while reading +2067052:M 12 Apr 2026 23:39:03.188 * Node ceb41590bdb53727bffcb314c80a81989180d13a () reported node 7cbfdaa3b3c0ebf9bf9116dc86ed8bceffd31db0 () as not reachable. +2067052:M 12 Apr 2026 23:39:03.188 * Marking node 7cbfdaa3b3c0ebf9bf9116dc86ed8bceffd31db0 () as failing (quorum reached). +2067052:M 12 Apr 2026 23:39:03.255 - Accepting cluster node connection from 127.0.0.1:47974 +2067052:M 12 Apr 2026 23:39:03.261 - Accepting cluster node connection from 127.0.0.1:47982 +2067052:M 12 Apr 2026 23:39:03.301 * Clear FAIL state for node 7cbfdaa3b3c0ebf9bf9116dc86ed8bceffd31db0 (): primary without slots is reachable again. +2067052:M 12 Apr 2026 23:39:03.387 * Node ceb41590bdb53727bffcb314c80a81989180d13a () reported node 7cbfdaa3b3c0ebf9bf9116dc86ed8bceffd31db0 () is back online. +2067052:M 12 Apr 2026 23:39:21.026 * NODE 71cce21972c3b30e1e72d50b69266c87048e6a42 () possibly failing. +2067052:M 12 Apr 2026 23:39:21.026 * NODE d24296bca8e44a56521bbf138bec235238274ad7 () possibly failing. +2067052:M 12 Apr 2026 23:39:21.026 * NODE 27c281c71471f64d57c2f90a6fce7ca7d11f3a9e () possibly failing. +2067052:M 12 Apr 2026 23:39:21.026 * NODE ceb41590bdb53727bffcb314c80a81989180d13a () possibly failing. +2067052:M 12 Apr 2026 23:39:21.026 * NODE 7cbfdaa3b3c0ebf9bf9116dc86ed8bceffd31db0 () possibly failing. +2067052:M 12 Apr 2026 23:39:21.026 # Cluster state changed: fail +2067052:M 12 Apr 2026 23:39:21.144 - Accepting cluster node connection from 127.0.0.1:54210 +2067052:M 12 Apr 2026 23:39:21.163 * Node ceb41590bdb53727bffcb314c80a81989180d13a () reported node 71cce21972c3b30e1e72d50b69266c87048e6a42 () as not reachable. +2067052:M 12 Apr 2026 23:39:21.163 * Node ceb41590bdb53727bffcb314c80a81989180d13a () reported node 27c281c71471f64d57c2f90a6fce7ca7d11f3a9e () as not reachable. +2067052:M 12 Apr 2026 23:39:21.163 * Node ceb41590bdb53727bffcb314c80a81989180d13a () reported node d24296bca8e44a56521bbf138bec235238274ad7 () as not reachable. +2067052:M 12 Apr 2026 23:39:21.163 * FAIL message received from ceb41590bdb53727bffcb314c80a81989180d13a () about 71cce21972c3b30e1e72d50b69266c87048e6a42 () +2067052:M 12 Apr 2026 23:39:21.163 * Node ceb41590bdb53727bffcb314c80a81989180d13a () reported node d24296bca8e44a56521bbf138bec235238274ad7 () is back online. +2067052:M 12 Apr 2026 23:39:21.163 * Node ceb41590bdb53727bffcb314c80a81989180d13a () reported node 27c281c71471f64d57c2f90a6fce7ca7d11f3a9e () is back online. +2067052:M 12 Apr 2026 23:39:21.331 * Clear FAIL state for node 71cce21972c3b30e1e72d50b69266c87048e6a42 (): replica is reachable again. +2067052:M 12 Apr 2026 23:39:21.333 * Node ceb41590bdb53727bffcb314c80a81989180d13a () reported node 71cce21972c3b30e1e72d50b69266c87048e6a42 () is back online. +===== End of server log (pid 2067052) ===== + + +===== Start of server stderr log (pid 2067052) ===== + + +===== End of server stderr log (pid 2067052) ===== + + +===== Start of server log (pid 2066947) ===== + +### Starting server for test +2066947:C 12 Apr 2026 23:36:26.555 # WARNING: Changing databases number from 16 to 1 since we are in cluster mode +2066947:C 12 Apr 2026 23:36:26.556 # WARNING Memory overcommit must be enabled! Without it, a background save or replication may fail under low memory condition. Being disabled, it can also cause failures without low memory condition, see https://github.com/jemalloc/jemalloc/issues/1328. To fix this issue add 'vm.overcommit_memory = 1' to /etc/sysctl.conf and then reboot or run the command 'sysctl vm.overcommit_memory=1' for this to take effect. +2066947:C 12 Apr 2026 23:36:26.556 * oO0OoO0OoO0Oo Valkey is starting oO0OoO0OoO0Oo +2066947:C 12 Apr 2026 23:36:26.556 * Valkey version=8.0.2, bits=64, commit=00000000, modified=0, pid=2066947, just started +2066947:C 12 Apr 2026 23:36:26.556 * Configuration loaded +2066947:M 12 Apr 2026 23:36:26.557 * Increased maximum number of open files to 10032 (it was originally set to 1024). +2066947:M 12 Apr 2026 23:36:26.557 * monotonic clock: POSIX clock_gettime +2066947:M 12 Apr 2026 23:36:26.559 # Failed to write PID file: Permission denied + .+^+. + .+#########+. + .+########+########+. Valkey 8.0.2 (00000000/0) 64 bit + .+########+' '+########+. + .########+' .+. '+########. Running in cluster mode + |####+' .+#######+. '+####| Port: 21696 + |###| .+###############+. |###| PID: 2066947 + |###| |#####*'' ''*#####| |###| + |###| |####' .-. '####| |###| + |###| |###( (@@@) )###| |###| https://valkey.io + |###| |####. '-' .####| |###| + |###| |#####*. .*#####| |###| + |###| '+#####| |#####+' |###| + |####+. +##| |#+' .+####| + '#######+ |##| .+########' + '+###| |##| .+########+' + '| |####+########+' + +#########+' + '+v+' + +2066947:M 12 Apr 2026 23:36:26.560 * No cluster configuration found, I'm d24296bca8e44a56521bbf138bec235238274ad7 +2066947:M 12 Apr 2026 23:36:26.581 * Server initialized +2066947:M 12 Apr 2026 23:36:26.582 * Ready to accept connections tcp +2066947:M 12 Apr 2026 23:36:26.582 * Ready to accept connections unix +2066947:M 12 Apr 2026 23:36:26.582 * Ready to accept connections tls +2066947:M 12 Apr 2026 23:36:26.648 - Accepted 127.0.0.1:37165 +2066947:M 12 Apr 2026 23:36:26.666 - Reading from client: error:0A000126:SSL routines::unexpected eof while reading +2066947:M 12 Apr 2026 23:36:26.680 - Accepted 127.0.0.1:41543 +2066947:M 12 Apr 2026 23:36:27.368 - Accepting cluster node connection from 127.0.0.1:33324 +2066947:M 12 Apr 2026 23:36:27.421 * IP address for this node updated to 127.0.0.1 +2066947:M 12 Apr 2026 23:36:27.539 * Successfully completed handshake with 7cbfdaa3b3c0ebf9bf9116dc86ed8bceffd31db0 () +2066947:M 12 Apr 2026 23:36:27.543 - Handshake: we already know node 27c281c71471f64d57c2f90a6fce7ca7d11f3a9e (), updating the address if needed. +2066947:M 12 Apr 2026 23:36:27.569 - Accepting cluster node connection from 127.0.0.1:33338 +2066947:M 12 Apr 2026 23:36:27.588 - Accepting cluster node connection from 127.0.0.1:33340 +2066947:M 12 Apr 2026 23:36:27.649 * Address updated for node 7cbfdaa3b3c0ebf9bf9116dc86ed8bceffd31db0 (), now 127.0.0.1:21691 +2066947:M 12 Apr 2026 23:36:27.685 - Accepting cluster node connection from 127.0.0.1:33342 +2066947:M 12 Apr 2026 23:36:27.688 - Accepting cluster node connection from 127.0.0.1:33346 +2066947:M 12 Apr 2026 23:36:27.722 - Accepting cluster node connection from 127.0.0.1:33360 +2066947:M 12 Apr 2026 23:36:27.726 - Accepting cluster node connection from 127.0.0.1:33362 +2066947:M 12 Apr 2026 23:36:27.957 * Node ceb41590bdb53727bffcb314c80a81989180d13a () is no longer primary of shard 7be85fc7a35a355988332b0ade6bb85642f90b76; removed all 0 slot(s) it used to own +2066947:M 12 Apr 2026 23:36:27.957 * Node ceb41590bdb53727bffcb314c80a81989180d13a () is now part of shard 0ccab0ca3bdf2df4fb1d88a6388f0d2e2e9492d9 +2066947:M 12 Apr 2026 23:36:27.957 * Node ceb41590bdb53727bffcb314c80a81989180d13a () is now a replica of node 27c281c71471f64d57c2f90a6fce7ca7d11f3a9e () in shard 0ccab0ca3bdf2df4fb1d88a6388f0d2e2e9492d9 +2066947:M 12 Apr 2026 23:36:27.972 * Node 7cbfdaa3b3c0ebf9bf9116dc86ed8bceffd31db0 () is no longer primary of shard 6d5a08da00fb925618c9429f0a0abf50cb184dee; removed all 0 slot(s) it used to own +2066947:M 12 Apr 2026 23:36:27.972 * Node 7cbfdaa3b3c0ebf9bf9116dc86ed8bceffd31db0 () is now part of shard fb1332a38fbff59dddf699f62b6c363413af8689 +2066947:M 12 Apr 2026 23:36:27.972 * Node 7cbfdaa3b3c0ebf9bf9116dc86ed8bceffd31db0 () is now a replica of node ad56f48bae1934f4d1c3141ffee95fa8bf2fa37c () in shard fb1332a38fbff59dddf699f62b6c363413af8689 +2066947:M 12 Apr 2026 23:36:27.975 - Accepted 127.0.0.1:47576 +2066947:M 12 Apr 2026 23:36:27.978 * Node 71cce21972c3b30e1e72d50b69266c87048e6a42 () is no longer primary of shard 797d84842d0558c353ef2fd200fec2d129d7c96f; removed all 0 slot(s) it used to own +2066947:M 12 Apr 2026 23:36:27.978 * Node 71cce21972c3b30e1e72d50b69266c87048e6a42 () is now part of shard b553502a57202365d35122f1f841b02047c3c23d +2066947:M 12 Apr 2026 23:36:27.978 * Node 71cce21972c3b30e1e72d50b69266c87048e6a42 () is now a replica of node d24296bca8e44a56521bbf138bec235238274ad7 () in shard b553502a57202365d35122f1f841b02047c3c23d +2066947:M 12 Apr 2026 23:36:28.007 # DEBUG LOG: ========== I am primary 2 ========== +2066947:M 12 Apr 2026 23:36:28.017 * Replica 127.0.0.1:21689 asks for synchronization +2066947:M 12 Apr 2026 23:36:28.017 * Full resync requested by replica 127.0.0.1:21689 +2066947:M 12 Apr 2026 23:36:28.018 * Replication backlog created, my new replication IDs are 'f9a64ac3e118f63770d16efaeec31459ca9e6a4b' and '0000000000000000000000000000000000000000' +2066947:M 12 Apr 2026 23:36:28.018 * Starting BGSAVE for SYNC with target: replicas sockets using: normal sync +2066947:M 12 Apr 2026 23:36:28.018 * Background RDB transfer started by pid 2067762 to pipe through parent process +2067762:C 12 Apr 2026 23:36:28.023 * Fork CoW for RDB: current 1 MB, peak 1 MB, average 1 MB +2066947:M 12 Apr 2026 23:36:28.031 * Diskless rdb transfer, done reading from pipe, 1 replicas still up. +2066947:M 12 Apr 2026 23:36:28.038 * Background RDB transfer terminated with success +2066947:M 12 Apr 2026 23:36:28.038 * Streamed RDB transfer with replica 127.0.0.1:21689 succeeded (socket). Waiting for REPLCONF ACK from replica to enable streaming +2066947:M 12 Apr 2026 23:36:28.038 * Synchronization with replica 127.0.0.1:21689 succeeded +2066947:M 12 Apr 2026 23:36:28.630 * Cluster state changed: ok +2066947:M 12 Apr 2026 23:36:35.764 * NODE 71cce21972c3b30e1e72d50b69266c87048e6a42 () possibly failing. +### Starting test Slot migration states are replicated in tests/unit/cluster/slot-migration.tcl +### Starting test Migration target is auto-updated after failover in target shard in tests/unit/cluster/slot-migration.tcl +2066947:M 12 Apr 2026 23:36:38.898 * NODE ad56f48bae1934f4d1c3141ffee95fa8bf2fa37c () possibly failing. +2066947:M 12 Apr 2026 23:36:39.108 * FAIL message received from 27c281c71471f64d57c2f90a6fce7ca7d11f3a9e () about ad56f48bae1934f4d1c3141ffee95fa8bf2fa37c () +2066947:M 12 Apr 2026 23:36:39.108 # Cluster state changed: fail +2066947:M 12 Apr 2026 23:36:39.149 * Node 27c281c71471f64d57c2f90a6fce7ca7d11f3a9e () reported node ad56f48bae1934f4d1c3141ffee95fa8bf2fa37c () as not reachable. +2066947:M 12 Apr 2026 23:36:39.929 * Failover auth granted to 7cbfdaa3b3c0ebf9bf9116dc86ed8bceffd31db0 () for epoch 6 +2066947:M 12 Apr 2026 23:36:39.981 * Node 7cbfdaa3b3c0ebf9bf9116dc86ed8bceffd31db0 () reported node ad56f48bae1934f4d1c3141ffee95fa8bf2fa37c () as not reachable. +2066947:M 12 Apr 2026 23:36:39.981 * Cluster state changed: ok +2066947:M 12 Apr 2026 23:36:40.925 - Node ad56f48bae1934f4d1c3141ffee95fa8bf2fa37c has old slots configuration, sending an UPDATE message about 7cbfdaa3b3c0ebf9bf9116dc86ed8bceffd31db0 +2066947:M 12 Apr 2026 23:36:41.177 * A failover occurred in shard fb1332a38fbff59dddf699f62b6c363413af8689; node ad56f48bae1934f4d1c3141ffee95fa8bf2fa37c () failed over to node 7cbfdaa3b3c0ebf9bf9116dc86ed8bceffd31db0 () with a config epoch of 6 +2066947:M 12 Apr 2026 23:36:41.177 * Node ad56f48bae1934f4d1c3141ffee95fa8bf2fa37c () is now a replica of node 7cbfdaa3b3c0ebf9bf9116dc86ed8bceffd31db0 () in shard fb1332a38fbff59dddf699f62b6c363413af8689 +2066947:M 12 Apr 2026 23:36:41.189 * Clear FAIL state for node ad56f48bae1934f4d1c3141ffee95fa8bf2fa37c (): replica is reachable again. +2066947:M 12 Apr 2026 23:36:41.260 * Node 27c281c71471f64d57c2f90a6fce7ca7d11f3a9e () reported node ad56f48bae1934f4d1c3141ffee95fa8bf2fa37c () is back online. +2066947:M 12 Apr 2026 23:36:41.292 * Node 7cbfdaa3b3c0ebf9bf9116dc86ed8bceffd31db0 () reported node ad56f48bae1934f4d1c3141ffee95fa8bf2fa37c () is back online. +2066947:M 12 Apr 2026 23:36:41.293 * Failover auth granted to ad56f48bae1934f4d1c3141ffee95fa8bf2fa37c () for epoch 7 +2066947:M 12 Apr 2026 23:36:41.491 * A failover occurred in shard fb1332a38fbff59dddf699f62b6c363413af8689; node 7cbfdaa3b3c0ebf9bf9116dc86ed8bceffd31db0 () failed over to node ad56f48bae1934f4d1c3141ffee95fa8bf2fa37c () with a config epoch of 7 +2066947:M 12 Apr 2026 23:36:41.491 * Node 7cbfdaa3b3c0ebf9bf9116dc86ed8bceffd31db0 () is now a replica of node ad56f48bae1934f4d1c3141ffee95fa8bf2fa37c () in shard fb1332a38fbff59dddf699f62b6c363413af8689 +### Starting test Migration source is auto-updated after failover in source shard in tests/unit/cluster/slot-migration.tcl +2066947:M 12 Apr 2026 23:36:42.712 * NODE 27c281c71471f64d57c2f90a6fce7ca7d11f3a9e () possibly failing. +2066947:M 12 Apr 2026 23:36:43.121 * Node ad56f48bae1934f4d1c3141ffee95fa8bf2fa37c () reported node 27c281c71471f64d57c2f90a6fce7ca7d11f3a9e () as not reachable. +2066947:M 12 Apr 2026 23:36:43.121 * Marking node 27c281c71471f64d57c2f90a6fce7ca7d11f3a9e () as failing (quorum reached). +2066947:M 12 Apr 2026 23:36:43.121 # Cluster state changed: fail +2066947:M 12 Apr 2026 23:36:45.746 * Clear FAIL state for node 27c281c71471f64d57c2f90a6fce7ca7d11f3a9e (): is reachable again and nobody is serving its slots after some time. +2066947:M 12 Apr 2026 23:36:45.746 * Cluster state changed: ok +2066947:M 12 Apr 2026 23:36:46.111 * Node ad56f48bae1934f4d1c3141ffee95fa8bf2fa37c () reported node 27c281c71471f64d57c2f90a6fce7ca7d11f3a9e () is back online. +2066947:M 12 Apr 2026 23:36:46.727 * NODE ceb41590bdb53727bffcb314c80a81989180d13a () possibly failing. +2066947:M 12 Apr 2026 23:36:47.077 * Node 27c281c71471f64d57c2f90a6fce7ca7d11f3a9e () reported node ceb41590bdb53727bffcb314c80a81989180d13a () as not reachable. +2066947:M 12 Apr 2026 23:36:47.077 * Marking node ceb41590bdb53727bffcb314c80a81989180d13a () as failing (quorum reached). +2066947:M 12 Apr 2026 23:36:47.124 * Node ad56f48bae1934f4d1c3141ffee95fa8bf2fa37c () reported node ceb41590bdb53727bffcb314c80a81989180d13a () as not reachable. +2066947:M 12 Apr 2026 23:36:47.227 # Failover auth denied to ceb41590bdb53727bffcb314c80a81989180d13a () for epoch 8: its primary is up +2066947:M 12 Apr 2026 23:36:47.267 * Clear FAIL state for node ceb41590bdb53727bffcb314c80a81989180d13a (): replica is reachable again. +2066947:M 12 Apr 2026 23:36:47.390 * Node ad56f48bae1934f4d1c3141ffee95fa8bf2fa37c () reported node ceb41590bdb53727bffcb314c80a81989180d13a () is back online. +2066947:M 12 Apr 2026 23:36:47.397 * Node 27c281c71471f64d57c2f90a6fce7ca7d11f3a9e () reported node ceb41590bdb53727bffcb314c80a81989180d13a () is back online. +2066947:M 12 Apr 2026 23:36:53.745 * NODE 7cbfdaa3b3c0ebf9bf9116dc86ed8bceffd31db0 () possibly failing. +2066947:M 12 Apr 2026 23:36:53.850 * Node 27c281c71471f64d57c2f90a6fce7ca7d11f3a9e () reported node 7cbfdaa3b3c0ebf9bf9116dc86ed8bceffd31db0 () as not reachable. +2066947:M 12 Apr 2026 23:36:53.850 * Marking node 7cbfdaa3b3c0ebf9bf9116dc86ed8bceffd31db0 () as failing (quorum reached). +2066947:M 12 Apr 2026 23:36:53.999 * Node ad56f48bae1934f4d1c3141ffee95fa8bf2fa37c () reported node 7cbfdaa3b3c0ebf9bf9116dc86ed8bceffd31db0 () as not reachable. +2066947:M 12 Apr 2026 23:36:54.012 * Clear FAIL state for node 7cbfdaa3b3c0ebf9bf9116dc86ed8bceffd31db0 (): replica is reachable again. +2066947:M 12 Apr 2026 23:36:54.249 * Node 27c281c71471f64d57c2f90a6fce7ca7d11f3a9e () reported node 7cbfdaa3b3c0ebf9bf9116dc86ed8bceffd31db0 () is back online. +2066947:M 12 Apr 2026 23:36:54.543 * Node ad56f48bae1934f4d1c3141ffee95fa8bf2fa37c () reported node 7cbfdaa3b3c0ebf9bf9116dc86ed8bceffd31db0 () is back online. +2066947:M 12 Apr 2026 23:36:55.169 - Accepting cluster node connection from 127.0.0.1:33018 +2066947:M 12 Apr 2026 23:36:55.235 - Accepting cluster node connection from 127.0.0.1:33030 +2066947:M 12 Apr 2026 23:36:55.247 - Accepting cluster node connection from 127.0.0.1:33044 +### Starting test Replica redirects key access in migrating slots in tests/unit/cluster/slot-migration.tcl +### Starting test Replica of migrating node returns ASK redirect after READONLY in tests/unit/cluster/slot-migration.tcl +### Starting test Replica of migrating node returns TRYAGAIN after READONLY in tests/unit/cluster/slot-migration.tcl +### Starting test Replica of importing node returns TRYAGAIN after READONLY and ASKING in tests/unit/cluster/slot-migration.tcl +### Starting test New replica inherits migrating slot in tests/unit/cluster/slot-migration.tcl +2066947:M 12 Apr 2026 23:36:59.030 * NODE ceb41590bdb53727bffcb314c80a81989180d13a () possibly failing. +2066947:M 12 Apr 2026 23:36:59.503 * Node 27c281c71471f64d57c2f90a6fce7ca7d11f3a9e () reported node ceb41590bdb53727bffcb314c80a81989180d13a () as not reachable. +2066947:M 12 Apr 2026 23:36:59.503 * Marking node ceb41590bdb53727bffcb314c80a81989180d13a () as failing (quorum reached). +2066947:M 12 Apr 2026 23:36:59.637 * Node ad56f48bae1934f4d1c3141ffee95fa8bf2fa37c () reported node ceb41590bdb53727bffcb314c80a81989180d13a () as not reachable. +2066947:M 12 Apr 2026 23:37:00.764 * Clear FAIL state for node ceb41590bdb53727bffcb314c80a81989180d13a (): replica is reachable again. +2066947:M 12 Apr 2026 23:37:00.952 * Node ad56f48bae1934f4d1c3141ffee95fa8bf2fa37c () reported node ceb41590bdb53727bffcb314c80a81989180d13a () is back online. +2066947:M 12 Apr 2026 23:37:00.956 * Node 27c281c71471f64d57c2f90a6fce7ca7d11f3a9e () reported node ceb41590bdb53727bffcb314c80a81989180d13a () is back online. +2066947:M 12 Apr 2026 23:37:01.648 - Accepting cluster node connection from 127.0.0.1:33046 +2066947:M 12 Apr 2026 23:37:02.169 * NODE ceb41590bdb53727bffcb314c80a81989180d13a () possibly failing. +2066947:M 12 Apr 2026 23:37:03.693 * NODE ad56f48bae1934f4d1c3141ffee95fa8bf2fa37c () possibly failing. +2066947:M 12 Apr 2026 23:37:03.711 * NODE 7cbfdaa3b3c0ebf9bf9116dc86ed8bceffd31db0 () possibly failing. +2066947:M 12 Apr 2026 23:37:03.711 * NODE 27c281c71471f64d57c2f90a6fce7ca7d11f3a9e () possibly failing. +2066947:M 12 Apr 2026 23:37:03.711 # Cluster state changed: fail +2066947:M 12 Apr 2026 23:37:03.711 * FAIL message received from 27c281c71471f64d57c2f90a6fce7ca7d11f3a9e () about ceb41590bdb53727bffcb314c80a81989180d13a () +2066947:M 12 Apr 2026 23:37:04.079 * Node 27c281c71471f64d57c2f90a6fce7ca7d11f3a9e () reported node ceb41590bdb53727bffcb314c80a81989180d13a () as not reachable. +2066947:M 12 Apr 2026 23:37:04.268 - Error accepting cluster node connection: Connection reset by peer +2066947:M 12 Apr 2026 23:37:04.567 - Accepting cluster node connection from 127.0.0.1:47448 +2066947:M 12 Apr 2026 23:37:04.592 - Accepting cluster node connection from 127.0.0.1:47452 +2066947:M 12 Apr 2026 23:37:04.594 * Node 27c281c71471f64d57c2f90a6fce7ca7d11f3a9e () reported node ceb41590bdb53727bffcb314c80a81989180d13a () is back online. +2066947:M 12 Apr 2026 23:37:05.036 * Cluster state changed: ok +2066947:M 12 Apr 2026 23:37:05.209 * Node ad56f48bae1934f4d1c3141ffee95fa8bf2fa37c () reported node 71cce21972c3b30e1e72d50b69266c87048e6a42 () as not reachable. +2066947:M 12 Apr 2026 23:37:05.245 * Node ad56f48bae1934f4d1c3141ffee95fa8bf2fa37c () reported node 71cce21972c3b30e1e72d50b69266c87048e6a42 () is back online. +2066947:M 12 Apr 2026 23:37:06.874 * Node ad56f48bae1934f4d1c3141ffee95fa8bf2fa37c () reported node ceb41590bdb53727bffcb314c80a81989180d13a () as not reachable. +2066947:M 12 Apr 2026 23:37:07.058 * Node 27c281c71471f64d57c2f90a6fce7ca7d11f3a9e () reported node ceb41590bdb53727bffcb314c80a81989180d13a () as not reachable. +2066947:M 12 Apr 2026 23:37:08.576 * Clear FAIL state for node ceb41590bdb53727bffcb314c80a81989180d13a (): primary without slots is reachable again. +2066947:M 12 Apr 2026 23:37:08.788 * Node 27c281c71471f64d57c2f90a6fce7ca7d11f3a9e () reported node ceb41590bdb53727bffcb314c80a81989180d13a () is back online. +2066947:M 12 Apr 2026 23:37:08.795 * Node ad56f48bae1934f4d1c3141ffee95fa8bf2fa37c () reported node ceb41590bdb53727bffcb314c80a81989180d13a () is back online. +2066947:M 12 Apr 2026 23:37:09.064 * A failover occurred in shard 0ccab0ca3bdf2df4fb1d88a6388f0d2e2e9492d9; node ceb41590bdb53727bffcb314c80a81989180d13a () failed over to node 27c281c71471f64d57c2f90a6fce7ca7d11f3a9e () with a config epoch of 5 +2066947:M 12 Apr 2026 23:37:09.065 * Node ceb41590bdb53727bffcb314c80a81989180d13a () is now a replica of node 27c281c71471f64d57c2f90a6fce7ca7d11f3a9e () in shard 0ccab0ca3bdf2df4fb1d88a6388f0d2e2e9492d9 +### Starting test New replica inherits importing slot in tests/unit/cluster/slot-migration.tcl +2066947:M 12 Apr 2026 23:37:12.556 * NODE 71cce21972c3b30e1e72d50b69266c87048e6a42 () possibly failing. +2066947:M 12 Apr 2026 23:37:12.556 * NODE 7cbfdaa3b3c0ebf9bf9116dc86ed8bceffd31db0 () possibly failing. +2066947:M 12 Apr 2026 23:37:12.763 * Node ad56f48bae1934f4d1c3141ffee95fa8bf2fa37c () reported node 71cce21972c3b30e1e72d50b69266c87048e6a42 () as not reachable. +2066947:M 12 Apr 2026 23:37:12.763 * Marking node 71cce21972c3b30e1e72d50b69266c87048e6a42 () as failing (quorum reached). +2066947:M 12 Apr 2026 23:37:12.763 * Node ad56f48bae1934f4d1c3141ffee95fa8bf2fa37c () reported node 7cbfdaa3b3c0ebf9bf9116dc86ed8bceffd31db0 () as not reachable. +2066947:M 12 Apr 2026 23:37:12.764 * Marking node 7cbfdaa3b3c0ebf9bf9116dc86ed8bceffd31db0 () as failing (quorum reached). +2066947:M 12 Apr 2026 23:37:12.837 * Clear FAIL state for node 71cce21972c3b30e1e72d50b69266c87048e6a42 (): replica is reachable again. +2066947:M 12 Apr 2026 23:37:12.837 * Clear FAIL state for node 7cbfdaa3b3c0ebf9bf9116dc86ed8bceffd31db0 (): replica is reachable again. +2066947:M 12 Apr 2026 23:37:12.837 * FAIL message received from 71cce21972c3b30e1e72d50b69266c87048e6a42 () about 7cbfdaa3b3c0ebf9bf9116dc86ed8bceffd31db0 () +2066947:M 12 Apr 2026 23:37:13.019 * Node ad56f48bae1934f4d1c3141ffee95fa8bf2fa37c () reported node 7cbfdaa3b3c0ebf9bf9116dc86ed8bceffd31db0 () is back online. +2066947:M 12 Apr 2026 23:37:13.019 * Node 27c281c71471f64d57c2f90a6fce7ca7d11f3a9e () reported node 7cbfdaa3b3c0ebf9bf9116dc86ed8bceffd31db0 () as not reachable. +2066947:M 12 Apr 2026 23:37:13.019 * Node 27c281c71471f64d57c2f90a6fce7ca7d11f3a9e () reported node 71cce21972c3b30e1e72d50b69266c87048e6a42 () as not reachable. +2066947:M 12 Apr 2026 23:37:13.031 * Clear FAIL state for node 7cbfdaa3b3c0ebf9bf9116dc86ed8bceffd31db0 (): primary without slots is reachable again. +2066947:M 12 Apr 2026 23:37:13.196 * Node 27c281c71471f64d57c2f90a6fce7ca7d11f3a9e () reported node 71cce21972c3b30e1e72d50b69266c87048e6a42 () is back online. +2066947:M 12 Apr 2026 23:37:13.196 * Node 27c281c71471f64d57c2f90a6fce7ca7d11f3a9e () reported node 7cbfdaa3b3c0ebf9bf9116dc86ed8bceffd31db0 () is back online. +2066947:M 12 Apr 2026 23:37:13.196 * Node ad56f48bae1934f4d1c3141ffee95fa8bf2fa37c () reported node 7cbfdaa3b3c0ebf9bf9116dc86ed8bceffd31db0 () as not reachable. +2066947:M 12 Apr 2026 23:37:13.196 * Node ad56f48bae1934f4d1c3141ffee95fa8bf2fa37c () reported node 7cbfdaa3b3c0ebf9bf9116dc86ed8bceffd31db0 () is back online. +2066947:M 12 Apr 2026 23:37:13.196 * Node ad56f48bae1934f4d1c3141ffee95fa8bf2fa37c () reported node 71cce21972c3b30e1e72d50b69266c87048e6a42 () is back online. +2066947:M 12 Apr 2026 23:37:22.553 - Accepting cluster node connection from 127.0.0.1:59818 +2066947:M 12 Apr 2026 23:37:22.657 * Node 27c281c71471f64d57c2f90a6fce7ca7d11f3a9e () reported node 71cce21972c3b30e1e72d50b69266c87048e6a42 () as not reachable. +2066947:M 12 Apr 2026 23:37:22.657 * Node 27c281c71471f64d57c2f90a6fce7ca7d11f3a9e () reported node ceb41590bdb53727bffcb314c80a81989180d13a () as not reachable. +2066947:M 12 Apr 2026 23:37:22.657 * Node 27c281c71471f64d57c2f90a6fce7ca7d11f3a9e () reported node 7cbfdaa3b3c0ebf9bf9116dc86ed8bceffd31db0 () as not reachable. +2066947:M 12 Apr 2026 23:37:22.657 * Node 27c281c71471f64d57c2f90a6fce7ca7d11f3a9e () reported node ad56f48bae1934f4d1c3141ffee95fa8bf2fa37c () as not reachable. +2066947:M 12 Apr 2026 23:37:22.759 * Node 27c281c71471f64d57c2f90a6fce7ca7d11f3a9e () reported node 7cbfdaa3b3c0ebf9bf9116dc86ed8bceffd31db0 () is back online. +2066947:M 12 Apr 2026 23:37:22.759 * Node 27c281c71471f64d57c2f90a6fce7ca7d11f3a9e () reported node ceb41590bdb53727bffcb314c80a81989180d13a () is back online. +2066947:M 12 Apr 2026 23:37:22.759 * Node 27c281c71471f64d57c2f90a6fce7ca7d11f3a9e () reported node ad56f48bae1934f4d1c3141ffee95fa8bf2fa37c () is back online. +2066947:M 12 Apr 2026 23:37:22.855 * Node 27c281c71471f64d57c2f90a6fce7ca7d11f3a9e () reported node 71cce21972c3b30e1e72d50b69266c87048e6a42 () is back online. +2066947:M 12 Apr 2026 23:37:30.086 - Accepting cluster node connection from 127.0.0.1:59828 +2066947:M 12 Apr 2026 23:37:30.089 * FAIL message received from ad56f48bae1934f4d1c3141ffee95fa8bf2fa37c () about 7cbfdaa3b3c0ebf9bf9116dc86ed8bceffd31db0 () +2066947:M 12 Apr 2026 23:37:30.182 * Node ad56f48bae1934f4d1c3141ffee95fa8bf2fa37c () reported node 7cbfdaa3b3c0ebf9bf9116dc86ed8bceffd31db0 () as not reachable. +2066947:M 12 Apr 2026 23:37:30.257 * Clear FAIL state for node 7cbfdaa3b3c0ebf9bf9116dc86ed8bceffd31db0 (): primary without slots is reachable again. +2066947:M 12 Apr 2026 23:37:30.380 * Node ad56f48bae1934f4d1c3141ffee95fa8bf2fa37c () reported node 7cbfdaa3b3c0ebf9bf9116dc86ed8bceffd31db0 () is back online. +2066947:M 12 Apr 2026 23:37:39.656 * NODE 7cbfdaa3b3c0ebf9bf9116dc86ed8bceffd31db0 () possibly failing. +2066947:M 12 Apr 2026 23:37:39.656 * NODE ceb41590bdb53727bffcb314c80a81989180d13a () possibly failing. +2066947:M 12 Apr 2026 23:37:39.776 - Accepting cluster node connection from 127.0.0.1:48386 +2066947:M 12 Apr 2026 23:37:39.839 - Accepting cluster node connection from 127.0.0.1:48388 +2066947:M 12 Apr 2026 23:37:39.843 * Node ad56f48bae1934f4d1c3141ffee95fa8bf2fa37c () reported node 27c281c71471f64d57c2f90a6fce7ca7d11f3a9e () as not reachable. +2066947:M 12 Apr 2026 23:37:39.843 * Node ad56f48bae1934f4d1c3141ffee95fa8bf2fa37c () reported node ceb41590bdb53727bffcb314c80a81989180d13a () as not reachable. +2066947:M 12 Apr 2026 23:37:39.843 * Marking node ceb41590bdb53727bffcb314c80a81989180d13a () as failing (quorum reached). +2066947:M 12 Apr 2026 23:37:39.878 * Node 27c281c71471f64d57c2f90a6fce7ca7d11f3a9e () reported node ceb41590bdb53727bffcb314c80a81989180d13a () as not reachable. +2066947:M 12 Apr 2026 23:37:39.878 * Node 27c281c71471f64d57c2f90a6fce7ca7d11f3a9e () reported node 7cbfdaa3b3c0ebf9bf9116dc86ed8bceffd31db0 () as not reachable. +2066947:M 12 Apr 2026 23:37:39.879 * FAIL message received from 27c281c71471f64d57c2f90a6fce7ca7d11f3a9e () about 7cbfdaa3b3c0ebf9bf9116dc86ed8bceffd31db0 () +2066947:M 12 Apr 2026 23:37:39.914 * Clear FAIL state for node ceb41590bdb53727bffcb314c80a81989180d13a (): replica is reachable again. +2066947:M 12 Apr 2026 23:37:40.009 * Node ad56f48bae1934f4d1c3141ffee95fa8bf2fa37c () reported node 27c281c71471f64d57c2f90a6fce7ca7d11f3a9e () is back online. +2066947:M 12 Apr 2026 23:37:40.009 * Node ad56f48bae1934f4d1c3141ffee95fa8bf2fa37c () reported node ceb41590bdb53727bffcb314c80a81989180d13a () is back online. +2066947:M 12 Apr 2026 23:37:40.009 * Clear FAIL state for node 7cbfdaa3b3c0ebf9bf9116dc86ed8bceffd31db0 (): primary without slots is reachable again. +2066947:M 12 Apr 2026 23:37:40.076 * Node ad56f48bae1934f4d1c3141ffee95fa8bf2fa37c () reported node 7cbfdaa3b3c0ebf9bf9116dc86ed8bceffd31db0 () as not reachable. +2066947:M 12 Apr 2026 23:37:40.217 * Node 27c281c71471f64d57c2f90a6fce7ca7d11f3a9e () reported node 7cbfdaa3b3c0ebf9bf9116dc86ed8bceffd31db0 () is back online. +2066947:M 12 Apr 2026 23:37:40.254 * Node 27c281c71471f64d57c2f90a6fce7ca7d11f3a9e () reported node ceb41590bdb53727bffcb314c80a81989180d13a () is back online. +2066947:M 12 Apr 2026 23:37:40.285 * Node ad56f48bae1934f4d1c3141ffee95fa8bf2fa37c () reported node 7cbfdaa3b3c0ebf9bf9116dc86ed8bceffd31db0 () is back online. +2066947:M 12 Apr 2026 23:37:49.345 * NODE 7cbfdaa3b3c0ebf9bf9116dc86ed8bceffd31db0 () possibly failing. +2066947:M 12 Apr 2026 23:37:57.841 * NODE 27c281c71471f64d57c2f90a6fce7ca7d11f3a9e () possibly failing. +2066947:M 12 Apr 2026 23:37:57.943 * FAIL message received from ad56f48bae1934f4d1c3141ffee95fa8bf2fa37c () about 27c281c71471f64d57c2f90a6fce7ca7d11f3a9e () +2066947:M 12 Apr 2026 23:37:57.943 # Cluster state changed: fail +2066947:M 12 Apr 2026 23:37:57.984 * Node ad56f48bae1934f4d1c3141ffee95fa8bf2fa37c () reported node 27c281c71471f64d57c2f90a6fce7ca7d11f3a9e () as not reachable. +2066947:M 12 Apr 2026 23:37:58.369 * NODE 71cce21972c3b30e1e72d50b69266c87048e6a42 () possibly failing. +2066947:M 12 Apr 2026 23:37:58.480 * FAIL message received from ad56f48bae1934f4d1c3141ffee95fa8bf2fa37c () about 71cce21972c3b30e1e72d50b69266c87048e6a42 () +2066947:M 12 Apr 2026 23:37:58.516 * Node ad56f48bae1934f4d1c3141ffee95fa8bf2fa37c () reported node 71cce21972c3b30e1e72d50b69266c87048e6a42 () as not reachable. +2066947:M 12 Apr 2026 23:37:58.574 * NODE 7cbfdaa3b3c0ebf9bf9116dc86ed8bceffd31db0 () possibly failing. +2066947:M 12 Apr 2026 23:37:58.574 * NODE ceb41590bdb53727bffcb314c80a81989180d13a () possibly failing. +2066947:M 12 Apr 2026 23:37:58.700 * Clear FAIL state for node 71cce21972c3b30e1e72d50b69266c87048e6a42 (): replica is reachable again. +2066947:M 12 Apr 2026 23:37:58.701 - Accepting cluster node connection from 127.0.0.1:47586 +2066947:M 12 Apr 2026 23:37:58.842 * Node ad56f48bae1934f4d1c3141ffee95fa8bf2fa37c () reported node 71cce21972c3b30e1e72d50b69266c87048e6a42 () is back online. +2066947:M 12 Apr 2026 23:37:59.569 * Failover auth granted to ceb41590bdb53727bffcb314c80a81989180d13a () for epoch 9 +2066947:M 12 Apr 2026 23:37:59.718 * Cluster state changed: ok +2066947:M 12 Apr 2026 23:37:59.761 * Node ceb41590bdb53727bffcb314c80a81989180d13a () reported node 27c281c71471f64d57c2f90a6fce7ca7d11f3a9e () as not reachable. +2066947:M 12 Apr 2026 23:37:59.804 * Clear FAIL state for node 27c281c71471f64d57c2f90a6fce7ca7d11f3a9e (): primary without slots is reachable again. +2066947:M 12 Apr 2026 23:37:59.804 * A failover occurred in shard 0ccab0ca3bdf2df4fb1d88a6388f0d2e2e9492d9; node 27c281c71471f64d57c2f90a6fce7ca7d11f3a9e () failed over to node ceb41590bdb53727bffcb314c80a81989180d13a () with a config epoch of 9 +2066947:M 12 Apr 2026 23:37:59.804 * Node 27c281c71471f64d57c2f90a6fce7ca7d11f3a9e () is now a replica of node ceb41590bdb53727bffcb314c80a81989180d13a () in shard 0ccab0ca3bdf2df4fb1d88a6388f0d2e2e9492d9 +2066947:M 12 Apr 2026 23:37:59.962 * Node ad56f48bae1934f4d1c3141ffee95fa8bf2fa37c () reported node 27c281c71471f64d57c2f90a6fce7ca7d11f3a9e () is back online. +2066947:M 12 Apr 2026 23:37:59.970 * Node ceb41590bdb53727bffcb314c80a81989180d13a () reported node 27c281c71471f64d57c2f90a6fce7ca7d11f3a9e () is back online. +2066947:M 12 Apr 2026 23:38:07.748 * NODE 71cce21972c3b30e1e72d50b69266c87048e6a42 () possibly failing. +2066947:M 12 Apr 2026 23:38:07.886 * Node ad56f48bae1934f4d1c3141ffee95fa8bf2fa37c () reported node 71cce21972c3b30e1e72d50b69266c87048e6a42 () as not reachable. +2066947:M 12 Apr 2026 23:38:07.886 * Marking node 71cce21972c3b30e1e72d50b69266c87048e6a42 () as failing (quorum reached). +2066947:M 12 Apr 2026 23:38:07.937 * Clear FAIL state for node 71cce21972c3b30e1e72d50b69266c87048e6a42 (): replica is reachable again. +2066947:M 12 Apr 2026 23:38:07.955 * Node ad56f48bae1934f4d1c3141ffee95fa8bf2fa37c () reported node 71cce21972c3b30e1e72d50b69266c87048e6a42 () is back online. +2066947:M 12 Apr 2026 23:38:16.456 * NODE 71cce21972c3b30e1e72d50b69266c87048e6a42 () possibly failing. +2066947:M 12 Apr 2026 23:38:16.456 * NODE ceb41590bdb53727bffcb314c80a81989180d13a () possibly failing. +2066947:M 12 Apr 2026 23:38:17.641 * NODE ad56f48bae1934f4d1c3141ffee95fa8bf2fa37c () possibly failing. +2066947:M 12 Apr 2026 23:38:17.641 # Cluster state changed: fail +2066947:M 12 Apr 2026 23:38:17.937 * Node ad56f48bae1934f4d1c3141ffee95fa8bf2fa37c () reported node 27c281c71471f64d57c2f90a6fce7ca7d11f3a9e () as not reachable. +2066947:M 12 Apr 2026 23:38:17.988 - Accepting cluster node connection from 127.0.0.1:33656 +2066947:M 12 Apr 2026 23:38:18.104 * Node ad56f48bae1934f4d1c3141ffee95fa8bf2fa37c () reported node 71cce21972c3b30e1e72d50b69266c87048e6a42 () as not reachable. +2066947:M 12 Apr 2026 23:38:18.104 * Node ad56f48bae1934f4d1c3141ffee95fa8bf2fa37c () reported node 7cbfdaa3b3c0ebf9bf9116dc86ed8bceffd31db0 () as not reachable. +2066947:M 12 Apr 2026 23:38:18.150 * Node ad56f48bae1934f4d1c3141ffee95fa8bf2fa37c () reported node 71cce21972c3b30e1e72d50b69266c87048e6a42 () is back online. +2066947:M 12 Apr 2026 23:38:18.319 * Node ad56f48bae1934f4d1c3141ffee95fa8bf2fa37c () reported node 27c281c71471f64d57c2f90a6fce7ca7d11f3a9e () is back online. +2066947:M 12 Apr 2026 23:38:18.319 * Node ad56f48bae1934f4d1c3141ffee95fa8bf2fa37c () reported node 7cbfdaa3b3c0ebf9bf9116dc86ed8bceffd31db0 () is back online. +2066947:M 12 Apr 2026 23:38:18.519 * Node ad56f48bae1934f4d1c3141ffee95fa8bf2fa37c () reported node ceb41590bdb53727bffcb314c80a81989180d13a () as not reachable. +2066947:M 12 Apr 2026 23:38:18.942 * Cluster state changed: ok +2066947:M 12 Apr 2026 23:38:20.073 * Node ad56f48bae1934f4d1c3141ffee95fa8bf2fa37c () reported node ceb41590bdb53727bffcb314c80a81989180d13a () is back online. +2066947:M 12 Apr 2026 23:38:26.712 * FAIL message received from ceb41590bdb53727bffcb314c80a81989180d13a () about 7cbfdaa3b3c0ebf9bf9116dc86ed8bceffd31db0 () +2066947:M 12 Apr 2026 23:38:26.759 * Node ad56f48bae1934f4d1c3141ffee95fa8bf2fa37c () reported node 7cbfdaa3b3c0ebf9bf9116dc86ed8bceffd31db0 () as not reachable. +2066947:M 12 Apr 2026 23:38:26.877 * Clear FAIL state for node 7cbfdaa3b3c0ebf9bf9116dc86ed8bceffd31db0 (): primary without slots is reachable again. +2066947:M 12 Apr 2026 23:38:26.950 * Node ad56f48bae1934f4d1c3141ffee95fa8bf2fa37c () reported node 7cbfdaa3b3c0ebf9bf9116dc86ed8bceffd31db0 () is back online. +2066947:M 12 Apr 2026 23:38:35.618 - Accepting cluster node connection from 127.0.0.1:41464 +2066947:M 12 Apr 2026 23:38:35.657 * Node ad56f48bae1934f4d1c3141ffee95fa8bf2fa37c () reported node ceb41590bdb53727bffcb314c80a81989180d13a () as not reachable. +2066947:M 12 Apr 2026 23:38:35.702 * Node ad56f48bae1934f4d1c3141ffee95fa8bf2fa37c () reported node ceb41590bdb53727bffcb314c80a81989180d13a () is back online. +2066947:M 12 Apr 2026 23:38:44.684 * NODE ad56f48bae1934f4d1c3141ffee95fa8bf2fa37c () possibly failing. +2066947:M 12 Apr 2026 23:38:44.684 * NODE 7cbfdaa3b3c0ebf9bf9116dc86ed8bceffd31db0 () possibly failing. +2066947:M 12 Apr 2026 23:38:44.684 * NODE ceb41590bdb53727bffcb314c80a81989180d13a () possibly failing. +2066947:M 12 Apr 2026 23:38:44.684 * NODE 27c281c71471f64d57c2f90a6fce7ca7d11f3a9e () possibly failing. +2066947:M 12 Apr 2026 23:38:44.684 # Cluster state changed: fail +2066947:M 12 Apr 2026 23:38:44.685 - Accepting cluster node connection from 127.0.0.1:49560 +2066947:M 12 Apr 2026 23:38:44.690 - Accepting cluster node connection from 127.0.0.1:49568 +2066947:M 12 Apr 2026 23:38:44.701 - Accepting cluster node connection from 127.0.0.1:49574 +2066947:M 12 Apr 2026 23:38:44.785 - Accepting cluster node connection from 127.0.0.1:49582 +2066947:M 12 Apr 2026 23:38:44.842 * Node ceb41590bdb53727bffcb314c80a81989180d13a () reported node 7cbfdaa3b3c0ebf9bf9116dc86ed8bceffd31db0 () as not reachable. +2066947:M 12 Apr 2026 23:38:44.842 * Marking node 7cbfdaa3b3c0ebf9bf9116dc86ed8bceffd31db0 () as failing (quorum reached). +2066947:M 12 Apr 2026 23:38:44.860 * Node ad56f48bae1934f4d1c3141ffee95fa8bf2fa37c () reported node 7cbfdaa3b3c0ebf9bf9116dc86ed8bceffd31db0 () as not reachable. +2066947:M 12 Apr 2026 23:38:44.900 * Clear FAIL state for node 7cbfdaa3b3c0ebf9bf9116dc86ed8bceffd31db0 (): primary without slots is reachable again. +2066947:M 12 Apr 2026 23:38:44.989 * Node ceb41590bdb53727bffcb314c80a81989180d13a () reported node 7cbfdaa3b3c0ebf9bf9116dc86ed8bceffd31db0 () is back online. +2066947:M 12 Apr 2026 23:38:45.616 * Node ad56f48bae1934f4d1c3141ffee95fa8bf2fa37c () reported node 7cbfdaa3b3c0ebf9bf9116dc86ed8bceffd31db0 () is back online. +2066947:M 12 Apr 2026 23:38:45.921 * Cluster state changed: ok +2066947:M 12 Apr 2026 23:38:54.106 * NODE 71cce21972c3b30e1e72d50b69266c87048e6a42 () possibly failing. +2066947:M 12 Apr 2026 23:38:54.106 * NODE ad56f48bae1934f4d1c3141ffee95fa8bf2fa37c () possibly failing. +2066947:M 12 Apr 2026 23:38:54.106 * NODE 7cbfdaa3b3c0ebf9bf9116dc86ed8bceffd31db0 () possibly failing. +2066947:M 12 Apr 2026 23:38:54.106 * NODE ceb41590bdb53727bffcb314c80a81989180d13a () possibly failing. +2066947:M 12 Apr 2026 23:38:54.106 * NODE 27c281c71471f64d57c2f90a6fce7ca7d11f3a9e () possibly failing. +2066947:M 12 Apr 2026 23:38:54.106 # Cluster state changed: fail +2066947:M 12 Apr 2026 23:38:55.319 * Cluster state changed: ok +2066947:M 12 Apr 2026 23:39:03.154 * NODE 7cbfdaa3b3c0ebf9bf9116dc86ed8bceffd31db0 () possibly failing. +2066947:M 12 Apr 2026 23:39:03.159 - Accepting cluster node connection from 127.0.0.1:47732 +2066947:M 12 Apr 2026 23:39:03.162 - Error accepting cluster node connection: error:0A000126:SSL routines::unexpected eof while reading +2066947:M 12 Apr 2026 23:39:03.167 - Accepting cluster node connection from 127.0.0.1:47744 +2066947:M 12 Apr 2026 23:39:03.170 - Error accepting cluster node connection: error:0A000126:SSL routines::unexpected eof while reading +2066947:M 12 Apr 2026 23:39:03.170 - Accepting cluster node connection from 127.0.0.1:47750 +2066947:M 12 Apr 2026 23:39:03.183 * Node ceb41590bdb53727bffcb314c80a81989180d13a () reported node 7cbfdaa3b3c0ebf9bf9116dc86ed8bceffd31db0 () as not reachable. +2066947:M 12 Apr 2026 23:39:03.183 * Marking node 7cbfdaa3b3c0ebf9bf9116dc86ed8bceffd31db0 () as failing (quorum reached). +2066947:M 12 Apr 2026 23:39:03.256 - Accepting cluster node connection from 127.0.0.1:60092 +2066947:M 12 Apr 2026 23:39:03.258 - Accepting cluster node connection from 127.0.0.1:60098 +2066947:M 12 Apr 2026 23:39:03.304 * Clear FAIL state for node 7cbfdaa3b3c0ebf9bf9116dc86ed8bceffd31db0 (): primary without slots is reachable again. +2066947:M 12 Apr 2026 23:39:03.375 * Node ceb41590bdb53727bffcb314c80a81989180d13a () reported node 7cbfdaa3b3c0ebf9bf9116dc86ed8bceffd31db0 () is back online. +2066947:M 12 Apr 2026 23:39:21.127 - Accepting cluster node connection from 127.0.0.1:50520 +2066947:M 12 Apr 2026 23:39:21.134 - Accepting cluster node connection from 127.0.0.1:50534 +2066947:M 12 Apr 2026 23:39:21.136 - Accepting cluster node connection from 127.0.0.1:50538 +2066947:M 12 Apr 2026 23:39:21.145 * Node ceb41590bdb53727bffcb314c80a81989180d13a () reported node 71cce21972c3b30e1e72d50b69266c87048e6a42 () as not reachable. +2066947:M 12 Apr 2026 23:39:21.145 * Node ceb41590bdb53727bffcb314c80a81989180d13a () reported node 7cbfdaa3b3c0ebf9bf9116dc86ed8bceffd31db0 () as not reachable. +2066947:M 12 Apr 2026 23:39:21.145 * Node ceb41590bdb53727bffcb314c80a81989180d13a () reported node ad56f48bae1934f4d1c3141ffee95fa8bf2fa37c () as not reachable. +2066947:M 12 Apr 2026 23:39:21.145 * Node ceb41590bdb53727bffcb314c80a81989180d13a () reported node 27c281c71471f64d57c2f90a6fce7ca7d11f3a9e () as not reachable. +2066947:M 12 Apr 2026 23:39:21.151 * Node ad56f48bae1934f4d1c3141ffee95fa8bf2fa37c () reported node 71cce21972c3b30e1e72d50b69266c87048e6a42 () as not reachable. +2066947:M 12 Apr 2026 23:39:21.151 * Node ad56f48bae1934f4d1c3141ffee95fa8bf2fa37c () reported node 27c281c71471f64d57c2f90a6fce7ca7d11f3a9e () as not reachable. +2066947:M 12 Apr 2026 23:39:21.151 * Node ad56f48bae1934f4d1c3141ffee95fa8bf2fa37c () reported node ceb41590bdb53727bffcb314c80a81989180d13a () as not reachable. +2066947:M 12 Apr 2026 23:39:21.151 * Node ad56f48bae1934f4d1c3141ffee95fa8bf2fa37c () reported node 7cbfdaa3b3c0ebf9bf9116dc86ed8bceffd31db0 () as not reachable. +2066947:M 12 Apr 2026 23:39:21.163 * FAIL message received from ceb41590bdb53727bffcb314c80a81989180d13a () about 71cce21972c3b30e1e72d50b69266c87048e6a42 () +2066947:M 12 Apr 2026 23:39:21.333 * Node ad56f48bae1934f4d1c3141ffee95fa8bf2fa37c () reported node 27c281c71471f64d57c2f90a6fce7ca7d11f3a9e () is back online. +2066947:M 12 Apr 2026 23:39:21.333 * Node ad56f48bae1934f4d1c3141ffee95fa8bf2fa37c () reported node 7cbfdaa3b3c0ebf9bf9116dc86ed8bceffd31db0 () is back online. +2066947:M 12 Apr 2026 23:39:21.336 * Clear FAIL state for node 71cce21972c3b30e1e72d50b69266c87048e6a42 (): replica is reachable again. +2066947:M 12 Apr 2026 23:39:21.336 * Node ad56f48bae1934f4d1c3141ffee95fa8bf2fa37c () reported node 71cce21972c3b30e1e72d50b69266c87048e6a42 () is back online. +2066947:M 12 Apr 2026 23:39:21.336 * Node ad56f48bae1934f4d1c3141ffee95fa8bf2fa37c () reported node ceb41590bdb53727bffcb314c80a81989180d13a () is back online. +2066947:M 12 Apr 2026 23:39:21.336 * Node ceb41590bdb53727bffcb314c80a81989180d13a () reported node 27c281c71471f64d57c2f90a6fce7ca7d11f3a9e () is back online. +2066947:M 12 Apr 2026 23:39:21.336 * Node ceb41590bdb53727bffcb314c80a81989180d13a () reported node ad56f48bae1934f4d1c3141ffee95fa8bf2fa37c () is back online. +2066947:M 12 Apr 2026 23:39:21.341 * Node ceb41590bdb53727bffcb314c80a81989180d13a () reported node 71cce21972c3b30e1e72d50b69266c87048e6a42 () is back online. +2066947:M 12 Apr 2026 23:39:21.538 * Node ceb41590bdb53727bffcb314c80a81989180d13a () reported node 7cbfdaa3b3c0ebf9bf9116dc86ed8bceffd31db0 () is back online. +2066947:M 12 Apr 2026 23:39:22.043 - Connection with Node 27c281c71471f64d57c2f90a6fce7ca7d11f3a9e at 127.0.0.1:31699 failed: (null) +===== End of server log (pid 2066947) ===== + + +===== Start of server stderr log (pid 2066947) ===== + + +===== End of server stderr log (pid 2066947) ===== + + +===== Start of server log (pid 2066855) ===== + +### Starting server for test +2066855:C 12 Apr 2026 23:36:26.367 # WARNING: Changing databases number from 16 to 1 since we are in cluster mode +2066855:C 12 Apr 2026 23:36:26.367 # WARNING Memory overcommit must be enabled! Without it, a background save or replication may fail under low memory condition. Being disabled, it can also cause failures without low memory condition, see https://github.com/jemalloc/jemalloc/issues/1328. To fix this issue add 'vm.overcommit_memory = 1' to /etc/sysctl.conf and then reboot or run the command 'sysctl vm.overcommit_memory=1' for this to take effect. +2066855:C 12 Apr 2026 23:36:26.367 * oO0OoO0OoO0Oo Valkey is starting oO0OoO0OoO0Oo +2066855:C 12 Apr 2026 23:36:26.367 * Valkey version=8.0.2, bits=64, commit=00000000, modified=0, pid=2066855, just started +2066855:C 12 Apr 2026 23:36:26.367 * Configuration loaded +2066855:M 12 Apr 2026 23:36:26.367 * Increased maximum number of open files to 10032 (it was originally set to 1024). +2066855:M 12 Apr 2026 23:36:26.367 * monotonic clock: POSIX clock_gettime +2066855:M 12 Apr 2026 23:36:26.368 # Failed to write PID file: Permission denied + .+^+. + .+#########+. + .+########+########+. Valkey 8.0.2 (00000000/0) 64 bit + .+########+' '+########+. + .########+' .+. '+########. Running in cluster mode + |####+' .+#######+. '+####| Port: 21694 + |###| .+###############+. |###| PID: 2066855 + |###| |#####*'' ''*#####| |###| + |###| |####' .-. '####| |###| + |###| |###( (@@@) )###| |###| https://valkey.io + |###| |####. '-' .####| |###| + |###| |#####*. .*#####| |###| + |###| '+#####| |#####+' |###| + |####+. +##| |#+' .+####| + '#######+ |##| .+########' + '+###| |##| .+########+' + '| |####+########+' + +#########+' + '+v+' + +2066855:M 12 Apr 2026 23:36:26.369 * No cluster configuration found, I'm ceb41590bdb53727bffcb314c80a81989180d13a +2066855:M 12 Apr 2026 23:36:26.388 * Server initialized +2066855:M 12 Apr 2026 23:36:26.388 * Ready to accept connections tcp +2066855:M 12 Apr 2026 23:36:26.388 * Ready to accept connections unix +2066855:M 12 Apr 2026 23:36:26.388 * Ready to accept connections tls +2066855:M 12 Apr 2026 23:36:26.488 - Accepted 127.0.0.1:34711 +2066855:M 12 Apr 2026 23:36:26.500 - Reading from client: error:0A000126:SSL routines::unexpected eof while reading +2066855:M 12 Apr 2026 23:36:26.511 - Accepted 127.0.0.1:38427 +2066855:M 12 Apr 2026 23:36:27.360 - Accepting cluster node connection from 127.0.0.1:47270 +2066855:M 12 Apr 2026 23:36:27.416 * IP address for this node updated to 127.0.0.1 +2066855:M 12 Apr 2026 23:36:27.465 - Accepting cluster node connection from 127.0.0.1:47276 +2066855:M 12 Apr 2026 23:36:27.538 * Successfully completed handshake with 71cce21972c3b30e1e72d50b69266c87048e6a42 () +2066855:M 12 Apr 2026 23:36:27.538 * Successfully completed handshake with ad56f48bae1934f4d1c3141ffee95fa8bf2fa37c () +2066855:M 12 Apr 2026 23:36:27.550 * Successfully completed handshake with 7cbfdaa3b3c0ebf9bf9116dc86ed8bceffd31db0 () +2066855:M 12 Apr 2026 23:36:27.591 - Accepting cluster node connection from 127.0.0.1:47282 +2066855:M 12 Apr 2026 23:36:27.594 - Accepting cluster node connection from 127.0.0.1:47296 +2066855:M 12 Apr 2026 23:36:27.614 * Address updated for node ad56f48bae1934f4d1c3141ffee95fa8bf2fa37c (), now 127.0.0.1:21697 +2066855:M 12 Apr 2026 23:36:27.621 - Accepting cluster node connection from 127.0.0.1:47300 +2066855:M 12 Apr 2026 23:36:27.651 * Address updated for node 7cbfdaa3b3c0ebf9bf9116dc86ed8bceffd31db0 (), now 127.0.0.1:21691 +2066855:M 12 Apr 2026 23:36:27.651 * Address updated for node 71cce21972c3b30e1e72d50b69266c87048e6a42 (), now 127.0.0.1:21689 +2066855:M 12 Apr 2026 23:36:27.684 - Accepting cluster node connection from 127.0.0.1:47310 +2066855:M 12 Apr 2026 23:36:27.688 - Accepting cluster node connection from 127.0.0.1:47316 +2066855:M 12 Apr 2026 23:36:27.717 - Accepting cluster node connection from 127.0.0.1:47324 +2066855:M 12 Apr 2026 23:36:27.719 - Accepting cluster node connection from 127.0.0.1:47332 +2066855:M 12 Apr 2026 23:36:27.739 * configEpoch collision with node d24296bca8e44a56521bbf138bec235238274ad7 (). configEpoch set to 3 +2066855:S 12 Apr 2026 23:36:27.953 * Connecting to PRIMARY 127.0.0.1:21699 +2066855:S 12 Apr 2026 23:36:27.953 * PRIMARY <-> REPLICA sync started +2066855:S 12 Apr 2026 23:36:27.953 * Cluster state changed: ok +2066855:S 12 Apr 2026 23:36:27.969 * Non blocking connect for SYNC fired the event. +2066855:S 12 Apr 2026 23:36:27.969 * Node 7cbfdaa3b3c0ebf9bf9116dc86ed8bceffd31db0 () is no longer primary of shard 6d5a08da00fb925618c9429f0a0abf50cb184dee; removed all 0 slot(s) it used to own +2066855:S 12 Apr 2026 23:36:27.969 * Node 7cbfdaa3b3c0ebf9bf9116dc86ed8bceffd31db0 () is now part of shard fb1332a38fbff59dddf699f62b6c363413af8689 +2066855:S 12 Apr 2026 23:36:27.969 * Node 7cbfdaa3b3c0ebf9bf9116dc86ed8bceffd31db0 () is now a replica of node ad56f48bae1934f4d1c3141ffee95fa8bf2fa37c () in shard fb1332a38fbff59dddf699f62b6c363413af8689 +2066855:S 12 Apr 2026 23:36:27.976 * Primary replied to PING, replication can continue... +2066855:S 12 Apr 2026 23:36:27.976 * Node 71cce21972c3b30e1e72d50b69266c87048e6a42 () is no longer primary of shard 797d84842d0558c353ef2fd200fec2d129d7c96f; removed all 0 slot(s) it used to own +2066855:S 12 Apr 2026 23:36:27.976 * Node 71cce21972c3b30e1e72d50b69266c87048e6a42 () is now part of shard b553502a57202365d35122f1f841b02047c3c23d +2066855:S 12 Apr 2026 23:36:27.976 * Node 71cce21972c3b30e1e72d50b69266c87048e6a42 () is now a replica of node d24296bca8e44a56521bbf138bec235238274ad7 () in shard b553502a57202365d35122f1f841b02047c3c23d +2066855:S 12 Apr 2026 23:36:27.981 * Partial resynchronization not possible (no cached primary) +2066855:S 12 Apr 2026 23:36:27.982 * Full resync from primary: 3434ac96bb7f1503c85d89d9ee94a4c782639d94:0 +2066855:S 12 Apr 2026 23:36:27.989 * PRIMARY <-> REPLICA sync: receiving streamed RDB from primary with EOF to disk +2066855:S 12 Apr 2026 23:36:28.003 * PRIMARY <-> REPLICA sync: Flushing old data +2066855:S 12 Apr 2026 23:36:28.003 * PRIMARY <-> REPLICA sync: Loading DB in memory +2066855:S 12 Apr 2026 23:36:28.003 * Loading RDB produced by Valkey version 8.0.2 +2066855:S 12 Apr 2026 23:36:28.003 * RDB age 1 seconds +2066855:S 12 Apr 2026 23:36:28.003 * RDB memory usage when created 2.83 Mb +2066855:S 12 Apr 2026 23:36:28.003 * Done loading RDB, keys loaded: 0, keys expired: 0. +2066855:S 12 Apr 2026 23:36:28.003 * PRIMARY <-> REPLICA sync: Finished with success +2066855:S 12 Apr 2026 23:36:28.007 # DEBUG LOG: ========== I am replica 3 ========== +2066855:S 12 Apr 2026 23:36:35.661 * NODE 71cce21972c3b30e1e72d50b69266c87048e6a42 () possibly failing. +### Starting test Slot migration states are replicated in tests/unit/cluster/slot-migration.tcl +2066855:S 12 Apr 2026 23:36:37.788 * Migrating slot 609 to node ad56f48bae1934f4d1c3141ffee95fa8bf2fa37c () +### Starting test Migration target is auto-updated after failover in target shard in tests/unit/cluster/slot-migration.tcl +2066855:S 12 Apr 2026 23:36:38.888 * NODE ad56f48bae1934f4d1c3141ffee95fa8bf2fa37c () possibly failing. +2066855:S 12 Apr 2026 23:36:39.092 * Node d24296bca8e44a56521bbf138bec235238274ad7 () reported node ad56f48bae1934f4d1c3141ffee95fa8bf2fa37c () as not reachable. +2066855:S 12 Apr 2026 23:36:39.107 * FAIL message received from 27c281c71471f64d57c2f90a6fce7ca7d11f3a9e () about ad56f48bae1934f4d1c3141ffee95fa8bf2fa37c () +2066855:S 12 Apr 2026 23:36:39.107 # Cluster state changed: fail +2066855:S 12 Apr 2026 23:36:39.149 * Node 27c281c71471f64d57c2f90a6fce7ca7d11f3a9e () reported node ad56f48bae1934f4d1c3141ffee95fa8bf2fa37c () as not reachable. +2066855:S 12 Apr 2026 23:36:39.979 * Failover occurred in migration target. Slot 609 is now being migrated to node 7cbfdaa3b3c0ebf9bf9116dc86ed8bceffd31db0 () in shard fb1332a38fbff59dddf699f62b6c363413af8689. +2066855:S 12 Apr 2026 23:36:39.980 * Node 7cbfdaa3b3c0ebf9bf9116dc86ed8bceffd31db0 () reported node ad56f48bae1934f4d1c3141ffee95fa8bf2fa37c () as not reachable. +2066855:S 12 Apr 2026 23:36:39.980 * Cluster state changed: ok +2066855:S 12 Apr 2026 23:36:40.925 - Node ad56f48bae1934f4d1c3141ffee95fa8bf2fa37c has old slots configuration, sending an UPDATE message about 7cbfdaa3b3c0ebf9bf9116dc86ed8bceffd31db0 +2066855:S 12 Apr 2026 23:36:41.181 * A failover occurred in shard fb1332a38fbff59dddf699f62b6c363413af8689; node ad56f48bae1934f4d1c3141ffee95fa8bf2fa37c () failed over to node 7cbfdaa3b3c0ebf9bf9116dc86ed8bceffd31db0 () with a config epoch of 6 +2066855:S 12 Apr 2026 23:36:41.181 * Node ad56f48bae1934f4d1c3141ffee95fa8bf2fa37c () is now a replica of node 7cbfdaa3b3c0ebf9bf9116dc86ed8bceffd31db0 () in shard fb1332a38fbff59dddf699f62b6c363413af8689 +2066855:S 12 Apr 2026 23:36:41.189 * Clear FAIL state for node ad56f48bae1934f4d1c3141ffee95fa8bf2fa37c (): replica is reachable again. +2066855:S 12 Apr 2026 23:36:41.215 * Node 7cbfdaa3b3c0ebf9bf9116dc86ed8bceffd31db0 () reported node ad56f48bae1934f4d1c3141ffee95fa8bf2fa37c () is back online. +2066855:S 12 Apr 2026 23:36:41.260 * Node 27c281c71471f64d57c2f90a6fce7ca7d11f3a9e () reported node ad56f48bae1934f4d1c3141ffee95fa8bf2fa37c () is back online. +2066855:S 12 Apr 2026 23:36:41.289 * Node d24296bca8e44a56521bbf138bec235238274ad7 () reported node ad56f48bae1934f4d1c3141ffee95fa8bf2fa37c () is back online. +2066855:S 12 Apr 2026 23:36:41.322 * Failover occurred in migration target. Slot 609 is now being migrated to node ad56f48bae1934f4d1c3141ffee95fa8bf2fa37c () in shard fb1332a38fbff59dddf699f62b6c363413af8689. +2066855:S 12 Apr 2026 23:36:41.322 - Node 7cbfdaa3b3c0ebf9bf9116dc86ed8bceffd31db0 has old slots configuration, sending an UPDATE message about ad56f48bae1934f4d1c3141ffee95fa8bf2fa37c +2066855:S 12 Apr 2026 23:36:41.417 * A failover occurred in shard fb1332a38fbff59dddf699f62b6c363413af8689; node 7cbfdaa3b3c0ebf9bf9116dc86ed8bceffd31db0 () failed over to node ad56f48bae1934f4d1c3141ffee95fa8bf2fa37c () with a config epoch of 7 +2066855:S 12 Apr 2026 23:36:41.417 * Node 7cbfdaa3b3c0ebf9bf9116dc86ed8bceffd31db0 () is now a replica of node ad56f48bae1934f4d1c3141ffee95fa8bf2fa37c () in shard fb1332a38fbff59dddf699f62b6c363413af8689 +### Starting test Migration source is auto-updated after failover in source shard in tests/unit/cluster/slot-migration.tcl +2066855:S 12 Apr 2026 23:36:42.032 - DB 0: 1 keys (0 volatile) in 4 slots HT. +2066855:S 12 Apr 2026 23:36:42.749 * NODE 27c281c71471f64d57c2f90a6fce7ca7d11f3a9e () possibly failing. +2066855:S 12 Apr 2026 23:36:42.859 * Node ad56f48bae1934f4d1c3141ffee95fa8bf2fa37c () reported node 27c281c71471f64d57c2f90a6fce7ca7d11f3a9e () as not reachable. +2066855:S 12 Apr 2026 23:36:42.859 * Node d24296bca8e44a56521bbf138bec235238274ad7 () reported node 27c281c71471f64d57c2f90a6fce7ca7d11f3a9e () as not reachable. +2066855:S 12 Apr 2026 23:36:42.859 * Marking node 27c281c71471f64d57c2f90a6fce7ca7d11f3a9e () as failing (quorum reached). +2066855:S 12 Apr 2026 23:36:42.859 # Cluster state changed: fail +2066855:S 12 Apr 2026 23:36:43.121 * Start of election delayed for 930 milliseconds (rank #0, offset 212). +2066855:S 12 Apr 2026 23:36:43.221 * Currently unable to failover: Waiting the delay before I can start a new failover. +2066855:S 12 Apr 2026 23:36:45.657 * Starting a failover election for epoch 8. +2066855:S 12 Apr 2026 23:36:47.228 - Accepting cluster node connection from 127.0.0.1:53458 +2066855:S 12 Apr 2026 23:36:47.230 - Accepting cluster node connection from 127.0.0.1:53456 +2066855:S 12 Apr 2026 23:36:47.240 - Accepting cluster node connection from 127.0.0.1:53466 +2066855:S 12 Apr 2026 23:36:47.243 - Accepting cluster node connection from 127.0.0.1:53472 +2066855:S 12 Apr 2026 23:36:47.249 - Accepting cluster node connection from 127.0.0.1:53478 +2066855:S 12 Apr 2026 23:36:47.251 * Currently unable to failover: Failover attempt expired. +2066855:S 12 Apr 2026 23:36:47.251 * Needed quorum: 2. Number of votes received so far: 0 +2066855:S 12 Apr 2026 23:36:47.251 * Node ad56f48bae1934f4d1c3141ffee95fa8bf2fa37c () reported node 27c281c71471f64d57c2f90a6fce7ca7d11f3a9e () is back online. +2066855:S 12 Apr 2026 23:36:47.397 * Clear FAIL state for node 27c281c71471f64d57c2f90a6fce7ca7d11f3a9e (): is reachable again and nobody is serving its slots after some time. +2066855:S 12 Apr 2026 23:36:47.397 * Cluster state changed: ok +2066855:S 12 Apr 2026 23:36:47.488 * Node d24296bca8e44a56521bbf138bec235238274ad7 () reported node 27c281c71471f64d57c2f90a6fce7ca7d11f3a9e () is back online. +2066855:S 12 Apr 2026 23:36:51.041 - DB 0: 1 keys (0 volatile) in 4 slots HT. +2066855:S 12 Apr 2026 23:36:53.761 * NODE 7cbfdaa3b3c0ebf9bf9116dc86ed8bceffd31db0 () possibly failing. +2066855:S 12 Apr 2026 23:36:54.004 * FAIL message received from d24296bca8e44a56521bbf138bec235238274ad7 () about 7cbfdaa3b3c0ebf9bf9116dc86ed8bceffd31db0 () +2066855:S 12 Apr 2026 23:36:54.152 * Node 27c281c71471f64d57c2f90a6fce7ca7d11f3a9e () reported node 7cbfdaa3b3c0ebf9bf9116dc86ed8bceffd31db0 () as not reachable. +2066855:S 12 Apr 2026 23:36:54.155 * Clear FAIL state for node 7cbfdaa3b3c0ebf9bf9116dc86ed8bceffd31db0 (): replica is reachable again. +2066855:S 12 Apr 2026 23:36:54.345 * Node ad56f48bae1934f4d1c3141ffee95fa8bf2fa37c () reported node 7cbfdaa3b3c0ebf9bf9116dc86ed8bceffd31db0 () as not reachable. +2066855:S 12 Apr 2026 23:36:54.348 * Node 27c281c71471f64d57c2f90a6fce7ca7d11f3a9e () reported node 7cbfdaa3b3c0ebf9bf9116dc86ed8bceffd31db0 () is back online. +2066855:S 12 Apr 2026 23:36:54.829 * Node ad56f48bae1934f4d1c3141ffee95fa8bf2fa37c () reported node 7cbfdaa3b3c0ebf9bf9116dc86ed8bceffd31db0 () is back online. +2066855:S 12 Apr 2026 23:36:57.913 - DB 0: 1 keys (0 volatile) in 4 slots HT. +### Starting test Replica redirects key access in migrating slots in tests/unit/cluster/slot-migration.tcl +### Starting test Replica of migrating node returns ASK redirect after READONLY in tests/unit/cluster/slot-migration.tcl +### Starting test Replica of migrating node returns TRYAGAIN after READONLY in tests/unit/cluster/slot-migration.tcl +### Starting test Replica of importing node returns TRYAGAIN after READONLY and ASKING in tests/unit/cluster/slot-migration.tcl +### Starting test New replica inherits migrating slot in tests/unit/cluster/slot-migration.tcl +2066855:M 12 Apr 2026 23:36:57.967 * Connection with primary lost. +2066855:M 12 Apr 2026 23:36:57.967 * Caching the disconnected primary state. +2066855:M 12 Apr 2026 23:36:57.967 * Discarding previously cached primary state. +2066855:M 12 Apr 2026 23:36:57.967 * Setting secondary replication ID to 3434ac96bb7f1503c85d89d9ee94a4c782639d94, valid up to offset: 227. New replication ID is 187b825c010b950b10f6d6f33e61544cfc947ae4 +2066855:M 12 Apr 2026 23:36:57.969 # Cluster state changed: fail +2066855:M 12 Apr 2026 23:37:00.545 - Accepting cluster node connection from 127.0.0.1:45064 +2066855:M 12 Apr 2026 23:37:00.596 - Error accepting cluster node connection: error:0A000126:SSL routines::unexpected eof while reading +2066855:M 12 Apr 2026 23:37:00.596 - Accepting cluster node connection from 127.0.0.1:45070 +2066855:M 12 Apr 2026 23:37:00.599 - Error accepting cluster node connection: error:0A000126:SSL routines::unexpected eof while reading +2066855:M 12 Apr 2026 23:37:00.605 - Accepting cluster node connection from 127.0.0.1:45074 +2066855:M 12 Apr 2026 23:37:00.609 - Error accepting cluster node connection: error:0A000126:SSL routines::unexpected eof while reading +2066855:M 12 Apr 2026 23:37:00.621 - Accepting cluster node connection from 127.0.0.1:45076 +2066855:M 12 Apr 2026 23:37:00.623 - Error accepting cluster node connection: error:0A000126:SSL routines::unexpected eof while reading +2066855:M 12 Apr 2026 23:37:00.624 - Accepting cluster node connection from 127.0.0.1:45088 +2066855:M 12 Apr 2026 23:37:00.642 - Error accepting cluster node connection: error:0A000126:SSL routines::unexpected eof while reading +2066855:M 12 Apr 2026 23:37:00.643 - Accepting cluster node connection from 127.0.0.1:45094 +2066855:M 12 Apr 2026 23:37:00.661 - Error accepting cluster node connection: error:0A000126:SSL routines::unexpected eof while reading +2066855:M 12 Apr 2026 23:37:00.662 - Accepting cluster node connection from 127.0.0.1:45098 +2066855:M 12 Apr 2026 23:37:00.664 - Error accepting cluster node connection: error:0A000126:SSL routines::unexpected eof while reading +2066855:M 12 Apr 2026 23:37:00.664 - Accepting cluster node connection from 127.0.0.1:45108 +2066855:M 12 Apr 2026 23:37:00.675 - Error accepting cluster node connection: error:0A000126:SSL routines::unexpected eof while reading +2066855:M 12 Apr 2026 23:37:00.675 - Accepting cluster node connection from 127.0.0.1:45122 +2066855:M 12 Apr 2026 23:37:00.681 - Error accepting cluster node connection: error:0A000126:SSL routines::unexpected eof while reading +2066855:M 12 Apr 2026 23:37:00.682 - Accepting cluster node connection from 127.0.0.1:45130 +2066855:M 12 Apr 2026 23:37:00.694 - Accepting cluster node connection from 127.0.0.1:45140 +2066855:M 12 Apr 2026 23:37:00.696 - Accepting cluster node connection from 127.0.0.1:45146 +2066855:M 12 Apr 2026 23:37:00.708 - Accepting cluster node connection from 127.0.0.1:45162 +2066855:M 12 Apr 2026 23:37:00.731 - Accepting cluster node connection from 127.0.0.1:45164 +2066855:M 12 Apr 2026 23:37:01.274 * Successfully completed handshake with 27c281c71471f64d57c2f90a6fce7ca7d11f3a9e () +2066855:M 12 Apr 2026 23:37:01.645 * Address updated for node 27c281c71471f64d57c2f90a6fce7ca7d11f3a9e (), now 127.0.0.1:21699 +2066855:M 12 Apr 2026 23:37:04.037 - Accepting cluster node connection from 127.0.0.1:45176 +2066855:M 12 Apr 2026 23:37:04.263 - Error accepting cluster node connection: error:0A000126:SSL routines::unexpected eof while reading +2066855:M 12 Apr 2026 23:37:04.264 * NODE 71cce21972c3b30e1e72d50b69266c87048e6a42 () possibly failing. +2066855:M 12 Apr 2026 23:37:04.264 * NODE 7cbfdaa3b3c0ebf9bf9116dc86ed8bceffd31db0 () possibly failing. +2066855:M 12 Apr 2026 23:37:04.264 * NODE ad56f48bae1934f4d1c3141ffee95fa8bf2fa37c () possibly failing. +2066855:M 12 Apr 2026 23:37:04.268 * NODE 27c281c71471f64d57c2f90a6fce7ca7d11f3a9e () possibly failing. +2066855:M 12 Apr 2026 23:37:04.268 * NODE d24296bca8e44a56521bbf138bec235238274ad7 () possibly failing. +2066855:M 12 Apr 2026 23:37:04.269 - Accepting cluster node connection from 127.0.0.1:45188 +2066855:M 12 Apr 2026 23:37:04.271 - Error accepting cluster node connection: error:0A000126:SSL routines::unexpected eof while reading +2066855:M 12 Apr 2026 23:37:04.271 - Accepting cluster node connection from 127.0.0.1:45202 +2066855:M 12 Apr 2026 23:37:04.284 - Error accepting cluster node connection: error:0A000126:SSL routines::unexpected eof while reading +2066855:M 12 Apr 2026 23:37:04.289 - Accepting cluster node connection from 127.0.0.1:45206 +2066855:M 12 Apr 2026 23:37:04.291 - Error accepting cluster node connection: error:0A000126:SSL routines::unexpected eof while reading +2066855:M 12 Apr 2026 23:37:04.291 - Accepting cluster node connection from 127.0.0.1:45210 +2066855:M 12 Apr 2026 23:37:04.300 - Error accepting cluster node connection: error:0A000126:SSL routines::unexpected eof while reading +2066855:M 12 Apr 2026 23:37:04.300 - Accepting cluster node connection from 127.0.0.1:45066 +2066855:M 12 Apr 2026 23:37:04.315 - Accepting cluster node connection from 127.0.0.1:45078 +2066855:M 12 Apr 2026 23:37:04.327 - Accepting cluster node connection from 127.0.0.1:45094 +2066855:M 12 Apr 2026 23:37:04.329 - Accepting cluster node connection from 127.0.0.1:45096 +2066855:M 12 Apr 2026 23:37:04.341 - Accepting cluster node connection from 127.0.0.1:45100 +2066855:M 12 Apr 2026 23:37:04.349 * Node 7cbfdaa3b3c0ebf9bf9116dc86ed8bceffd31db0 () is now a replica of node ad56f48bae1934f4d1c3141ffee95fa8bf2fa37c () in shard fb1332a38fbff59dddf699f62b6c363413af8689 +2066855:M 12 Apr 2026 23:37:05.213 * Node 71cce21972c3b30e1e72d50b69266c87048e6a42 () is now a replica of node d24296bca8e44a56521bbf138bec235238274ad7 () in shard 91538f19e53855196ac7be00d216108415f3d90b +2066855:M 12 Apr 2026 23:37:05.214 * Node d24296bca8e44a56521bbf138bec235238274ad7 () reported node ad56f48bae1934f4d1c3141ffee95fa8bf2fa37c () as not reachable. +2066855:M 12 Apr 2026 23:37:05.215 - Accepting cluster node connection from 127.0.0.1:45106 +2066855:M 12 Apr 2026 23:37:05.221 - Accepting cluster node connection from 127.0.0.1:45118 +2066855:M 12 Apr 2026 23:37:07.877 * Node d24296bca8e44a56521bbf138bec235238274ad7 () reported node ad56f48bae1934f4d1c3141ffee95fa8bf2fa37c () is back online. +2066855:M 12 Apr 2026 23:37:08.004 - Accepting cluster node connection from 127.0.0.1:45124 +2066855:M 12 Apr 2026 23:37:08.013 - Error accepting cluster node connection: error:0A000126:SSL routines::unexpected eof while reading +2066855:M 12 Apr 2026 23:37:08.029 * Cluster state changed: ok +2066855:M 12 Apr 2026 23:37:08.110 - Accepting cluster node connection from 127.0.0.1:45130 +2066855:M 12 Apr 2026 23:37:08.186 - Error accepting cluster node connection: error:0A000126:SSL routines::unexpected eof while reading +2066855:M 12 Apr 2026 23:37:08.186 * Error converting peer IP to string: (null) +2066855:M 12 Apr 2026 23:37:08.186 * Error converting peer IP to string: (null) +2066855:M 12 Apr 2026 23:37:08.192 - Accepting cluster node connection from 127.0.0.1:45138 +2066855:M 12 Apr 2026 23:37:08.204 - Error accepting cluster node connection: error:0A000126:SSL routines::unexpected eof while reading +2066855:M 12 Apr 2026 23:37:08.205 - Accepting cluster node connection from 127.0.0.1:45140 +2066855:M 12 Apr 2026 23:37:08.216 - Error accepting cluster node connection: error:0A000126:SSL routines::unexpected eof while reading +2066855:M 12 Apr 2026 23:37:08.216 - Accepting cluster node connection from 127.0.0.1:45152 +2066855:M 12 Apr 2026 23:37:08.238 - Error accepting cluster node connection: error:0A000126:SSL routines::unexpected eof while reading +2066855:M 12 Apr 2026 23:37:08.239 - Accepting cluster node connection from 127.0.0.1:45158 +2066855:M 12 Apr 2026 23:37:08.248 - Error accepting cluster node connection: error:0A000126:SSL routines::unexpected eof while reading +2066855:M 12 Apr 2026 23:37:08.446 - Accepting cluster node connection from 127.0.0.1:45164 +2066855:M 12 Apr 2026 23:37:08.449 - Error accepting cluster node connection: error:0A000126:SSL routines::unexpected eof while reading +2066855:M 12 Apr 2026 23:37:08.461 - Accepting cluster node connection from 127.0.0.1:45176 +2066855:M 12 Apr 2026 23:37:08.464 - Accepting cluster node connection from 127.0.0.1:45188 +2066855:M 12 Apr 2026 23:37:08.494 - Accepting cluster node connection from 127.0.0.1:45192 +2066855:M 12 Apr 2026 23:37:08.515 - Accepting cluster node connection from 127.0.0.1:45206 +2066855:M 12 Apr 2026 23:37:08.517 - Accepting cluster node connection from 127.0.0.1:45208 +2066855:S 12 Apr 2026 23:37:08.852 * Connecting to PRIMARY 127.0.0.1:21699 +2066855:S 12 Apr 2026 23:37:08.852 * PRIMARY <-> REPLICA sync started +2066855:S 12 Apr 2026 23:37:09.080 * Non blocking connect for SYNC fired the event. +2066855:S 12 Apr 2026 23:37:09.183 * Primary replied to PING, replication can continue... +2066855:S 12 Apr 2026 23:37:09.189 * Partial resynchronization not possible (no cached primary) +2066855:S 12 Apr 2026 23:37:09.274 * Full resync from primary: 3434ac96bb7f1503c85d89d9ee94a4c782639d94:226 +2066855:S 12 Apr 2026 23:37:09.285 * PRIMARY <-> REPLICA sync: receiving streamed RDB from primary with EOF to disk +2066855:S 12 Apr 2026 23:37:10.381 * PRIMARY <-> REPLICA sync: Flushing old data +2066855:S 12 Apr 2026 23:37:10.381 * PRIMARY <-> REPLICA sync: Loading DB in memory +2066855:S 12 Apr 2026 23:37:10.381 * Loading RDB produced by Valkey version 8.0.2 +2066855:S 12 Apr 2026 23:37:10.381 * RDB age 1 seconds +2066855:S 12 Apr 2026 23:37:10.381 * RDB memory usage when created 3.00 Mb +2066855:S 12 Apr 2026 23:37:10.381 * Done loading RDB, keys loaded: 1, keys expired: 0. +2066855:S 12 Apr 2026 23:37:10.381 * PRIMARY <-> REPLICA sync: Finished with success +2066855:S 12 Apr 2026 23:37:10.382 - Accepting cluster node connection from 127.0.0.1:45212 +2066855:S 12 Apr 2026 23:37:10.393 - Accepting cluster node connection from 127.0.0.1:45222 +2066855:S 12 Apr 2026 23:37:10.401 - Accepting cluster node connection from 127.0.0.1:45228 +2066855:S 12 Apr 2026 23:37:10.403 - Accepting cluster node connection from 127.0.0.1:45240 +2066855:S 12 Apr 2026 23:37:10.411 - Accepting cluster node connection from 127.0.0.1:45242 +### Starting test New replica inherits importing slot in tests/unit/cluster/slot-migration.tcl +2066855:S 12 Apr 2026 23:37:12.498 * NODE 71cce21972c3b30e1e72d50b69266c87048e6a42 () possibly failing. +2066855:S 12 Apr 2026 23:37:12.498 * NODE 7cbfdaa3b3c0ebf9bf9116dc86ed8bceffd31db0 () possibly failing. +2066855:S 12 Apr 2026 23:37:12.653 * Node 27c281c71471f64d57c2f90a6fce7ca7d11f3a9e () reported node 71cce21972c3b30e1e72d50b69266c87048e6a42 () as not reachable. +2066855:S 12 Apr 2026 23:37:12.653 * Node 27c281c71471f64d57c2f90a6fce7ca7d11f3a9e () reported node 7cbfdaa3b3c0ebf9bf9116dc86ed8bceffd31db0 () as not reachable. +2066855:S 12 Apr 2026 23:37:12.761 * Node d24296bca8e44a56521bbf138bec235238274ad7 () reported node 71cce21972c3b30e1e72d50b69266c87048e6a42 () as not reachable. +2066855:S 12 Apr 2026 23:37:12.761 * Node d24296bca8e44a56521bbf138bec235238274ad7 () reported node 7cbfdaa3b3c0ebf9bf9116dc86ed8bceffd31db0 () as not reachable. +2066855:S 12 Apr 2026 23:37:12.761 * FAIL message received from ad56f48bae1934f4d1c3141ffee95fa8bf2fa37c () about 71cce21972c3b30e1e72d50b69266c87048e6a42 () +2066855:S 12 Apr 2026 23:37:13.197 * FAIL message received from 71cce21972c3b30e1e72d50b69266c87048e6a42 () about 7cbfdaa3b3c0ebf9bf9116dc86ed8bceffd31db0 () +2066855:S 12 Apr 2026 23:37:13.197 * Node ad56f48bae1934f4d1c3141ffee95fa8bf2fa37c () reported node 71cce21972c3b30e1e72d50b69266c87048e6a42 () as not reachable. +2066855:S 12 Apr 2026 23:37:13.198 * Clear FAIL state for node 7cbfdaa3b3c0ebf9bf9116dc86ed8bceffd31db0 (): primary without slots is reachable again. +2066855:S 12 Apr 2026 23:37:13.198 * Clear FAIL state for node 71cce21972c3b30e1e72d50b69266c87048e6a42 (): replica is reachable again. +2066855:S 12 Apr 2026 23:37:13.299 * Node d24296bca8e44a56521bbf138bec235238274ad7 () reported node 7cbfdaa3b3c0ebf9bf9116dc86ed8bceffd31db0 () is back online. +2066855:S 12 Apr 2026 23:37:13.299 * Node d24296bca8e44a56521bbf138bec235238274ad7 () reported node 71cce21972c3b30e1e72d50b69266c87048e6a42 () is back online. +2066855:S 12 Apr 2026 23:37:13.299 * Node 27c281c71471f64d57c2f90a6fce7ca7d11f3a9e () reported node 7cbfdaa3b3c0ebf9bf9116dc86ed8bceffd31db0 () is back online. +2066855:S 12 Apr 2026 23:37:13.398 * Node ad56f48bae1934f4d1c3141ffee95fa8bf2fa37c () reported node 71cce21972c3b30e1e72d50b69266c87048e6a42 () is back online. +2066855:S 12 Apr 2026 23:37:13.399 * Node 27c281c71471f64d57c2f90a6fce7ca7d11f3a9e () reported node 71cce21972c3b30e1e72d50b69266c87048e6a42 () is back online. +2066855:S 12 Apr 2026 23:37:15.640 - DB 0: 1 keys (0 volatile) in 4 slots HT. +2066855:S 12 Apr 2026 23:37:22.551 - Accepting cluster node connection from 127.0.0.1:42858 +2066855:S 12 Apr 2026 23:37:22.653 * Node 27c281c71471f64d57c2f90a6fce7ca7d11f3a9e () reported node 71cce21972c3b30e1e72d50b69266c87048e6a42 () as not reachable. +2066855:S 12 Apr 2026 23:37:22.653 * Node 27c281c71471f64d57c2f90a6fce7ca7d11f3a9e () reported node d24296bca8e44a56521bbf138bec235238274ad7 () as not reachable. +2066855:S 12 Apr 2026 23:37:22.653 * Node 27c281c71471f64d57c2f90a6fce7ca7d11f3a9e () reported node 7cbfdaa3b3c0ebf9bf9116dc86ed8bceffd31db0 () as not reachable. +2066855:S 12 Apr 2026 23:37:22.653 * Node 27c281c71471f64d57c2f90a6fce7ca7d11f3a9e () reported node ad56f48bae1934f4d1c3141ffee95fa8bf2fa37c () as not reachable. +2066855:S 12 Apr 2026 23:37:22.839 * Node 27c281c71471f64d57c2f90a6fce7ca7d11f3a9e () reported node ad56f48bae1934f4d1c3141ffee95fa8bf2fa37c () is back online. +2066855:S 12 Apr 2026 23:37:22.839 * Node 27c281c71471f64d57c2f90a6fce7ca7d11f3a9e () reported node 71cce21972c3b30e1e72d50b69266c87048e6a42 () is back online. +2066855:S 12 Apr 2026 23:37:22.839 * Node 27c281c71471f64d57c2f90a6fce7ca7d11f3a9e () reported node d24296bca8e44a56521bbf138bec235238274ad7 () is back online. +2066855:S 12 Apr 2026 23:37:22.855 * Node 27c281c71471f64d57c2f90a6fce7ca7d11f3a9e () reported node 7cbfdaa3b3c0ebf9bf9116dc86ed8bceffd31db0 () is back online. +2066855:S 12 Apr 2026 23:37:23.347 - DB 0: 1 keys (0 volatile) in 4 slots HT. +2066855:S 12 Apr 2026 23:37:28.626 - DB 0: 1 keys (0 volatile) in 4 slots HT. +2066855:S 12 Apr 2026 23:37:30.084 * NODE 7cbfdaa3b3c0ebf9bf9116dc86ed8bceffd31db0 () possibly failing. +2066855:S 12 Apr 2026 23:37:30.105 * Node 27c281c71471f64d57c2f90a6fce7ca7d11f3a9e () reported node 7cbfdaa3b3c0ebf9bf9116dc86ed8bceffd31db0 () as not reachable. +2066855:S 12 Apr 2026 23:37:30.105 * FAIL message received from ad56f48bae1934f4d1c3141ffee95fa8bf2fa37c () about 7cbfdaa3b3c0ebf9bf9116dc86ed8bceffd31db0 () +2066855:S 12 Apr 2026 23:37:30.149 * Node ad56f48bae1934f4d1c3141ffee95fa8bf2fa37c () reported node 7cbfdaa3b3c0ebf9bf9116dc86ed8bceffd31db0 () as not reachable. +2066855:S 12 Apr 2026 23:37:30.249 * Clear FAIL state for node 7cbfdaa3b3c0ebf9bf9116dc86ed8bceffd31db0 (): primary without slots is reachable again. +2066855:S 12 Apr 2026 23:37:30.297 * Node 27c281c71471f64d57c2f90a6fce7ca7d11f3a9e () reported node 7cbfdaa3b3c0ebf9bf9116dc86ed8bceffd31db0 () is back online. +2066855:S 12 Apr 2026 23:37:30.313 * Node ad56f48bae1934f4d1c3141ffee95fa8bf2fa37c () reported node 7cbfdaa3b3c0ebf9bf9116dc86ed8bceffd31db0 () is back online. +2066855:S 12 Apr 2026 23:37:33.938 - DB 0: 1 keys (0 volatile) in 4 slots HT. +2066855:S 12 Apr 2026 23:37:39.785 - Accepting cluster node connection from 127.0.0.1:33318 +2066855:S 12 Apr 2026 23:37:39.787 - Accepting cluster node connection from 127.0.0.1:33326 +2066855:S 12 Apr 2026 23:37:39.834 - Accepting cluster node connection from 127.0.0.1:33332 +2066855:S 12 Apr 2026 23:37:39.856 - Accepting cluster node connection from 127.0.0.1:33340 +2066855:S 12 Apr 2026 23:37:39.881 * Node ad56f48bae1934f4d1c3141ffee95fa8bf2fa37c () reported node d24296bca8e44a56521bbf138bec235238274ad7 () as not reachable. +2066855:S 12 Apr 2026 23:37:39.881 * Node ad56f48bae1934f4d1c3141ffee95fa8bf2fa37c () reported node 27c281c71471f64d57c2f90a6fce7ca7d11f3a9e () as not reachable. +2066855:S 12 Apr 2026 23:37:39.884 * FAIL message received from 27c281c71471f64d57c2f90a6fce7ca7d11f3a9e () about 7cbfdaa3b3c0ebf9bf9116dc86ed8bceffd31db0 () +2066855:S 12 Apr 2026 23:37:39.884 * Node d24296bca8e44a56521bbf138bec235238274ad7 () reported node 7cbfdaa3b3c0ebf9bf9116dc86ed8bceffd31db0 () as not reachable. +2066855:S 12 Apr 2026 23:37:39.912 * Node ad56f48bae1934f4d1c3141ffee95fa8bf2fa37c () reported node d24296bca8e44a56521bbf138bec235238274ad7 () is back online. +2066855:S 12 Apr 2026 23:37:39.912 * Node ad56f48bae1934f4d1c3141ffee95fa8bf2fa37c () reported node 27c281c71471f64d57c2f90a6fce7ca7d11f3a9e () is back online. +2066855:S 12 Apr 2026 23:37:39.914 * Clear FAIL state for node 7cbfdaa3b3c0ebf9bf9116dc86ed8bceffd31db0 (): primary without slots is reachable again. +2066855:S 12 Apr 2026 23:37:39.928 * FAIL message received from 71cce21972c3b30e1e72d50b69266c87048e6a42 () about 7cbfdaa3b3c0ebf9bf9116dc86ed8bceffd31db0 () +2066855:S 12 Apr 2026 23:37:40.053 * Node 27c281c71471f64d57c2f90a6fce7ca7d11f3a9e () reported node 7cbfdaa3b3c0ebf9bf9116dc86ed8bceffd31db0 () as not reachable. +2066855:S 12 Apr 2026 23:37:40.077 * Node d24296bca8e44a56521bbf138bec235238274ad7 () reported node 7cbfdaa3b3c0ebf9bf9116dc86ed8bceffd31db0 () is back online. +2066855:S 12 Apr 2026 23:37:40.077 * Clear FAIL state for node 7cbfdaa3b3c0ebf9bf9116dc86ed8bceffd31db0 (): primary without slots is reachable again. +2066855:S 12 Apr 2026 23:37:40.388 * Node 27c281c71471f64d57c2f90a6fce7ca7d11f3a9e () reported node 7cbfdaa3b3c0ebf9bf9116dc86ed8bceffd31db0 () is back online. +2066855:S 12 Apr 2026 23:37:40.797 - DB 0: 1 keys (0 volatile) in 4 slots HT. +2066855:S 12 Apr 2026 23:37:45.956 - DB 0: 1 keys (0 volatile) in 4 slots HT. +2066855:S 12 Apr 2026 23:37:49.342 * NODE 7cbfdaa3b3c0ebf9bf9116dc86ed8bceffd31db0 () possibly failing. +2066855:S 12 Apr 2026 23:37:49.446 * Node 27c281c71471f64d57c2f90a6fce7ca7d11f3a9e () reported node 7cbfdaa3b3c0ebf9bf9116dc86ed8bceffd31db0 () as not reachable. +2066855:S 12 Apr 2026 23:37:49.666 * Node 27c281c71471f64d57c2f90a6fce7ca7d11f3a9e () reported node 7cbfdaa3b3c0ebf9bf9116dc86ed8bceffd31db0 () is back online. +2066855:S 12 Apr 2026 23:37:52.574 - DB 0: 1 keys (0 volatile) in 4 slots HT. +2066855:S 12 Apr 2026 23:37:58.581 * NODE 71cce21972c3b30e1e72d50b69266c87048e6a42 () possibly failing. +2066855:S 12 Apr 2026 23:37:58.581 * NODE 7cbfdaa3b3c0ebf9bf9116dc86ed8bceffd31db0 () possibly failing. +2066855:S 12 Apr 2026 23:37:58.581 * NODE ad56f48bae1934f4d1c3141ffee95fa8bf2fa37c () possibly failing. +2066855:S 12 Apr 2026 23:37:58.581 * NODE 27c281c71471f64d57c2f90a6fce7ca7d11f3a9e () possibly failing. +2066855:S 12 Apr 2026 23:37:58.581 * NODE d24296bca8e44a56521bbf138bec235238274ad7 () possibly failing. +2066855:S 12 Apr 2026 23:37:58.581 # Cluster state changed: fail +2066855:S 12 Apr 2026 23:37:58.581 * FAIL message received from ad56f48bae1934f4d1c3141ffee95fa8bf2fa37c () about 27c281c71471f64d57c2f90a6fce7ca7d11f3a9e () +2066855:S 12 Apr 2026 23:37:58.581 - Accepting cluster node connection from 127.0.0.1:46036 +2066855:S 12 Apr 2026 23:37:58.613 - Accepting cluster node connection from 127.0.0.1:46048 +2066855:S 12 Apr 2026 23:37:58.637 * FAIL message received from ad56f48bae1934f4d1c3141ffee95fa8bf2fa37c () about 71cce21972c3b30e1e72d50b69266c87048e6a42 () +2066855:S 12 Apr 2026 23:37:58.637 * Node ad56f48bae1934f4d1c3141ffee95fa8bf2fa37c () reported node 71cce21972c3b30e1e72d50b69266c87048e6a42 () as not reachable. +2066855:S 12 Apr 2026 23:37:58.637 * Node d24296bca8e44a56521bbf138bec235238274ad7 () reported node 71cce21972c3b30e1e72d50b69266c87048e6a42 () as not reachable. +2066855:S 12 Apr 2026 23:37:58.637 * Node d24296bca8e44a56521bbf138bec235238274ad7 () reported node 27c281c71471f64d57c2f90a6fce7ca7d11f3a9e () as not reachable. +2066855:S 12 Apr 2026 23:37:58.693 * Start of election delayed for 784 milliseconds (rank #0, offset 282). +2066855:S 12 Apr 2026 23:37:58.819 - DB 0: 1 keys (0 volatile) in 4 slots HT. +2066855:S 12 Apr 2026 23:37:58.819 * Currently unable to failover: Waiting the delay before I can start a new failover. +2066855:S 12 Apr 2026 23:37:58.828 * Node d24296bca8e44a56521bbf138bec235238274ad7 () reported node 71cce21972c3b30e1e72d50b69266c87048e6a42 () is back online. +2066855:S 12 Apr 2026 23:37:58.859 * Clear FAIL state for node 71cce21972c3b30e1e72d50b69266c87048e6a42 (): replica is reachable again. +2066855:S 12 Apr 2026 23:37:58.860 * Node ad56f48bae1934f4d1c3141ffee95fa8bf2fa37c () reported node 71cce21972c3b30e1e72d50b69266c87048e6a42 () is back online. +2066855:S 12 Apr 2026 23:37:58.877 * Node ad56f48bae1934f4d1c3141ffee95fa8bf2fa37c () reported node 27c281c71471f64d57c2f90a6fce7ca7d11f3a9e () as not reachable. +2066855:S 12 Apr 2026 23:37:59.027 * Currently unable to failover: Waiting the delay before I can start a new failover. +2066855:S 12 Apr 2026 23:37:59.553 * Starting a failover election for epoch 9. +2066855:S 12 Apr 2026 23:37:59.596 * Currently unable to failover: Waiting for votes, but majority still not reached. +2066855:S 12 Apr 2026 23:37:59.673 * Needed quorum: 2. Number of votes received so far: 1 +2066855:S 12 Apr 2026 23:37:59.676 * Failover election won: I'm the new primary. +2066855:S 12 Apr 2026 23:37:59.676 * configEpoch set to 9 after successful failover +2066855:M 12 Apr 2026 23:37:59.676 * Connection with primary lost. +2066855:M 12 Apr 2026 23:37:59.676 * Caching the disconnected primary state. +2066855:M 12 Apr 2026 23:37:59.676 * Discarding previously cached primary state. +2066855:M 12 Apr 2026 23:37:59.676 * Setting secondary replication ID to 3434ac96bb7f1503c85d89d9ee94a4c782639d94, valid up to offset: 283. New replication ID is 0b3f1854e94de8dad9f58c42bfd4fbae2be7ceef +2066855:M 12 Apr 2026 23:37:59.723 - Accepted 127.0.0.1:55354 +2066855:M 12 Apr 2026 23:37:59.813 * Clear FAIL state for node 27c281c71471f64d57c2f90a6fce7ca7d11f3a9e (): primary without slots is reachable again. +2066855:M 12 Apr 2026 23:37:59.813 * A failover occurred in shard 0ccab0ca3bdf2df4fb1d88a6388f0d2e2e9492d9; node 27c281c71471f64d57c2f90a6fce7ca7d11f3a9e () failed over to node ceb41590bdb53727bffcb314c80a81989180d13a () with a config epoch of 9 +2066855:M 12 Apr 2026 23:37:59.813 * Node 27c281c71471f64d57c2f90a6fce7ca7d11f3a9e () is now a replica of node ceb41590bdb53727bffcb314c80a81989180d13a () in shard 0ccab0ca3bdf2df4fb1d88a6388f0d2e2e9492d9 +2066855:M 12 Apr 2026 23:37:59.820 * Replica 127.0.0.1:21699 asks for synchronization +2066855:M 12 Apr 2026 23:37:59.820 * Partial resynchronization request from 127.0.0.1:21699 accepted. Sending 0 bytes of backlog starting from offset 283. +2066855:M 12 Apr 2026 23:37:59.884 * Node ad56f48bae1934f4d1c3141ffee95fa8bf2fa37c () reported node 27c281c71471f64d57c2f90a6fce7ca7d11f3a9e () is back online. +2066855:M 12 Apr 2026 23:37:59.890 * Cluster state changed: ok +2066855:M 12 Apr 2026 23:37:59.963 * Node d24296bca8e44a56521bbf138bec235238274ad7 () reported node 27c281c71471f64d57c2f90a6fce7ca7d11f3a9e () is back online. +2066855:M 12 Apr 2026 23:38:04.074 - DB 0: 1 keys (0 volatile) in 4 slots HT. +2066855:M 12 Apr 2026 23:38:07.775 * NODE 71cce21972c3b30e1e72d50b69266c87048e6a42 () possibly failing. +2066855:M 12 Apr 2026 23:38:07.868 * Node ad56f48bae1934f4d1c3141ffee95fa8bf2fa37c () reported node 71cce21972c3b30e1e72d50b69266c87048e6a42 () as not reachable. +2066855:M 12 Apr 2026 23:38:07.868 * Marking node 71cce21972c3b30e1e72d50b69266c87048e6a42 () as failing (quorum reached). +2066855:M 12 Apr 2026 23:38:07.915 * Node d24296bca8e44a56521bbf138bec235238274ad7 () reported node 71cce21972c3b30e1e72d50b69266c87048e6a42 () as not reachable. +2066855:M 12 Apr 2026 23:38:07.953 * Clear FAIL state for node 71cce21972c3b30e1e72d50b69266c87048e6a42 (): replica is reachable again. +2066855:M 12 Apr 2026 23:38:07.954 * Node d24296bca8e44a56521bbf138bec235238274ad7 () reported node 71cce21972c3b30e1e72d50b69266c87048e6a42 () is back online. +2066855:M 12 Apr 2026 23:38:08.081 * Node ad56f48bae1934f4d1c3141ffee95fa8bf2fa37c () reported node 71cce21972c3b30e1e72d50b69266c87048e6a42 () is back online. +2066855:M 12 Apr 2026 23:38:09.319 - DB 0: 1 keys (0 volatile) in 4 slots HT. +2066855:M 12 Apr 2026 23:38:14.490 - DB 0: 1 keys (0 volatile) in 4 slots HT. +2066855:M 12 Apr 2026 23:38:17.926 - Accepting cluster node connection from 127.0.0.1:38358 +2066855:M 12 Apr 2026 23:38:17.928 - Error accepting cluster node connection: error:0A000126:SSL routines::unexpected eof while reading +2066855:M 12 Apr 2026 23:38:17.957 - Accepting cluster node connection from 127.0.0.1:38362 +2066855:M 12 Apr 2026 23:38:17.959 - Error accepting cluster node connection: error:0A000126:SSL routines::unexpected eof while reading +2066855:M 12 Apr 2026 23:38:17.960 * Node ad56f48bae1934f4d1c3141ffee95fa8bf2fa37c () reported node d24296bca8e44a56521bbf138bec235238274ad7 () as not reachable. +2066855:M 12 Apr 2026 23:38:17.960 * Node ad56f48bae1934f4d1c3141ffee95fa8bf2fa37c () reported node 27c281c71471f64d57c2f90a6fce7ca7d11f3a9e () as not reachable. +2066855:M 12 Apr 2026 23:38:17.960 - Accepting cluster node connection from 127.0.0.1:38364 +2066855:M 12 Apr 2026 23:38:17.991 - Accepting cluster node connection from 127.0.0.1:38366 +2066855:M 12 Apr 2026 23:38:18.026 - Accepting cluster node connection from 127.0.0.1:38378 +2066855:M 12 Apr 2026 23:38:18.085 - Error accepting cluster node connection: Connection reset by peer +2066855:M 12 Apr 2026 23:38:18.162 * Node ad56f48bae1934f4d1c3141ffee95fa8bf2fa37c () reported node 7cbfdaa3b3c0ebf9bf9116dc86ed8bceffd31db0 () as not reachable. +2066855:M 12 Apr 2026 23:38:18.162 - Accepting cluster node connection from 127.0.0.1:38394 +2066855:M 12 Apr 2026 23:38:18.319 * Node ad56f48bae1934f4d1c3141ffee95fa8bf2fa37c () reported node 27c281c71471f64d57c2f90a6fce7ca7d11f3a9e () is back online. +2066855:M 12 Apr 2026 23:38:18.319 * Node ad56f48bae1934f4d1c3141ffee95fa8bf2fa37c () reported node d24296bca8e44a56521bbf138bec235238274ad7 () is back online. +2066855:M 12 Apr 2026 23:38:18.521 * Node ad56f48bae1934f4d1c3141ffee95fa8bf2fa37c () reported node 7cbfdaa3b3c0ebf9bf9116dc86ed8bceffd31db0 () is back online. +2066855:M 12 Apr 2026 23:38:22.254 - DB 0: 1 keys (0 volatile) in 4 slots HT. +2066855:M 12 Apr 2026 23:38:26.473 * NODE 7cbfdaa3b3c0ebf9bf9116dc86ed8bceffd31db0 () possibly failing. +2066855:M 12 Apr 2026 23:38:26.661 * Node ad56f48bae1934f4d1c3141ffee95fa8bf2fa37c () reported node 7cbfdaa3b3c0ebf9bf9116dc86ed8bceffd31db0 () as not reachable. +2066855:M 12 Apr 2026 23:38:26.661 * Marking node 7cbfdaa3b3c0ebf9bf9116dc86ed8bceffd31db0 () as failing (quorum reached). +2066855:M 12 Apr 2026 23:38:26.712 * Clear FAIL state for node 7cbfdaa3b3c0ebf9bf9116dc86ed8bceffd31db0 (): primary without slots is reachable again. +2066855:M 12 Apr 2026 23:38:26.774 * Node d24296bca8e44a56521bbf138bec235238274ad7 () reported node 7cbfdaa3b3c0ebf9bf9116dc86ed8bceffd31db0 () as not reachable. +2066855:M 12 Apr 2026 23:38:26.961 * Node d24296bca8e44a56521bbf138bec235238274ad7 () reported node 7cbfdaa3b3c0ebf9bf9116dc86ed8bceffd31db0 () is back online. +2066855:M 12 Apr 2026 23:38:27.061 * Node ad56f48bae1934f4d1c3141ffee95fa8bf2fa37c () reported node 7cbfdaa3b3c0ebf9bf9116dc86ed8bceffd31db0 () is back online. +2066855:M 12 Apr 2026 23:38:29.017 - DB 0: 1 keys (0 volatile) in 4 slots HT. +2066855:M 12 Apr 2026 23:38:35.502 - Accepting cluster node connection from 127.0.0.1:56778 +2066855:M 12 Apr 2026 23:38:35.600 - Accepting cluster node connection from 127.0.0.1:56788 +2066855:M 12 Apr 2026 23:38:35.611 - Accepting cluster node connection from 127.0.0.1:56798 +2066855:M 12 Apr 2026 23:38:35.642 - Accepting cluster node connection from 127.0.0.1:56806 +2066855:M 12 Apr 2026 23:38:35.833 - DB 0: 1 keys (0 volatile) in 4 slots HT. +2066855:M 12 Apr 2026 23:38:40.949 - DB 0: 1 keys (0 volatile) in 4 slots HT. +2066855:M 12 Apr 2026 23:38:44.683 * NODE 7cbfdaa3b3c0ebf9bf9116dc86ed8bceffd31db0 () possibly failing. +2066855:M 12 Apr 2026 23:38:44.683 * NODE d24296bca8e44a56521bbf138bec235238274ad7 () possibly failing. +2066855:M 12 Apr 2026 23:38:44.808 - Accepting cluster node connection from 127.0.0.1:43728 +2066855:M 12 Apr 2026 23:38:44.832 * Node ad56f48bae1934f4d1c3141ffee95fa8bf2fa37c () reported node 7cbfdaa3b3c0ebf9bf9116dc86ed8bceffd31db0 () as not reachable. +2066855:M 12 Apr 2026 23:38:44.832 * Marking node 7cbfdaa3b3c0ebf9bf9116dc86ed8bceffd31db0 () as failing (quorum reached). +2066855:M 12 Apr 2026 23:38:44.864 * Node d24296bca8e44a56521bbf138bec235238274ad7 () reported node ad56f48bae1934f4d1c3141ffee95fa8bf2fa37c () as not reachable. +2066855:M 12 Apr 2026 23:38:44.864 * Node d24296bca8e44a56521bbf138bec235238274ad7 () reported node 7cbfdaa3b3c0ebf9bf9116dc86ed8bceffd31db0 () as not reachable. +2066855:M 12 Apr 2026 23:38:44.900 * Clear FAIL state for node 7cbfdaa3b3c0ebf9bf9116dc86ed8bceffd31db0 (): primary without slots is reachable again. +2066855:M 12 Apr 2026 23:38:44.989 * Node d24296bca8e44a56521bbf138bec235238274ad7 () reported node 7cbfdaa3b3c0ebf9bf9116dc86ed8bceffd31db0 () is back online. +2066855:M 12 Apr 2026 23:38:45.005 * Node d24296bca8e44a56521bbf138bec235238274ad7 () reported node ad56f48bae1934f4d1c3141ffee95fa8bf2fa37c () is back online. +2066855:M 12 Apr 2026 23:38:45.090 * Node ad56f48bae1934f4d1c3141ffee95fa8bf2fa37c () reported node 7cbfdaa3b3c0ebf9bf9116dc86ed8bceffd31db0 () is back online. +2066855:M 12 Apr 2026 23:38:47.371 - DB 0: 1 keys (0 volatile) in 4 slots HT. +2066855:M 12 Apr 2026 23:38:54.195 - DB 0: 1 keys (0 volatile) in 4 slots HT. +2066855:M 12 Apr 2026 23:38:54.218 - Accepting cluster node connection from 127.0.0.1:50778 +2066855:M 12 Apr 2026 23:38:54.233 * Node d24296bca8e44a56521bbf138bec235238274ad7 () reported node ad56f48bae1934f4d1c3141ffee95fa8bf2fa37c () as not reachable. +2066855:M 12 Apr 2026 23:38:54.233 * Node d24296bca8e44a56521bbf138bec235238274ad7 () reported node 7cbfdaa3b3c0ebf9bf9116dc86ed8bceffd31db0 () as not reachable. +2066855:M 12 Apr 2026 23:38:54.233 * Node d24296bca8e44a56521bbf138bec235238274ad7 () reported node 27c281c71471f64d57c2f90a6fce7ca7d11f3a9e () as not reachable. +2066855:M 12 Apr 2026 23:38:54.306 * Node d24296bca8e44a56521bbf138bec235238274ad7 () reported node 7cbfdaa3b3c0ebf9bf9116dc86ed8bceffd31db0 () is back online. +2066855:M 12 Apr 2026 23:38:54.306 * Node d24296bca8e44a56521bbf138bec235238274ad7 () reported node ad56f48bae1934f4d1c3141ffee95fa8bf2fa37c () is back online. +2066855:M 12 Apr 2026 23:38:54.410 * Node d24296bca8e44a56521bbf138bec235238274ad7 () reported node 27c281c71471f64d57c2f90a6fce7ca7d11f3a9e () is back online. +2066855:M 12 Apr 2026 23:38:59.290 - DB 0: 1 keys (0 volatile) in 4 slots HT. +2066855:M 12 Apr 2026 23:39:01.938 * NODE 7cbfdaa3b3c0ebf9bf9116dc86ed8bceffd31db0 () possibly failing. +2066855:M 12 Apr 2026 23:39:03.191 * FAIL message received from d24296bca8e44a56521bbf138bec235238274ad7 () about 7cbfdaa3b3c0ebf9bf9116dc86ed8bceffd31db0 () +2066855:M 12 Apr 2026 23:39:03.299 * Clear FAIL state for node 7cbfdaa3b3c0ebf9bf9116dc86ed8bceffd31db0 (): primary without slots is reachable again. +2066855:M 12 Apr 2026 23:39:05.409 - DB 0: 1 keys (0 volatile) in 4 slots HT. +2066855:M 12 Apr 2026 23:39:12.381 - DB 0: 1 keys (0 volatile) in 4 slots HT. +2066855:M 12 Apr 2026 23:39:17.448 - DB 0: 1 keys (0 volatile) in 4 slots HT. +2066855:M 12 Apr 2026 23:39:21.027 * NODE 71cce21972c3b30e1e72d50b69266c87048e6a42 () possibly failing. +2066855:M 12 Apr 2026 23:39:21.027 * NODE 7cbfdaa3b3c0ebf9bf9116dc86ed8bceffd31db0 () possibly failing. +2066855:M 12 Apr 2026 23:39:21.027 * NODE ad56f48bae1934f4d1c3141ffee95fa8bf2fa37c () possibly failing. +2066855:M 12 Apr 2026 23:39:21.028 * NODE 27c281c71471f64d57c2f90a6fce7ca7d11f3a9e () possibly failing. +2066855:M 12 Apr 2026 23:39:21.028 * NODE d24296bca8e44a56521bbf138bec235238274ad7 () possibly failing. +2066855:M 12 Apr 2026 23:39:21.028 # Cluster state changed: fail +2066855:M 12 Apr 2026 23:39:21.132 - Accepting cluster node connection from 127.0.0.1:49506 +2066855:M 12 Apr 2026 23:39:21.162 * Node ad56f48bae1934f4d1c3141ffee95fa8bf2fa37c () reported node 71cce21972c3b30e1e72d50b69266c87048e6a42 () as not reachable. +2066855:M 12 Apr 2026 23:39:21.162 * Marking node 71cce21972c3b30e1e72d50b69266c87048e6a42 () as failing (quorum reached). +2066855:M 12 Apr 2026 23:39:21.162 * Node ad56f48bae1934f4d1c3141ffee95fa8bf2fa37c () reported node d24296bca8e44a56521bbf138bec235238274ad7 () as not reachable. +2066855:M 12 Apr 2026 23:39:21.162 * Node ad56f48bae1934f4d1c3141ffee95fa8bf2fa37c () reported node 27c281c71471f64d57c2f90a6fce7ca7d11f3a9e () as not reachable. +2066855:M 12 Apr 2026 23:39:21.162 * Node ad56f48bae1934f4d1c3141ffee95fa8bf2fa37c () reported node 7cbfdaa3b3c0ebf9bf9116dc86ed8bceffd31db0 () as not reachable. +2066855:M 12 Apr 2026 23:39:21.163 * Node ad56f48bae1934f4d1c3141ffee95fa8bf2fa37c () reported node 27c281c71471f64d57c2f90a6fce7ca7d11f3a9e () is back online. +2066855:M 12 Apr 2026 23:39:21.163 * Node ad56f48bae1934f4d1c3141ffee95fa8bf2fa37c () reported node 71cce21972c3b30e1e72d50b69266c87048e6a42 () is back online. +2066855:M 12 Apr 2026 23:39:21.163 * Node ad56f48bae1934f4d1c3141ffee95fa8bf2fa37c () reported node 7cbfdaa3b3c0ebf9bf9116dc86ed8bceffd31db0 () is back online. +2066855:M 12 Apr 2026 23:39:21.164 * Clear FAIL state for node 71cce21972c3b30e1e72d50b69266c87048e6a42 (): replica is reachable again. +2066855:M 12 Apr 2026 23:39:21.333 * Node ad56f48bae1934f4d1c3141ffee95fa8bf2fa37c () reported node 71cce21972c3b30e1e72d50b69266c87048e6a42 () as not reachable. +2066855:M 12 Apr 2026 23:39:21.333 * Node ad56f48bae1934f4d1c3141ffee95fa8bf2fa37c () reported node d24296bca8e44a56521bbf138bec235238274ad7 () is back online. +2066855:M 12 Apr 2026 23:39:21.342 * Node ad56f48bae1934f4d1c3141ffee95fa8bf2fa37c () reported node 71cce21972c3b30e1e72d50b69266c87048e6a42 () is back online. +2066855:M 12 Apr 2026 23:39:21.948 - Reading from client: error:0A000126:SSL routines::unexpected eof while reading +2066855:M 12 Apr 2026 23:39:21.948 * Connection with replica client id #328 lost. +2066855:M 12 Apr 2026 23:39:22.045 - Connection with Node 27c281c71471f64d57c2f90a6fce7ca7d11f3a9e at 127.0.0.1:31699 failed: (null) +2066855:M 12 Apr 2026 23:39:22.145 - Connection with Node ad56f48bae1934f4d1c3141ffee95fa8bf2fa37c at 127.0.0.1:31697 failed: (null) +2066855:M 12 Apr 2026 23:39:22.145 - Connection with Node 27c281c71471f64d57c2f90a6fce7ca7d11f3a9e at 127.0.0.1:31699 failed: (null) +===== End of server log (pid 2066855) ===== + + +===== Start of server stderr log (pid 2066855) ===== + + +===== End of server stderr log (pid 2066855) ===== + + +===== Start of server log (pid 2066480) ===== + +### Starting server for test +2066480:C 12 Apr 2026 23:36:26.197 # WARNING: Changing databases number from 16 to 1 since we are in cluster mode +2066480:C 12 Apr 2026 23:36:26.197 # WARNING Memory overcommit must be enabled! Without it, a background save or replication may fail under low memory condition. Being disabled, it can also cause failures without low memory condition, see https://github.com/jemalloc/jemalloc/issues/1328. To fix this issue add 'vm.overcommit_memory = 1' to /etc/sysctl.conf and then reboot or run the command 'sysctl vm.overcommit_memory=1' for this to take effect. +2066480:C 12 Apr 2026 23:36:26.197 * oO0OoO0OoO0Oo Valkey is starting oO0OoO0OoO0Oo +2066480:C 12 Apr 2026 23:36:26.197 * Valkey version=8.0.2, bits=64, commit=00000000, modified=0, pid=2066480, just started +2066480:C 12 Apr 2026 23:36:26.197 * Configuration loaded +2066480:M 12 Apr 2026 23:36:26.198 * Increased maximum number of open files to 10032 (it was originally set to 1024). +2066480:M 12 Apr 2026 23:36:26.198 * monotonic clock: POSIX clock_gettime +2066480:M 12 Apr 2026 23:36:26.199 # Failed to write PID file: Permission denied + .+^+. + .+#########+. + .+########+########+. Valkey 8.0.2 (00000000/0) 64 bit + .+########+' '+########+. + .########+' .+. '+########. Running in cluster mode + |####+' .+#######+. '+####| Port: 21692 + |###| .+###############+. |###| PID: 2066480 + |###| |#####*'' ''*#####| |###| + |###| |####' .-. '####| |###| + |###| |###( (@@@) )###| |###| https://valkey.io + |###| |####. '-' .####| |###| + |###| |#####*. .*#####| |###| + |###| '+#####| |#####+' |###| + |####+. +##| |#+' .+####| + '#######+ |##| .+########' + '+###| |##| .+########+' + '| |####+########+' + +#########+' + '+v+' + +2066480:M 12 Apr 2026 23:36:26.200 * No cluster configuration found, I'm 7cbfdaa3b3c0ebf9bf9116dc86ed8bceffd31db0 +2066480:M 12 Apr 2026 23:36:26.214 * Server initialized +2066480:M 12 Apr 2026 23:36:26.214 * Ready to accept connections tcp +2066480:M 12 Apr 2026 23:36:26.214 * Ready to accept connections unix +2066480:M 12 Apr 2026 23:36:26.214 * Ready to accept connections tls +2066480:M 12 Apr 2026 23:36:26.308 - Accepted 127.0.0.1:41063 +2066480:M 12 Apr 2026 23:36:26.327 - Reading from client: error:0A000126:SSL routines::unexpected eof while reading +2066480:M 12 Apr 2026 23:36:26.336 - Accepted 127.0.0.1:41779 +2066480:M 12 Apr 2026 23:36:27.358 - Accepting cluster node connection from 127.0.0.1:49940 +2066480:M 12 Apr 2026 23:36:27.409 * IP address for this node updated to 127.0.0.1 +2066480:M 12 Apr 2026 23:36:27.505 - Accepting cluster node connection from 127.0.0.1:49944 +2066480:M 12 Apr 2026 23:36:27.520 - Accepting cluster node connection from 127.0.0.1:49946 +2066480:M 12 Apr 2026 23:36:27.569 - Accepting cluster node connection from 127.0.0.1:49958 +2066480:M 12 Apr 2026 23:36:27.599 - Accepting cluster node connection from 127.0.0.1:49964 +2066480:M 12 Apr 2026 23:36:27.615 - Accepting cluster node connection from 127.0.0.1:49974 +2066480:M 12 Apr 2026 23:36:27.658 - Error accepting cluster node connection: error:0A000126:SSL routines::unexpected eof while reading +2066480:M 12 Apr 2026 23:36:27.674 * Successfully completed handshake with ad56f48bae1934f4d1c3141ffee95fa8bf2fa37c () +2066480:M 12 Apr 2026 23:36:27.675 * Successfully completed handshake with 71cce21972c3b30e1e72d50b69266c87048e6a42 () +2066480:M 12 Apr 2026 23:36:27.675 - Handshake: we already know node d24296bca8e44a56521bbf138bec235238274ad7 (), updating the address if needed. +2066480:M 12 Apr 2026 23:36:27.676 - Handshake: we already know node ceb41590bdb53727bffcb314c80a81989180d13a (), updating the address if needed. +2066480:M 12 Apr 2026 23:36:27.676 * Error converting peer IP to string: (null) +2066480:M 12 Apr 2026 23:36:27.696 - Accepting cluster node connection from 127.0.0.1:49980 +2066480:M 12 Apr 2026 23:36:27.720 * configEpoch collision with node ceb41590bdb53727bffcb314c80a81989180d13a (). configEpoch set to 3 +2066480:M 12 Apr 2026 23:36:27.722 * Address updated for node ad56f48bae1934f4d1c3141ffee95fa8bf2fa37c (), now 127.0.0.1:21697 +2066480:M 12 Apr 2026 23:36:27.729 - Accepting cluster node connection from 127.0.0.1:49986 +2066480:M 12 Apr 2026 23:36:27.731 - Accepting cluster node connection from 127.0.0.1:49988 +2066480:M 12 Apr 2026 23:36:27.734 - Accepting cluster node connection from 127.0.0.1:49996 +2066480:M 12 Apr 2026 23:36:27.762 * configEpoch collision with node ceb41590bdb53727bffcb314c80a81989180d13a (). configEpoch set to 4 +2066480:M 12 Apr 2026 23:36:27.770 * Address updated for node 71cce21972c3b30e1e72d50b69266c87048e6a42 (), now 127.0.0.1:21689 +2066480:M 12 Apr 2026 23:36:27.965 * Node ceb41590bdb53727bffcb314c80a81989180d13a () is no longer primary of shard 7be85fc7a35a355988332b0ade6bb85642f90b76; removed all 0 slot(s) it used to own +2066480:M 12 Apr 2026 23:36:27.965 * Node ceb41590bdb53727bffcb314c80a81989180d13a () is now part of shard 0ccab0ca3bdf2df4fb1d88a6388f0d2e2e9492d9 +2066480:M 12 Apr 2026 23:36:27.965 * Node ceb41590bdb53727bffcb314c80a81989180d13a () is now a replica of node 27c281c71471f64d57c2f90a6fce7ca7d11f3a9e () in shard 0ccab0ca3bdf2df4fb1d88a6388f0d2e2e9492d9 +2066480:S 12 Apr 2026 23:36:27.965 * Connecting to PRIMARY 127.0.0.1:21697 +2066480:S 12 Apr 2026 23:36:27.965 * PRIMARY <-> REPLICA sync started +2066480:S 12 Apr 2026 23:36:27.965 * Cluster state changed: ok +2066480:S 12 Apr 2026 23:36:27.980 * Non blocking connect for SYNC fired the event. +2066480:S 12 Apr 2026 23:36:27.980 * Node 71cce21972c3b30e1e72d50b69266c87048e6a42 () is no longer primary of shard 797d84842d0558c353ef2fd200fec2d129d7c96f; removed all 0 slot(s) it used to own +2066480:S 12 Apr 2026 23:36:27.980 * Node 71cce21972c3b30e1e72d50b69266c87048e6a42 () is now part of shard b553502a57202365d35122f1f841b02047c3c23d +2066480:S 12 Apr 2026 23:36:27.980 * Node 71cce21972c3b30e1e72d50b69266c87048e6a42 () is now a replica of node d24296bca8e44a56521bbf138bec235238274ad7 () in shard b553502a57202365d35122f1f841b02047c3c23d +2066480:S 12 Apr 2026 23:36:28.006 * Primary replied to PING, replication can continue... +2066480:S 12 Apr 2026 23:36:28.007 # DEBUG LOG: ========== I am replica 4 ========== +2066480:S 12 Apr 2026 23:36:28.007 * Partial resynchronization not possible (no cached primary) +2066480:S 12 Apr 2026 23:36:28.009 * Full resync from primary: 950dbe04066d7028f8350da2375aa7b0709610d8:0 +2066480:S 12 Apr 2026 23:36:28.017 * PRIMARY <-> REPLICA sync: receiving streamed RDB from primary with EOF to disk +2066480:S 12 Apr 2026 23:36:28.034 * PRIMARY <-> REPLICA sync: Flushing old data +2066480:S 12 Apr 2026 23:36:28.035 * PRIMARY <-> REPLICA sync: Loading DB in memory +2066480:S 12 Apr 2026 23:36:28.035 * Loading RDB produced by Valkey version 8.0.2 +2066480:S 12 Apr 2026 23:36:28.035 * RDB age 0 seconds +2066480:S 12 Apr 2026 23:36:28.035 * RDB memory usage when created 2.81 Mb +2066480:S 12 Apr 2026 23:36:28.035 * Done loading RDB, keys loaded: 0, keys expired: 0. +2066480:S 12 Apr 2026 23:36:28.035 * PRIMARY <-> REPLICA sync: Finished with success +2066480:S 12 Apr 2026 23:36:35.769 * NODE 71cce21972c3b30e1e72d50b69266c87048e6a42 () possibly failing. +### Starting test Slot migration states are replicated in tests/unit/cluster/slot-migration.tcl +2066480:S 12 Apr 2026 23:36:37.790 * Importing slot 609 from node 27c281c71471f64d57c2f90a6fce7ca7d11f3a9e () +### Starting test Migration target is auto-updated after failover in target shard in tests/unit/cluster/slot-migration.tcl +2066480:S 12 Apr 2026 23:36:38.988 * NODE ad56f48bae1934f4d1c3141ffee95fa8bf2fa37c () possibly failing. +2066480:S 12 Apr 2026 23:36:39.001 * Node d24296bca8e44a56521bbf138bec235238274ad7 () reported node ad56f48bae1934f4d1c3141ffee95fa8bf2fa37c () as not reachable. +2066480:S 12 Apr 2026 23:36:39.106 * FAIL message received from 27c281c71471f64d57c2f90a6fce7ca7d11f3a9e () about ad56f48bae1934f4d1c3141ffee95fa8bf2fa37c () +2066480:S 12 Apr 2026 23:36:39.106 # Cluster state changed: fail +2066480:S 12 Apr 2026 23:36:39.149 * Node 27c281c71471f64d57c2f90a6fce7ca7d11f3a9e () reported node ad56f48bae1934f4d1c3141ffee95fa8bf2fa37c () as not reachable. +2066480:S 12 Apr 2026 23:36:39.193 * Start of election delayed for 700 milliseconds (rank #0, offset 175). +2066480:S 12 Apr 2026 23:36:39.299 * Currently unable to failover: Waiting the delay before I can start a new failover. +2066480:S 12 Apr 2026 23:36:39.920 * Starting a failover election for epoch 6. +2066480:S 12 Apr 2026 23:36:39.957 * Failover election won: I'm the new primary. +2066480:S 12 Apr 2026 23:36:39.957 * configEpoch set to 6 after successful failover +2066480:M 12 Apr 2026 23:36:39.957 * Connection with primary lost. +2066480:M 12 Apr 2026 23:36:39.957 * Caching the disconnected primary state. +2066480:M 12 Apr 2026 23:36:39.957 * Discarding previously cached primary state. +2066480:M 12 Apr 2026 23:36:39.957 * Setting secondary replication ID to 950dbe04066d7028f8350da2375aa7b0709610d8, valid up to offset: 176. New replication ID is 04bc904ae99c9565f5e08319325b2bf2948666fc +2066480:M 12 Apr 2026 23:36:39.957 * Cluster state changed: ok +2066480:M 12 Apr 2026 23:36:40.926 - Node ad56f48bae1934f4d1c3141ffee95fa8bf2fa37c has old slots configuration, sending an UPDATE message about 7cbfdaa3b3c0ebf9bf9116dc86ed8bceffd31db0 +2066480:M 12 Apr 2026 23:36:40.982 - Accepted 127.0.0.1:42338 +2066480:M 12 Apr 2026 23:36:41.091 * Replica 127.0.0.1:21697 asks for synchronization +2066480:M 12 Apr 2026 23:36:41.091 * Partial resynchronization request from 127.0.0.1:21697 accepted. Sending 0 bytes of backlog starting from offset 176. +2066480:M 12 Apr 2026 23:36:41.179 * A failover occurred in shard fb1332a38fbff59dddf699f62b6c363413af8689; node ad56f48bae1934f4d1c3141ffee95fa8bf2fa37c () failed over to node 7cbfdaa3b3c0ebf9bf9116dc86ed8bceffd31db0 () with a config epoch of 6 +2066480:M 12 Apr 2026 23:36:41.179 * Node ad56f48bae1934f4d1c3141ffee95fa8bf2fa37c () is now a replica of node 7cbfdaa3b3c0ebf9bf9116dc86ed8bceffd31db0 () in shard fb1332a38fbff59dddf699f62b6c363413af8689 +2066480:M 12 Apr 2026 23:36:41.180 * Clear FAIL state for node ad56f48bae1934f4d1c3141ffee95fa8bf2fa37c (): replica is reachable again. +2066480:M 12 Apr 2026 23:36:41.259 * Node 27c281c71471f64d57c2f90a6fce7ca7d11f3a9e () reported node ad56f48bae1934f4d1c3141ffee95fa8bf2fa37c () is back online. +2066480:M 12 Apr 2026 23:36:41.291 * Node d24296bca8e44a56521bbf138bec235238274ad7 () reported node ad56f48bae1934f4d1c3141ffee95fa8bf2fa37c () is back online. +2066480:M 12 Apr 2026 23:36:41.292 * Manual failover requested by replica ad56f48bae1934f4d1c3141ffee95fa8bf2fa37c (). +2066480:M 12 Apr 2026 23:36:41.310 - Client closed connection id=8 addr=127.0.0.1:42338 laddr=127.0.0.1:21691 fd=24 name= age=1 idle=0 flags=S db=0 sub=0 psub=0 ssub=0 multi=-1 watch=0 qbuf=0 qbuf-free=20474 argv-mem=0 multi-mem=0 rbs=1024 rbp=7 obl=0 oll=1 omem=20504 tot-mem=42904 events=r cmd=psync user=default redir=-1 resp=2 lib-name= lib-ver= tot-net-in=236 tot-net-out=22 tot-cmds=5 +2066480:M 12 Apr 2026 23:36:41.310 * Connection with replica 127.0.0.1:21697 lost. +2066480:M 12 Apr 2026 23:36:41.337 * Failover auth granted to ad56f48bae1934f4d1c3141ffee95fa8bf2fa37c () for epoch 7 +2066480:M 12 Apr 2026 23:36:41.345 * Configuration change detected. Reconfiguring myself as a replica of node ad56f48bae1934f4d1c3141ffee95fa8bf2fa37c () in shard fb1332a38fbff59dddf699f62b6c363413af8689 +2066480:S 12 Apr 2026 23:36:41.345 * Before turning into a replica, using my own primary parameters to synthesize a cached primary: I may be able to synchronize with the new primary with just a partial transfer. +2066480:S 12 Apr 2026 23:36:41.345 * Connecting to PRIMARY 127.0.0.1:21697 +2066480:S 12 Apr 2026 23:36:41.345 * PRIMARY <-> REPLICA sync started +2066480:S 12 Apr 2026 23:36:41.384 * Non blocking connect for SYNC fired the event. +2066480:S 12 Apr 2026 23:36:41.393 * Primary replied to PING, replication can continue... +2066480:S 12 Apr 2026 23:36:41.397 * Trying a partial resynchronization (request 04bc904ae99c9565f5e08319325b2bf2948666fc:176). +2066480:S 12 Apr 2026 23:36:41.397 * Successful partial resynchronization with primary. +2066480:S 12 Apr 2026 23:36:41.397 * Primary replication ID changed to 823749a7ecd27acd70ba1ca1af68e1d71a98c40e +2066480:S 12 Apr 2026 23:36:41.397 * PRIMARY <-> REPLICA sync: Primary accepted a Partial Resynchronization. +### Starting test Migration source is auto-updated after failover in source shard in tests/unit/cluster/slot-migration.tcl +2066480:S 12 Apr 2026 23:36:42.748 * NODE 27c281c71471f64d57c2f90a6fce7ca7d11f3a9e () possibly failing. +2066480:S 12 Apr 2026 23:36:42.857 * Node d24296bca8e44a56521bbf138bec235238274ad7 () reported node 27c281c71471f64d57c2f90a6fce7ca7d11f3a9e () as not reachable. +2066480:S 12 Apr 2026 23:36:42.857 * Node ad56f48bae1934f4d1c3141ffee95fa8bf2fa37c () reported node 27c281c71471f64d57c2f90a6fce7ca7d11f3a9e () as not reachable. +2066480:S 12 Apr 2026 23:36:42.857 * Marking node 27c281c71471f64d57c2f90a6fce7ca7d11f3a9e () as failing (quorum reached). +2066480:S 12 Apr 2026 23:36:42.857 # Cluster state changed: fail +2066480:S 12 Apr 2026 23:36:46.110 * Node d24296bca8e44a56521bbf138bec235238274ad7 () reported node 27c281c71471f64d57c2f90a6fce7ca7d11f3a9e () is back online. +2066480:S 12 Apr 2026 23:36:46.111 * Node ad56f48bae1934f4d1c3141ffee95fa8bf2fa37c () reported node 27c281c71471f64d57c2f90a6fce7ca7d11f3a9e () is back online. +2066480:S 12 Apr 2026 23:36:46.126 * Clear FAIL state for node 27c281c71471f64d57c2f90a6fce7ca7d11f3a9e (): is reachable again and nobody is serving its slots after some time. +2066480:S 12 Apr 2026 23:36:46.126 * Cluster state changed: ok +2066480:S 12 Apr 2026 23:36:46.699 * NODE ceb41590bdb53727bffcb314c80a81989180d13a () possibly failing. +2066480:S 12 Apr 2026 23:36:46.904 * Node d24296bca8e44a56521bbf138bec235238274ad7 () reported node ceb41590bdb53727bffcb314c80a81989180d13a () as not reachable. +2066480:S 12 Apr 2026 23:36:47.076 * FAIL message received from 27c281c71471f64d57c2f90a6fce7ca7d11f3a9e () about ceb41590bdb53727bffcb314c80a81989180d13a () +2066480:S 12 Apr 2026 23:36:47.077 * Node 27c281c71471f64d57c2f90a6fce7ca7d11f3a9e () reported node ceb41590bdb53727bffcb314c80a81989180d13a () as not reachable. +2066480:S 12 Apr 2026 23:36:47.282 * Clear FAIL state for node ceb41590bdb53727bffcb314c80a81989180d13a (): replica is reachable again. +2066480:S 12 Apr 2026 23:36:47.398 * Node 27c281c71471f64d57c2f90a6fce7ca7d11f3a9e () reported node ceb41590bdb53727bffcb314c80a81989180d13a () is back online. +2066480:S 12 Apr 2026 23:36:47.499 * Node d24296bca8e44a56521bbf138bec235238274ad7 () reported node ceb41590bdb53727bffcb314c80a81989180d13a () is back online. +2066480:S 12 Apr 2026 23:36:53.910 - Accepting cluster node connection from 127.0.0.1:47522 +2066480:S 12 Apr 2026 23:36:53.917 - Accepting cluster node connection from 127.0.0.1:47530 +2066480:S 12 Apr 2026 23:36:53.917 - Accepting cluster node connection from 127.0.0.1:47534 +2066480:S 12 Apr 2026 23:36:53.930 - Accepting cluster node connection from 127.0.0.1:47538 +2066480:S 12 Apr 2026 23:36:53.939 - Accepting cluster node connection from 127.0.0.1:47546 +### Starting test Replica redirects key access in migrating slots in tests/unit/cluster/slot-migration.tcl +### Starting test Replica of migrating node returns ASK redirect after READONLY in tests/unit/cluster/slot-migration.tcl +### Starting test Replica of migrating node returns TRYAGAIN after READONLY in tests/unit/cluster/slot-migration.tcl +### Starting test Replica of importing node returns TRYAGAIN after READONLY and ASKING in tests/unit/cluster/slot-migration.tcl +### Starting test New replica inherits migrating slot in tests/unit/cluster/slot-migration.tcl +2066480:S 12 Apr 2026 23:36:59.030 * NODE ceb41590bdb53727bffcb314c80a81989180d13a () possibly failing. +2066480:S 12 Apr 2026 23:36:59.637 * FAIL message received from ad56f48bae1934f4d1c3141ffee95fa8bf2fa37c () about ceb41590bdb53727bffcb314c80a81989180d13a () +2066480:S 12 Apr 2026 23:36:59.638 * Node d24296bca8e44a56521bbf138bec235238274ad7 () reported node ceb41590bdb53727bffcb314c80a81989180d13a () as not reachable. +2066480:S 12 Apr 2026 23:36:59.638 * Node 27c281c71471f64d57c2f90a6fce7ca7d11f3a9e () reported node ceb41590bdb53727bffcb314c80a81989180d13a () as not reachable. +2066480:S 12 Apr 2026 23:36:59.711 - Accepting cluster node connection from 127.0.0.1:47560 +2066480:S 12 Apr 2026 23:36:59.740 - Accepting cluster node connection from 127.0.0.1:47564 +2066480:S 12 Apr 2026 23:37:00.066 * Node ad56f48bae1934f4d1c3141ffee95fa8bf2fa37c () reported node ceb41590bdb53727bffcb314c80a81989180d13a () as not reachable. +2066480:S 12 Apr 2026 23:37:00.765 * Clear FAIL state for node ceb41590bdb53727bffcb314c80a81989180d13a (): replica is reachable again. +2066480:S 12 Apr 2026 23:37:00.956 * Node 27c281c71471f64d57c2f90a6fce7ca7d11f3a9e () reported node ceb41590bdb53727bffcb314c80a81989180d13a () is back online. +2066480:S 12 Apr 2026 23:37:01.057 * Node ad56f48bae1934f4d1c3141ffee95fa8bf2fa37c () reported node ceb41590bdb53727bffcb314c80a81989180d13a () is back online. +2066480:S 12 Apr 2026 23:37:01.150 * Node d24296bca8e44a56521bbf138bec235238274ad7 () reported node ceb41590bdb53727bffcb314c80a81989180d13a () is back online. +2066480:S 12 Apr 2026 23:37:01.647 - Accepting cluster node connection from 127.0.0.1:47572 +2066480:S 12 Apr 2026 23:37:02.160 * NODE ceb41590bdb53727bffcb314c80a81989180d13a () possibly failing. +2066480:S 12 Apr 2026 23:37:02.269 * Node 27c281c71471f64d57c2f90a6fce7ca7d11f3a9e () reported node ceb41590bdb53727bffcb314c80a81989180d13a () as not reachable. +2066480:S 12 Apr 2026 23:37:02.271 * Node d24296bca8e44a56521bbf138bec235238274ad7 () reported node ceb41590bdb53727bffcb314c80a81989180d13a () as not reachable. +2066480:S 12 Apr 2026 23:37:02.271 * Marking node ceb41590bdb53727bffcb314c80a81989180d13a () as failing (quorum reached). +2066480:S 12 Apr 2026 23:37:03.737 * NODE ad56f48bae1934f4d1c3141ffee95fa8bf2fa37c () possibly failing. +2066480:S 12 Apr 2026 23:37:04.052 - Accepting cluster node connection from 127.0.0.1:41132 +2066480:S 12 Apr 2026 23:37:04.055 - Accepting cluster node connection from 127.0.0.1:41144 +2066480:S 12 Apr 2026 23:37:04.058 * Node d24296bca8e44a56521bbf138bec235238274ad7 () reported node ad56f48bae1934f4d1c3141ffee95fa8bf2fa37c () as not reachable. +2066480:S 12 Apr 2026 23:37:04.270 - Error accepting cluster node connection: Connection reset by peer +2066480:S 12 Apr 2026 23:37:04.270 * Node d24296bca8e44a56521bbf138bec235238274ad7 () reported node 27c281c71471f64d57c2f90a6fce7ca7d11f3a9e () as not reachable. +2066480:S 12 Apr 2026 23:37:04.348 * Node ad56f48bae1934f4d1c3141ffee95fa8bf2fa37c () reported node ceb41590bdb53727bffcb314c80a81989180d13a () as not reachable. +2066480:S 12 Apr 2026 23:37:04.430 * Node 27c281c71471f64d57c2f90a6fce7ca7d11f3a9e () reported node ceb41590bdb53727bffcb314c80a81989180d13a () is back online. +2066480:S 12 Apr 2026 23:37:04.430 * Clear FAIL state for node ceb41590bdb53727bffcb314c80a81989180d13a (): primary without slots is reachable again. +2066480:S 12 Apr 2026 23:37:04.600 - Accepting cluster node connection from 127.0.0.1:41154 +2066480:S 12 Apr 2026 23:37:04.620 * Node d24296bca8e44a56521bbf138bec235238274ad7 () reported node 27c281c71471f64d57c2f90a6fce7ca7d11f3a9e () is back online. +2066480:S 12 Apr 2026 23:37:05.038 * Node ad56f48bae1934f4d1c3141ffee95fa8bf2fa37c () reported node ceb41590bdb53727bffcb314c80a81989180d13a () is back online. +2066480:S 12 Apr 2026 23:37:05.189 * Node d24296bca8e44a56521bbf138bec235238274ad7 () reported node ad56f48bae1934f4d1c3141ffee95fa8bf2fa37c () is back online. +2066480:S 12 Apr 2026 23:37:05.207 * Node ad56f48bae1934f4d1c3141ffee95fa8bf2fa37c () reported node 71cce21972c3b30e1e72d50b69266c87048e6a42 () as not reachable. +2066480:S 12 Apr 2026 23:37:05.309 * Node ad56f48bae1934f4d1c3141ffee95fa8bf2fa37c () reported node 71cce21972c3b30e1e72d50b69266c87048e6a42 () is back online. +2066480:S 12 Apr 2026 23:37:06.049 - Accepting cluster node connection from 127.0.0.1:41170 +2066480:S 12 Apr 2026 23:37:06.256 * NODE ceb41590bdb53727bffcb314c80a81989180d13a () possibly failing. +2066480:S 12 Apr 2026 23:37:06.863 * FAIL message received from 27c281c71471f64d57c2f90a6fce7ca7d11f3a9e () about ceb41590bdb53727bffcb314c80a81989180d13a () +2066480:S 12 Apr 2026 23:37:06.969 * Node ad56f48bae1934f4d1c3141ffee95fa8bf2fa37c () reported node ceb41590bdb53727bffcb314c80a81989180d13a () as not reachable. +2066480:S 12 Apr 2026 23:37:06.972 * Node 27c281c71471f64d57c2f90a6fce7ca7d11f3a9e () reported node ceb41590bdb53727bffcb314c80a81989180d13a () as not reachable. +2066480:S 12 Apr 2026 23:37:08.577 * Clear FAIL state for node ceb41590bdb53727bffcb314c80a81989180d13a (): primary without slots is reachable again. +2066480:S 12 Apr 2026 23:37:08.783 * Node 27c281c71471f64d57c2f90a6fce7ca7d11f3a9e () reported node ceb41590bdb53727bffcb314c80a81989180d13a () is back online. +2066480:S 12 Apr 2026 23:37:08.797 * Node ad56f48bae1934f4d1c3141ffee95fa8bf2fa37c () reported node ceb41590bdb53727bffcb314c80a81989180d13a () is back online. +2066480:S 12 Apr 2026 23:37:08.986 * Node d24296bca8e44a56521bbf138bec235238274ad7 () reported node ceb41590bdb53727bffcb314c80a81989180d13a () is back online. +2066480:S 12 Apr 2026 23:37:09.063 * A failover occurred in shard 0ccab0ca3bdf2df4fb1d88a6388f0d2e2e9492d9; node ceb41590bdb53727bffcb314c80a81989180d13a () failed over to node 27c281c71471f64d57c2f90a6fce7ca7d11f3a9e () with a config epoch of 5 +2066480:S 12 Apr 2026 23:37:09.063 * Node ceb41590bdb53727bffcb314c80a81989180d13a () is now a replica of node 27c281c71471f64d57c2f90a6fce7ca7d11f3a9e () in shard 0ccab0ca3bdf2df4fb1d88a6388f0d2e2e9492d9 +### Starting test New replica inherits importing slot in tests/unit/cluster/slot-migration.tcl +2066480:M 12 Apr 2026 23:37:10.535 * Connection with primary lost. +2066480:M 12 Apr 2026 23:37:10.535 * Caching the disconnected primary state. +2066480:M 12 Apr 2026 23:37:10.536 * Discarding previously cached primary state. +2066480:M 12 Apr 2026 23:37:10.536 * Setting secondary replication ID to 823749a7ecd27acd70ba1ca1af68e1d71a98c40e, valid up to offset: 204. New replication ID is 696841950d9a4a37c63d0149923302a91d34dba4 +2066480:M 12 Apr 2026 23:37:10.538 # Cluster state changed: fail +2066480:M 12 Apr 2026 23:37:10.909 - Accepting cluster node connection from 127.0.0.1:41176 +2066480:M 12 Apr 2026 23:37:10.921 - Accepting cluster node connection from 127.0.0.1:41180 +2066480:M 12 Apr 2026 23:37:10.928 - Accepting cluster node connection from 127.0.0.1:41188 +2066480:M 12 Apr 2026 23:37:10.975 - Accepting cluster node connection from 127.0.0.1:41190 +2066480:M 12 Apr 2026 23:37:10.978 - Accepting cluster node connection from 127.0.0.1:41200 +2066480:M 12 Apr 2026 23:37:12.498 - Error accepting cluster node connection: Connection reset by peer +2066480:M 12 Apr 2026 23:37:12.565 - Error accepting cluster node connection: Connection reset by peer +2066480:M 12 Apr 2026 23:37:12.621 - Accepting cluster node connection from 127.0.0.1:48642 +2066480:M 12 Apr 2026 23:37:12.666 - Accepting cluster node connection from 127.0.0.1:48644 +2066480:M 12 Apr 2026 23:37:12.696 - Accepting cluster node connection from 127.0.0.1:48646 +2066480:M 12 Apr 2026 23:37:12.734 - Accepting cluster node connection from 127.0.0.1:48658 +2066480:M 12 Apr 2026 23:37:12.765 - Accepting cluster node connection from 127.0.0.1:48670 +2066480:M 12 Apr 2026 23:37:12.796 * Ignoring FAIL message from unknown node ad56f48bae1934f4d1c3141ffee95fa8bf2fa37c about 71cce21972c3b30e1e72d50b69266c87048e6a42 +2066480:M 12 Apr 2026 23:37:12.796 * Ignoring FAIL message from unknown node ad56f48bae1934f4d1c3141ffee95fa8bf2fa37c about 7cbfdaa3b3c0ebf9bf9116dc86ed8bceffd31db0 +2066480:M 12 Apr 2026 23:37:12.821 * Ignoring FAIL message from unknown node 71cce21972c3b30e1e72d50b69266c87048e6a42 about 7cbfdaa3b3c0ebf9bf9116dc86ed8bceffd31db0 +2066480:M 12 Apr 2026 23:37:12.837 * Ignoring FAIL message from unknown node d24296bca8e44a56521bbf138bec235238274ad7 about 71cce21972c3b30e1e72d50b69266c87048e6a42 +2066480:M 12 Apr 2026 23:37:12.837 * Ignoring FAIL message from unknown node d24296bca8e44a56521bbf138bec235238274ad7 about 7cbfdaa3b3c0ebf9bf9116dc86ed8bceffd31db0 +2066480:M 12 Apr 2026 23:37:12.838 * Ignoring FAIL message from unknown node 27c281c71471f64d57c2f90a6fce7ca7d11f3a9e about 7cbfdaa3b3c0ebf9bf9116dc86ed8bceffd31db0 +2066480:M 12 Apr 2026 23:37:22.560 - Accepting cluster node connection from 127.0.0.1:54504 +2066480:M 12 Apr 2026 23:37:30.040 - Accepting cluster node connection from 127.0.0.1:54506 +2066480:M 12 Apr 2026 23:37:30.097 - Accepting cluster node connection from 127.0.0.1:54516 +2066480:M 12 Apr 2026 23:37:30.156 - Accepting cluster node connection from 127.0.0.1:54524 +2066480:M 12 Apr 2026 23:37:30.183 - Accepting cluster node connection from 127.0.0.1:54532 +2066480:M 12 Apr 2026 23:37:30.185 - Accepting cluster node connection from 127.0.0.1:54542 +2066480:M 12 Apr 2026 23:37:30.234 * Ignoring FAIL message from unknown node ad56f48bae1934f4d1c3141ffee95fa8bf2fa37c about 7cbfdaa3b3c0ebf9bf9116dc86ed8bceffd31db0 +2066480:M 12 Apr 2026 23:37:39.658 - Accepting cluster node connection from 127.0.0.1:49056 +2066480:M 12 Apr 2026 23:37:39.780 - Accepting cluster node connection from 127.0.0.1:49058 +2066480:M 12 Apr 2026 23:37:39.825 - Accepting cluster node connection from 127.0.0.1:49060 +2066480:M 12 Apr 2026 23:37:39.854 - Accepting cluster node connection from 127.0.0.1:49066 +2066480:M 12 Apr 2026 23:37:39.873 * Ignoring FAIL message from unknown node d24296bca8e44a56521bbf138bec235238274ad7 about ceb41590bdb53727bffcb314c80a81989180d13a +2066480:M 12 Apr 2026 23:37:39.873 * Ignoring FAIL message from unknown node ad56f48bae1934f4d1c3141ffee95fa8bf2fa37c about ceb41590bdb53727bffcb314c80a81989180d13a +2066480:M 12 Apr 2026 23:37:39.903 * Ignoring FAIL message from unknown node 27c281c71471f64d57c2f90a6fce7ca7d11f3a9e about 7cbfdaa3b3c0ebf9bf9116dc86ed8bceffd31db0 +2066480:M 12 Apr 2026 23:37:39.903 * Ignoring FAIL message from unknown node 27c281c71471f64d57c2f90a6fce7ca7d11f3a9e about ceb41590bdb53727bffcb314c80a81989180d13a +2066480:M 12 Apr 2026 23:37:39.924 * Ignoring FAIL message from unknown node 71cce21972c3b30e1e72d50b69266c87048e6a42 about 7cbfdaa3b3c0ebf9bf9116dc86ed8bceffd31db0 +2066480:M 12 Apr 2026 23:37:49.366 - Accepting cluster node connection from 127.0.0.1:35738 +2066480:M 12 Apr 2026 23:37:49.366 - Error accepting cluster node connection: error:0A000126:SSL routines::unexpected eof while reading +2066480:M 12 Apr 2026 23:37:49.366 - Accepting cluster node connection from 127.0.0.1:35746 +2066480:M 12 Apr 2026 23:37:49.379 - Accepting cluster node connection from 127.0.0.1:35760 +2066480:M 12 Apr 2026 23:37:49.461 - Accepting cluster node connection from 127.0.0.1:35770 +2066480:M 12 Apr 2026 23:37:49.464 - Accepting cluster node connection from 127.0.0.1:35778 +2066480:M 12 Apr 2026 23:37:50.040 - Accepting cluster node connection from 127.0.0.1:35780 +2066480:M 12 Apr 2026 23:37:58.580 * Ignoring FAIL message from unknown node ad56f48bae1934f4d1c3141ffee95fa8bf2fa37c about 27c281c71471f64d57c2f90a6fce7ca7d11f3a9e +2066480:M 12 Apr 2026 23:37:58.580 - Accepting cluster node connection from 127.0.0.1:60498 +2066480:M 12 Apr 2026 23:37:58.614 - Accepting cluster node connection from 127.0.0.1:60510 +2066480:M 12 Apr 2026 23:37:58.649 * Ignoring FAIL message from unknown node ad56f48bae1934f4d1c3141ffee95fa8bf2fa37c about 71cce21972c3b30e1e72d50b69266c87048e6a42 +2066480:M 12 Apr 2026 23:37:58.696 - Accepting cluster node connection from 127.0.0.1:60524 +2066480:M 12 Apr 2026 23:37:58.764 * Ignoring FAIL message from unknown node 71cce21972c3b30e1e72d50b69266c87048e6a42 about 27c281c71471f64d57c2f90a6fce7ca7d11f3a9e +2066480:M 12 Apr 2026 23:38:07.869 * Ignoring FAIL message from unknown node ceb41590bdb53727bffcb314c80a81989180d13a about 71cce21972c3b30e1e72d50b69266c87048e6a42 +2066480:M 12 Apr 2026 23:38:07.874 * Ignoring FAIL message from unknown node 27c281c71471f64d57c2f90a6fce7ca7d11f3a9e about 71cce21972c3b30e1e72d50b69266c87048e6a42 +2066480:M 12 Apr 2026 23:38:07.887 * Ignoring FAIL message from unknown node d24296bca8e44a56521bbf138bec235238274ad7 about 71cce21972c3b30e1e72d50b69266c87048e6a42 +2066480:M 12 Apr 2026 23:38:08.051 - Accepting cluster node connection from 127.0.0.1:43884 +2066480:M 12 Apr 2026 23:38:17.988 - Accepting cluster node connection from 127.0.0.1:55644 +2066480:M 12 Apr 2026 23:38:26.654 - Accepting cluster node connection from 127.0.0.1:42376 +2066480:M 12 Apr 2026 23:38:26.662 - Accepting cluster node connection from 127.0.0.1:42386 +2066480:M 12 Apr 2026 23:38:26.698 - Accepting cluster node connection from 127.0.0.1:42396 +2066480:M 12 Apr 2026 23:38:26.717 * Ignoring FAIL message from unknown node ceb41590bdb53727bffcb314c80a81989180d13a about 7cbfdaa3b3c0ebf9bf9116dc86ed8bceffd31db0 +2066480:M 12 Apr 2026 23:38:35.508 - Accepting cluster node connection from 127.0.0.1:46236 +2066480:M 12 Apr 2026 23:38:44.790 - Accepting cluster node connection from 127.0.0.1:48818 +2066480:M 12 Apr 2026 23:38:44.792 - Accepting cluster node connection from 127.0.0.1:48832 +2066480:M 12 Apr 2026 23:38:44.794 - Accepting cluster node connection from 127.0.0.1:48842 +2066480:M 12 Apr 2026 23:38:44.794 - Accepting cluster node connection from 127.0.0.1:48854 +2066480:M 12 Apr 2026 23:38:44.820 - Accepting cluster node connection from 127.0.0.1:48868 +2066480:M 12 Apr 2026 23:38:44.882 * Ignoring FAIL message from unknown node 27c281c71471f64d57c2f90a6fce7ca7d11f3a9e about 7cbfdaa3b3c0ebf9bf9116dc86ed8bceffd31db0 +2066480:M 12 Apr 2026 23:38:44.882 * Ignoring FAIL message from unknown node ad56f48bae1934f4d1c3141ffee95fa8bf2fa37c about 7cbfdaa3b3c0ebf9bf9116dc86ed8bceffd31db0 +2066480:M 12 Apr 2026 23:38:44.894 * Ignoring FAIL message from unknown node d24296bca8e44a56521bbf138bec235238274ad7 about 7cbfdaa3b3c0ebf9bf9116dc86ed8bceffd31db0 +2066480:M 12 Apr 2026 23:38:44.896 * Ignoring FAIL message from unknown node ceb41590bdb53727bffcb314c80a81989180d13a about 7cbfdaa3b3c0ebf9bf9116dc86ed8bceffd31db0 +2066480:M 12 Apr 2026 23:38:54.153 - Accepting cluster node connection from 127.0.0.1:45146 +2066480:M 12 Apr 2026 23:38:54.157 - Error accepting cluster node connection: error:0A000126:SSL routines::unexpected eof while reading +2066480:M 12 Apr 2026 23:38:54.216 - Accepting cluster node connection from 127.0.0.1:45148 +2066480:M 12 Apr 2026 23:38:54.251 - Accepting cluster node connection from 127.0.0.1:45162 +2066480:M 12 Apr 2026 23:38:54.256 - Accepting cluster node connection from 127.0.0.1:45176 +2066480:M 12 Apr 2026 23:39:03.152 - Accepting cluster node connection from 127.0.0.1:45182 +2066480:M 12 Apr 2026 23:39:03.155 - Accepting cluster node connection from 127.0.0.1:45186 +2066480:M 12 Apr 2026 23:39:03.164 - Error accepting cluster node connection: error:0A000126:SSL routines::unexpected eof while reading +2066480:M 12 Apr 2026 23:39:03.164 - Accepting cluster node connection from 127.0.0.1:45190 +2066480:M 12 Apr 2026 23:39:03.167 - Error accepting cluster node connection: error:0A000126:SSL routines::unexpected eof while reading +2066480:M 12 Apr 2026 23:39:03.260 - Accepting cluster node connection from 127.0.0.1:43674 +2066480:M 12 Apr 2026 23:39:03.263 - Accepting cluster node connection from 127.0.0.1:43678 +2066480:M 12 Apr 2026 23:39:03.269 - Accepting cluster node connection from 127.0.0.1:43688 +2066480:M 12 Apr 2026 23:39:03.279 - Accepting cluster node connection from 127.0.0.1:43698 +2066480:M 12 Apr 2026 23:39:03.285 - Accepting cluster node connection from 127.0.0.1:43708 +2066480:M 12 Apr 2026 23:39:21.128 - Accepting cluster node connection from 127.0.0.1:58674 +2066480:M 12 Apr 2026 23:39:21.130 - Accepting cluster node connection from 127.0.0.1:58686 +2066480:M 12 Apr 2026 23:39:21.136 - Accepting cluster node connection from 127.0.0.1:58690 +2066480:M 12 Apr 2026 23:39:21.163 * Ignoring FAIL message from unknown node ceb41590bdb53727bffcb314c80a81989180d13a about 71cce21972c3b30e1e72d50b69266c87048e6a42 +===== End of server log (pid 2066480) ===== + + +===== Start of server stderr log (pid 2066480) ===== + + +===== End of server stderr log (pid 2066480) ===== + + +===== Start of server log (pid 2066403) ===== + +### Starting server for test +2066403:C 12 Apr 2026 23:36:26.050 # WARNING: Changing databases number from 16 to 1 since we are in cluster mode +2066403:C 12 Apr 2026 23:36:26.050 # WARNING Memory overcommit must be enabled! Without it, a background save or replication may fail under low memory condition. Being disabled, it can also cause failures without low memory condition, see https://github.com/jemalloc/jemalloc/issues/1328. To fix this issue add 'vm.overcommit_memory = 1' to /etc/sysctl.conf and then reboot or run the command 'sysctl vm.overcommit_memory=1' for this to take effect. +2066403:C 12 Apr 2026 23:36:26.050 * oO0OoO0OoO0Oo Valkey is starting oO0OoO0OoO0Oo +2066403:C 12 Apr 2026 23:36:26.050 * Valkey version=8.0.2, bits=64, commit=00000000, modified=0, pid=2066403, just started +2066403:C 12 Apr 2026 23:36:26.050 * Configuration loaded +2066403:M 12 Apr 2026 23:36:26.050 * Increased maximum number of open files to 10032 (it was originally set to 1024). +2066403:M 12 Apr 2026 23:36:26.050 * monotonic clock: POSIX clock_gettime +2066403:M 12 Apr 2026 23:36:26.051 # Failed to write PID file: Permission denied + .+^+. + .+#########+. + .+########+########+. Valkey 8.0.2 (00000000/0) 64 bit + .+########+' '+########+. + .########+' .+. '+########. Running in cluster mode + |####+' .+#######+. '+####| Port: 21690 + |###| .+###############+. |###| PID: 2066403 + |###| |#####*'' ''*#####| |###| + |###| |####' .-. '####| |###| + |###| |###( (@@@) )###| |###| https://valkey.io + |###| |####. '-' .####| |###| + |###| |#####*. .*#####| |###| + |###| '+#####| |#####+' |###| + |####+. +##| |#+' .+####| + '#######+ |##| .+########' + '+###| |##| .+########+' + '| |####+########+' + +#########+' + '+v+' + +2066403:M 12 Apr 2026 23:36:26.052 * No cluster configuration found, I'm 71cce21972c3b30e1e72d50b69266c87048e6a42 +2066403:M 12 Apr 2026 23:36:26.065 * Server initialized +2066403:M 12 Apr 2026 23:36:26.065 * Ready to accept connections tcp +2066403:M 12 Apr 2026 23:36:26.065 * Ready to accept connections unix +2066403:M 12 Apr 2026 23:36:26.065 * Ready to accept connections tls +2066403:M 12 Apr 2026 23:36:26.156 - Accepted 127.0.0.1:36665 +2066403:M 12 Apr 2026 23:36:26.169 - Reading from client: error:0A000126:SSL routines::unexpected eof while reading +2066403:M 12 Apr 2026 23:36:26.175 - Accepted 127.0.0.1:41247 +2066403:M 12 Apr 2026 23:36:27.358 - Accepting cluster node connection from 127.0.0.1:38680 +2066403:M 12 Apr 2026 23:36:27.410 * IP address for this node updated to 127.0.0.1 +2066403:M 12 Apr 2026 23:36:27.522 - Accepting cluster node connection from 127.0.0.1:38684 +2066403:M 12 Apr 2026 23:36:27.573 - Accepting cluster node connection from 127.0.0.1:38692 +2066403:M 12 Apr 2026 23:36:27.585 - Accepting cluster node connection from 127.0.0.1:38708 +2066403:M 12 Apr 2026 23:36:27.587 - Accepting cluster node connection from 127.0.0.1:38724 +2066403:M 12 Apr 2026 23:36:27.617 * Successfully completed handshake with ad56f48bae1934f4d1c3141ffee95fa8bf2fa37c () +2066403:M 12 Apr 2026 23:36:27.617 * Address updated for node ad56f48bae1934f4d1c3141ffee95fa8bf2fa37c (), now 127.0.0.1:21697 +2066403:M 12 Apr 2026 23:36:27.625 * configEpoch collision with node ad56f48bae1934f4d1c3141ffee95fa8bf2fa37c (). configEpoch set to 1 +2066403:M 12 Apr 2026 23:36:27.644 * Successfully completed handshake with 7cbfdaa3b3c0ebf9bf9116dc86ed8bceffd31db0 () +2066403:M 12 Apr 2026 23:36:27.644 * Address updated for node 7cbfdaa3b3c0ebf9bf9116dc86ed8bceffd31db0 (), now 127.0.0.1:21691 +2066403:M 12 Apr 2026 23:36:27.668 - Handshake: we already know node ceb41590bdb53727bffcb314c80a81989180d13a (), updating the address if needed. +2066403:M 12 Apr 2026 23:36:27.715 - Accepting cluster node connection from 127.0.0.1:38732 +2066403:M 12 Apr 2026 23:36:27.742 - Accepting cluster node connection from 127.0.0.1:38744 +2066403:M 12 Apr 2026 23:36:27.788 - Accepting cluster node connection from 127.0.0.1:38758 +2066403:M 12 Apr 2026 23:36:27.791 - Accepting cluster node connection from 127.0.0.1:38774 +2066403:M 12 Apr 2026 23:36:27.956 * Node ceb41590bdb53727bffcb314c80a81989180d13a () is no longer primary of shard 7be85fc7a35a355988332b0ade6bb85642f90b76; removed all 0 slot(s) it used to own +2066403:M 12 Apr 2026 23:36:27.956 * Node ceb41590bdb53727bffcb314c80a81989180d13a () is now part of shard 0ccab0ca3bdf2df4fb1d88a6388f0d2e2e9492d9 +2066403:M 12 Apr 2026 23:36:27.956 * Node ceb41590bdb53727bffcb314c80a81989180d13a () is now a replica of node 27c281c71471f64d57c2f90a6fce7ca7d11f3a9e () in shard 0ccab0ca3bdf2df4fb1d88a6388f0d2e2e9492d9 +2066403:M 12 Apr 2026 23:36:27.971 * Node 7cbfdaa3b3c0ebf9bf9116dc86ed8bceffd31db0 () is no longer primary of shard 6d5a08da00fb925618c9429f0a0abf50cb184dee; removed all 0 slot(s) it used to own +2066403:M 12 Apr 2026 23:36:27.971 * Node 7cbfdaa3b3c0ebf9bf9116dc86ed8bceffd31db0 () is now part of shard fb1332a38fbff59dddf699f62b6c363413af8689 +2066403:M 12 Apr 2026 23:36:27.971 * Node 7cbfdaa3b3c0ebf9bf9116dc86ed8bceffd31db0 () is now a replica of node ad56f48bae1934f4d1c3141ffee95fa8bf2fa37c () in shard fb1332a38fbff59dddf699f62b6c363413af8689 +2066403:S 12 Apr 2026 23:36:27.973 * Connecting to PRIMARY 127.0.0.1:21695 +2066403:S 12 Apr 2026 23:36:27.974 * PRIMARY <-> REPLICA sync started +2066403:S 12 Apr 2026 23:36:27.974 * Cluster state changed: ok +2066403:S 12 Apr 2026 23:36:27.982 * Non blocking connect for SYNC fired the event. +2066403:S 12 Apr 2026 23:36:28.013 * Primary replied to PING, replication can continue... +2066403:S 12 Apr 2026 23:36:28.013 # DEBUG LOG: ========== I am replica 5 ========== +2066403:S 12 Apr 2026 23:36:28.017 * Partial resynchronization not possible (no cached primary) +2066403:S 12 Apr 2026 23:36:28.021 * Full resync from primary: f9a64ac3e118f63770d16efaeec31459ca9e6a4b:0 +2066403:S 12 Apr 2026 23:36:28.030 * PRIMARY <-> REPLICA sync: receiving streamed RDB from primary with EOF to disk +2066403:S 12 Apr 2026 23:36:28.036 * PRIMARY <-> REPLICA sync: Flushing old data +2066403:S 12 Apr 2026 23:36:28.036 * PRIMARY <-> REPLICA sync: Loading DB in memory +2066403:S 12 Apr 2026 23:36:28.038 * Loading RDB produced by Valkey version 8.0.2 +2066403:S 12 Apr 2026 23:36:28.038 * RDB age 0 seconds +2066403:S 12 Apr 2026 23:36:28.038 * RDB memory usage when created 2.81 Mb +2066403:S 12 Apr 2026 23:36:28.038 * Done loading RDB, keys loaded: 0, keys expired: 0. +2066403:S 12 Apr 2026 23:36:28.038 * PRIMARY <-> REPLICA sync: Finished with success +2066403:S 12 Apr 2026 23:36:35.913 - Accepting cluster node connection from 127.0.0.1:54396 +2066403:S 12 Apr 2026 23:36:35.923 - Accepting cluster node connection from 127.0.0.1:54404 +2066403:S 12 Apr 2026 23:36:35.926 - Accepting cluster node connection from 127.0.0.1:54406 +2066403:S 12 Apr 2026 23:36:35.931 - Accepting cluster node connection from 127.0.0.1:54420 +2066403:S 12 Apr 2026 23:36:35.937 - Accepting cluster node connection from 127.0.0.1:54426 +### Starting test Slot migration states are replicated in tests/unit/cluster/slot-migration.tcl +### Starting test Migration target is auto-updated after failover in target shard in tests/unit/cluster/slot-migration.tcl +2066403:S 12 Apr 2026 23:36:38.887 * NODE ad56f48bae1934f4d1c3141ffee95fa8bf2fa37c () possibly failing. +2066403:S 12 Apr 2026 23:36:39.000 * Node d24296bca8e44a56521bbf138bec235238274ad7 () reported node ad56f48bae1934f4d1c3141ffee95fa8bf2fa37c () as not reachable. +2066403:S 12 Apr 2026 23:36:39.107 * FAIL message received from 27c281c71471f64d57c2f90a6fce7ca7d11f3a9e () about ad56f48bae1934f4d1c3141ffee95fa8bf2fa37c () +2066403:S 12 Apr 2026 23:36:39.107 # Cluster state changed: fail +2066403:S 12 Apr 2026 23:36:39.149 * Node 27c281c71471f64d57c2f90a6fce7ca7d11f3a9e () reported node ad56f48bae1934f4d1c3141ffee95fa8bf2fa37c () as not reachable. +2066403:S 12 Apr 2026 23:36:39.978 * Cluster state changed: ok +2066403:S 12 Apr 2026 23:36:40.125 * Node 7cbfdaa3b3c0ebf9bf9116dc86ed8bceffd31db0 () reported node ad56f48bae1934f4d1c3141ffee95fa8bf2fa37c () as not reachable. +2066403:S 12 Apr 2026 23:36:40.926 - Node ad56f48bae1934f4d1c3141ffee95fa8bf2fa37c has old slots configuration, sending an UPDATE message about 7cbfdaa3b3c0ebf9bf9116dc86ed8bceffd31db0 +2066403:S 12 Apr 2026 23:36:41.176 * A failover occurred in shard fb1332a38fbff59dddf699f62b6c363413af8689; node ad56f48bae1934f4d1c3141ffee95fa8bf2fa37c () failed over to node 7cbfdaa3b3c0ebf9bf9116dc86ed8bceffd31db0 () with a config epoch of 6 +2066403:S 12 Apr 2026 23:36:41.176 * Node ad56f48bae1934f4d1c3141ffee95fa8bf2fa37c () is now a replica of node 7cbfdaa3b3c0ebf9bf9116dc86ed8bceffd31db0 () in shard fb1332a38fbff59dddf699f62b6c363413af8689 +2066403:S 12 Apr 2026 23:36:41.182 * Clear FAIL state for node ad56f48bae1934f4d1c3141ffee95fa8bf2fa37c (): replica is reachable again. +2066403:S 12 Apr 2026 23:36:41.205 * Node d24296bca8e44a56521bbf138bec235238274ad7 () reported node ad56f48bae1934f4d1c3141ffee95fa8bf2fa37c () is back online. +2066403:S 12 Apr 2026 23:36:41.205 * Node 7cbfdaa3b3c0ebf9bf9116dc86ed8bceffd31db0 () reported node ad56f48bae1934f4d1c3141ffee95fa8bf2fa37c () is back online. +2066403:S 12 Apr 2026 23:36:41.308 * Node 27c281c71471f64d57c2f90a6fce7ca7d11f3a9e () reported node ad56f48bae1934f4d1c3141ffee95fa8bf2fa37c () is back online. +2066403:S 12 Apr 2026 23:36:41.406 * A failover occurred in shard fb1332a38fbff59dddf699f62b6c363413af8689; node 7cbfdaa3b3c0ebf9bf9116dc86ed8bceffd31db0 () failed over to node ad56f48bae1934f4d1c3141ffee95fa8bf2fa37c () with a config epoch of 7 +2066403:S 12 Apr 2026 23:36:41.406 * Node 7cbfdaa3b3c0ebf9bf9116dc86ed8bceffd31db0 () is now a replica of node ad56f48bae1934f4d1c3141ffee95fa8bf2fa37c () in shard fb1332a38fbff59dddf699f62b6c363413af8689 +### Starting test Migration source is auto-updated after failover in source shard in tests/unit/cluster/slot-migration.tcl +2066403:S 12 Apr 2026 23:36:42.735 * NODE 27c281c71471f64d57c2f90a6fce7ca7d11f3a9e () possibly failing. +2066403:S 12 Apr 2026 23:36:42.845 * Node ad56f48bae1934f4d1c3141ffee95fa8bf2fa37c () reported node 27c281c71471f64d57c2f90a6fce7ca7d11f3a9e () as not reachable. +2066403:S 12 Apr 2026 23:36:42.845 * Node d24296bca8e44a56521bbf138bec235238274ad7 () reported node 27c281c71471f64d57c2f90a6fce7ca7d11f3a9e () as not reachable. +2066403:S 12 Apr 2026 23:36:42.845 * Marking node 27c281c71471f64d57c2f90a6fce7ca7d11f3a9e () as failing (quorum reached). +2066403:S 12 Apr 2026 23:36:42.845 # Cluster state changed: fail +2066403:S 12 Apr 2026 23:36:45.746 * Clear FAIL state for node 27c281c71471f64d57c2f90a6fce7ca7d11f3a9e (): is reachable again and nobody is serving its slots after some time. +2066403:S 12 Apr 2026 23:36:45.746 * Cluster state changed: ok +2066403:S 12 Apr 2026 23:36:46.111 * Node d24296bca8e44a56521bbf138bec235238274ad7 () reported node 27c281c71471f64d57c2f90a6fce7ca7d11f3a9e () is back online. +2066403:S 12 Apr 2026 23:36:46.112 * Node ad56f48bae1934f4d1c3141ffee95fa8bf2fa37c () reported node 27c281c71471f64d57c2f90a6fce7ca7d11f3a9e () is back online. +2066403:S 12 Apr 2026 23:36:46.728 * NODE ceb41590bdb53727bffcb314c80a81989180d13a () possibly failing. +2066403:S 12 Apr 2026 23:36:46.904 * Node d24296bca8e44a56521bbf138bec235238274ad7 () reported node ceb41590bdb53727bffcb314c80a81989180d13a () as not reachable. +2066403:S 12 Apr 2026 23:36:47.075 * Node ad56f48bae1934f4d1c3141ffee95fa8bf2fa37c () reported node ceb41590bdb53727bffcb314c80a81989180d13a () as not reachable. +2066403:S 12 Apr 2026 23:36:47.075 * Marking node ceb41590bdb53727bffcb314c80a81989180d13a () as failing (quorum reached). +2066403:S 12 Apr 2026 23:36:47.120 * Node 27c281c71471f64d57c2f90a6fce7ca7d11f3a9e () reported node ceb41590bdb53727bffcb314c80a81989180d13a () as not reachable. +2066403:S 12 Apr 2026 23:36:47.272 * Clear FAIL state for node ceb41590bdb53727bffcb314c80a81989180d13a (): replica is reachable again. +2066403:S 12 Apr 2026 23:36:47.398 * Node 27c281c71471f64d57c2f90a6fce7ca7d11f3a9e () reported node ceb41590bdb53727bffcb314c80a81989180d13a () is back online. +2066403:S 12 Apr 2026 23:36:47.488 * Node d24296bca8e44a56521bbf138bec235238274ad7 () reported node ceb41590bdb53727bffcb314c80a81989180d13a () is back online. +2066403:S 12 Apr 2026 23:36:47.593 * Node ad56f48bae1934f4d1c3141ffee95fa8bf2fa37c () reported node ceb41590bdb53727bffcb314c80a81989180d13a () is back online. +2066403:S 12 Apr 2026 23:36:53.757 * NODE 7cbfdaa3b3c0ebf9bf9116dc86ed8bceffd31db0 () possibly failing. +2066403:S 12 Apr 2026 23:36:53.861 * Node ad56f48bae1934f4d1c3141ffee95fa8bf2fa37c () reported node 7cbfdaa3b3c0ebf9bf9116dc86ed8bceffd31db0 () as not reachable. +2066403:S 12 Apr 2026 23:36:54.000 * FAIL message received from d24296bca8e44a56521bbf138bec235238274ad7 () about 7cbfdaa3b3c0ebf9bf9116dc86ed8bceffd31db0 () +2066403:S 12 Apr 2026 23:36:54.000 * Node d24296bca8e44a56521bbf138bec235238274ad7 () reported node 7cbfdaa3b3c0ebf9bf9116dc86ed8bceffd31db0 () as not reachable. +2066403:S 12 Apr 2026 23:36:54.246 * Node 27c281c71471f64d57c2f90a6fce7ca7d11f3a9e () reported node 7cbfdaa3b3c0ebf9bf9116dc86ed8bceffd31db0 () as not reachable. +2066403:S 12 Apr 2026 23:36:54.247 * Node d24296bca8e44a56521bbf138bec235238274ad7 () reported node 7cbfdaa3b3c0ebf9bf9116dc86ed8bceffd31db0 () is back online. +2066403:S 12 Apr 2026 23:36:54.247 * Clear FAIL state for node 7cbfdaa3b3c0ebf9bf9116dc86ed8bceffd31db0 (): replica is reachable again. +2066403:S 12 Apr 2026 23:36:54.361 * Node ad56f48bae1934f4d1c3141ffee95fa8bf2fa37c () reported node 7cbfdaa3b3c0ebf9bf9116dc86ed8bceffd31db0 () is back online. +2066403:S 12 Apr 2026 23:36:54.449 * Node 27c281c71471f64d57c2f90a6fce7ca7d11f3a9e () reported node 7cbfdaa3b3c0ebf9bf9116dc86ed8bceffd31db0 () is back online. +2066403:S 12 Apr 2026 23:36:55.146 - Accepting cluster node connection from 127.0.0.1:53574 +2066403:S 12 Apr 2026 23:36:55.148 - Accepting cluster node connection from 127.0.0.1:53582 +2066403:S 12 Apr 2026 23:36:55.234 - Accepting cluster node connection from 127.0.0.1:53590 +2066403:S 12 Apr 2026 23:36:55.248 - Accepting cluster node connection from 127.0.0.1:53606 +### Starting test Replica redirects key access in migrating slots in tests/unit/cluster/slot-migration.tcl +### Starting test Replica of migrating node returns ASK redirect after READONLY in tests/unit/cluster/slot-migration.tcl +### Starting test Replica of migrating node returns TRYAGAIN after READONLY in tests/unit/cluster/slot-migration.tcl +### Starting test Replica of importing node returns TRYAGAIN after READONLY and ASKING in tests/unit/cluster/slot-migration.tcl +### Starting test New replica inherits migrating slot in tests/unit/cluster/slot-migration.tcl +2066403:S 12 Apr 2026 23:36:59.031 * NODE ceb41590bdb53727bffcb314c80a81989180d13a () possibly failing. +2066403:S 12 Apr 2026 23:36:59.402 * Node 27c281c71471f64d57c2f90a6fce7ca7d11f3a9e () reported node ceb41590bdb53727bffcb314c80a81989180d13a () as not reachable. +2066403:S 12 Apr 2026 23:36:59.637 * FAIL message received from ad56f48bae1934f4d1c3141ffee95fa8bf2fa37c () about ceb41590bdb53727bffcb314c80a81989180d13a () +2066403:S 12 Apr 2026 23:36:59.680 * Node d24296bca8e44a56521bbf138bec235238274ad7 () reported node ceb41590bdb53727bffcb314c80a81989180d13a () as not reachable. +2066403:S 12 Apr 2026 23:36:59.914 * Node ad56f48bae1934f4d1c3141ffee95fa8bf2fa37c () reported node ceb41590bdb53727bffcb314c80a81989180d13a () as not reachable. +2066403:S 12 Apr 2026 23:37:00.763 * Clear FAIL state for node ceb41590bdb53727bffcb314c80a81989180d13a (): replica is reachable again. +2066403:S 12 Apr 2026 23:37:00.951 * Node ad56f48bae1934f4d1c3141ffee95fa8bf2fa37c () reported node ceb41590bdb53727bffcb314c80a81989180d13a () is back online. +2066403:S 12 Apr 2026 23:37:00.956 * Node 27c281c71471f64d57c2f90a6fce7ca7d11f3a9e () reported node ceb41590bdb53727bffcb314c80a81989180d13a () is back online. +2066403:S 12 Apr 2026 23:37:01.554 * Node d24296bca8e44a56521bbf138bec235238274ad7 () reported node ceb41590bdb53727bffcb314c80a81989180d13a () is back online. +2066403:S 12 Apr 2026 23:37:01.649 - Accepting cluster node connection from 127.0.0.1:53612 +2066403:S 12 Apr 2026 23:37:02.169 * NODE ceb41590bdb53727bffcb314c80a81989180d13a () possibly failing. +2066403:S 12 Apr 2026 23:37:02.272 * Node d24296bca8e44a56521bbf138bec235238274ad7 () reported node ceb41590bdb53727bffcb314c80a81989180d13a () as not reachable. +2066403:S 12 Apr 2026 23:37:03.694 * NODE 7cbfdaa3b3c0ebf9bf9116dc86ed8bceffd31db0 () possibly failing. +2066403:S 12 Apr 2026 23:37:03.717 * NODE 27c281c71471f64d57c2f90a6fce7ca7d11f3a9e () possibly failing. +2066403:S 12 Apr 2026 23:37:03.717 * NODE ad56f48bae1934f4d1c3141ffee95fa8bf2fa37c () possibly failing. +2066403:S 12 Apr 2026 23:37:03.717 # Cluster state changed: fail +2066403:S 12 Apr 2026 23:37:03.717 * FAIL message received from 27c281c71471f64d57c2f90a6fce7ca7d11f3a9e () about ceb41590bdb53727bffcb314c80a81989180d13a () +2066403:S 12 Apr 2026 23:37:04.041 * Node d24296bca8e44a56521bbf138bec235238274ad7 () reported node ad56f48bae1934f4d1c3141ffee95fa8bf2fa37c () as not reachable. +2066403:S 12 Apr 2026 23:37:04.041 * Node 27c281c71471f64d57c2f90a6fce7ca7d11f3a9e () reported node ceb41590bdb53727bffcb314c80a81989180d13a () as not reachable. +2066403:S 12 Apr 2026 23:37:04.264 * Cluster state changed: ok +2066403:S 12 Apr 2026 23:37:04.469 - Error accepting cluster node connection: Connection reset by peer +2066403:S 12 Apr 2026 23:37:04.469 * Node d24296bca8e44a56521bbf138bec235238274ad7 () reported node 7cbfdaa3b3c0ebf9bf9116dc86ed8bceffd31db0 () as not reachable. +2066403:S 12 Apr 2026 23:37:04.480 - Accepting cluster node connection from 127.0.0.1:55114 +2066403:S 12 Apr 2026 23:37:04.493 - Accepting cluster node connection from 127.0.0.1:55130 +2066403:S 12 Apr 2026 23:37:04.500 * Node 27c281c71471f64d57c2f90a6fce7ca7d11f3a9e () reported node ceb41590bdb53727bffcb314c80a81989180d13a () is back online. +2066403:S 12 Apr 2026 23:37:04.617 * Node d24296bca8e44a56521bbf138bec235238274ad7 () reported node 7cbfdaa3b3c0ebf9bf9116dc86ed8bceffd31db0 () is back online. +2066403:S 12 Apr 2026 23:37:05.214 * Node d24296bca8e44a56521bbf138bec235238274ad7 () reported node ad56f48bae1934f4d1c3141ffee95fa8bf2fa37c () is back online. +2066403:S 12 Apr 2026 23:37:06.865 * Node ad56f48bae1934f4d1c3141ffee95fa8bf2fa37c () reported node ceb41590bdb53727bffcb314c80a81989180d13a () as not reachable. +2066403:S 12 Apr 2026 23:37:06.904 * Node 27c281c71471f64d57c2f90a6fce7ca7d11f3a9e () reported node ceb41590bdb53727bffcb314c80a81989180d13a () as not reachable. +2066403:S 12 Apr 2026 23:37:08.588 * Clear FAIL state for node ceb41590bdb53727bffcb314c80a81989180d13a (): primary without slots is reachable again. +2066403:S 12 Apr 2026 23:37:08.794 * Node 27c281c71471f64d57c2f90a6fce7ca7d11f3a9e () reported node ceb41590bdb53727bffcb314c80a81989180d13a () is back online. +2066403:S 12 Apr 2026 23:37:08.990 * Node d24296bca8e44a56521bbf138bec235238274ad7 () reported node ceb41590bdb53727bffcb314c80a81989180d13a () is back online. +2066403:S 12 Apr 2026 23:37:09.065 * A failover occurred in shard 0ccab0ca3bdf2df4fb1d88a6388f0d2e2e9492d9; node ceb41590bdb53727bffcb314c80a81989180d13a () failed over to node 27c281c71471f64d57c2f90a6fce7ca7d11f3a9e () with a config epoch of 5 +2066403:S 12 Apr 2026 23:37:09.065 * Node ceb41590bdb53727bffcb314c80a81989180d13a () is now a replica of node 27c281c71471f64d57c2f90a6fce7ca7d11f3a9e () in shard 0ccab0ca3bdf2df4fb1d88a6388f0d2e2e9492d9 +2066403:S 12 Apr 2026 23:37:09.141 * Node ad56f48bae1934f4d1c3141ffee95fa8bf2fa37c () reported node ceb41590bdb53727bffcb314c80a81989180d13a () is back online. +### Starting test New replica inherits importing slot in tests/unit/cluster/slot-migration.tcl +2066403:S 12 Apr 2026 23:37:12.558 * NODE 7cbfdaa3b3c0ebf9bf9116dc86ed8bceffd31db0 () possibly failing. +2066403:S 12 Apr 2026 23:37:12.605 - Accepting cluster node connection from 127.0.0.1:40244 +2066403:S 12 Apr 2026 23:37:12.669 - Accepting cluster node connection from 127.0.0.1:40250 +2066403:S 12 Apr 2026 23:37:12.671 - Accepting cluster node connection from 127.0.0.1:40262 +2066403:S 12 Apr 2026 23:37:12.704 * Node 27c281c71471f64d57c2f90a6fce7ca7d11f3a9e () reported node 7cbfdaa3b3c0ebf9bf9116dc86ed8bceffd31db0 () as not reachable. +2066403:S 12 Apr 2026 23:37:12.742 * Node d24296bca8e44a56521bbf138bec235238274ad7 () reported node 7cbfdaa3b3c0ebf9bf9116dc86ed8bceffd31db0 () as not reachable. +2066403:S 12 Apr 2026 23:37:12.742 * Marking node 7cbfdaa3b3c0ebf9bf9116dc86ed8bceffd31db0 () as failing (quorum reached). +2066403:S 12 Apr 2026 23:37:12.776 - Accepting cluster node connection from 127.0.0.1:40264 +2066403:S 12 Apr 2026 23:37:12.848 * Clear FAIL state for node 7cbfdaa3b3c0ebf9bf9116dc86ed8bceffd31db0 (): replica is reachable again. +2066403:S 12 Apr 2026 23:37:12.902 * FAIL message received from d24296bca8e44a56521bbf138bec235238274ad7 () about 7cbfdaa3b3c0ebf9bf9116dc86ed8bceffd31db0 () +2066403:S 12 Apr 2026 23:37:13.020 * Node 27c281c71471f64d57c2f90a6fce7ca7d11f3a9e () reported node 7cbfdaa3b3c0ebf9bf9116dc86ed8bceffd31db0 () is back online. +2066403:S 12 Apr 2026 23:37:13.021 * Clear FAIL state for node 7cbfdaa3b3c0ebf9bf9116dc86ed8bceffd31db0 (): primary without slots is reachable again. +2066403:S 12 Apr 2026 23:37:13.196 * Node d24296bca8e44a56521bbf138bec235238274ad7 () reported node 7cbfdaa3b3c0ebf9bf9116dc86ed8bceffd31db0 () is back online. +2066403:S 12 Apr 2026 23:37:22.556 - Accepting cluster node connection from 127.0.0.1:56258 +2066403:S 12 Apr 2026 23:37:22.657 * Node 27c281c71471f64d57c2f90a6fce7ca7d11f3a9e () reported node ceb41590bdb53727bffcb314c80a81989180d13a () as not reachable. +2066403:S 12 Apr 2026 23:37:22.657 * Node 27c281c71471f64d57c2f90a6fce7ca7d11f3a9e () reported node d24296bca8e44a56521bbf138bec235238274ad7 () as not reachable. +2066403:S 12 Apr 2026 23:37:22.657 * Node 27c281c71471f64d57c2f90a6fce7ca7d11f3a9e () reported node 7cbfdaa3b3c0ebf9bf9116dc86ed8bceffd31db0 () as not reachable. +2066403:S 12 Apr 2026 23:37:22.657 * Node 27c281c71471f64d57c2f90a6fce7ca7d11f3a9e () reported node ad56f48bae1934f4d1c3141ffee95fa8bf2fa37c () as not reachable. +2066403:S 12 Apr 2026 23:37:22.657 * Node 27c281c71471f64d57c2f90a6fce7ca7d11f3a9e () reported node ceb41590bdb53727bffcb314c80a81989180d13a () is back online. +2066403:S 12 Apr 2026 23:37:22.769 * Node 27c281c71471f64d57c2f90a6fce7ca7d11f3a9e () reported node ad56f48bae1934f4d1c3141ffee95fa8bf2fa37c () is back online. +2066403:S 12 Apr 2026 23:37:22.784 * Node 27c281c71471f64d57c2f90a6fce7ca7d11f3a9e () reported node d24296bca8e44a56521bbf138bec235238274ad7 () is back online. +2066403:S 12 Apr 2026 23:37:22.960 * Node 27c281c71471f64d57c2f90a6fce7ca7d11f3a9e () reported node 7cbfdaa3b3c0ebf9bf9116dc86ed8bceffd31db0 () is back online. +2066403:S 12 Apr 2026 23:37:30.058 * NODE 7cbfdaa3b3c0ebf9bf9116dc86ed8bceffd31db0 () possibly failing. +2066403:S 12 Apr 2026 23:37:30.058 - Accepting cluster node connection from 127.0.0.1:56274 +2066403:S 12 Apr 2026 23:37:30.129 - Accepting cluster node connection from 127.0.0.1:56276 +2066403:S 12 Apr 2026 23:37:30.131 - Accepting cluster node connection from 127.0.0.1:56286 +2066403:S 12 Apr 2026 23:37:30.194 * FAIL message received from ad56f48bae1934f4d1c3141ffee95fa8bf2fa37c () about 7cbfdaa3b3c0ebf9bf9116dc86ed8bceffd31db0 () +2066403:S 12 Apr 2026 23:37:30.194 * Node ad56f48bae1934f4d1c3141ffee95fa8bf2fa37c () reported node 7cbfdaa3b3c0ebf9bf9116dc86ed8bceffd31db0 () as not reachable. +2066403:S 12 Apr 2026 23:37:30.196 * Node 27c281c71471f64d57c2f90a6fce7ca7d11f3a9e () reported node 7cbfdaa3b3c0ebf9bf9116dc86ed8bceffd31db0 () as not reachable. +2066403:S 12 Apr 2026 23:37:30.248 * Clear FAIL state for node 7cbfdaa3b3c0ebf9bf9116dc86ed8bceffd31db0 (): primary without slots is reachable again. +2066403:S 12 Apr 2026 23:37:30.382 * Node ad56f48bae1934f4d1c3141ffee95fa8bf2fa37c () reported node 7cbfdaa3b3c0ebf9bf9116dc86ed8bceffd31db0 () is back online. +2066403:S 12 Apr 2026 23:37:30.398 * Node 27c281c71471f64d57c2f90a6fce7ca7d11f3a9e () reported node 7cbfdaa3b3c0ebf9bf9116dc86ed8bceffd31db0 () is back online. +2066403:S 12 Apr 2026 23:37:39.670 * NODE d24296bca8e44a56521bbf138bec235238274ad7 () possibly failing. +2066403:S 12 Apr 2026 23:37:39.670 * NODE 7cbfdaa3b3c0ebf9bf9116dc86ed8bceffd31db0 () possibly failing. +2066403:S 12 Apr 2026 23:37:39.671 * NODE 27c281c71471f64d57c2f90a6fce7ca7d11f3a9e () possibly failing. +2066403:S 12 Apr 2026 23:37:39.671 * NODE ceb41590bdb53727bffcb314c80a81989180d13a () possibly failing. +2066403:S 12 Apr 2026 23:37:39.671 # Cluster state changed: fail +2066403:S 12 Apr 2026 23:37:39.805 * Node d24296bca8e44a56521bbf138bec235238274ad7 () reported node 7cbfdaa3b3c0ebf9bf9116dc86ed8bceffd31db0 () as not reachable. +2066403:S 12 Apr 2026 23:37:39.805 * Node d24296bca8e44a56521bbf138bec235238274ad7 () reported node ceb41590bdb53727bffcb314c80a81989180d13a () as not reachable. +2066403:S 12 Apr 2026 23:37:39.828 * Node ad56f48bae1934f4d1c3141ffee95fa8bf2fa37c () reported node d24296bca8e44a56521bbf138bec235238274ad7 () as not reachable. +2066403:S 12 Apr 2026 23:37:39.828 * Node ad56f48bae1934f4d1c3141ffee95fa8bf2fa37c () reported node 27c281c71471f64d57c2f90a6fce7ca7d11f3a9e () as not reachable. +2066403:S 12 Apr 2026 23:37:39.896 * FAIL message received from ad56f48bae1934f4d1c3141ffee95fa8bf2fa37c () about ceb41590bdb53727bffcb314c80a81989180d13a () +2066403:S 12 Apr 2026 23:37:39.916 * Node 27c281c71471f64d57c2f90a6fce7ca7d11f3a9e () reported node ceb41590bdb53727bffcb314c80a81989180d13a () as not reachable. +2066403:S 12 Apr 2026 23:37:39.917 * Node 27c281c71471f64d57c2f90a6fce7ca7d11f3a9e () reported node 7cbfdaa3b3c0ebf9bf9116dc86ed8bceffd31db0 () as not reachable. +2066403:S 12 Apr 2026 23:37:39.917 * Marking node 7cbfdaa3b3c0ebf9bf9116dc86ed8bceffd31db0 () as failing (quorum reached). +2066403:S 12 Apr 2026 23:37:39.926 * Clear FAIL state for node 7cbfdaa3b3c0ebf9bf9116dc86ed8bceffd31db0 (): primary without slots is reachable again. +2066403:S 12 Apr 2026 23:37:39.926 * Clear FAIL state for node ceb41590bdb53727bffcb314c80a81989180d13a (): replica is reachable again. +2066403:S 12 Apr 2026 23:37:39.929 * Node d24296bca8e44a56521bbf138bec235238274ad7 () reported node ceb41590bdb53727bffcb314c80a81989180d13a () is back online. +2066403:S 12 Apr 2026 23:37:39.930 * Cluster state changed: ok +2066403:S 12 Apr 2026 23:37:40.025 * Node ad56f48bae1934f4d1c3141ffee95fa8bf2fa37c () reported node 27c281c71471f64d57c2f90a6fce7ca7d11f3a9e () is back online. +2066403:S 12 Apr 2026 23:37:40.025 * Node ad56f48bae1934f4d1c3141ffee95fa8bf2fa37c () reported node d24296bca8e44a56521bbf138bec235238274ad7 () is back online. +2066403:S 12 Apr 2026 23:37:40.076 * Node ad56f48bae1934f4d1c3141ffee95fa8bf2fa37c () reported node 7cbfdaa3b3c0ebf9bf9116dc86ed8bceffd31db0 () as not reachable. +2066403:S 12 Apr 2026 23:37:40.141 * Node 27c281c71471f64d57c2f90a6fce7ca7d11f3a9e () reported node 7cbfdaa3b3c0ebf9bf9116dc86ed8bceffd31db0 () is back online. +2066403:S 12 Apr 2026 23:37:40.141 * Node 27c281c71471f64d57c2f90a6fce7ca7d11f3a9e () reported node ceb41590bdb53727bffcb314c80a81989180d13a () is back online. +2066403:S 12 Apr 2026 23:37:40.177 * Node ad56f48bae1934f4d1c3141ffee95fa8bf2fa37c () reported node 7cbfdaa3b3c0ebf9bf9116dc86ed8bceffd31db0 () is back online. +2066403:S 12 Apr 2026 23:37:40.216 * Node d24296bca8e44a56521bbf138bec235238274ad7 () reported node 7cbfdaa3b3c0ebf9bf9116dc86ed8bceffd31db0 () is back online. +2066403:S 12 Apr 2026 23:37:49.343 * NODE 7cbfdaa3b3c0ebf9bf9116dc86ed8bceffd31db0 () possibly failing. +2066403:S 12 Apr 2026 23:37:49.448 * Node 27c281c71471f64d57c2f90a6fce7ca7d11f3a9e () reported node 7cbfdaa3b3c0ebf9bf9116dc86ed8bceffd31db0 () as not reachable. +2066403:S 12 Apr 2026 23:37:49.553 * Node 27c281c71471f64d57c2f90a6fce7ca7d11f3a9e () reported node 7cbfdaa3b3c0ebf9bf9116dc86ed8bceffd31db0 () is back online. +2066403:S 12 Apr 2026 23:37:58.605 * NODE 27c281c71471f64d57c2f90a6fce7ca7d11f3a9e () possibly failing. +2066403:S 12 Apr 2026 23:37:58.606 - Accepting cluster node connection from 127.0.0.1:43520 +2066403:S 12 Apr 2026 23:37:58.642 - Accepting cluster node connection from 127.0.0.1:43526 +2066403:S 12 Apr 2026 23:37:58.674 * Node d24296bca8e44a56521bbf138bec235238274ad7 () reported node 27c281c71471f64d57c2f90a6fce7ca7d11f3a9e () as not reachable. +2066403:S 12 Apr 2026 23:37:58.700 - Accepting cluster node connection from 127.0.0.1:43536 +2066403:S 12 Apr 2026 23:37:58.740 * Node d24296bca8e44a56521bbf138bec235238274ad7 () reported node 7cbfdaa3b3c0ebf9bf9116dc86ed8bceffd31db0 () as not reachable. +2066403:S 12 Apr 2026 23:37:58.741 * Node ad56f48bae1934f4d1c3141ffee95fa8bf2fa37c () reported node 27c281c71471f64d57c2f90a6fce7ca7d11f3a9e () as not reachable. +2066403:S 12 Apr 2026 23:37:58.741 * Marking node 27c281c71471f64d57c2f90a6fce7ca7d11f3a9e () as failing (quorum reached). +2066403:S 12 Apr 2026 23:37:58.741 # Cluster state changed: fail +2066403:S 12 Apr 2026 23:37:58.837 * Node d24296bca8e44a56521bbf138bec235238274ad7 () reported node 7cbfdaa3b3c0ebf9bf9116dc86ed8bceffd31db0 () is back online. +2066403:S 12 Apr 2026 23:37:59.722 * Node ceb41590bdb53727bffcb314c80a81989180d13a () reported node 27c281c71471f64d57c2f90a6fce7ca7d11f3a9e () as not reachable. +2066403:S 12 Apr 2026 23:37:59.723 * Cluster state changed: ok +2066403:S 12 Apr 2026 23:37:59.802 * Clear FAIL state for node 27c281c71471f64d57c2f90a6fce7ca7d11f3a9e (): primary without slots is reachable again. +2066403:S 12 Apr 2026 23:37:59.802 * A failover occurred in shard 0ccab0ca3bdf2df4fb1d88a6388f0d2e2e9492d9; node 27c281c71471f64d57c2f90a6fce7ca7d11f3a9e () failed over to node ceb41590bdb53727bffcb314c80a81989180d13a () with a config epoch of 9 +2066403:S 12 Apr 2026 23:37:59.802 * Node 27c281c71471f64d57c2f90a6fce7ca7d11f3a9e () is now a replica of node ceb41590bdb53727bffcb314c80a81989180d13a () in shard 0ccab0ca3bdf2df4fb1d88a6388f0d2e2e9492d9 +2066403:S 12 Apr 2026 23:37:59.885 * Node ceb41590bdb53727bffcb314c80a81989180d13a () reported node 27c281c71471f64d57c2f90a6fce7ca7d11f3a9e () is back online. +2066403:S 12 Apr 2026 23:37:59.985 * Node ad56f48bae1934f4d1c3141ffee95fa8bf2fa37c () reported node 27c281c71471f64d57c2f90a6fce7ca7d11f3a9e () is back online. +2066403:S 12 Apr 2026 23:37:59.987 * Node d24296bca8e44a56521bbf138bec235238274ad7 () reported node 27c281c71471f64d57c2f90a6fce7ca7d11f3a9e () is back online. +2066403:S 12 Apr 2026 23:38:07.837 - Accepting cluster node connection from 127.0.0.1:43594 +2066403:S 12 Apr 2026 23:38:07.858 - Accepting cluster node connection from 127.0.0.1:43608 +2066403:S 12 Apr 2026 23:38:07.894 - Accepting cluster node connection from 127.0.0.1:43610 +2066403:S 12 Apr 2026 23:38:07.918 - Accepting cluster node connection from 127.0.0.1:43622 +2066403:S 12 Apr 2026 23:38:07.933 * NODE 7cbfdaa3b3c0ebf9bf9116dc86ed8bceffd31db0 () possibly failing. +2066403:S 12 Apr 2026 23:38:07.933 * NODE 27c281c71471f64d57c2f90a6fce7ca7d11f3a9e () possibly failing. +2066403:S 12 Apr 2026 23:38:17.545 - Accepting cluster node connection from 127.0.0.1:44762 +2066403:S 12 Apr 2026 23:38:17.553 - Error accepting cluster node connection: error:0A000126:SSL routines::unexpected eof while reading +2066403:S 12 Apr 2026 23:38:17.557 - Accepting cluster node connection from 127.0.0.1:44772 +2066403:S 12 Apr 2026 23:38:17.559 - Accepting cluster node connection from 127.0.0.1:44774 +2066403:S 12 Apr 2026 23:38:17.609 * Node d24296bca8e44a56521bbf138bec235238274ad7 () reported node ceb41590bdb53727bffcb314c80a81989180d13a () as not reachable. +2066403:S 12 Apr 2026 23:38:17.642 - Accepting cluster node connection from 127.0.0.1:44788 +2066403:S 12 Apr 2026 23:38:17.675 * Node d24296bca8e44a56521bbf138bec235238274ad7 () reported node ad56f48bae1934f4d1c3141ffee95fa8bf2fa37c () as not reachable. +2066403:S 12 Apr 2026 23:38:17.884 - Error accepting cluster node connection: Connection reset by peer +2066403:S 12 Apr 2026 23:38:17.901 * Node ad56f48bae1934f4d1c3141ffee95fa8bf2fa37c () reported node d24296bca8e44a56521bbf138bec235238274ad7 () as not reachable. +2066403:S 12 Apr 2026 23:38:17.901 * Node ad56f48bae1934f4d1c3141ffee95fa8bf2fa37c () reported node 27c281c71471f64d57c2f90a6fce7ca7d11f3a9e () as not reachable. +2066403:S 12 Apr 2026 23:38:17.985 - Accepting cluster node connection from 127.0.0.1:44796 +2066403:S 12 Apr 2026 23:38:18.058 - Error accepting cluster node connection: Connection reset by peer +2066403:S 12 Apr 2026 23:38:18.058 * Node ad56f48bae1934f4d1c3141ffee95fa8bf2fa37c () reported node 7cbfdaa3b3c0ebf9bf9116dc86ed8bceffd31db0 () as not reachable. +2066403:S 12 Apr 2026 23:38:18.087 * Node d24296bca8e44a56521bbf138bec235238274ad7 () reported node ad56f48bae1934f4d1c3141ffee95fa8bf2fa37c () is back online. +2066403:S 12 Apr 2026 23:38:18.129 * Node d24296bca8e44a56521bbf138bec235238274ad7 () reported node ceb41590bdb53727bffcb314c80a81989180d13a () is back online. +2066403:S 12 Apr 2026 23:38:18.159 - Accepting cluster node connection from 127.0.0.1:44806 +2066403:S 12 Apr 2026 23:38:18.229 * Node ad56f48bae1934f4d1c3141ffee95fa8bf2fa37c () reported node 27c281c71471f64d57c2f90a6fce7ca7d11f3a9e () is back online. +2066403:S 12 Apr 2026 23:38:18.229 * Node ad56f48bae1934f4d1c3141ffee95fa8bf2fa37c () reported node d24296bca8e44a56521bbf138bec235238274ad7 () is back online. +2066403:S 12 Apr 2026 23:38:18.229 * Node ad56f48bae1934f4d1c3141ffee95fa8bf2fa37c () reported node 7cbfdaa3b3c0ebf9bf9116dc86ed8bceffd31db0 () is back online. +2066403:S 12 Apr 2026 23:38:18.241 * Node ad56f48bae1934f4d1c3141ffee95fa8bf2fa37c () reported node ceb41590bdb53727bffcb314c80a81989180d13a () as not reachable. +2066403:S 12 Apr 2026 23:38:20.384 * Node ad56f48bae1934f4d1c3141ffee95fa8bf2fa37c () reported node ceb41590bdb53727bffcb314c80a81989180d13a () is back online. +2066403:S 12 Apr 2026 23:38:26.461 * NODE 7cbfdaa3b3c0ebf9bf9116dc86ed8bceffd31db0 () possibly failing. +2066403:S 12 Apr 2026 23:38:26.710 * FAIL message received from ceb41590bdb53727bffcb314c80a81989180d13a () about 7cbfdaa3b3c0ebf9bf9116dc86ed8bceffd31db0 () +2066403:S 12 Apr 2026 23:38:26.753 * Clear FAIL state for node 7cbfdaa3b3c0ebf9bf9116dc86ed8bceffd31db0 (): primary without slots is reachable again. +2066403:S 12 Apr 2026 23:38:26.804 * Node ad56f48bae1934f4d1c3141ffee95fa8bf2fa37c () reported node 7cbfdaa3b3c0ebf9bf9116dc86ed8bceffd31db0 () as not reachable. +2066403:S 12 Apr 2026 23:38:26.805 * Node d24296bca8e44a56521bbf138bec235238274ad7 () reported node 7cbfdaa3b3c0ebf9bf9116dc86ed8bceffd31db0 () as not reachable. +2066403:S 12 Apr 2026 23:38:26.953 * Node d24296bca8e44a56521bbf138bec235238274ad7 () reported node 7cbfdaa3b3c0ebf9bf9116dc86ed8bceffd31db0 () is back online. +2066403:S 12 Apr 2026 23:38:26.953 * Node ad56f48bae1934f4d1c3141ffee95fa8bf2fa37c () reported node 7cbfdaa3b3c0ebf9bf9116dc86ed8bceffd31db0 () is back online. +2066403:S 12 Apr 2026 23:38:35.500 * NODE ceb41590bdb53727bffcb314c80a81989180d13a () possibly failing. +2066403:S 12 Apr 2026 23:38:35.501 - Accepting cluster node connection from 127.0.0.1:44532 +2066403:S 12 Apr 2026 23:38:35.631 * Node ad56f48bae1934f4d1c3141ffee95fa8bf2fa37c () reported node ceb41590bdb53727bffcb314c80a81989180d13a () as not reachable. +2066403:S 12 Apr 2026 23:38:35.725 * Node ad56f48bae1934f4d1c3141ffee95fa8bf2fa37c () reported node ceb41590bdb53727bffcb314c80a81989180d13a () is back online. +2066403:S 12 Apr 2026 23:38:44.687 * NODE 7cbfdaa3b3c0ebf9bf9116dc86ed8bceffd31db0 () possibly failing. +2066403:S 12 Apr 2026 23:38:44.798 * Node d24296bca8e44a56521bbf138bec235238274ad7 () reported node ad56f48bae1934f4d1c3141ffee95fa8bf2fa37c () as not reachable. +2066403:S 12 Apr 2026 23:38:44.798 * Node d24296bca8e44a56521bbf138bec235238274ad7 () reported node 7cbfdaa3b3c0ebf9bf9116dc86ed8bceffd31db0 () as not reachable. +2066403:S 12 Apr 2026 23:38:44.798 * Node d24296bca8e44a56521bbf138bec235238274ad7 () reported node ceb41590bdb53727bffcb314c80a81989180d13a () as not reachable. +2066403:S 12 Apr 2026 23:38:44.798 * Node d24296bca8e44a56521bbf138bec235238274ad7 () reported node 27c281c71471f64d57c2f90a6fce7ca7d11f3a9e () as not reachable. +2066403:S 12 Apr 2026 23:38:44.836 * FAIL message received from ceb41590bdb53727bffcb314c80a81989180d13a () about 7cbfdaa3b3c0ebf9bf9116dc86ed8bceffd31db0 () +2066403:S 12 Apr 2026 23:38:44.885 * Node ceb41590bdb53727bffcb314c80a81989180d13a () reported node 7cbfdaa3b3c0ebf9bf9116dc86ed8bceffd31db0 () as not reachable. +2066403:S 12 Apr 2026 23:38:44.902 * Clear FAIL state for node 7cbfdaa3b3c0ebf9bf9116dc86ed8bceffd31db0 (): primary without slots is reachable again. +2066403:S 12 Apr 2026 23:38:44.902 * Node d24296bca8e44a56521bbf138bec235238274ad7 () reported node 27c281c71471f64d57c2f90a6fce7ca7d11f3a9e () is back online. +2066403:S 12 Apr 2026 23:38:44.902 * Node d24296bca8e44a56521bbf138bec235238274ad7 () reported node 7cbfdaa3b3c0ebf9bf9116dc86ed8bceffd31db0 () is back online. +2066403:S 12 Apr 2026 23:38:44.902 * Node d24296bca8e44a56521bbf138bec235238274ad7 () reported node ceb41590bdb53727bffcb314c80a81989180d13a () is back online. +2066403:S 12 Apr 2026 23:38:44.995 * Node d24296bca8e44a56521bbf138bec235238274ad7 () reported node ad56f48bae1934f4d1c3141ffee95fa8bf2fa37c () is back online. +2066403:S 12 Apr 2026 23:38:45.098 * Node ceb41590bdb53727bffcb314c80a81989180d13a () reported node 7cbfdaa3b3c0ebf9bf9116dc86ed8bceffd31db0 () is back online. +2066403:S 12 Apr 2026 23:38:54.148 * NODE 7cbfdaa3b3c0ebf9bf9116dc86ed8bceffd31db0 () possibly failing. +2066403:S 12 Apr 2026 23:38:54.149 * NODE ad56f48bae1934f4d1c3141ffee95fa8bf2fa37c () possibly failing. +2066403:S 12 Apr 2026 23:38:54.149 - Accepting cluster node connection from 127.0.0.1:41016 +2066403:S 12 Apr 2026 23:38:54.152 - Error accepting cluster node connection: error:0A000126:SSL routines::unexpected eof while reading +2066403:S 12 Apr 2026 23:38:54.209 - Accepting cluster node connection from 127.0.0.1:41020 +2066403:S 12 Apr 2026 23:38:54.224 * Node d24296bca8e44a56521bbf138bec235238274ad7 () reported node ad56f48bae1934f4d1c3141ffee95fa8bf2fa37c () as not reachable. +2066403:S 12 Apr 2026 23:38:54.224 * Node d24296bca8e44a56521bbf138bec235238274ad7 () reported node 7cbfdaa3b3c0ebf9bf9116dc86ed8bceffd31db0 () as not reachable. +2066403:S 12 Apr 2026 23:38:54.224 * Node d24296bca8e44a56521bbf138bec235238274ad7 () reported node ceb41590bdb53727bffcb314c80a81989180d13a () as not reachable. +2066403:S 12 Apr 2026 23:38:54.224 * Node d24296bca8e44a56521bbf138bec235238274ad7 () reported node 27c281c71471f64d57c2f90a6fce7ca7d11f3a9e () as not reachable. +2066403:S 12 Apr 2026 23:38:54.253 - Accepting cluster node connection from 127.0.0.1:41022 +2066403:S 12 Apr 2026 23:38:54.284 * Node d24296bca8e44a56521bbf138bec235238274ad7 () reported node ceb41590bdb53727bffcb314c80a81989180d13a () is back online. +2066403:S 12 Apr 2026 23:38:54.285 * Node d24296bca8e44a56521bbf138bec235238274ad7 () reported node ad56f48bae1934f4d1c3141ffee95fa8bf2fa37c () is back online. +2066403:S 12 Apr 2026 23:38:54.285 * Node d24296bca8e44a56521bbf138bec235238274ad7 () reported node 7cbfdaa3b3c0ebf9bf9116dc86ed8bceffd31db0 () is back online. +2066403:S 12 Apr 2026 23:38:54.409 * Node d24296bca8e44a56521bbf138bec235238274ad7 () reported node 27c281c71471f64d57c2f90a6fce7ca7d11f3a9e () is back online. +2066403:S 12 Apr 2026 23:39:02.039 * Node ceb41590bdb53727bffcb314c80a81989180d13a () reported node 7cbfdaa3b3c0ebf9bf9116dc86ed8bceffd31db0 () as not reachable. +2066403:S 12 Apr 2026 23:39:02.050 * NODE 7cbfdaa3b3c0ebf9bf9116dc86ed8bceffd31db0 () possibly failing. +2066403:S 12 Apr 2026 23:39:03.153 * NODE d24296bca8e44a56521bbf138bec235238274ad7 () possibly failing. +2066403:S 12 Apr 2026 23:39:03.153 * NODE ad56f48bae1934f4d1c3141ffee95fa8bf2fa37c () possibly failing. +2066403:S 12 Apr 2026 23:39:03.153 # Cluster state changed: fail +2066403:S 12 Apr 2026 23:39:03.184 * FAIL message received from d24296bca8e44a56521bbf138bec235238274ad7 () about 7cbfdaa3b3c0ebf9bf9116dc86ed8bceffd31db0 () +2066403:S 12 Apr 2026 23:39:03.283 * Node d24296bca8e44a56521bbf138bec235238274ad7 () reported node 7cbfdaa3b3c0ebf9bf9116dc86ed8bceffd31db0 () as not reachable. +2066403:S 12 Apr 2026 23:39:03.299 * Cluster state changed: ok +2066403:S 12 Apr 2026 23:39:03.299 * Node ad56f48bae1934f4d1c3141ffee95fa8bf2fa37c () reported node 7cbfdaa3b3c0ebf9bf9116dc86ed8bceffd31db0 () as not reachable. +2066403:S 12 Apr 2026 23:39:03.303 * Clear FAIL state for node 7cbfdaa3b3c0ebf9bf9116dc86ed8bceffd31db0 (): primary without slots is reachable again. +2066403:S 12 Apr 2026 23:39:03.385 * Node ceb41590bdb53727bffcb314c80a81989180d13a () reported node 7cbfdaa3b3c0ebf9bf9116dc86ed8bceffd31db0 () is back online. +2066403:S 12 Apr 2026 23:39:03.459 * Node d24296bca8e44a56521bbf138bec235238274ad7 () reported node 7cbfdaa3b3c0ebf9bf9116dc86ed8bceffd31db0 () is back online. +2066403:S 12 Apr 2026 23:39:03.459 * Node ad56f48bae1934f4d1c3141ffee95fa8bf2fa37c () reported node 7cbfdaa3b3c0ebf9bf9116dc86ed8bceffd31db0 () is back online. +2066403:S 12 Apr 2026 23:39:21.025 * NODE d24296bca8e44a56521bbf138bec235238274ad7 () possibly failing. +2066403:S 12 Apr 2026 23:39:21.025 * NODE 7cbfdaa3b3c0ebf9bf9116dc86ed8bceffd31db0 () possibly failing. +2066403:S 12 Apr 2026 23:39:21.026 * NODE 27c281c71471f64d57c2f90a6fce7ca7d11f3a9e () possibly failing. +2066403:S 12 Apr 2026 23:39:21.128 - Accepting cluster node connection from 127.0.0.1:38538 +2066403:S 12 Apr 2026 23:39:21.143 - Accepting cluster node connection from 127.0.0.1:38548 +2066403:S 12 Apr 2026 23:39:21.147 * Node ad56f48bae1934f4d1c3141ffee95fa8bf2fa37c () reported node d24296bca8e44a56521bbf138bec235238274ad7 () as not reachable. +2066403:S 12 Apr 2026 23:39:21.147 * Node ad56f48bae1934f4d1c3141ffee95fa8bf2fa37c () reported node 27c281c71471f64d57c2f90a6fce7ca7d11f3a9e () as not reachable. +2066403:S 12 Apr 2026 23:39:21.147 * Node ad56f48bae1934f4d1c3141ffee95fa8bf2fa37c () reported node ceb41590bdb53727bffcb314c80a81989180d13a () as not reachable. +2066403:S 12 Apr 2026 23:39:21.147 * Node ad56f48bae1934f4d1c3141ffee95fa8bf2fa37c () reported node 7cbfdaa3b3c0ebf9bf9116dc86ed8bceffd31db0 () as not reachable. +2066403:S 12 Apr 2026 23:39:21.163 * Node ceb41590bdb53727bffcb314c80a81989180d13a () reported node ad56f48bae1934f4d1c3141ffee95fa8bf2fa37c () as not reachable. +2066403:S 12 Apr 2026 23:39:21.163 * Node ceb41590bdb53727bffcb314c80a81989180d13a () reported node 27c281c71471f64d57c2f90a6fce7ca7d11f3a9e () as not reachable. +2066403:S 12 Apr 2026 23:39:21.163 * Node ceb41590bdb53727bffcb314c80a81989180d13a () reported node d24296bca8e44a56521bbf138bec235238274ad7 () as not reachable. +2066403:S 12 Apr 2026 23:39:21.229 * Node ad56f48bae1934f4d1c3141ffee95fa8bf2fa37c () reported node d24296bca8e44a56521bbf138bec235238274ad7 () is back online. +2066403:S 12 Apr 2026 23:39:21.229 * Node ad56f48bae1934f4d1c3141ffee95fa8bf2fa37c () reported node 27c281c71471f64d57c2f90a6fce7ca7d11f3a9e () is back online. +2066403:S 12 Apr 2026 23:39:21.229 * Node ad56f48bae1934f4d1c3141ffee95fa8bf2fa37c () reported node 7cbfdaa3b3c0ebf9bf9116dc86ed8bceffd31db0 () is back online. +2066403:S 12 Apr 2026 23:39:21.229 * Node ceb41590bdb53727bffcb314c80a81989180d13a () reported node 27c281c71471f64d57c2f90a6fce7ca7d11f3a9e () is back online. +2066403:S 12 Apr 2026 23:39:21.229 * Node ceb41590bdb53727bffcb314c80a81989180d13a () reported node ad56f48bae1934f4d1c3141ffee95fa8bf2fa37c () is back online. +2066403:S 12 Apr 2026 23:39:21.331 * Node ad56f48bae1934f4d1c3141ffee95fa8bf2fa37c () reported node ceb41590bdb53727bffcb314c80a81989180d13a () is back online. +2066403:S 12 Apr 2026 23:39:21.541 * Node ceb41590bdb53727bffcb314c80a81989180d13a () reported node d24296bca8e44a56521bbf138bec235238274ad7 () is back online. +2066403:S 12 Apr 2026 23:39:22.042 - Connection with Node 27c281c71471f64d57c2f90a6fce7ca7d11f3a9e at 127.0.0.1:31699 failed: (null) +2066403:S 12 Apr 2026 23:39:22.144 - Connection with Node 27c281c71471f64d57c2f90a6fce7ca7d11f3a9e at 127.0.0.1:31699 failed: (null) +2066403:S 12 Apr 2026 23:39:22.144 - Connection with Node ad56f48bae1934f4d1c3141ffee95fa8bf2fa37c at 127.0.0.1:31697 failed: (null) +2066403:S 12 Apr 2026 23:39:22.154 - Reading from client: error:0A000126:SSL routines::unexpected eof while reading +2066403:S 12 Apr 2026 23:39:22.154 * Connection with primary lost. +2066403:S 12 Apr 2026 23:39:22.154 * Caching the disconnected primary state. +2066403:S 12 Apr 2026 23:39:22.154 * Reconnecting to PRIMARY 127.0.0.1:21695 +2066403:S 12 Apr 2026 23:39:22.154 * PRIMARY <-> REPLICA sync started +2066403:S 12 Apr 2026 23:39:22.154 # Error condition on socket for SYNC: (null) +2066403:S 12 Apr 2026 23:39:22.246 - Connection with Node d24296bca8e44a56521bbf138bec235238274ad7 at 127.0.0.1:31695 failed: (null) +2066403:S 12 Apr 2026 23:39:22.246 - Connection with Node 27c281c71471f64d57c2f90a6fce7ca7d11f3a9e at 127.0.0.1:31699 failed: (null) +2066403:S 12 Apr 2026 23:39:22.246 - Connection with Node ad56f48bae1934f4d1c3141ffee95fa8bf2fa37c at 127.0.0.1:31697 failed: (null) +2066403:S 12 Apr 2026 23:39:22.348 - Connection with Node d24296bca8e44a56521bbf138bec235238274ad7 at 127.0.0.1:31695 failed: (null) +2066403:S 12 Apr 2026 23:39:22.348 - Connection with Node 27c281c71471f64d57c2f90a6fce7ca7d11f3a9e at 127.0.0.1:31699 failed: (null) +2066403:S 12 Apr 2026 23:39:22.349 - Connection with Node ad56f48bae1934f4d1c3141ffee95fa8bf2fa37c at 127.0.0.1:31697 failed: (null) +2066403:S 12 Apr 2026 23:39:22.349 - Connection with Node ceb41590bdb53727bffcb314c80a81989180d13a at 127.0.0.1:31693 failed: (null) +===== End of server log (pid 2066403) ===== + + +===== Start of server stderr log (pid 2066403) ===== + + +===== End of server stderr log (pid 2066403) ===== + === (external:skip cluster) Starting server 127.0.0.1:21701 ok === (external:skip cluster) Starting server 127.0.0.1:21703 ok === (external:skip cluster) Starting server 127.0.0.1:21705 ok === (external:skip cluster) Starting server 127.0.0.1:21707 ok -[ok]: Cluster is up (59 ms) -[ok]: Cluster is writable (1287 ms) -[ok]: Killing the first primary node (155 ms) -[ok]: Wait for failover (4446 ms) -[ok]: Cluster should eventually be up again (1949 ms) -[ok]: Restarting the previously killed primary node (12 ms) -[ok]: Instance #0 gets converted into a replica (319 ms) -[ok]: Make sure the replicas always get the different ranks (35 ms) -[64/124 done]: unit/cluster/failover (53 seconds) -Testing unit/cluster/cluster-multiple-meets -[ignore]: Not supported in tls mode -[65/124 done]: unit/cluster/cluster-multiple-meets (0 seconds) -Testing unit/cluster/failover2 === (external:skip cluster) Starting server 127.0.0.1:21709 ok === (external:skip cluster) Starting server 127.0.0.1:21711 ok +[ok]: Multiple slot migration states are replicated (157 ms) +[ok]: New replica inherits multiple migrating slots (887 ms) +[ok]: Slot finalization succeeds on both primary and replicas (62 ms) === (external:skip cluster) Starting server 127.0.0.1:21713 ok === (external:skip cluster) Starting server 127.0.0.1:21715 ok === (external:skip cluster) Starting server 127.0.0.1:21717 ok === (external:skip cluster) Starting server 127.0.0.1:21719 ok === (external:skip cluster) Starting server 127.0.0.1:21721 ok -[ok]: Cluster is up (42 ms) -[ok]: Cluster is writable (1115 ms) -[ok]: Killing one primary node (223 ms) -[ok]: Wait for failover (6494 ms) -[ok]: Killing the new primary node (259 ms) -[ok]: Cluster should eventually be up again (1403 ms) -[ok]: wait for new failover (5128 ms) -[ok]: Restarting the previously killed primary nodes (45 ms) -[ok]: Make sure there is no failover timeout (42 ms) -[66/124 done]: unit/cluster/failover2 (30 seconds) -Testing unit/cluster/slave-stop-cond === (external:skip cluster) Starting server 127.0.0.1:21723 ok +[ok]: Slot is auto-claimed by target after source relinquishes ownership (1063 ms) === (external:skip cluster) Starting server 127.0.0.1:21725 ok === (external:skip cluster) Starting server 127.0.0.1:21727 ok === (external:skip cluster) Starting server 127.0.0.1:21729 ok === (external:skip cluster) Starting server 127.0.0.1:21731 ok === (external:skip cluster) Starting server 127.0.0.1:21733 ok === (external:skip cluster) Starting server 127.0.0.1:21735 ok +[ok]: CLUSTER SETSLOT with invalid timeouts (19 ms) === (external:skip cluster) Starting server 127.0.0.1:21737 ok === (external:skip cluster) Starting server 127.0.0.1:21739 ok === (external:skip cluster) Starting server 127.0.0.1:21741 ok -[ok]: Cluster is up (59 ms) -[ok]: The first master has actually one slave (7 ms) -[ok]: Slaves of #0 is instance #5 as expected (0 ms) -[ok]: Instance #5 synced with the master (1 ms) -[ok]: Lower the slave validity factor of #5 to the value of 2 (0 ms) -[ok]: Break master-slave link and prevent further reconnections (20281 ms) -[ok]: Slave #5 is reachable and alive (18 ms) -[ok]: Slave #5 should not be able to failover (10013 ms) -[ok]: Cluster should be down (1938 ms) -[67/124 done]: unit/cluster/slave-stop-cond (49 seconds) -Testing unit/cluster/replica-in-sync === (external:skip cluster) Starting server 127.0.0.1:21743 ok === (external:skip cluster) Starting server 127.0.0.1:21745 ok -[ok]: Cluster is up (8 ms) -[ok]: Cluster is writable (1940 ms) -[ok]: Fill up primary with data (10 ms) -[ok]: Add new node as replica (20 ms) -[ok]: Check digest and replica state (19 ms) -[ok]: Replica in loading state is hidden (9354 ms) -[ok]: Check disconnected replica not hidden from slots (282 ms) -[68/124 done]: unit/cluster/replica-in-sync (23 seconds) -Testing unit/cluster/replica-migration === (external:skip cluster) Starting server 127.0.0.1:21747 ok +[ok]: CLUSTER SETSLOT with an explicit timeout (3095 ms) +[ignore]: Not supported in tls mode === (external:skip cluster) Starting server 127.0.0.1:21749 ok === (external:skip cluster) Starting server 127.0.0.1:21751 ok === (external:skip cluster) Starting server 127.0.0.1:21753 ok @@ -5161,43 +7357,60 @@ === (external:skip cluster) Starting server 127.0.0.1:21757 ok === (external:skip cluster) Starting server 127.0.0.1:21759 ok === (external:skip cluster) Starting server 127.0.0.1:21761 ok -[ok]: Migrated replica reports zero repl offset and rank, and fails to win election - shutdown (28637 ms) === (external:skip cluster) Starting server 127.0.0.1:21763 ok === (external:skip cluster) Starting server 127.0.0.1:21765 ok +[ok]: Slot migration is ok when the replicas are down (277 ms) +[64/124 done]: unit/cluster/slot-migration (254 seconds) +Testing unit/cluster/no-failover-option === (external:skip cluster) Starting server 127.0.0.1:21767 ok === (external:skip cluster) Starting server 127.0.0.1:21769 ok === (external:skip cluster) Starting server 127.0.0.1:21771 ok === (external:skip cluster) Starting server 127.0.0.1:21773 ok === (external:skip cluster) Starting server 127.0.0.1:21775 ok === (external:skip cluster) Starting server 127.0.0.1:21777 ok -[ok]: Migrated replica reports zero repl offset and rank, and fails to win election - sigstop (23075 ms) +[ok]: Cluster is up (7 ms) +[ok]: Instance #3 is a replica (2 ms) +[ok]: Instance #3 synced with the master (1 ms) +[ok]: The nofailover flag is propagated (178 ms) +[ok]: Killing one master node (301 ms) +[ok]: Cluster should be still down after some time (3114 ms) +[ok]: Instance #3 is still a replica (2 ms) +[ok]: Restarting the previously killed master node (7 ms) +[65/124 done]: unit/cluster/no-failover-option (18 seconds) +Testing unit/cluster/slot-ownership === (external:skip cluster) Starting server 127.0.0.1:21779 ok === (external:skip cluster) Starting server 127.0.0.1:21781 ok === (external:skip cluster) Starting server 127.0.0.1:21783 ok === (external:skip cluster) Starting server 127.0.0.1:21785 ok +[ok]: Verify that slot ownership transfer through gossip propagates deletes to replicas (208 ms) === (external:skip cluster) Starting server 127.0.0.1:21787 ok === (external:skip cluster) Starting server 127.0.0.1:21789 ok === (external:skip cluster) Starting server 127.0.0.1:21791 ok === (external:skip cluster) Starting server 127.0.0.1:21793 ok -[ok]: New non-empty replica reports zero repl offset and rank, and fails to win election - shutdown (30144 ms) +[ok]: Primary lost a slot during the shutdown waiting (7229 ms) === (external:skip cluster) Starting server 127.0.0.1:21795 ok === (external:skip cluster) Starting server 127.0.0.1:21797 ok === (external:skip cluster) Starting server 127.0.0.1:21799 ok === (external:skip cluster) Starting server 127.0.0.1:21801 ok +[ok]: Primary lost a slot during the manual failover pausing (93 ms) === (external:skip cluster) Starting server 127.0.0.1:21803 ok === (external:skip cluster) Starting server 127.0.0.1:21805 ok === (external:skip cluster) Starting server 127.0.0.1:21807 ok === (external:skip cluster) Starting server 127.0.0.1:21809 ok -[ok]: New non-empty replica reports zero repl offset and rank, and fails to win election - sigstop (37300 ms) +[ok]: Primary lost a slot during the client pause command (72 ms) +[66/124 done]: unit/cluster/slot-ownership (62 seconds) +Testing unit/cluster/announced-endpoints === (external:skip cluster) Starting server 127.0.0.1:21811 ok === (external:skip cluster) Starting server 127.0.0.1:21813 ok === (external:skip cluster) Starting server 127.0.0.1:21815 ok === (external:skip cluster) Starting server 127.0.0.1:21817 ok -=== (external:skip cluster) Starting server 127.0.0.1:21819 ok +[ok]: Test change cluster-announce-port and cluster-announce-tls-port at runtime (253 ms) +[ok]: Test change cluster-announce-bus-port at runtime (135 ms) +[67/124 done]: unit/cluster/announced-endpoints (14 seconds) +Testing unit/cluster/slave-selection === (external:skip cluster) Starting server 127.0.0.1:21821 ok === (external:skip cluster) Starting server 127.0.0.1:21823 ok === (external:skip cluster) Starting server 127.0.0.1:21825 ok -[ok]: Sub-replica reports zero repl offset and rank, and fails to win election - shutdown (40372 ms) === (external:skip cluster) Starting server 127.0.0.1:21827 ok === (external:skip cluster) Starting server 127.0.0.1:21829 ok === (external:skip cluster) Starting server 127.0.0.1:21831 ok @@ -5206,16 +7419,25 @@ === (external:skip cluster) Starting server 127.0.0.1:21837 ok === (external:skip cluster) Starting server 127.0.0.1:21839 ok === (external:skip cluster) Starting server 127.0.0.1:21841 ok -[ok]: Sub-replica reports zero repl offset and rank, and fails to win election - sigstop (14246 ms) === (external:skip cluster) Starting server 127.0.0.1:21843 ok === (external:skip cluster) Starting server 127.0.0.1:21845 ok === (external:skip cluster) Starting server 127.0.0.1:21847 ok === (external:skip cluster) Starting server 127.0.0.1:21849 ok +[ok]: Cluster is up (104 ms) +[ok]: The first master has actually two slaves (15 ms) +[ok]: CLUSTER SLAVES and CLUSTER REPLICAS output is consistent (28 ms) +[ok]: Slaves of #0 are instance #5 and #10 as expected (31 ms) +[ok]: Instance #5 and #10 synced with the master (6 ms) +[ok]: Slaves are both able to receive and acknowledge writes (533 ms) +[ok]: Write data while slave #10 is paused and can't receive it (10347 ms) +[ok]: Wait for instance #5 (and not #10) to turn into a master (7195 ms) +[ok]: Wait for the node #10 to return alive before ending the test (36 ms) +[ok]: Cluster should eventually be up again (4582 ms) +[ok]: Node #10 should eventually replicate node #5 (4 ms) === (external:skip cluster) Starting server 127.0.0.1:21851 ok === (external:skip cluster) Starting server 127.0.0.1:21853 ok === (external:skip cluster) Starting server 127.0.0.1:21855 ok === (external:skip cluster) Starting server 127.0.0.1:21857 ok -[ok]: valkey-cli make source node ignores NOREPLICAS error when doing the last CLUSTER SETSLOT - gossip (2836 ms) === (external:skip cluster) Starting server 127.0.0.1:21859 ok === (external:skip cluster) Starting server 127.0.0.1:21861 ok === (external:skip cluster) Starting server 127.0.0.1:21863 ok @@ -5224,1967 +7446,194 @@ === (external:skip cluster) Starting server 127.0.0.1:21869 ok === (external:skip cluster) Starting server 127.0.0.1:21871 ok === (external:skip cluster) Starting server 127.0.0.1:21873 ok -[ok]: valkey-cli make source node ignores NOREPLICAS error when doing the last CLUSTER SETSLOT - setslot (5972 ms) -[69/124 done]: unit/cluster/replica-migration (309 seconds) -Testing unit/cluster/consistency-check === (external:skip cluster) Starting server 127.0.0.1:21875 ok === (external:skip cluster) Starting server 127.0.0.1:21877 ok === (external:skip cluster) Starting server 127.0.0.1:21879 ok === (external:skip cluster) Starting server 127.0.0.1:21881 ok === (external:skip cluster) Starting server 127.0.0.1:21883 ok === (external:skip cluster) Starting server 127.0.0.1:21885 ok +[ok]: Cluster is up (28 ms) +[ok]: The first master has actually 5 slaves (7 ms) +[ok]: Slaves of #0 are instance #3, #6, #9, #12 and #15 as expected (8 ms) +[ok]: Instance #3, #6, #9, #12 and #15 synced with the master (14 ms) +[ok]: New Master down consecutively (48844 ms) +[68/124 done]: unit/cluster/slave-selection (121 seconds) +Testing unit/cluster/failure-marking === (external:skip cluster) Starting server 127.0.0.1:21887 ok === (external:skip cluster) Starting server 127.0.0.1:21889 ok +[ok]: Verify that single primary marks replica as failed (3214 ms) === (external:skip cluster) Starting server 127.0.0.1:21891 ok === (external:skip cluster) Starting server 127.0.0.1:21893 ok -[ok]: Cluster should start ok (59 ms) -[ok]: Cluster is writable (2490 ms) -[ok]: Slave expired keys is loaded when restarted: appendonly=no (5105 ms) -[ok]: Slave expired keys is loaded when restarted: appendonly=yes (6269 ms) -[70/124 done]: unit/cluster/consistency-check (32 seconds) -Testing unit/cluster/cli -[ignore]: Not supported in tls mode -[71/124 done]: unit/cluster/cli (0 seconds) -Testing unit/cluster/base -[ignore]: Not supported in tls mode -[72/124 done]: unit/cluster/base (0 seconds) -Testing unit/cluster/info === (external:skip cluster) Starting server 127.0.0.1:21895 ok +[ok]: Verify that multiple primaries mark replica as failed (4181 ms) +[ignore]: Not supported in tls mode +[69/124 done]: unit/cluster/failure-marking (31 seconds) +Testing unit/cluster/faildet === (external:skip cluster) Starting server 127.0.0.1:21897 ok -[ok]: Cluster should start ok (15 ms) -[ok]: errorstats: rejected call due to MOVED Redirection (67 ms) -[73/124 done]: unit/cluster/info (4 seconds) -Testing unit/cluster/announced-endpoints === (external:skip cluster) Starting server 127.0.0.1:21899 ok === (external:skip cluster) Starting server 127.0.0.1:21901 ok === (external:skip cluster) Starting server 127.0.0.1:21903 ok === (external:skip cluster) Starting server 127.0.0.1:21905 ok -[ok]: Test change cluster-announce-port and cluster-announce-tls-port at runtime (396 ms) -[ok]: Test change cluster-announce-bus-port at runtime (211 ms) -[74/124 done]: unit/cluster/announced-endpoints (15 seconds) -Testing unit/cluster/slot-ownership +=== (external:skip cluster) Starting server 127.0.0.1:21907 ok === (external:skip cluster) Starting server 127.0.0.1:21909 ok === (external:skip cluster) Starting server 127.0.0.1:21911 ok === (external:skip cluster) Starting server 127.0.0.1:21913 ok === (external:skip cluster) Starting server 127.0.0.1:21915 ok -[ok]: Verify that slot ownership transfer through gossip propagates deletes to replicas (1000 ms) +[ok]: Cluster should start ok (18 ms) +[ok]: Killing two slave nodes (225 ms) +[ok]: Cluster should be still up (1573 ms) +[ok]: Killing one master node (118 ms) +[ok]: Cluster should be down now (2754 ms) +[ok]: Restarting master node (167 ms) +[ok]: Cluster should be up again (1425 ms) +Waiting for process 2274593 to exit... +Waiting for process 2274593 to exit... +Waiting for process 2274593 to exit... +Waiting for process 2274593 to exit... +Waiting for process 2274593 to exit... +Waiting for process 2274488 to exit... +Waiting for process 2274488 to exit... +Waiting for process 2274488 to exit... +Waiting for process 2274488 to exit... +Waiting for process 2274488 to exit... +[70/124 done]: unit/cluster/faildet (41 seconds) +Testing unit/cluster/slot-stats === (external:skip cluster) Starting server 127.0.0.1:21917 ok +[ok]: CLUSTER SLOT-STATS cpu-usec reset upon CONFIG RESETSTAT. (10601 ms) +[ok]: CLUSTER SLOT-STATS cpu-usec reset upon slot migration. (20624 ms) +[ok]: CLUSTER SLOT-STATS cpu-usec for non-slot specific commands. (11194 ms) +[ok]: CLUSTER SLOT-STATS cpu-usec for slot specific commands. (8520 ms) +[ok]: CLUSTER SLOT-STATS cpu-usec for blocking commands, unblocked on keyspace update. (11055 ms) +[ok]: CLUSTER SLOT-STATS cpu-usec for blocking commands, unblocked on timeout. (16263 ms) +[ok]: CLUSTER SLOT-STATS cpu-usec for transactions. (34354 ms) +[ok]: CLUSTER SLOT-STATS cpu-usec for lua-scripts, without cross-slot keys. (9282 ms) +[ok]: CLUSTER SLOT-STATS cpu-usec for lua-scripts, with cross-slot keys. (9381 ms) +[ok]: CLUSTER SLOT-STATS cpu-usec for functions, without cross-slot keys. (6116 ms) +[ok]: CLUSTER SLOT-STATS cpu-usec for functions, with cross-slot keys. (5833 ms) === (external:skip cluster) Starting server 127.0.0.1:21919 ok +[ok]: CLUSTER SLOT-STATS network-bytes-in, multi bulk buffer processing. (4501 ms) +[ok]: CLUSTER SLOT-STATS network-bytes-in, in-line buffer processing. (8081 ms) +[ok]: CLUSTER SLOT-STATS network-bytes-in, blocking command. (27944 ms) +[ok]: CLUSTER SLOT-STATS network-bytes-in, multi-exec transaction. (19586 ms) +[ok]: CLUSTER SLOT-STATS network-bytes-in, non slot specific command. (13617 ms) +[ok]: CLUSTER SLOT-STATS network-bytes-in, pub/sub. (9354 ms) === (external:skip cluster) Starting server 127.0.0.1:21921 ok === (external:skip cluster) Starting server 127.0.0.1:21923 ok -[ok]: Primary lost a slot during the shutdown waiting (5480 ms) +[ok]: CLUSTER SLOT-STATS network-bytes-in, sharded pub/sub. (5780 ms) +Waiting for process 2381090 to exit... +Waiting for process 2381090 to exit... +Waiting for process 2381090 to exit... +Waiting for process 2381090 to exit... +Waiting for process 2381090 to exit... +Waiting for process 2381090 to exit... +Waiting for process 2381090 to exit... +Waiting for process 2381090 to exit... === (external:skip cluster) Starting server 127.0.0.1:21925 ok +[ok]: CLUSTER SLOT-STATS network-bytes-out, for non-slot specific commands. (5041 ms) +[ok]: CLUSTER SLOT-STATS network-bytes-out, for slot specific commands. (3665 ms) +[ok]: CLUSTER SLOT-STATS network-bytes-out, blocking commands. (14740 ms) === (external:skip cluster) Starting server 127.0.0.1:21927 ok === (external:skip cluster) Starting server 127.0.0.1:21929 ok +[ok]: CLUSTER SLOT-STATS network-bytes-out, replication stream egress. (14861 ms) === (external:skip cluster) Starting server 127.0.0.1:21931 ok -[ok]: Primary lost a slot during the manual failover pausing (152 ms) === (external:skip cluster) Starting server 127.0.0.1:21933 ok +[ok]: CLUSTER SLOT-STATS network-bytes-out, sharded pub/sub, single channel. (27263 ms) +[ok]: CLUSTER SLOT-STATS network-bytes-out, sharded pub/sub, cross-slot channels. (18587 ms) === (external:skip cluster) Starting server 127.0.0.1:21935 ok +[ok]: CLUSTER SLOT-STATS contains default value upon valkey-server startup (11090 ms) +[ok]: CLUSTER SLOT-STATS contains correct metrics upon key introduction (9339 ms) +[ok]: CLUSTER SLOT-STATS contains correct metrics upon key mutation (8914 ms) +[ok]: CLUSTER SLOT-STATS contains correct metrics upon key deletion (11869 ms) +[ok]: CLUSTER SLOT-STATS slot visibility based on slot ownership changes (24674 ms) === (external:skip cluster) Starting server 127.0.0.1:21937 ok +[ok]: CLUSTER SLOT-STATS SLOTSRANGE all slots present (0 ms) +[ok]: CLUSTER SLOT-STATS SLOTSRANGE some slots missing (2 ms) === (external:skip cluster) Starting server 127.0.0.1:21939 ok -[ok]: Primary lost a slot during the client pause command (127 ms) -[75/124 done]: unit/cluster/slot-ownership (62 seconds) -Testing unit/cluster/manual-takeover +[ok]: CLUSTER SLOT-STATS ORDERBY DESC correct ordering (65674 ms) +[ok]: CLUSTER SLOT-STATS ORDERBY ASC correct ordering (55334 ms) +[ok]: CLUSTER SLOT-STATS ORDERBY LIMIT correct response pagination, where limit is less than number of assigned slots (64 ms) +[ok]: CLUSTER SLOT-STATS ORDERBY LIMIT correct response pagination, where limit is greater than number of assigned slots (33 ms) +[ok]: CLUSTER SLOT-STATS ORDERBY arg sanity check. (10 ms) === (external:skip cluster) Starting server 127.0.0.1:21941 ok === (external:skip cluster) Starting server 127.0.0.1:21943 ok +[ok]: CLUSTER SLOT-STATS metrics replication for new keys (28930 ms) +[ok]: CLUSTER SLOT-STATS metrics replication for existing keys (16223 ms) +[ok]: CLUSTER SLOT-STATS metrics replication for deleting keys (9498 ms) +[71/124 done]: unit/cluster/slot-stats (653 seconds) +Testing unit/cluster/transactions-on-replica === (external:skip cluster) Starting server 127.0.0.1:21945 ok === (external:skip cluster) Starting server 127.0.0.1:21947 ok -=== (external:skip cluster) Starting server 127.0.0.1:21949 ok -=== (external:skip cluster) Starting server 127.0.0.1:21951 ok -=== (external:skip cluster) Starting server 127.0.0.1:21953 ok -=== (external:skip cluster) Starting server 127.0.0.1:21955 ok -=== (external:skip cluster) Starting server 127.0.0.1:21957 ok -=== (external:skip cluster) Starting server 127.0.0.1:21959 ok -[ok]: Cluster is up (62 ms) -[ok]: Cluster is writable (1719 ms) -[ok]: Killing majority of master nodes (794 ms) -[ok]: Cluster should eventually be down (4418 ms) -[ok]: Use takeover to bring slaves back (127 ms) -[ok]: Cluster should eventually be up again (4232 ms) -[ok]: Cluster is writable (2201 ms) -[ok]: Instance #5, #6, #7 are now masters (36 ms) -[ok]: Restarting the previously killed master nodes (47 ms) -[ok]: Instance #0, #1, #2 gets converted into a slaves (129 ms) -[76/124 done]: unit/cluster/manual-takeover (31 seconds) -Testing unit/cluster/transactions-on-replica -=== (external:skip cluster) Starting server 127.0.0.1:21961 ok -=== (external:skip cluster) Starting server 127.0.0.1:21963 ok -[ok]: Cluster should start ok (11 ms) -[ok]: Can't read from replica without READONLY (31 ms) -[ok]: Can't read from replica after READWRITE (10 ms) -[ok]: Can read from replica after READONLY (11 ms) -[ok]: Can perform HSET primary and HGET from replica (64 ms) -[ok]: Can MULTI-EXEC transaction of HGET operations from replica (34 ms) -[ok]: MULTI-EXEC with write operations is MOVED (24 ms) -[ok]: write command is QUEUED, then EXEC should be MOVED after failover (2144 ms) -[ok]: read-only blocking operations from replica (64 ms) -[ok]: reply MOVED when eval from replica for update (1 ms) -[77/124 done]: unit/cluster/transactions-on-replica (15 seconds) -Testing unit/cluster/many-slot-migration -=== (external:skip cluster) Starting server 127.0.0.1:21965 ok -=== (external:skip cluster) Starting server 127.0.0.1:21967 ok -=== (external:skip cluster) Starting server 127.0.0.1:21969 ok -=== (external:skip cluster) Starting server 127.0.0.1:21971 ok -=== (external:skip cluster) Starting server 127.0.0.1:21973 ok -=== (external:skip cluster) Starting server 127.0.0.1:21975 ok -=== (external:skip cluster) Starting server 127.0.0.1:21977 ok -=== (external:skip cluster) Starting server 127.0.0.1:21979 ok -=== (external:skip cluster) Starting server 127.0.0.1:21981 ok -=== (external:skip cluster) Starting server 127.0.0.1:21983 ok -[ok]: Cluster is up (75 ms) -[ok]: Set many keys (260652 ms) -[ok]: Keys are accessible (267174 ms) -[ok]: Init migration of many slots (15559 ms) -[ok]: Fix cluster (289619 ms) -[ok]: Keys are accessible (264440 ms) -[78/124 done]: unit/cluster/many-slot-migration (1109 seconds) -Testing unit/cluster/failure-marking -=== (external:skip cluster) Starting server 127.0.0.1:21985 ok -=== (external:skip cluster) Starting server 127.0.0.1:21987 ok -[ok]: Verify that single primary marks replica as failed (3096 ms) -=== (external:skip cluster) Starting server 127.0.0.1:21989 ok -=== (external:skip cluster) Starting server 127.0.0.1:21991 ok -=== (external:skip cluster) Starting server 127.0.0.1:21993 ok -[ok]: Verify that multiple primaries mark replica as failed (3298 ms) -[ignore]: Not supported in tls mode -[79/124 done]: unit/cluster/failure-marking (30 seconds) -Testing unit/cluster/hostnames -=== (external:skip cluster) Starting server 127.0.0.1:21995 ok -=== (external:skip cluster) Starting server 127.0.0.1:21997 ok -=== (external:skip cluster) Starting server 127.0.0.1:21999 ok -=== (external:skip cluster) Starting server 127.0.0.1:22001 ok -=== (external:skip cluster) Starting server 127.0.0.1:22003 ok -=== (external:skip cluster) Starting server 127.0.0.1:22005 ok -=== (external:skip cluster) Starting server 127.0.0.1:22007 ok -[ok]: Set cluster hostnames and verify they are propagated (223 ms) -[ok]: Update hostnames and make sure they are all eventually propagated (272 ms) -[ok]: Remove hostnames and make sure they are all eventually propagated (354 ms) -[ok]: Verify cluster-preferred-endpoint-type behavior for redirects and info (266 ms) -[ok]: Verify the nodes configured with prefer hostname only show hostname for new nodes (1223 ms) -[ok]: Test restart will keep hostname information (628 ms) -[ok]: Test hostname validation (18 ms) -[80/124 done]: unit/cluster/hostnames (17 seconds) -Testing unit/cluster/pubsubshard-slot-migration -=== (external:skip cluster) Starting server 127.0.0.1:22009 ok -=== (external:skip cluster) Starting server 127.0.0.1:22011 ok -=== (external:skip cluster) Starting server 127.0.0.1:22013 ok -=== (external:skip cluster) Starting server 127.0.0.1:22015 ok -=== (external:skip cluster) Starting server 127.0.0.1:22017 ok -=== (external:skip cluster) Starting server 127.0.0.1:22019 ok -[ok]: Cluster is up (26 ms) -[ok]: Migrate a slot, verify client receives sunsubscribe on primary serving the slot. (156 ms) -[ok]: Client subscribes to multiple channels, migrate a slot, verify client receives sunsubscribe on primary serving the slot. (163 ms) -[ok]: Migrate a slot, verify client receives sunsubscribe on replica serving the slot. (327 ms) -[ok]: Move a replica to another primary, verify client receives sunsubscribe on replica serving the slot. (283 ms) -[ok]: Delete a slot, verify sunsubscribe message (39 ms) -[ok]: Reset cluster, verify sunsubscribe message (151 ms) -[81/124 done]: unit/cluster/pubsubshard-slot-migration (15 seconds) -Testing unit/cluster/cluster-nodes-slots -=== (external:skip cluster) Starting server 127.0.0.1:22021 ok -=== (external:skip cluster) Starting server 127.0.0.1:22023 ok -=== (external:skip cluster) Starting server 127.0.0.1:22025 ok -=== (external:skip cluster) Starting server 127.0.0.1:22027 ok -[ok]: Cluster should start ok (32 ms) -[ok]: Continuous slots distribution (79 ms) -[ok]: Discontinuous slots distribution (107 ms) -[82/124 done]: unit/cluster/cluster-nodes-slots (13 seconds) -Testing unit/cluster/cluster-reliable-meet -[ignore]: Not supported in tls mode -[83/124 done]: unit/cluster/cluster-reliable-meet (0 seconds) -Testing unit/cluster/links -=== (external:skip cluster) Starting server 127.0.0.1:22029 ok -=== (external:skip cluster) Starting server 127.0.0.1:22031 ok -=== (external:skip cluster) Starting server 127.0.0.1:22033 ok -[ok]: Broadcast message across a cluster shard while a cluster link is down (244 ms) -=== (external:skip cluster) Starting server 127.0.0.1:22035 ok -=== (external:skip cluster) Starting server 127.0.0.1:22037 ok -=== (external:skip cluster) Starting server 127.0.0.1:22039 ok -[ok]: Each node has two links with each peer (72 ms) -[ok]: Validate cluster links format (7 ms) -[ok]: Disconnect link when send buffer limit reached (9003 ms) -[ok]: Link memory increases with publishes (836 ms) -[ok]: Link memory resets after publish messages flush (646 ms) -[84/124 done]: unit/cluster/links (27 seconds) -Testing unit/cluster/cluster-shards -=== (external:skip cluster) Starting server 127.0.0.1:22041 ok -=== (external:skip cluster) Starting server 127.0.0.1:22043 ok -=== (external:skip cluster) Starting server 127.0.0.1:22045 ok -=== (external:skip cluster) Starting server 127.0.0.1:22047 ok -=== (external:skip cluster) Starting server 127.0.0.1:22049 ok -=== (external:skip cluster) Starting server 127.0.0.1:22051 ok -[ok]: Cluster should start ok (47 ms) -[ok]: Cluster shards response is ok for shard 0 (12 ms) -[ok]: Kill a node and tell the replica to immediately takeover (190 ms) -[ok]: Verify health as fail for killed node (3113 ms) -[ok]: CLUSTER SHARDS slot response is non-empty when primary node fails (8 ms) -[85/124 done]: unit/cluster/cluster-shards (17 seconds) -Testing unit/cluster/cluster-slots -=== (external:skip cluster) Starting server 127.0.0.1:22053 ok -=== (external:skip cluster) Starting server 127.0.0.1:22055 ok -=== (external:skip cluster) Starting server 127.0.0.1:22057 ok -=== (external:skip cluster) Starting server 127.0.0.1:22059 ok -=== (external:skip cluster) Starting server 127.0.0.1:22061 ok -=== (external:skip cluster) Starting server 127.0.0.1:22063 ok -=== (external:skip cluster) Starting server 127.0.0.1:22065 ok -=== (external:skip cluster) Starting server 127.0.0.1:22067 ok -=== (external:skip cluster) Starting server 127.0.0.1:22069 ok -=== (external:skip cluster) Starting server 127.0.0.1:22071 ok -=== (external:skip cluster) Starting server 127.0.0.1:22073 ok -=== (external:skip cluster) Starting server 127.0.0.1:22075 ok -=== (external:skip cluster) Starting server 127.0.0.1:22077 ok -=== (external:skip cluster) Starting server 127.0.0.1:22079 ok -=== (external:skip cluster) Starting server 127.0.0.1:22081 ok -[ok]: Cluster is up (107 ms) -[ok]: Cluster is writable (1680 ms) -[ok]: Instance #5 is a slave (7 ms) -[ok]: client do not break when cluster slot (7922 ms) -[ok]: client can handle keys with hash tag (461 ms) -[ok]: slot migration is valid from primary to another primary (473 ms) -[ok]: slot migration is invalid from primary to replica (452 ms) -[ok]: slot must be unbound on the owner when it is deleted (33906 ms) -[ok]: CLUSTER SLOTS from non-TLS client in TLS cluster (14562 ms) -[86/124 done]: unit/cluster/cluster-slots (354 seconds) -Testing unit/cluster/pubsubshard -=== (external:skip cluster) Starting server 127.0.0.1:22083 ok -=== (external:skip cluster) Starting server 127.0.0.1:22085 ok -=== (external:skip cluster) Starting server 127.0.0.1:22087 ok -=== (external:skip cluster) Starting server 127.0.0.1:22089 ok -=== (external:skip cluster) Starting server 127.0.0.1:22091 ok -=== (external:skip cluster) Starting server 127.0.0.1:22093 ok -[ok]: Pub/Sub shard basics (175 ms) -[ok]: client can't subscribe to multiple shard channels across different slots in same call (32 ms) -[ok]: client can subscribe to multiple shard channels across different slots in separate call (31 ms) -[ok]: sunsubscribe without specifying any channel would unsubscribe all shard channels subscribed (76 ms) -[ok]: Verify Pub/Sub and Pub/Sub shard no overlap (137 ms) -[ok]: PUBSUB channels/shardchannels (141 ms) -[87/124 done]: unit/cluster/pubsubshard (14 seconds) -Testing unit/cluster/scripting -=== (external:skip cluster) Starting server 127.0.0.1:22095 ok -[ok]: Eval scripts with shebangs and functions default to no cross slots (2 ms) -[ok]: Cross slot commands are allowed by default for eval scripts and with allow-cross-slot-keys flag (2 ms) -[ok]: Cross slot commands are also blocked if they disagree with pre-declared keys (0 ms) -[ok]: Cross slot commands are allowed by default if they disagree with pre-declared keys (1 ms) -[ok]: Function no-cluster flag (3 ms) -[ok]: Script no-cluster flag (0 ms) -[ok]: Calling cluster slots in scripts is OK (4 ms) -[ok]: Calling cluster shards in scripts is OK (2 ms) -[88/124 done]: unit/cluster/scripting (3 seconds) -Testing unit/cluster/faildet -=== (external:skip cluster) Starting server 127.0.0.1:22097 ok -=== (external:skip cluster) Starting server 127.0.0.1:22099 ok -=== (external:skip cluster) Starting server 127.0.0.1:22101 ok -=== (external:skip cluster) Starting server 127.0.0.1:22103 ok -=== (external:skip cluster) Starting server 127.0.0.1:22105 ok -=== (external:skip cluster) Starting server 127.0.0.1:22107 ok -=== (external:skip cluster) Starting server 127.0.0.1:22109 ok -=== (external:skip cluster) Starting server 127.0.0.1:22111 ok -=== (external:skip cluster) Starting server 127.0.0.1:22113 ok -=== (external:skip cluster) Starting server 127.0.0.1:22115 ok -[ok]: Cluster should start ok (76 ms) -[ok]: Killing two slave nodes (304 ms) -[ok]: Cluster should be still up (1324 ms) -[ok]: Killing one master node (176 ms) -[ok]: Cluster should be down now (1516 ms) -[ok]: Restarting master node (157 ms) -[ok]: Cluster should be up again (1258 ms) -Waiting for process 2603080 to exit... -Waiting for process 2603080 to exit... -Waiting for process 2603080 to exit... -Waiting for process 2603039 to exit... -Waiting for process 2603039 to exit... -Waiting for process 2603039 to exit... -[89/124 done]: unit/cluster/faildet (40 seconds) -Testing unit/cluster/update-msg -=== (external:skip cluster) Starting server 127.0.0.1:22117 ok -=== (external:skip cluster) Starting server 127.0.0.1:22119 ok -=== (external:skip cluster) Starting server 127.0.0.1:22121 ok -=== (external:skip cluster) Starting server 127.0.0.1:22123 ok -=== (external:skip cluster) Starting server 127.0.0.1:22125 ok -=== (external:skip cluster) Starting server 127.0.0.1:22127 ok -=== (external:skip cluster) Starting server 127.0.0.1:22129 ok -=== (external:skip cluster) Starting server 127.0.0.1:22131 ok -=== (external:skip cluster) Starting server 127.0.0.1:22133 ok -=== (external:skip cluster) Starting server 127.0.0.1:22135 ok -[ok]: Cluster is up (22 ms) -[ok]: Cluster is writable (1114 ms) -[ok]: Instance #5 is a slave (4 ms) -[ok]: Instance #5 synced with the master (1 ms) -[ok]: Killing one master node (146 ms) -[ok]: Wait for failover (3746 ms) -[ok]: Cluster should eventually be up again (1419 ms) -[ok]: Cluster is writable (1555 ms) -[ok]: Instance #5 is now a master (7 ms) -[ok]: Killing the new master #5 (151 ms) -[ok]: Cluster should be down now (4256 ms) -[ok]: Restarting the old master node (18 ms) -[ok]: Instance #0 gets converted into a slave (335 ms) -[ok]: Restarting the new master node (15 ms) -[ok]: Cluster is up again (4669 ms) -[90/124 done]: unit/cluster/update-msg (33 seconds) -Testing unit/cluster/slave-selection -=== (external:skip cluster) Starting server 127.0.0.1:22137 ok -=== (external:skip cluster) Starting server 127.0.0.1:22139 ok -=== (external:skip cluster) Starting server 127.0.0.1:22141 ok -=== (external:skip cluster) Starting server 127.0.0.1:22143 ok -=== (external:skip cluster) Starting server 127.0.0.1:22145 ok -=== (external:skip cluster) Starting server 127.0.0.1:22147 ok -=== (external:skip cluster) Starting server 127.0.0.1:22149 ok -=== (external:skip cluster) Starting server 127.0.0.1:22151 ok -=== (external:skip cluster) Starting server 127.0.0.1:22153 ok -=== (external:skip cluster) Starting server 127.0.0.1:22155 ok -=== (external:skip cluster) Starting server 127.0.0.1:22157 ok -=== (external:skip cluster) Starting server 127.0.0.1:22159 ok -=== (external:skip cluster) Starting server 127.0.0.1:22161 ok -=== (external:skip cluster) Starting server 127.0.0.1:22163 ok -=== (external:skip cluster) Starting server 127.0.0.1:22165 ok -[ok]: Cluster is up (102 ms) -[ok]: The first master has actually two slaves (19 ms) -[ok]: CLUSTER SLAVES and CLUSTER REPLICAS output is consistent (35 ms) -[ok]: Slaves of #0 are instance #5 and #10 as expected (7 ms) -[ok]: Instance #5 and #10 synced with the master (16 ms) -[ok]: Slaves are both able to receive and acknowledge writes (774 ms) -[ok]: Write data while slave #10 is paused and can't receive it (10160 ms) -[ok]: Wait for instance #5 (and not #10) to turn into a master (5758 ms) -[ok]: Wait for the node #10 to return alive before ending the test (792 ms) -[ok]: Cluster should eventually be up again (2776 ms) -[ok]: Node #10 should eventually replicate node #5 (15 ms) -=== (external:skip cluster) Starting server 127.0.0.1:22167 ok -=== (external:skip cluster) Starting server 127.0.0.1:22169 ok -=== (external:skip cluster) Starting server 127.0.0.1:22171 ok -=== (external:skip cluster) Starting server 127.0.0.1:22173 ok -=== (external:skip cluster) Starting server 127.0.0.1:22175 ok -=== (external:skip cluster) Starting server 127.0.0.1:22177 ok -=== (external:skip cluster) Starting server 127.0.0.1:22179 ok -=== (external:skip cluster) Starting server 127.0.0.1:22181 ok -=== (external:skip cluster) Starting server 127.0.0.1:22183 ok -=== (external:skip cluster) Starting server 127.0.0.1:22185 ok -=== (external:skip cluster) Starting server 127.0.0.1:22187 ok -=== (external:skip cluster) Starting server 127.0.0.1:22189 ok -=== (external:skip cluster) Starting server 127.0.0.1:22191 ok -=== (external:skip cluster) Starting server 127.0.0.1:22193 ok -=== (external:skip cluster) Starting server 127.0.0.1:22195 ok -=== (external:skip cluster) Starting server 127.0.0.1:22197 ok -=== (external:skip cluster) Starting server 127.0.0.1:22199 ok -=== (external:skip cluster) Starting server 127.0.0.1:22201 ok -[ok]: Cluster is up (111 ms) -[ok]: The first master has actually 5 slaves (7 ms) -[ok]: Slaves of #0 are instance #3, #6, #9, #12 and #15 as expected (35 ms) -[ok]: Instance #3, #6, #9, #12 and #15 synced with the master (32 ms) -[ok]: New Master down consecutively (33908 ms) -[91/124 done]: unit/cluster/slave-selection (94 seconds) -Testing unit/cluster/pubsub -=== (external:skip cluster) Starting server 127.0.0.1:22203 ok -=== (external:skip cluster) Starting server 127.0.0.1:22205 ok -=== (external:skip cluster) Starting server 127.0.0.1:22207 ok -=== (external:skip cluster) Starting server 127.0.0.1:22209 ok -=== (external:skip cluster) Starting server 127.0.0.1:22211 ok -=== (external:skip cluster) Starting server 127.0.0.1:22213 ok -=== (external:skip cluster) Starting server 127.0.0.1:22215 ok -=== (external:skip cluster) Starting server 127.0.0.1:22217 ok -=== (external:skip cluster) Starting server 127.0.0.1:22219 ok -=== (external:skip cluster) Starting server 127.0.0.1:22221 ok -[ok]: Test publishing to master (104 ms) -[ok]: Test publishing to slave (82 ms) -[92/124 done]: unit/cluster/pubsub (16 seconds) -Testing unit/cluster/announce-client-ip -=== (external:skip cluster) Starting server 127.0.0.1:22223 ok -=== (external:skip cluster) Starting server 127.0.0.1:22225 ok -=== (external:skip cluster) Starting server 127.0.0.1:22227 ok -=== (external:skip cluster) Starting server 127.0.0.1:22229 ok -[ok]: Set cluster announced IPv4 to invalid IP (0 ms) -[ok]: Set cluster announced IPv4 and check that it propagates (355 ms) -[ok]: Clear announced client IPv4 and check that it propagates (321 ms) -=== (external:skip cluster ipv6) Starting server 127.0.0.1:22231 ok -=== (external:skip cluster ipv6) Starting server 127.0.0.1:22233 ok -=== (external:skip cluster ipv6) Starting server 127.0.0.1:22235 ok -=== (external:skip cluster ipv6) Starting server 127.0.0.1:22237 ok -[ok]: Set cluster announced IPv6 to invalid IP (4 ms) -[ok]: Set cluster announced IPv6 and check that it propagates (393 ms) -[ok]: Clear announced client IPv6 and check that it propagates (244 ms) -=== (external:skip cluster) Starting server 127.0.0.1:22239 ok -[ok]: Load cluster announced IPv4 config on server start (308 ms) -=== (external:skip cluster ipv6) Starting server 127.0.0.1:22241 ok -[ok]: Load cluster announced IPv6 config on server start (405 ms) -[93/124 done]: unit/cluster/announce-client-ip (32 seconds) -Testing unit/cluster/slot-migration -=== (external:skip cluster) Starting server 127.0.0.1:22243 ok -=== (external:skip cluster) Starting server 127.0.0.1:22245 ok -=== (external:skip cluster) Starting server 127.0.0.1:22247 ok -=== (external:skip cluster) Starting server 127.0.0.1:22249 ok -=== (external:skip cluster) Starting server 127.0.0.1:22251 ok -=== (external:skip cluster) Starting server 127.0.0.1:22253 ok -[ok]: Slot migration states are replicated (358 ms) -[ok]: Migration target is auto-updated after failover in target shard (3916 ms) -[ok]: Migration source is auto-updated after failover in source shard (3812 ms) -[ok]: Replica redirects key access in migrating slots (27 ms) -[ok]: Replica of migrating node returns ASK redirect after READONLY (48 ms) -[ok]: Replica of migrating node returns TRYAGAIN after READONLY (47 ms) -[ok]: Replica of importing node returns TRYAGAIN after READONLY and ASKING (26 ms) -[ok]: New replica inherits migrating slot (865 ms) -[ok]: New replica inherits importing slot (600 ms) -=== (external:skip cluster) Starting server 127.0.0.1:22255 ok -=== (external:skip cluster) Starting server 127.0.0.1:22257 ok -=== (external:skip cluster) Starting server 127.0.0.1:22259 ok -=== (external:skip cluster) Starting server 127.0.0.1:22261 ok -=== (external:skip cluster) Starting server 127.0.0.1:22263 ok -=== (external:skip cluster) Starting server 127.0.0.1:22265 ok -[ok]: Multiple slot migration states are replicated (192 ms) -[ok]: New replica inherits multiple migrating slots (783 ms) -[ok]: Slot finalization succeeds on both primary and replicas (111 ms) -=== (external:skip cluster) Starting server 127.0.0.1:22267 ok -=== (external:skip cluster) Starting server 127.0.0.1:22269 ok -=== (external:skip cluster) Starting server 127.0.0.1:22271 ok -=== (external:skip cluster) Starting server 127.0.0.1:22273 ok -=== (external:skip cluster) Starting server 127.0.0.1:22275 ok -=== (external:skip cluster) Starting server 127.0.0.1:22277 ok -[ok]: Slot is auto-claimed by target after source relinquishes ownership (1041 ms) -=== (external:skip cluster) Starting server 127.0.0.1:22279 ok -=== (external:skip cluster) Starting server 127.0.0.1:22281 ok -=== (external:skip cluster) Starting server 127.0.0.1:22283 ok -=== (external:skip cluster) Starting server 127.0.0.1:22285 ok -=== (external:skip cluster) Starting server 127.0.0.1:22287 ok -=== (external:skip cluster) Starting server 127.0.0.1:22289 ok -[ok]: CLUSTER SETSLOT with invalid timeouts (40 ms) -=== (external:skip cluster) Starting server 127.0.0.1:22291 ok -=== (external:skip cluster) Starting server 127.0.0.1:22293 ok -=== (external:skip cluster) Starting server 127.0.0.1:22295 ok -=== (external:skip cluster) Starting server 127.0.0.1:22297 ok -=== (external:skip cluster) Starting server 127.0.0.1:22299 ok -=== (external:skip cluster) Starting server 127.0.0.1:22301 ok -[ok]: CLUSTER SETSLOT with an explicit timeout (3188 ms) -[ignore]: Not supported in tls mode -=== (external:skip cluster) Starting server 127.0.0.1:22303 ok -=== (external:skip cluster) Starting server 127.0.0.1:22305 ok -=== (external:skip cluster) Starting server 127.0.0.1:22307 ok -=== (external:skip cluster) Starting server 127.0.0.1:22309 ok -=== (external:skip cluster) Starting server 127.0.0.1:22311 ok -=== (external:skip cluster) Starting server 127.0.0.1:22313 ok -=== (external:skip cluster) Starting server 127.0.0.1:22315 ok -=== (external:skip cluster) Starting server 127.0.0.1:22317 ok -=== (external:skip cluster) Starting server 127.0.0.1:22319 ok -[ok]: Slot migration is ok when the replicas are down (393 ms) -[94/124 done]: unit/cluster/slot-migration (108 seconds) -Testing unit/cluster/manual-failover -=== (external:skip cluster) Starting server 127.0.0.1:22321 ok -=== (external:skip cluster) Starting server 127.0.0.1:22323 ok -=== (external:skip cluster) Starting server 127.0.0.1:22325 ok -=== (external:skip cluster) Starting server 127.0.0.1:22327 ok -=== (external:skip cluster) Starting server 127.0.0.1:22329 ok -=== (external:skip cluster) Starting server 127.0.0.1:22331 ok -=== (external:skip cluster) Starting server 127.0.0.1:22333 ok -=== (external:skip cluster) Starting server 127.0.0.1:22335 ok -=== (external:skip cluster) Starting server 127.0.0.1:22337 ok -=== (external:skip cluster) Starting server 127.0.0.1:22339 ok -[ok]: Cluster is up (55 ms) -[ok]: Cluster is writable (1352 ms) -[ok]: Instance #5 is a slave (8 ms) -[ok]: Instance #5 synced with the master (8 ms) -WWWWWWWWWW[ok]: Send CLUSTER FAILOVER to #5, during load (53002 ms) -[ok]: Wait for failover (64 ms) -[ok]: Cluster should eventually be up again (13 ms) -[ok]: Cluster is writable (1236 ms) -[ok]: Instance #5 is now a master (7 ms) -[ok]: Verify 50000 keys for consistency with logical content (29745 ms) -[ok]: Instance #0 gets converted into a slave (115 ms) -=== (external:skip cluster) Starting server 127.0.0.1:22341 ok -=== (external:skip cluster) Starting server 127.0.0.1:22343 ok -=== (external:skip cluster) Starting server 127.0.0.1:22345 ok -=== (external:skip cluster) Starting server 127.0.0.1:22347 ok -=== (external:skip cluster) Starting server 127.0.0.1:22349 ok -=== (external:skip cluster) Starting server 127.0.0.1:22351 ok -=== (external:skip cluster) Starting server 127.0.0.1:22353 ok -=== (external:skip cluster) Starting server 127.0.0.1:22355 ok -=== (external:skip cluster) Starting server 127.0.0.1:22357 ok -=== (external:skip cluster) Starting server 127.0.0.1:22359 ok -[ok]: Cluster is up (23 ms) -[ok]: Cluster is writable (1510 ms) -[ok]: Instance #5 is a slave (5 ms) -[ok]: Instance #5 synced with the master (8 ms) -[ok]: Make instance #0 unreachable without killing it (0 ms) -[ok]: Send CLUSTER FAILOVER to instance #5 (12 ms) -[ok]: Instance #5 is still a slave after some time (no failover) (5012 ms) -[ok]: Wait for instance #0 to return back alive (4970 ms) -=== (external:skip cluster) Starting server 127.0.0.1:22361 ok -=== (external:skip cluster) Starting server 127.0.0.1:22363 ok -=== (external:skip cluster) Starting server 127.0.0.1:22365 ok -=== (external:skip cluster) Starting server 127.0.0.1:22367 ok -=== (external:skip cluster) Starting server 127.0.0.1:22369 ok -=== (external:skip cluster) Starting server 127.0.0.1:22371 ok -=== (external:skip cluster) Starting server 127.0.0.1:22373 ok -=== (external:skip cluster) Starting server 127.0.0.1:22375 ok -=== (external:skip cluster) Starting server 127.0.0.1:22377 ok -=== (external:skip cluster) Starting server 127.0.0.1:22379 ok -=== (external:skip cluster) Starting server 127.0.0.1:22381 ok -=== (external:skip cluster) Starting server 127.0.0.1:22383 ok -=== (external:skip cluster) Starting server 127.0.0.1:22385 ok -=== (external:skip cluster) Starting server 127.0.0.1:22387 ok -=== (external:skip cluster) Starting server 127.0.0.1:22389 ok -[ok]: Cluster is up (95 ms) -[ok]: Cluster is writable (1686 ms) -[ok]: Instance #5 is a slave (19 ms) -[ok]: Instance #5 synced with the master (4 ms) -[ok]: Make instance #0 unreachable without killing it (1 ms) -[ok]: Send CLUSTER FAILOVER to instance #5 (15 ms) -[ok]: Instance #5 is a master after some time (117 ms) -[ok]: Wait for instance #0 to return back alive (9797 ms) -[95/124 done]: unit/cluster/manual-failover (165 seconds) -Testing unit/cluster/multi-slot-operations -=== (external:skip cluster) Starting server 127.0.0.1:22391 ok -=== (external:skip cluster) Starting server 127.0.0.1:22393 ok -=== (external:skip cluster) Starting server 127.0.0.1:22395 ok -=== (external:skip cluster) Starting server 127.0.0.1:22397 ok -=== (external:skip cluster) Starting server 127.0.0.1:22399 ok -[ok]: Continuous slots distribution (308 ms) -[ok]: ADDSLOTS command with several boundary conditions test suite (40 ms) -[ok]: ADDSLOTSRANGE command with several boundary conditions test suite (51 ms) -[ok]: DELSLOTSRANGE command with several boundary conditions test suite (116 ms) -[96/124 done]: unit/cluster/multi-slot-operations (5 seconds) -Testing integration/corrupt-dump-fuzzer -=== () Starting server 127.0.0.1:22401 ok -Done 18 cycles in 10 seconds. -RESTORE: successful: 11, rejected: 7 -Total commands sent in traffic: 1938, crashes during traffic: 2 (0 by signal). -[ok]: Fuzzer corrupt restore payloads - sanitize_dump: no (9616 ms) -=== () Starting server 127.0.0.1:22403 ok -Done 15 cycles in 10 seconds. -RESTORE: successful: 10, rejected: 5 -Total commands sent in traffic: 3560, crashes during traffic: 0 (0 by signal). -[ok]: Fuzzer corrupt restore payloads - sanitize_dump: yes (9936 ms) -[97/124 done]: integration/corrupt-dump-fuzzer (20 seconds) -Testing integration/valkey-cli -=== (cli) Starting server 127.0.0.1:22405 ok -[ok]: Interactive CLI: INFO response should be printed raw (129 ms) -[ok]: Interactive CLI: Status reply (111 ms) -[ok]: Interactive CLI: Integer reply (128 ms) -[ok]: Interactive CLI: Bulk reply (124 ms) -[ok]: Interactive CLI: Multi-bulk reply (128 ms) -[ok]: Interactive CLI: Parsing quotes (828 ms) -[ok]: Interactive CLI: Subscribed mode (1093 ms) -[ok]: Interactive non-TTY CLI: Subscribed mode (775 ms) -[ok]: Non-interactive TTY CLI: Status reply (87 ms) -[ok]: Non-interactive TTY CLI: Integer reply (92 ms) -[ok]: Non-interactive TTY CLI: Bulk reply (98 ms) -[ok]: Non-interactive TTY CLI: Multi-bulk reply (104 ms) -[ok]: Non-interactive TTY CLI: Read last argument from pipe (249 ms) -[ok]: Non-interactive TTY CLI: Read last argument from file (166 ms) -[ok]: Non-interactive TTY CLI: Escape character in JSON mode (867 ms) -[ok]: Non-interactive non-TTY CLI: Status reply (88 ms) -[ok]: Non-interactive non-TTY CLI: Integer reply (96 ms) -[ok]: Non-interactive non-TTY CLI: Bulk reply (93 ms) -[ok]: Non-interactive non-TTY CLI: Multi-bulk reply (101 ms) -[ok]: Non-interactive non-TTY CLI: Quoted input arguments (100 ms) -[ok]: Non-interactive non-TTY CLI: No accidental unquoting of input arguments (208 ms) -[ok]: Non-interactive non-TTY CLI: Invalid quoted input arguments (177 ms) -[ok]: Non-interactive non-TTY CLI: Read last argument from pipe (222 ms) -[ok]: Non-interactive non-TTY CLI: Read last argument from file (180 ms) -[ok]: Non-interactive non-TTY CLI: Test command-line hinting - latest server (107 ms) -[ok]: Non-interactive non-TTY CLI: Test command-line hinting - no server (28 ms) -[ok]: Non-interactive non-TTY CLI: Test command-line hinting - old server (191 ms) -[ok]: Dumping an RDB - functions only: no (6373 ms) -[ok]: Dumping an RDB - functions only: yes (1491 ms) -[ok]: Scan mode (441 ms) -[ok]: Connecting as a replica (2422 ms) -[ok]: Piping raw protocol (3839 ms) -[ok]: Options -X with illegal argument (144 ms) -[ok]: DUMP RESTORE with -x option (496 ms) -[ok]: DUMP RESTORE with -X option (543 ms) -[ok]: Valid Connection Scheme: redis:// (79 ms) -[ok]: Valid Connection Scheme: valkey:// (198 ms) -[ok]: Valid Connection Scheme: rediss:// (87 ms) -[ok]: Valid Connection Scheme: valkeys:// (75 ms) -[98/124 done]: integration/valkey-cli (23 seconds) -Testing integration/shutdown -=== () Starting server 127.0.0.1:22407 ok -=== () Starting server 127.0.0.1:22409 ok -[ok]: Shutting down master waits for replica to catch up (sigterm) (4395 ms) -=== () Starting server 127.0.0.1:22411 ok -=== () Starting server 127.0.0.1:22413 ok -[ok]: Shutting down master waits for replica to catch up (shutdown) (4812 ms) -=== () Starting server 127.0.0.1:22415 ok -=== () Starting server 127.0.0.1:22417 ok -[ok]: Shutting down master waits for replica timeout (4300 ms) -=== () Starting server 127.0.0.1:22419 ok -=== () Starting server 127.0.0.1:22421 ok -[ok]: Shutting down master waits for replica then fails (2052 ms) -=== () Starting server 127.0.0.1:22423 ok -=== () Starting server 127.0.0.1:22425 ok -[ok]: Shutting down master waits for replica then aborted (1443 ms) -[99/124 done]: integration/shutdown (17 seconds) -Testing integration/replication-3 -=== (repl external:skip) Starting server 127.0.0.1:22427 ok -=== () Starting server 127.0.0.1:22429 ok -[ok]: First server should have role slave after SLAVEOF (127 ms) -[ok]: MASTER and SLAVE consistency with expire (11956 ms) -[ok]: Master can replicate command longer than client-query-buffer-limit on replica (207 ms) -[ok]: Slave is able to evict keys created in writable slaves (6009 ms) -[ok]: Writable replica doesn't return expired keys (137 ms) -[ok]: PFCOUNT updates cache on readonly replica (84 ms) -[ok]: PFCOUNT doesn't use expired key on readonly replica (60 ms) -[100/124 done]: integration/replication-3 (20 seconds) -Testing integration/failover -=== (failover external:skip) Starting server 127.0.0.1:22431 ok -=== () Starting server 127.0.0.1:22433 ok -=== () Starting server 127.0.0.1:22435 ok -[ok]: failover command fails without connected replica (11 ms) -[ok]: setup replication for following tests (263 ms) -[ok]: failover command fails with invalid host (3 ms) -[ok]: failover command fails with invalid port (6 ms) -[ok]: failover command fails with just force and timeout (8 ms) -[ok]: failover command fails when sent to a replica (12 ms) -[ok]: failover command fails with force without timeout (4 ms) -[ok]: failover command to specific replica works (959 ms) -[ok]: failover command to any replica works (1408 ms) -[ok]: failover to a replica with force works (692 ms) -[ok]: failover with timeout aborts if replica never catches up (899 ms) -[ok]: failovers can be aborted (251 ms) -[ok]: failover aborts if target rejects sync request (657 ms) -[101/124 done]: integration/failover (8 seconds) -Testing integration/rdb -=== () Starting server 127.0.0.1:22437 ok -[ok]: test old version rdb file (5 ms) -=== () Starting server 127.0.0.1:22439 ok -[ok]: RDB encoding loading test (32 ms) -=== () Starting server 127.0.0.1:22441 ok -[ok]: Server started empty with non-existing RDB file (8 ms) -=== () Starting server 127.0.0.1:22443 ok -[ok]: Server started empty with empty RDB file (7 ms) -=== () Starting server 127.0.0.1:22445 ok -[ok]: Test RDB stream encoding (621 ms) -[ok]: Test RDB stream encoding - sanitize dump (34 ms) -=== () Starting server 127.0.0.1:22447 -[ok]: Server should not start if RDB file can't be open (12 ms) -=== () Starting server 127.0.0.1:22449 -[ok]: Server should not start if RDB is corrupted (20 ms) -=== () Starting server 127.0.0.1:22451 ok -[ok]: Test FLUSHALL aborts bgsave (2374 ms) -[ok]: bgsave resets the change counter (140 ms) -=== () Starting server 127.0.0.1:22453 ok -[ok]: client freed during loading (3990 ms) -=== () Starting server 127.0.0.1:22455 ok -[ok]: Test RDB load info (2493 ms) -=== () Starting server 127.0.0.1:22457 ok -Keys before bgsave start: 10000 -COW info before copy-on-write: 0 -current_save_keys_processed info : 0 -COW info before copy-on-write: 1204224 -current_save_keys_processed info : 1 -[ok]: Test child sending info (5084 ms) -=== () Starting server 127.0.0.1:22459 ok -[ok]: script won't load anymore if it's in rdb (4 ms) -=== () Starting server 127.0.0.1:22461 ok - -Waiting for background save to finish... -Waiting for background save to finish... -Waiting for background save to finish... -Waiting for background save to finish... [ok]: failed bgsave prevents writes (963 ms) -[102/124 done]: integration/rdb (22 seconds) -Testing integration/convert-ziplist-hash-on-load -=== () Starting server 127.0.0.1:22463 ok -[ok]: RDB load ziplist hash: converts to listpack when RDB loading (32 ms) -=== () Starting server 127.0.0.1:22465 ok -[ok]: RDB load ziplist hash: converts to hash table when hash-max-ziplist-entries is exceeded (31 ms) -[103/124 done]: integration/convert-ziplist-hash-on-load (1 seconds) -Testing integration/replication-4 -=== (repl network external:skip singledb:skip) Starting server 127.0.0.1:22467 ok -=== () Starting server 127.0.0.1:22469 ok -[ok]: First server should have role slave after SLAVEOF (12 ms) -[ok]: Test replication with parallel clients writing in different DBs (5077 ms) -=== (repl external:skip) Starting server 127.0.0.1:22471 ok -=== () Starting server 127.0.0.1:22473 ok -[ok]: First server should have role slave after SLAVEOF (239 ms) -[ok]: With min-slaves-to-write (1,3): master should be writable (32 ms) -[ok]: With min-slaves-to-write (2,3): master should not be writable (32 ms) -[ok]: With min-slaves-to-write function without no-write flag (16 ms) -[ok]: With not enough good slaves, read in Lua script is still accepted (40 ms) -[ok]: With min-slaves-to-write: master not writable with lagged slave (2632 ms) -=== (repl external:skip) Starting server 127.0.0.1:22475 ok -=== () Starting server 127.0.0.1:22477 ok -[ok]: First server should have role slave after SLAVEOF (128 ms) -Replication of an expired key does not delete the expired key test attempts: 0 -[ok]: Replication of an expired key does not delete the expired key (759 ms) -=== (repl external:skip) Starting server 127.0.0.1:22479 ok -=== () Starting server 127.0.0.1:22481 ok -[ok]: First server should have role slave after SLAVEOF (16 ms) -[ok]: Replication: commands with many arguments (issue #1221) (30185 ms) -[ok]: spopwithcount rewrite srem command (303 ms) -[ok]: Replication of SPOP command -- alsoPropagate() API (2210 ms) -=== (repl external:skip) Starting server 127.0.0.1:22483 ok -=== () Starting server 127.0.0.1:22485 ok -[ok]: First server should have role slave after SLAVEOF (148 ms) -[ok]: Data divergence can happen under default conditions (95 ms) -[ok]: Data divergence is allowed on writable replicas (80 ms) -[104/124 done]: integration/replication-4 (49 seconds) -Testing integration/valkey-benchmark -=== () Starting server 127.0.0.1:22487 ok -[ok]: benchmark: set,get (284 ms) -[ok]: benchmark: connecting using URI set,get (220 ms) -[ok]: benchmark: connecting using URI with authentication set,get (271 ms) -[ok]: benchmark: full test suite (4136 ms) -[ok]: benchmark: multi-thread set,get (780 ms) -[ok]: benchmark: pipelined full set,get (3228 ms) -[ok]: benchmark: arbitrary command (425 ms) -[ok]: benchmark: keyspace length (1245 ms) -[ok]: benchmark: clients idle mode should return error when reached maxclients limit (174 ms) -[ok]: benchmark: read last argument from stdin (419 ms) -[ok]: benchmark: specific tls-ciphers (1577 ms) -[ok]: benchmark: tls connecting using URI with authentication set,get (349 ms) -[ok]: benchmark: specific tls-ciphersuites (1217 ms) -[105/124 done]: integration/valkey-benchmark (15 seconds) -Testing integration/psync2-reg -=== (psync2 external:skip) Starting server 127.0.0.1:22489 ok -=== () Starting server 127.0.0.1:22491 ok -=== () Starting server 127.0.0.1:22493 ok -[ok]: PSYNC2 #3899 regression: setup (1091 ms) -[ok]: PSYNC2 #3899 regression: kill first replica (7 ms) -[ok]: PSYNC2 #3899 regression: kill chained replica (6 ms) -[ok]: PSYNC2 #3899 regression: kill first replica (4 ms) -[ok]: PSYNC2 #3899 regression: kill first replica (8 ms) -[ok]: PSYNC2 #3899 regression: kill chained replica (8 ms) -[ok]: PSYNC2 #3899 regression: kill first replica (8 ms) -[ok]: PSYNC2 #3899 regression: kill chained replica (7 ms) -[ok]: PSYNC2 #3899 regression: kill chained replica (8 ms) -[ok]: PSYNC2 #3899 regression: kill chained replica (7 ms) -[ok]: PSYNC2 #3899 regression: kill first replica (8 ms) -[ok]: PSYNC2 #3899 regression: kill chained replica (8 ms) -[ok]: PSYNC2 #3899 regression: kill chained replica (6 ms) -[ok]: PSYNC2 #3899 regression: kill chained replica (8 ms) -[ok]: PSYNC2 #3899 regression: kill first replica (8 ms) -[ok]: PSYNC2 #3899 regression: kill first replica (5 ms) -[ok]: PSYNC2 #3899 regression: kill chained replica (6 ms) -[ok]: PSYNC2 #3899 regression: kill first replica (15 ms) -[ok]: PSYNC2 #3899 regression: verify consistency (71 ms) -[106/124 done]: integration/psync2-reg (24 seconds) -Testing integration/corrupt-dump -=== () Starting server 127.0.0.1:22495 ok -[ok]: corrupt payload: #7445 - with sanitize (435 ms) -=== () Starting server 127.0.0.1:22497 ok -[ok]: corrupt payload: hash with valid zip list header, invalid entry len (552 ms) -=== () Starting server 127.0.0.1:22499 ok -[ok]: corrupt payload: invalid zlbytes header (520 ms) -=== () Starting server 127.0.0.1:22501 ok -[ok]: corrupt payload: valid zipped hash header, dup records (732 ms) -=== () Starting server 127.0.0.1:22503 ok -[ok]: corrupt payload: quicklist big ziplist prev len (568 ms) -=== () Starting server 127.0.0.1:22505 ok -[ok]: corrupt payload: quicklist small ziplist prev len (543 ms) -=== () Starting server 127.0.0.1:22507 ok -[ok]: corrupt payload: quicklist ziplist wrong count (512 ms) -=== () Starting server 127.0.0.1:22509 ok -[ok]: corrupt payload: #3080 - quicklist (661 ms) -=== () Starting server 127.0.0.1:22511 ok -[ok]: corrupt payload: quicklist with empty ziplist (748 ms) -=== () Starting server 127.0.0.1:22513 ok -[ok]: corrupt payload: quicklist encoded_len is 0 (560 ms) -=== () Starting server 127.0.0.1:22515 ok -[ok]: corrupt payload: quicklist listpack entry start with EOF (465 ms) -=== () Starting server 127.0.0.1:22517 ok -[ok]: corrupt payload: #3080 - ziplist (893 ms) -=== () Starting server 127.0.0.1:22519 -[ok]: corrupt payload: load corrupted rdb with no CRC - #3505 (410 ms) -=== () Starting server 127.0.0.1:22521 ok -[ok]: corrupt payload: load corrupted rdb with empty keys (873 ms) -=== () Starting server 127.0.0.1:22523 ok -[ok]: corrupt payload: load corrupted rdb with empty keys (572 ms) -=== () Starting server 127.0.0.1:22525 ok -[ok]: corrupt payload: listpack invalid size header (1019 ms) -=== () Starting server 127.0.0.1:22527 ok -[ok]: corrupt payload: listpack too long entry len (732 ms) -=== () Starting server 127.0.0.1:22529 ok -[ok]: corrupt payload: listpack very long entry len (816 ms) -=== () Starting server 127.0.0.1:22531 ok -[ok]: corrupt payload: listpack too long entry prev len (1580 ms) -=== () Starting server 127.0.0.1:22533 ok -[ok]: corrupt payload: stream with duplicate consumers (1250 ms) -=== () Starting server 127.0.0.1:22535 ok -[ok]: corrupt payload: hash ziplist with duplicate records (1882 ms) -=== () Starting server 127.0.0.1:22537 ok -[ok]: corrupt payload: hash listpack with duplicate records (1328 ms) -=== () Starting server 127.0.0.1:22539 ok -[ok]: corrupt payload: hash listpack with duplicate records - convert (1819 ms) -=== () Starting server 127.0.0.1:22541 ok -[ok]: corrupt payload: hash ziplist uneven record count (1130 ms) -=== () Starting server 127.0.0.1:22543 ok -[ok]: corrupt payload: hash duplicate records (626 ms) -=== () Starting server 127.0.0.1:22545 ok -[ok]: corrupt payload: hash empty zipmap (636 ms) -=== () Starting server 127.0.0.1:22547 ok -[ok]: corrupt payload: fuzzer findings - NPD in streamIteratorGetID (1100 ms) -=== () Starting server 127.0.0.1:22549 ok -[ok]: corrupt payload: fuzzer findings - listpack NPD on invalid stream (485 ms) -=== () Starting server 127.0.0.1:22551 ok -[ok]: corrupt payload: fuzzer findings - NPD in quicklistIndex (619 ms) -=== () Starting server 127.0.0.1:22553 ok -[ok]: corrupt payload: fuzzer findings - encoded entry header reach outside the allocation (561 ms) -=== () Starting server 127.0.0.1:22555 ok -[ok]: corrupt payload: fuzzer findings - invalid ziplist encoding (605 ms) -=== () Starting server 127.0.0.1:22557 ok -[ok]: corrupt payload: fuzzer findings - hash crash (460 ms) -=== () Starting server 127.0.0.1:22559 ok -[ok]: corrupt payload: fuzzer findings - uneven entry count in hash (500 ms) -=== () Starting server 127.0.0.1:22561 ok -[ok]: corrupt payload: fuzzer findings - invalid read in lzf_decompress (504 ms) -=== () Starting server 127.0.0.1:22563 ok -[ok]: corrupt payload: fuzzer findings - leak in rdbloading due to dup entry in set (610 ms) -=== () Starting server 127.0.0.1:22565 ok -[ok]: corrupt payload: fuzzer findings - empty intset (605 ms) -=== () Starting server 127.0.0.1:22567 ok -[ok]: corrupt payload: fuzzer findings - zset ziplist entry lensize is 0 (668 ms) -=== () Starting server 127.0.0.1:22569 ok -[ok]: corrupt payload: fuzzer findings - valgrind ziplist prevlen reaches outside the ziplist (695 ms) -=== () Starting server 127.0.0.1:22571 ok -[ok]: corrupt payload: fuzzer findings - valgrind - bad rdbLoadDoubleValue (527 ms) -=== () Starting server 127.0.0.1:22573 ok -[ok]: corrupt payload: fuzzer findings - valgrind ziplist prev too big (538 ms) -=== () Starting server 127.0.0.1:22575 ok -[ok]: corrupt payload: fuzzer findings - lzf decompression fails, avoid valgrind invalid read (498 ms) -=== () Starting server 127.0.0.1:22577 ok -[ok]: corrupt payload: fuzzer findings - stream bad lp_count (453 ms) -=== () Starting server 127.0.0.1:22579 ok -[ok]: corrupt payload: fuzzer findings - stream bad lp_count - unsanitized (457 ms) -=== () Starting server 127.0.0.1:22581 ok -[ok]: corrupt payload: fuzzer findings - stream integrity check issue (484 ms) -=== () Starting server 127.0.0.1:22583 ok -[ok]: corrupt payload: fuzzer findings - infinite loop (832 ms) -=== () Starting server 127.0.0.1:22585 ok -[ok]: corrupt payload: fuzzer findings - hash ziplist too long entry len (708 ms) -=== () Starting server 127.0.0.1:22587 ok -[ok]: corrupt payload: OOM in rdbGenericLoadStringObject (609 ms) -=== () Starting server 127.0.0.1:22589 ok -[ok]: corrupt payload: fuzzer findings - OOM in dictExpand (534 ms) -=== () Starting server 127.0.0.1:22591 ok -[ok]: corrupt payload: fuzzer findings - zset ziplist invalid tail offset (756 ms) -=== () Starting server 127.0.0.1:22593 ok -[ok]: corrupt payload: fuzzer findings - negative reply length (956 ms) -=== () Starting server 127.0.0.1:22595 ok -[ok]: corrupt payload: fuzzer findings - valgrind negative malloc (792 ms) -=== () Starting server 127.0.0.1:22597 ok -[ok]: corrupt payload: fuzzer findings - valgrind invalid read (1652 ms) -=== () Starting server 127.0.0.1:22599 ok -[ok]: corrupt payload: fuzzer findings - empty hash ziplist (836 ms) -=== () Starting server 127.0.0.1:22601 ok -[ok]: corrupt payload: fuzzer findings - stream with no records (516 ms) -=== () Starting server 127.0.0.1:22603 ok -[ok]: corrupt payload: fuzzer findings - quicklist ziplist tail followed by extra data which start with 0xff (980 ms) -=== () Starting server 127.0.0.1:22605 ok -[ok]: corrupt payload: fuzzer findings - dict init to huge size (1016 ms) -=== () Starting server 127.0.0.1:22607 ok -[ok]: corrupt payload: fuzzer findings - huge string (866 ms) -=== () Starting server 127.0.0.1:22609 ok -[ok]: corrupt payload: fuzzer findings - stream PEL without consumer (838 ms) -=== () Starting server 127.0.0.1:22611 ok -[ok]: corrupt payload: fuzzer findings - stream listpack valgrind issue (599 ms) -=== () Starting server 127.0.0.1:22613 ok -[ok]: corrupt payload: fuzzer findings - stream with bad lpFirst (564 ms) -=== () Starting server 127.0.0.1:22615 ok -[ok]: corrupt payload: fuzzer findings - stream listpack lpPrev valgrind issue (512 ms) -=== () Starting server 127.0.0.1:22617 ok -[ok]: corrupt payload: fuzzer findings - stream with non-integer entry id (512 ms) -=== () Starting server 127.0.0.1:22619 ok -[ok]: corrupt payload: fuzzer findings - empty quicklist (616 ms) -=== () Starting server 127.0.0.1:22621 ok -[ok]: corrupt payload: fuzzer findings - empty zset (615 ms) -=== () Starting server 127.0.0.1:22623 ok -[ok]: corrupt payload: fuzzer findings - hash with len of 0 (667 ms) -=== () Starting server 127.0.0.1:22625 ok -[ok]: corrupt payload: fuzzer findings - hash listpack first element too long entry len (632 ms) -=== () Starting server 127.0.0.1:22627 ok -[ok]: corrupt payload: fuzzer findings - stream double free listpack when insert dup node to rax returns 0 (672 ms) -=== () Starting server 127.0.0.1:22629 ok -[ok]: corrupt payload: fuzzer findings - LCS OOM (628 ms) -=== () Starting server 127.0.0.1:22631 ok -[ok]: corrupt payload: fuzzer findings - gcc asan reports false leak on assert (541 ms) -=== () Starting server 127.0.0.1:22633 ok -[ok]: corrupt payload: fuzzer findings - lpFind invalid access (412 ms) -=== () Starting server 127.0.0.1:22635 ok -[ok]: corrupt payload: fuzzer findings - invalid access in ziplist tail prevlen decoding (534 ms) -=== () Starting server 127.0.0.1:22637 ok -[ok]: corrupt payload: fuzzer findings - zset zslInsert with a NAN score (604 ms) -=== () Starting server 127.0.0.1:22639 ok -[ok]: corrupt payload: fuzzer findings - streamLastValidID panic (644 ms) -=== () Starting server 127.0.0.1:22641 ok -[ok]: corrupt payload: fuzzer findings - valgrind fishy value warning (596 ms) -=== () Starting server 127.0.0.1:22643 ok -[ok]: corrupt payload: fuzzer findings - empty set listpack (544 ms) -=== () Starting server 127.0.0.1:22645 ok -[ok]: corrupt payload: fuzzer findings - set with duplicate elements causes sdiff to hang (679 ms) -[107/124 done]: integration/corrupt-dump (55 seconds) -Testing integration/replication-2 -=== (repl external:skip) Starting server 127.0.0.1:22647 ok -=== () Starting server 127.0.0.1:22649 ok -[ok]: First server should have role slave after SLAVEOF (134 ms) -[ok]: If min-slaves-to-write is honored, write is accepted (28 ms) -[ok]: No write if min-slaves-to-write is < attached slaves (23 ms) -[ok]: If min-slaves-to-write is honored, write is accepted (again) (28 ms) -[ok]: No write if min-slaves-max-lag is > of the slave lag (3799 ms) -[ok]: min-slaves-to-write is ignored by slaves (47 ms) -[ok]: MASTER and SLAVE dataset should be identical after complex ops (23263 ms) -[108/124 done]: integration/replication-2 (28 seconds) -Testing integration/psync2-pingoff -=== (psync2 external:skip) Starting server 127.0.0.1:22651 ok -=== () Starting server 127.0.0.1:22653 ok -[ok]: PSYNC2 pingoff: setup (1034 ms) -[ok]: PSYNC2 pingoff: write and wait replication (2 ms) -[ok]: PSYNC2 pingoff: pause replica and promote it (5016 ms) -[ok]: Make the old master a replica of the new one and check conditions (1032 ms) -=== (psync2 external:skip) Starting server 127.0.0.1:22655 ok -=== () Starting server 127.0.0.1:22657 ok -=== () Starting server 127.0.0.1:22659 ok -=== () Starting server 127.0.0.1:22661 ok -=== () Starting server 127.0.0.1:22663 ok -[ok]: test various edge cases of repl topology changes with missing pings at the end (4405 ms) -=== (psync2 external:skip) Starting server 127.0.0.1:22665 ok -=== () Starting server 127.0.0.1:22667 ok -=== () Starting server 127.0.0.1:22669 ok -[ok]: Chained replicas disconnect when replica re-connect with the same master (3019 ms) -[109/124 done]: integration/psync2-pingoff (19 seconds) -Testing integration/replication-psync -=== (repl) Starting server 127.0.0.1:22671 ok -=== () Starting server 127.0.0.1:22673 ok -[ok]: Slave should be able to synchronize with the master (161 ms) -[ok]: Detect write load to master (1005 ms) -[ok]: Test replication partial resync: no reconnection, just sync (diskless: no, disabled, dual-channel: yes, reconnect: 0) (35 ms) -=== (repl) Starting server 127.0.0.1:22675 ok -=== () Starting server 127.0.0.1:22677 ok -[ok]: Slave should be able to synchronize with the master (220 ms) -[ok]: Detect write load to master (1013 ms) -[ok]: Test replication partial resync: ok psync (diskless: no, disabled, dual-channel: yes, reconnect: 1) (6295 ms) -=== (repl) Starting server 127.0.0.1:22679 ok -=== () Starting server 127.0.0.1:22681 ok -[ok]: Slave should be able to synchronize with the master (217 ms) -[ok]: Detect write load to master (1003 ms) -[ok]: Test replication partial resync: no backlog (diskless: no, disabled, dual-channel: yes, reconnect: 1) (7949 ms) -=== (repl) Starting server 127.0.0.1:22683 ok -=== () Starting server 127.0.0.1:22685 ok -[ok]: Slave should be able to synchronize with the master (139 ms) -[ok]: Detect write load to master (3027 ms) -[ok]: Test replication partial resync: ok after delay (diskless: no, disabled, dual-channel: yes, reconnect: 1) (9332 ms) -=== (repl) Starting server 127.0.0.1:22687 ok -=== () Starting server 127.0.0.1:22689 ok -[ok]: Slave should be able to synchronize with the master (156 ms) -[ok]: Detect write load to master (4035 ms) -[ok]: Test replication partial resync: backlog expired (diskless: no, disabled, dual-channel: yes, reconnect: 1) (9347 ms) -=== (repl) Starting server 127.0.0.1:22691 ok -=== () Starting server 127.0.0.1:22693 ok -[ok]: Slave should be able to synchronize with the master (132 ms) -[ok]: Detect write load to master (1031 ms) -[ok]: Test replication partial resync: no reconnection, just sync (diskless: no, disabled, dual-channel: no, reconnect: 0) (100 ms) -=== (repl) Starting server 127.0.0.1:22695 ok -=== () Starting server 127.0.0.1:22697 ok -[ok]: Slave should be able to synchronize with the master (196 ms) -[ok]: Detect write load to master (1012 ms) -[ok]: Test replication partial resync: ok psync (diskless: no, disabled, dual-channel: no, reconnect: 1) (6390 ms) -=== (repl) Starting server 127.0.0.1:22699 ok -=== () Starting server 127.0.0.1:22701 ok -[ok]: Slave should be able to synchronize with the master (245 ms) -[ok]: Detect write load to master (2018 ms) -[ok]: Test replication partial resync: no backlog (diskless: no, disabled, dual-channel: no, reconnect: 1) (7913 ms) -=== (repl) Starting server 127.0.0.1:22703 ok -=== () Starting server 127.0.0.1:22705 ok -[ok]: Slave should be able to synchronize with the master (136 ms) -[ok]: Detect write load to master (1015 ms) -[ok]: Test replication partial resync: ok after delay (diskless: no, disabled, dual-channel: no, reconnect: 1) (9344 ms) -=== (repl) Starting server 127.0.0.1:22707 ok -=== () Starting server 127.0.0.1:22709 ok -[ok]: Slave should be able to synchronize with the master (154 ms) -[ok]: Detect write load to master (1030 ms) -[ok]: Test replication partial resync: backlog expired (diskless: no, disabled, dual-channel: no, reconnect: 1) (9383 ms) -=== (repl) Starting server 127.0.0.1:22711 ok -=== () Starting server 127.0.0.1:22713 ok -[ok]: Slave should be able to synchronize with the master (255 ms) -[ok]: Detect write load to master (2023 ms) -[ok]: Test replication partial resync: no reconnection, just sync (diskless: no, swapdb, dual-channel: yes, reconnect: 0) (79 ms) -=== (repl) Starting server 127.0.0.1:22715 ok -=== () Starting server 127.0.0.1:22717 ok -[ok]: Slave should be able to synchronize with the master (392 ms) -[ok]: Detect write load to master (2016 ms) -[ok]: Test replication partial resync: ok psync (diskless: no, swapdb, dual-channel: yes, reconnect: 1) (6355 ms) -=== (repl) Starting server 127.0.0.1:22719 ok -=== () Starting server 127.0.0.1:22721 ok -[ok]: Slave should be able to synchronize with the master (1068 ms) -[ok]: Detect write load to master (3 ms) -[ok]: Test replication partial resync: no backlog (diskless: no, swapdb, dual-channel: yes, reconnect: 1) (8476 ms) -=== (repl) Starting server 127.0.0.1:22723 ok -=== () Starting server 127.0.0.1:22725 ok -[ok]: Slave should be able to synchronize with the master (1363 ms) -[ok]: Detect write load to master (7 ms) -[ok]: Test replication partial resync: ok after delay (diskless: no, swapdb, dual-channel: yes, reconnect: 1) (9553 ms) -=== (repl) Starting server 127.0.0.1:22727 ok -=== () Starting server 127.0.0.1:22729 ok -[ok]: Slave should be able to synchronize with the master (172 ms) -[ok]: Detect write load to master (2025 ms) -[ok]: Test replication partial resync: backlog expired (diskless: no, swapdb, dual-channel: yes, reconnect: 1) (9354 ms) -=== (repl) Starting server 127.0.0.1:22731 ok -=== () Starting server 127.0.0.1:22733 ok -[ok]: Slave should be able to synchronize with the master (260 ms) -[ok]: Detect write load to master (3028 ms) -[ok]: Test replication partial resync: no reconnection, just sync (diskless: no, swapdb, dual-channel: no, reconnect: 0) (127 ms) -=== (repl) Starting server 127.0.0.1:22735 ok -=== () Starting server 127.0.0.1:22737 ok -[ok]: Slave should be able to synchronize with the master (244 ms) -[ok]: Detect write load to master (2019 ms) -[ok]: Test replication partial resync: ok psync (diskless: no, swapdb, dual-channel: no, reconnect: 1) (6370 ms) -=== (repl) Starting server 127.0.0.1:22739 ok -=== () Starting server 127.0.0.1:22741 ok -[ok]: Slave should be able to synchronize with the master (244 ms) -[ok]: Detect write load to master (4038 ms) -[ok]: Test replication partial resync: no backlog (diskless: no, swapdb, dual-channel: no, reconnect: 1) (7846 ms) -=== (repl) Starting server 127.0.0.1:22743 ok -=== () Starting server 127.0.0.1:22745 ok -[ok]: Slave should be able to synchronize with the master (247 ms) -[ok]: Detect write load to master (1017 ms) -[ok]: Test replication partial resync: ok after delay (diskless: no, swapdb, dual-channel: no, reconnect: 1) (9336 ms) -=== (repl) Starting server 127.0.0.1:22747 ok -=== () Starting server 127.0.0.1:22749 ok -[ok]: Slave should be able to synchronize with the master (244 ms) -[ok]: Detect write load to master (2018 ms) -[ok]: Test replication partial resync: backlog expired (diskless: no, swapdb, dual-channel: no, reconnect: 1) (9339 ms) -=== (repl) Starting server 127.0.0.1:22751 ok -=== () Starting server 127.0.0.1:22753 ok -[ok]: Slave should be able to synchronize with the master (1535 ms) -[ok]: Detect write load to master (1016 ms) -[ok]: Test replication partial resync: no reconnection, just sync (diskless: yes, disabled, dual-channel: yes, reconnect: 0) (99 ms) -=== (repl) Starting server 127.0.0.1:22755 ok -=== () Starting server 127.0.0.1:22757 ok -[ok]: Slave should be able to synchronize with the master (1500 ms) -[ok]: Detect write load to master (1006 ms) -[ok]: Test replication partial resync: ok psync (diskless: yes, disabled, dual-channel: yes, reconnect: 1) (6377 ms) -=== (repl) Starting server 127.0.0.1:22759 ok -=== () Starting server 127.0.0.1:22761 ok -[ok]: Slave should be able to synchronize with the master (1481 ms) -[ok]: Detect write load to master (2022 ms) -[ok]: Test replication partial resync: no backlog (diskless: yes, disabled, dual-channel: yes, reconnect: 1) (8016 ms) -=== (repl) Starting server 127.0.0.1:22763 ok -=== () Starting server 127.0.0.1:22765 ok -[ok]: Slave should be able to synchronize with the master (1321 ms) -[ok]: Detect write load to master (3 ms) -[ok]: Test replication partial resync: ok after delay (diskless: yes, disabled, dual-channel: yes, reconnect: 1) (9302 ms) -=== (repl) Starting server 127.0.0.1:22767 ok -=== () Starting server 127.0.0.1:22769 ok -[ok]: Slave should be able to synchronize with the master (474 ms) -[ok]: Detect write load to master (3027 ms) -[ok]: Test replication partial resync: backlog expired (diskless: yes, disabled, dual-channel: yes, reconnect: 1) (9479 ms) -=== (repl) Starting server 127.0.0.1:22771 ok -=== () Starting server 127.0.0.1:22773 ok -[ok]: Slave should be able to synchronize with the master (495 ms) -[ok]: Detect write load to master (3024 ms) -[ok]: Test replication partial resync: no reconnection, just sync (diskless: yes, disabled, dual-channel: no, reconnect: 0) (85 ms) -=== (repl) Starting server 127.0.0.1:22775 ok -=== () Starting server 127.0.0.1:22777 ok -[ok]: Slave should be able to synchronize with the master (496 ms) -[ok]: Detect write load to master (2030 ms) -[ok]: Test replication partial resync: ok psync (diskless: yes, disabled, dual-channel: no, reconnect: 1) (6360 ms) -=== (repl) Starting server 127.0.0.1:22779 ok -=== () Starting server 127.0.0.1:22781 ok -[ok]: Slave should be able to synchronize with the master (475 ms) -[ok]: Detect write load to master (1019 ms) -[ok]: Test replication partial resync: no backlog (diskless: yes, disabled, dual-channel: no, reconnect: 1) (7931 ms) -=== (repl) Starting server 127.0.0.1:22783 ok -=== () Starting server 127.0.0.1:22785 ok -[ok]: Slave should be able to synchronize with the master (1536 ms) -[ok]: Detect write load to master (1011 ms) -[ok]: Test replication partial resync: ok after delay (diskless: yes, disabled, dual-channel: no, reconnect: 1) (9263 ms) -=== (repl) Starting server 127.0.0.1:22787 ok -=== () Starting server 127.0.0.1:22789 ok -[ok]: Slave should be able to synchronize with the master (1504 ms) -[ok]: Detect write load to master (8 ms) -[ok]: Test replication partial resync: backlog expired (diskless: yes, disabled, dual-channel: no, reconnect: 1) (9301 ms) -=== (repl) Starting server 127.0.0.1:22791 ok -=== () Starting server 127.0.0.1:22793 ok -[ok]: Slave should be able to synchronize with the master (1563 ms) -[ok]: Detect write load to master (2020 ms) -[ok]: Test replication partial resync: no reconnection, just sync (diskless: yes, swapdb, dual-channel: yes, reconnect: 0) (99 ms) -=== (repl) Starting server 127.0.0.1:22795 ok -=== () Starting server 127.0.0.1:22797 ok -[ok]: Slave should be able to synchronize with the master (1552 ms) -[ok]: Detect write load to master (2023 ms) -[ok]: Test replication partial resync: ok psync (diskless: yes, swapdb, dual-channel: yes, reconnect: 1) (6460 ms) -=== (repl) Starting server 127.0.0.1:22799 ok -=== () Starting server 127.0.0.1:22801 ok -[ok]: Slave should be able to synchronize with the master (1176 ms) -[ok]: Detect write load to master (1011 ms) -[ok]: Test replication partial resync: no backlog (diskless: yes, swapdb, dual-channel: yes, reconnect: 1) (7891 ms) -=== (repl) Starting server 127.0.0.1:22803 ok -=== () Starting server 127.0.0.1:22805 ok -[ok]: Slave should be able to synchronize with the master (1547 ms) -[ok]: Detect write load to master (0 ms) -[ok]: Test replication partial resync: ok after delay (diskless: yes, swapdb, dual-channel: yes, reconnect: 1) (9334 ms) -=== (repl) Starting server 127.0.0.1:22807 ok -=== () Starting server 127.0.0.1:22809 ok -[ok]: Slave should be able to synchronize with the master (1534 ms) -[ok]: Detect write load to master (2016 ms) -[ok]: Test replication partial resync: backlog expired (diskless: yes, swapdb, dual-channel: yes, reconnect: 1) (9418 ms) -=== (repl) Starting server 127.0.0.1:22811 ok -=== () Starting server 127.0.0.1:22813 ok -[ok]: Slave should be able to synchronize with the master (1540 ms) -[ok]: Detect write load to master (3036 ms) -[ok]: Test replication partial resync: no reconnection, just sync (diskless: yes, swapdb, dual-channel: no, reconnect: 0) (164 ms) -=== (repl) Starting server 127.0.0.1:22815 ok -=== () Starting server 127.0.0.1:22817 ok -[ok]: Slave should be able to synchronize with the master (1554 ms) -[ok]: Detect write load to master (7 ms) -[ok]: Test replication partial resync: ok psync (diskless: yes, swapdb, dual-channel: no, reconnect: 1) (6372 ms) -=== (repl) Starting server 127.0.0.1:22819 ok -=== () Starting server 127.0.0.1:22821 ok -[ok]: Slave should be able to synchronize with the master (1576 ms) -[ok]: Detect write load to master (7 ms) -[ok]: Test replication partial resync: no backlog (diskless: yes, swapdb, dual-channel: no, reconnect: 1) (7936 ms) -=== (repl) Starting server 127.0.0.1:22823 ok -=== () Starting server 127.0.0.1:22825 ok -[ok]: Slave should be able to synchronize with the master (1411 ms) -[ok]: Detect write load to master (0 ms) -[ok]: Test replication partial resync: ok after delay (diskless: yes, swapdb, dual-channel: no, reconnect: 1) (9252 ms) -=== (repl) Starting server 127.0.0.1:22827 ok -=== () Starting server 127.0.0.1:22829 ok -[ok]: Slave should be able to synchronize with the master (1478 ms) -[ok]: Detect write load to master (1010 ms) -[ok]: Test replication partial resync: backlog expired (diskless: yes, swapdb, dual-channel: no, reconnect: 1) (9423 ms) -[110/124 done]: integration/replication-psync (406 seconds) -Testing integration/logging -=== () Starting server 127.0.0.1:22831 ok -{2786626:M 09 Mar 2025 13:21:13.174 # valkey 8.0.2 crashed by signal: 6, si_code: 0} 40 -[ok]: Crash report generated on SIGABRT (132 ms) -=== () Starting server 127.0.0.1:22833 ok -{2786732:M 09 Mar 2025 13:21:13.583 # valkey 8.0.2 crashed by signal: 11, si_code: 2} 40 -[ok]: Crash report generated on DEBUG SEGFAULT (36 ms) -=== () Starting server 127.0.0.1:22835 ok -{Received SIGALRM} 38 -[ok]: Stacktraces generated on SIGALRM (139 ms) -[111/124 done]: integration/logging (2 seconds) -Testing integration/block-repl -=== (repl external:skip) Starting server 127.0.0.1:22837 ok -=== () Starting server 127.0.0.1:22839 ok -[ok]: First server should have role slave after SLAVEOF (1019 ms) -[ok]: Test replication with blocking lists and sorted sets operations (25239 ms) -[112/124 done]: integration/block-repl (28 seconds) -Testing integration/replication -=== (repl network external:skip) Starting server 127.0.0.1:22841 ok -=== () Starting server 127.0.0.1:22843 ok -[ok]: Slave enters handshake (1020 ms) -[ok]: Slave enters wait_bgsave (8 ms) -[ok]: Slave is able to detect timeout during handshake (0 ms) -=== (repl external:skip) Starting server 127.0.0.1:22845 ok -=== () Starting server 127.0.0.1:22847 ok -[ok]: Set instance A as slave of B (147 ms) -[ok]: INCRBYFLOAT replication, should not remove expire (48 ms) -[ok]: GETSET replication (34 ms) -[ok]: BRPOPLPUSH replication, when blocking against empty list (44 ms) -[ok]: BRPOPLPUSH replication, list exists (1094 ms) -[ok]: BLMOVE (left, left) replication, when blocking against empty list (64 ms) -[ok]: BLMOVE (left, left) replication, list exists (1055 ms) -[ok]: BLMOVE (left, right) replication, when blocking against empty list (60 ms) -[ok]: BLMOVE (left, right) replication, list exists (1064 ms) -[ok]: BLMOVE (right, left) replication, when blocking against empty list (64 ms) -[ok]: BLMOVE (right, left) replication, list exists (1096 ms) -[ok]: BLMOVE (right, right) replication, when blocking against empty list (79 ms) -[ok]: BLMOVE (right, right) replication, list exists (1093 ms) -[ok]: BLPOP followed by role change, issue #2473 (347 ms) -[ok]: Replica output bytes metric (15 ms) -=== (repl external:skip) Starting server 127.0.0.1:22849 ok -=== () Starting server 127.0.0.1:22851 ok -[ok]: Second server should have role master at first (6 ms) -[ok]: SLAVEOF should start with link status "down" (20 ms) -[ok]: The role should immediately be changed to "replica" (11 ms) -[ok]: Sync should have transferred keys from master (6 ms) -[ok]: The link status should be up (4 ms) -[ok]: SET on the master should immediately propagate (6 ms) -[ok]: FLUSHDB / FLUSHALL should replicate (335 ms) -[ok]: ROLE in master reports master with a slave (1 ms) -[ok]: ROLE in slave reports slave in connected state (1 ms) -=== (repl external:skip) Starting server 127.0.0.1:22853 ok -=== () Starting server 127.0.0.1:22855 ok -=== () Starting server 127.0.0.1:22857 ok -=== () Starting server 127.0.0.1:22859 ok -[ok]: Connect multiple replicas at the same time (issue #141), master diskless=no, replica diskless=disabled dual-channel-replication-enabled=no (11614 ms) -=== (repl external:skip) Starting server 127.0.0.1:22861 ok -=== () Starting server 127.0.0.1:22863 ok -=== () Starting server 127.0.0.1:22865 ok -=== () Starting server 127.0.0.1:22867 ok -[ok]: Connect multiple replicas at the same time (issue #141), master diskless=no, replica diskless=swapdb dual-channel-replication-enabled=no (13534 ms) -=== (repl external:skip) Starting server 127.0.0.1:22869 ok -=== () Starting server 127.0.0.1:22871 ok -=== () Starting server 127.0.0.1:22873 ok -=== () Starting server 127.0.0.1:22875 ok -[ok]: Connect multiple replicas at the same time (issue #141), master diskless=yes, replica diskless=disabled dual-channel-replication-enabled=yes (14066 ms) -=== (repl external:skip) Starting server 127.0.0.1:22877 ok -=== () Starting server 127.0.0.1:22879 ok -=== () Starting server 127.0.0.1:22881 ok -=== () Starting server 127.0.0.1:22883 ok -[ok]: Connect multiple replicas at the same time (issue #141), master diskless=yes, replica diskless=swapdb dual-channel-replication-enabled=yes (13834 ms) -=== (repl external:skip) Starting server 127.0.0.1:22885 ok -=== () Starting server 127.0.0.1:22887 ok -[ok]: Master stream is correctly processed while the replica has a script in -BUSY state (46114 ms) -=== (repl external:skip) Starting server 127.0.0.1:22889 ok -=== () Starting server 127.0.0.1:22891 ok -[ok]: Diskless load swapdb (different replid): new database is exposed after swapping dual-channel=yes (12 ms) -=== (repl external:skip) Starting server 127.0.0.1:22893 ok -=== () Starting server 127.0.0.1:22895 ok -[ok]: Diskless load swapdb (different replid): replica enter loading dual-channel-replication-enabled=no (748 ms) -[ok]: Diskless load swapdb (different replid): old database is exposed after replication fails dual-channel=no (24 ms) -=== (repl external:skip) Starting server 127.0.0.1:22897 ok -=== () Starting server 127.0.0.1:22899 ok -[ok]: Diskless load swapdb (async_loading): new database is exposed after swapping (19 ms) -=== (repl external:skip) Starting server 127.0.0.1:22901 ok -=== () Starting server 127.0.0.1:22903 ok -[ok]: Diskless load swapdb (async_loading): replica enter async_loading (700 ms) -[ok]: Diskless load swapdb (async_loading): old database is exposed while async replication is in progress (1357 ms) -[ok]: Busy script during async loading (1760 ms) -[ok]: Blocked commands and configs during async-loading (647 ms) -[ok]: Diskless load swapdb (async_loading): old database is exposed after async replication fails (20 ms) -Waiting for process 2804682 to exit... -Waiting for process 2804682 to exit... -=== (repl) Starting server 127.0.0.1:22905 ok -=== () Starting server 127.0.0.1:22907 ok -filling took 9288 ms (TODO: use pipeline) -{2806938:S 09 Mar 2025 13:24:29.916 # Internal error in RDB reading offset 8211478, function at rdb.c:3390 -> Unexpected EOF reading RDB file. Failure loading rdb format from socket, assuming connection error, resuming operation.} 50 -{2806938:S 09 Mar 2025 13:24:30.086 # Internal error in RDB reading offset 8211478, function at rdb.c:3390 -> Unexpected EOF reading RDB file. Failure loading rdb format from socket, assuming connection error, resuming operation.} 65 -{2806938:S 09 Mar 2025 13:24:30.198 # Internal error in RDB reading offset 0, function at rdb.c:3390 -> Unexpected EOF reading RDB file. Failure loading rdb format from socket, assuming connection error, resuming operation.} 80 -{2806938:S 09 Mar 2025 13:24:30.318 # Internal error in RDB reading offset 0, function at rdb.c:3390 -> Unexpected EOF reading RDB file. Failure loading rdb format from socket, assuming connection error, resuming operation.} 95 -{2806938:S 09 Mar 2025 13:24:30.494 # Internal error in RDB reading offset 0, function at rdb.c:3390 -> Unexpected EOF reading RDB file. Failure loading rdb format from socket, assuming connection error, resuming operation.} 110 -{2806938:S 09 Mar 2025 13:24:30.710 # Internal error in RDB reading offset 0, function at rdb.c:3390 -> Unexpected EOF reading RDB file. Failure loading rdb format from socket, assuming connection error, resuming operation.} 125 -{2806938:S 09 Mar 2025 13:24:30.838 # Internal error in RDB reading offset 0, function at rdb.c:3390 -> Unexpected EOF reading RDB file. Failure loading rdb format from socket, assuming connection error, resuming operation.} 140 -{2806938:S 09 Mar 2025 13:24:30.988 # Internal error in RDB reading offset 0, function at rdb.c:3390 -> Unexpected EOF reading RDB file. Failure loading rdb format from socket, assuming connection error, resuming operation.} 155 -{2806938:S 09 Mar 2025 13:24:31.115 # Internal error in RDB reading offset 0, function at rdb.c:3390 -> Unexpected EOF reading RDB file. Failure loading rdb format from socket, assuming connection error, resuming operation.} 170 -{2806938:S 09 Mar 2025 13:24:31.231 # Internal error in RDB reading offset 0, function at rdb.c:3390 -> Unexpected EOF reading RDB file. Failure loading rdb format from socket, assuming connection error, resuming operation.} 185 -{2806938:S 09 Mar 2025 13:24:31.352 # Internal error in RDB reading offset 0, function at rdb.c:3390 -> Unexpected EOF reading RDB file. Failure loading rdb format from socket, assuming connection error, resuming operation.} 200 -{2806938:S 09 Mar 2025 13:24:31.476 # Internal error in RDB reading offset 0, function at rdb.c:3390 -> Unexpected EOF reading RDB file. Failure loading rdb format from socket, assuming connection error, resuming operation.} 215 -{2806938:S 09 Mar 2025 13:24:31.619 # Internal error in RDB reading offset 3877902, function at rdb.c:3390 -> Unexpected EOF reading RDB file. Failure loading rdb format from socket, assuming connection error, resuming operation.} 230 -{2806938:S 09 Mar 2025 13:24:31.704 # Internal error in RDB reading offset 0, function at rdb.c:3390 -> Unexpected EOF reading RDB file. Failure loading rdb format from socket, assuming connection error, resuming operation.} 245 -{2806938:S 09 Mar 2025 13:24:31.848 # Internal error in RDB reading offset 0, function at rdb.c:3390 -> Unexpected EOF reading RDB file. Failure loading rdb format from socket, assuming connection error, resuming operation.} 260 -{2806938:S 09 Mar 2025 13:24:32.000 # Internal error in RDB reading offset 0, function at rdb.c:3390 -> Unexpected EOF reading RDB file. Failure loading rdb format from socket, assuming connection error, resuming operation.} 275 -{2806938:S 09 Mar 2025 13:24:32.122 # Internal error in RDB reading offset 0, function at rdb.c:3390 -> Unexpected EOF reading RDB file. Failure loading rdb format from socket, assuming connection error, resuming operation.} 290 -{2806938:S 09 Mar 2025 13:24:32.241 # Internal error in RDB reading offset 0, function at rdb.c:3390 -> Unexpected EOF reading RDB file. Failure loading rdb format from socket, assuming connection error, resuming operation.} 305 -{2806938:S 09 Mar 2025 13:24:32.350 # Internal error in RDB reading offset 0, function at rdb.c:3390 -> Unexpected EOF reading RDB file. Failure loading rdb format from socket, assuming connection error, resuming operation.} 320 -{2806938:S 09 Mar 2025 13:24:32.658 # Internal error in RDB reading offset 0, function at rdb.c:3390 -> Unexpected EOF reading RDB file. Failure loading rdb format from socket, assuming connection error, resuming operation.} 335 -{2806938:S 09 Mar 2025 13:24:32.987 # Internal error in RDB reading offset 0, function at rdb.c:3390 -> Unexpected EOF reading RDB file. Failure loading rdb format from socket, assuming connection error, resuming operation.} 350 -{2806938:S 09 Mar 2025 13:24:33.246 # Internal error in RDB reading offset 0, function at rdb.c:3390 -> Unexpected EOF reading RDB file. Failure loading rdb format from socket, assuming connection error, resuming operation.} 365 -{2806938:S 09 Mar 2025 13:24:33.368 # Internal error in RDB reading offset 0, function at rdb.c:3390 -> Unexpected EOF reading RDB file. Failure loading rdb format from socket, assuming connection error, resuming operation.} 380 -{2806938:S 09 Mar 2025 13:24:33.519 # Internal error in RDB reading offset 6052388, function at rdb.c:3390 -> Unexpected EOF reading RDB file. Failure loading rdb format from socket, assuming connection error, resuming operation.} 395 -{2806938:S 09 Mar 2025 13:24:33.643 # Internal error in RDB reading offset 0, function at rdb.c:3390 -> Unexpected EOF reading RDB file. Failure loading rdb format from socket, assuming connection error, resuming operation.} 410 -{2806938:S 09 Mar 2025 13:24:33.761 # Internal error in RDB reading offset 0, function at rdb.c:3390 -> Unexpected EOF reading RDB file. Failure loading rdb format from socket, assuming connection error, resuming operation.} 425 -{2806938:S 09 Mar 2025 13:24:33.879 # Internal error in RDB reading offset 0, function at rdb.c:3390 -> Unexpected EOF reading RDB file. Failure loading rdb format from socket, assuming connection error, resuming operation.} 440 -{2806938:S 09 Mar 2025 13:24:34.001 # Internal error in RDB reading offset 0, function at rdb.c:3390 -> Unexpected EOF reading RDB file. Failure loading rdb format from socket, assuming connection error, resuming operation.} 455 -{2806938:S 09 Mar 2025 13:24:34.132 # Internal error in RDB reading offset 2035644, function at rdb.c:3390 -> Unexpected EOF reading RDB file. Failure loading rdb format from socket, assuming connection error, resuming operation.} 470 -{2806938:S 09 Mar 2025 13:24:34.230 # Internal error in RDB reading offset 0, function at rdb.c:3390 -> Unexpected EOF reading RDB file. Failure loading rdb format from socket, assuming connection error, resuming operation.} 485 -{2806938:S 09 Mar 2025 13:24:34.330 # Internal error in RDB reading offset 0, function at rdb.c:3390 -> Unexpected EOF reading RDB file. Failure loading rdb format from socket, assuming connection error, resuming operation.} 500 -{2806938:S 09 Mar 2025 13:24:34.442 # Internal error in RDB reading offset 0, function at rdb.c:3390 -> Unexpected EOF reading RDB file. Failure loading rdb format from socket, assuming connection error, resuming operation.} 515 -{2806938:S 09 Mar 2025 13:24:34.566 # Internal error in RDB reading offset 0, function at rdb.c:3390 -> Unexpected EOF reading RDB file. Failure loading rdb format from socket, assuming connection error, resuming operation.} 530 -{2806938:S 09 Mar 2025 13:24:34.659 # Internal error in RDB reading offset 0, function at rdb.c:3390 -> Unexpected EOF reading RDB file. Failure loading rdb format from socket, assuming connection error, resuming operation.} 545 -{2806938:S 09 Mar 2025 13:24:34.771 # Internal error in RDB reading offset 0, function at rdb.c:3390 -> Unexpected EOF reading RDB file. Failure loading rdb format from socket, assuming connection error, resuming operation.} 560 -{2806938:S 09 Mar 2025 13:24:34.869 # Internal error in RDB reading offset 0, function at rdb.c:3390 -> Unexpected EOF reading RDB file. Failure loading rdb format from socket, assuming connection error, resuming operation.} 575 -{2806938:S 09 Mar 2025 13:24:35.026 # Internal error in RDB reading offset 0, function at rdb.c:3390 -> Unexpected EOF reading RDB file. Failure loading rdb format from socket, assuming connection error, resuming operation.} 590 -{2806938:S 09 Mar 2025 13:24:35.191 # Internal error in RDB reading offset 0, function at rdb.c:3390 -> Unexpected EOF reading RDB file. Failure loading rdb format from socket, assuming connection error, resuming operation.} 605 -{2806938:S 09 Mar 2025 13:24:35.328 # Internal error in RDB reading offset 0, function at rdb.c:3390 -> Unexpected EOF reading RDB file. Failure loading rdb format from socket, assuming connection error, resuming operation.} 620 -{2806938:S 09 Mar 2025 13:24:35.419 # Internal error in RDB reading offset 0, function at rdb.c:3390 -> Unexpected EOF reading RDB file. Failure loading rdb format from socket, assuming connection error, resuming operation.} 635 -{2806938:S 09 Mar 2025 13:24:35.546 # Internal error in RDB reading offset 0, function at rdb.c:3390 -> Unexpected EOF reading RDB file. Failure loading rdb format from socket, assuming connection error, resuming operation.} 650 -{2806938:S 09 Mar 2025 13:24:35.676 # Internal error in RDB reading offset 0, function at rdb.c:3390 -> Unexpected EOF reading RDB file. Failure loading rdb format from socket, assuming connection error, resuming operation.} 665 -{2806938:S 09 Mar 2025 13:24:35.782 # Internal error in RDB reading offset 0, function at rdb.c:3390 -> Unexpected EOF reading RDB file. Failure loading rdb format from socket, assuming connection error, resuming operation.} 680 -{2806938:S 09 Mar 2025 13:24:35.875 # Internal error in RDB reading offset 0, function at rdb.c:3390 -> Unexpected EOF reading RDB file. Failure loading rdb format from socket, assuming connection error, resuming operation.} 695 -{2806938:S 09 Mar 2025 13:24:35.990 # Internal error in RDB reading offset 0, function at rdb.c:3390 -> Unexpected EOF reading RDB file. Failure loading rdb format from socket, assuming connection error, resuming operation.} 710 -{2806938:S 09 Mar 2025 13:24:36.126 # Internal error in RDB reading offset 0, function at rdb.c:3390 -> Unexpected EOF reading RDB file. Failure loading rdb format from socket, assuming connection error, resuming operation.} 725 -{2806938:S 09 Mar 2025 13:24:36.230 # Internal error in RDB reading offset 0, function at rdb.c:3390 -> Unexpected EOF reading RDB file. Failure loading rdb format from socket, assuming connection error, resuming operation.} 740 -{2806938:S 09 Mar 2025 13:24:36.343 # Internal error in RDB reading offset 0, function at rdb.c:3390 -> Unexpected EOF reading RDB file. Failure loading rdb format from socket, assuming connection error, resuming operation.} 755 -{2806938:S 09 Mar 2025 13:24:36.441 # Internal error in RDB reading offset 2035644, function at rdb.c:3390 -> Unexpected EOF reading RDB file. Failure loading rdb format from socket, assuming connection error, resuming operation.} 770 -{2806938:S 09 Mar 2025 13:24:36.527 # Internal error in RDB reading offset 0, function at rdb.c:3390 -> Unexpected EOF reading RDB file. Failure loading rdb format from socket, assuming connection error, resuming operation.} 785 -{2806938:S 09 Mar 2025 13:24:36.635 # Internal error in RDB reading offset 0, function at rdb.c:3390 -> Unexpected EOF reading RDB file. Failure loading rdb format from socket, assuming connection error, resuming operation.} 800 -{2806938:S 09 Mar 2025 13:24:36.771 # Internal error in RDB reading offset 0, function at rdb.c:3390 -> Unexpected EOF reading RDB file. Failure loading rdb format from socket, assuming connection error, resuming operation.} 815 -{2806938:S 09 Mar 2025 13:24:36.902 # Internal error in RDB reading offset 0, function at rdb.c:3390 -> Unexpected EOF reading RDB file. Failure loading rdb format from socket, assuming connection error, resuming operation.} 830 -{2806938:S 09 Mar 2025 13:24:37.042 # Internal error in RDB reading offset 0, function at rdb.c:3390 -> Unexpected EOF reading RDB file. Failure loading rdb format from socket, assuming connection error, resuming operation.} 845 -{2806938:S 09 Mar 2025 13:24:37.191 # Internal error in RDB reading offset 2035644, function at rdb.c:3390 -> Unexpected EOF reading RDB file. Failure loading rdb format from socket, assuming connection error, resuming operation.} 860 -{2806938:S 09 Mar 2025 13:24:37.298 # Internal error in RDB reading offset 0, function at rdb.c:3390 -> Unexpected EOF reading RDB file. Failure loading rdb format from socket, assuming connection error, resuming operation.} 875 -{2806938:S 09 Mar 2025 13:24:37.446 # Internal error in RDB reading offset 0, function at rdb.c:3390 -> Unexpected EOF reading RDB file. Failure loading rdb format from socket, assuming connection error, resuming operation.} 890 -{2806938:S 09 Mar 2025 13:24:37.586 # Internal error in RDB reading offset 0, function at rdb.c:3390 -> Unexpected EOF reading RDB file. Failure loading rdb format from socket, assuming connection error, resuming operation.} 905 -{2806938:S 09 Mar 2025 13:24:37.730 # Internal error in RDB reading offset 0, function at rdb.c:3390 -> Unexpected EOF reading RDB file. Failure loading rdb format from socket, assuming connection error, resuming operation.} 920 -{2806938:S 09 Mar 2025 13:24:37.871 # Internal error in RDB reading offset 0, function at rdb.c:3390 -> Unexpected EOF reading RDB file. Failure loading rdb format from socket, assuming connection error, resuming operation.} 935 -{2806938:S 09 Mar 2025 13:24:37.984 # Internal error in RDB reading offset 0, function at rdb.c:3390 -> Unexpected EOF reading RDB file. Failure loading rdb format from socket, assuming connection error, resuming operation.} 950 -{2806938:S 09 Mar 2025 13:24:38.107 # Internal error in RDB reading offset 0, function at rdb.c:3390 -> Unexpected EOF reading RDB file. Failure loading rdb format from socket, assuming connection error, resuming operation.} 965 -{2806938:S 09 Mar 2025 13:24:38.205 # Internal error in RDB reading offset 0, function at rdb.c:3390 -> Unexpected EOF reading RDB file. Failure loading rdb format from socket, assuming connection error, resuming operation.} 980 -{2806938:S 09 Mar 2025 13:24:38.316 # Internal error in RDB reading offset 0, function at rdb.c:3390 -> Unexpected EOF reading RDB file. Failure loading rdb format from socket, assuming connection error, resuming operation.} 995 -{2806938:S 09 Mar 2025 13:24:38.479 # Internal error in RDB reading offset 0, function at rdb.c:3390 -> Unexpected EOF reading RDB file. Failure loading rdb format from socket, assuming connection error, resuming operation.} 1010 -{2806938:S 09 Mar 2025 13:24:38.602 # Internal error in RDB reading offset 0, function at rdb.c:3390 -> Unexpected EOF reading RDB file. Failure loading rdb format from socket, assuming connection error, resuming operation.} 1025 -{2806938:S 09 Mar 2025 13:24:38.749 # Internal error in RDB reading offset 0, function at rdb.c:3390 -> Unexpected EOF reading RDB file. Failure loading rdb format from socket, assuming connection error, resuming operation.} 1040 -{2806938:S 09 Mar 2025 13:24:38.866 # Internal error in RDB reading offset 0, function at rdb.c:3390 -> Unexpected EOF reading RDB file. Failure loading rdb format from socket, assuming connection error, resuming operation.} 1055 -{2806938:S 09 Mar 2025 13:24:38.987 # Internal error in RDB reading offset 0, function at rdb.c:3390 -> Unexpected EOF reading RDB file. Failure loading rdb format from socket, assuming connection error, resuming operation.} 1070 -{2806938:S 09 Mar 2025 13:24:39.080 # Internal error in RDB reading offset 0, function at rdb.c:3390 -> Unexpected EOF reading RDB file. Failure loading rdb format from socket, assuming connection error, resuming operation.} 1085 -{2806938:S 09 Mar 2025 13:24:39.215 # Internal error in RDB reading offset 0, function at rdb.c:3390 -> Unexpected EOF reading RDB file. Failure loading rdb format from socket, assuming connection error, resuming operation.} 1100 -{2806938:S 09 Mar 2025 13:24:39.311 # Internal error in RDB reading offset 0, function at rdb.c:3390 -> Unexpected EOF reading RDB file. Failure loading rdb format from socket, assuming connection error, resuming operation.} 1115 -{2806938:S 09 Mar 2025 13:24:39.419 # Internal error in RDB reading offset 0, function at rdb.c:3390 -> Unexpected EOF reading RDB file. Failure loading rdb format from socket, assuming connection error, resuming operation.} 1130 -{2806938:S 09 Mar 2025 13:24:39.562 # Internal error in RDB reading offset 0, function at rdb.c:3390 -> Unexpected EOF reading RDB file. Failure loading rdb format from socket, assuming connection error, resuming operation.} 1145 -{2806938:S 09 Mar 2025 13:24:39.653 # Internal error in RDB reading offset 2035644, function at rdb.c:3390 -> Unexpected EOF reading RDB file. Failure loading rdb format from socket, assuming connection error, resuming operation.} 1160 -{2806938:S 09 Mar 2025 13:24:39.817 # Internal error in RDB reading offset 2035644, function at rdb.c:3390 -> Unexpected EOF reading RDB file. Failure loading rdb format from socket, assuming connection error, resuming operation.} 1175 -{2806938:S 09 Mar 2025 13:24:40.704 # Internal error in RDB reading offset 10139628, function at rdb.c:3390 -> Unexpected EOF reading RDB file. Failure loading rdb format from socket, assuming connection error, resuming operation.} 1190 -{2806938:S 09 Mar 2025 13:24:40.859 # Internal error in RDB reading offset 0, function at rdb.c:3390 -> Unexpected EOF reading RDB file. Failure loading rdb format from socket, assuming connection error, resuming operation.} 1205 -{2806938:S 09 Mar 2025 13:24:41.027 # Internal error in RDB reading offset 0, function at rdb.c:3390 -> Unexpected EOF reading RDB file. Failure loading rdb format from socket, assuming connection error, resuming operation.} 1220 -{2806938:S 09 Mar 2025 13:24:41.208 # Internal error in RDB reading offset 3877902, function at rdb.c:3390 -> Unexpected EOF reading RDB file. Failure loading rdb format from socket, assuming connection error, resuming operation.} 1235 -{2806938:S 09 Mar 2025 13:24:41.273 # Internal error in RDB reading offset 0, function at rdb.c:3390 -> Unexpected EOF reading RDB file. Failure loading rdb format from socket, assuming connection error, resuming operation.} 1250 -{2806938:S 09 Mar 2025 13:24:41.378 # Internal error in RDB reading offset 3877902, function at rdb.c:3390 -> Unexpected EOF reading RDB file. Failure loading rdb format from socket, assuming connection error, resuming operation.} 1265 -{2806938:S 09 Mar 2025 13:24:41.440 # Internal error in RDB reading offset 0, function at rdb.c:3390 -> Unexpected EOF reading RDB file. Failure loading rdb format from socket, assuming connection error, resuming operation.} 1280 -{2806938:S 09 Mar 2025 13:24:41.556 # Internal error in RDB reading offset 0, function at rdb.c:3390 -> Unexpected EOF reading RDB file. Failure loading rdb format from socket, assuming connection error, resuming operation.} 1295 -{2806938:S 09 Mar 2025 13:24:41.668 # Internal error in RDB reading offset 0, function at rdb.c:3390 -> Unexpected EOF reading RDB file. Failure loading rdb format from socket, assuming connection error, resuming operation.} 1310 -{2806938:S 09 Mar 2025 13:24:41.733 # Internal error in RDB reading offset 0, function at rdb.c:3390 -> Unexpected EOF reading RDB file. Failure loading rdb format from socket, assuming connection error, resuming operation.} 1325 -{2806938:S 09 Mar 2025 13:24:41.809 # Internal error in RDB reading offset 0, function at rdb.c:3390 -> Unexpected EOF reading RDB file. Failure loading rdb format from socket, assuming connection error, resuming operation.} 1340 -{2806938:S 09 Mar 2025 13:24:41.923 # Internal error in RDB reading offset 0, function at rdb.c:3390 -> Unexpected EOF reading RDB file. Failure loading rdb format from socket, assuming connection error, resuming operation.} 1355 -{2806938:S 09 Mar 2025 13:24:42.071 # Internal error in RDB reading offset 0, function at rdb.c:3390 -> Unexpected EOF reading RDB file. Failure loading rdb format from socket, assuming connection error, resuming operation.} 1370 -{2806938:S 09 Mar 2025 13:24:42.174 # Internal error in RDB reading offset 0, function at rdb.c:3390 -> Unexpected EOF reading RDB file. Failure loading rdb format from socket, assuming connection error, resuming operation.} 1385 -{2806938:S 09 Mar 2025 13:24:42.251 # Internal error in RDB reading offset 0, function at rdb.c:3390 -> Unexpected EOF reading RDB file. Failure loading rdb format from socket, assuming connection error, resuming operation.} 1400 -{2806938:S 09 Mar 2025 13:24:42.359 # Internal error in RDB reading offset 0, function at rdb.c:3390 -> Unexpected EOF reading RDB file. Failure loading rdb format from socket, assuming connection error, resuming operation.} 1415 -{2806938:S 09 Mar 2025 13:24:42.495 # Internal error in RDB reading offset 0, function at rdb.c:3390 -> Unexpected EOF reading RDB file. Failure loading rdb format from socket, assuming connection error, resuming operation.} 1430 -{2806938:S 09 Mar 2025 13:24:42.618 # Internal error in RDB reading offset 0, function at rdb.c:3390 -> Unexpected EOF reading RDB file. Failure loading rdb format from socket, assuming connection error, resuming operation.} 1445 -{2806938:S 09 Mar 2025 13:24:42.747 # Internal error in RDB reading offset 0, function at rdb.c:3390 -> Unexpected EOF reading RDB file. Failure loading rdb format from socket, assuming connection error, resuming operation.} 1460 -{2806938:S 09 Mar 2025 13:24:42.875 # Internal error in RDB reading offset 0, function at rdb.c:3390 -> Unexpected EOF reading RDB file. Failure loading rdb format from socket, assuming connection error, resuming operation.} 1475 -{2806938:S 09 Mar 2025 13:24:42.985 # Internal error in RDB reading offset 0, function at rdb.c:3390 -> Unexpected EOF reading RDB file. Failure loading rdb format from socket, assuming connection error, resuming operation.} 1490 -{2806938:S 09 Mar 2025 13:24:43.116 # Internal error in RDB reading offset 0, function at rdb.c:3390 -> Unexpected EOF reading RDB file. Failure loading rdb format from socket, assuming connection error, resuming operation.} 1505 -{2806938:S 09 Mar 2025 13:24:43.281 # Internal error in RDB reading offset 0, function at rdb.c:3390 -> Unexpected EOF reading RDB file. Failure loading rdb format from socket, assuming connection error, resuming operation.} 1520 -{2806938:S 09 Mar 2025 13:24:43.486 # Internal error in RDB reading offset 0, function at rdb.c:3390 -> Unexpected EOF reading RDB file. Failure loading rdb format from socket, assuming connection error, resuming operation.} 1535 -test took 14609 ms -[ok]: diskless loading short read (24872 ms) -=== (repl external:skip) Starting server 127.0.0.1:22909 ok -=== () Starting server 127.0.0.1:22911 ok -=== () Starting server 127.0.0.1:22913 ok -[ok]: diskless no replicas drop during rdb pipe (38828 ms) -=== () Starting server 127.0.0.1:22915 ok -=== () Starting server 127.0.0.1:22917 ok -[ok]: diskless slow replicas drop during rdb pipe (11901 ms) -=== () Starting server 127.0.0.1:22919 ok -=== () Starting server 127.0.0.1:22921 ok -[ok]: diskless fast replicas drop during rdb pipe (43814 ms) -=== () Starting server 127.0.0.1:22923 ok -=== () Starting server 127.0.0.1:22925 ok -[ok]: diskless all replicas drop during rdb pipe (2582 ms) -=== () Starting server 127.0.0.1:22927 ok -=== () Starting server 127.0.0.1:22929 ok -[ok]: diskless timeout replicas drop during rdb pipe (19617 ms) -=== (repl) Starting server 127.0.0.1:22931 ok -=== () Starting server 127.0.0.1:22933 ok -[ok]: diskless replication child being killed is collected (3022 ms) -=== (repl) Starting server 127.0.0.1:22935 ok -=== () Starting server 127.0.0.1:22937 ok -[ok]: replication child dies when parent is killed - diskless: yes dual-channel-replication-enabled: yes (4132 ms) -=== (repl) Starting server 127.0.0.1:22939 ok -=== () Starting server 127.0.0.1:22941 ok -[ok]: replication child dies when parent is killed - diskless: no dual-channel-replication-enabled: no (4952 ms) -=== (repl) Starting server 127.0.0.1:22943 ok -=== () Starting server 127.0.0.1:22945 ok -[ok]: diskless replication read pipe cleanup (2184 ms) -=== (repl) Starting server 127.0.0.1:22947 ok -=== () Starting server 127.0.0.1:22949 ok -=== () Starting server 127.0.0.1:22951 ok -[ok]: replicaof right after disconnection (2792 ms) -=== (repl) Starting server 127.0.0.1:22953 ok -=== () Starting server 127.0.0.1:22955 ok -=== () Starting server 127.0.0.1:22957 ok -[ok]: Kill rdb child process if its dumping RDB is not useful (2264 ms) -=== (repl external:skip) Starting server 127.0.0.1:22959 ok -=== () Starting server 127.0.0.1:22961 ok -=== () Starting server 127.0.0.1:22963 ok -=== () Starting server 127.0.0.1:22965 ok -[ok]: Don't disconnect with replicas before loading transferred RDB when full sync with dual-channel-replication yes (29 ms) -[ok]: Discard cache master before loading transferred RDB when full sync with dual-channel-replication yes (200 ms) -=== (repl external:skip) Starting server 127.0.0.1:22967 ok -=== () Starting server 127.0.0.1:22969 ok -=== () Starting server 127.0.0.1:22971 ok -=== () Starting server 127.0.0.1:22973 ok -[ok]: Don't disconnect with replicas before loading transferred RDB when full sync with dual-channel-replication no (37 ms) -[ok]: Discard cache master before loading transferred RDB when full sync with dual-channel-replication no (204 ms) -=== (repl) Starting server 127.0.0.1:22975 ok -=== () Starting server 127.0.0.1:22977 ok -[ok]: replica can handle EINTR if use diskless load (1692 ms) -=== (repl external:skip) Starting server 127.0.0.1:22979 ok -SYNC _addReplyToBufferOrList: I/O error reading reply - -Waiting for background save to finish... [ok]: replica do not write the reply to the replication link - SYNC (_addReplyToBufferOrList) (196 ms) -SYNC addReplyDeferredLen: I/O error reading reply - -Waiting for background save to finish... [ok]: replica do not write the reply to the replication link - SYNC (addReplyDeferredLen) (196 ms) -PSYNC _addReplyToBufferOrList: I/O error reading reply - -Waiting for background save to finish... [ok]: replica do not write the reply to the replication link - PSYNC (_addReplyToBufferOrList) (188 ms) -PSYNC addReplyDeferredLen: I/O error reading reply -[ok]: replica do not write the reply to the replication link - PSYNC (addReplyDeferredLen) (113 ms) -[ok]: PSYNC with wrong offset should throw error (26 ms) -=== (repl external:skip) Starting server 127.0.0.1:22981 ok -=== () Starting server 127.0.0.1:22983 ok -[ok]: Test replication with lazy expire (191 ms) -=== (repl external:skip) Starting server 127.0.0.1:22985 ok -=== () Starting server 127.0.0.1:22987 ok -[ok]: Replica keep the old data if RDB file save fails in disk-based replication (293 ms) -[113/124 done]: integration/replication (330 seconds) -Testing integration/aof-race -=== () Starting server 127.0.0.1:22989 ok -=== () Starting server 127.0.0.1:22991 ok -[114/124 done]: integration/aof-race (6 seconds) -Testing integration/dual-channel-replication -=== (dual-channel-replication external:skip) Starting server 127.0.0.1:22993 ok -=== () Starting server 127.0.0.1:22995 ok -[ok]: Test dual-channel-replication-enabled replica enters handshake (12 ms) -[ok]: Test dual-channel-replication-enabled enters wait_bgsave (1016 ms) -[ok]: Test dual-channel-replication-enabled replica is able to sync (930 ms) -=== (dual-channel-replication external:skip) Starting server 127.0.0.1:22997 ok -=== () Starting server 127.0.0.1:22999 ok -[ok]: Primary memory usage does not increase during dual-channel-replication sync (5267 ms) -[ok]: Steady state after dual channel sync (1048 ms) -[ok]: Dual channel replication sync doesn't impair subsequent normal syncs (1983 ms) -=== (dual-channel-replication external:skip) Starting server 127.0.0.1:23001 ok -=== () Starting server 127.0.0.1:23003 ok -[ok]: Toggle dual-channel-replication-enabled: yes start (2236 ms) -[ok]: Online toggle dual-channel-replication-enabled on primary, starting with 'yes' (2712 ms) -[ok]: Online toggle dual-channel-replication-enabled on replica, starting with 'yes' (11803 ms) -[ok]: Toggle dual-channel-replication-enabled: no start (12951 ms) -[ok]: Online toggle dual-channel-replication-enabled on primary, starting with 'no' (2563 ms) -[ok]: Online toggle dual-channel-replication-enabled on replica, starting with 'no' (2604 ms) -=== (dual-channel-replication external:skip) Starting server 127.0.0.1:23005 ok -=== () Starting server 127.0.0.1:23007 ok -=== () Starting server 127.0.0.1:23009 ok -[ok]: dual-channel-replication with multiple replicas (5439 ms) -[ok]: Test diverse replica sync: dual-channel on/off (5564 ms) -[ok]: Test replica's buffer limit reached (5440 ms) -[ok]: dual-channel-replication fails when primary diskless disabled (1049 ms) -=== (dual-channel-replication external:skip) Starting server 127.0.0.1:23011 ok -=== () Starting server 127.0.0.1:23013 ok -[ok]: Test dual-channel-replication sync- psync established after rdb load (388 ms) -=== (dual-channel-replication external:skip) Starting server 127.0.0.1:23015 ok -=== () Starting server 127.0.0.1:23017 ok -[ok]: dual-channel-replication: Primary COB growth with inactive replica (33474 ms) -=== (dual-channel-replication external:skip) Starting server 127.0.0.1:23019 ok -=== () Starting server 127.0.0.1:23021 ok -=== () Starting server 127.0.0.1:23023 ok -[ok]: Test dual-channel: primary tracking replica backlog refcount - start with empty backlog (4095 ms) -[ok]: Test dual-channel: primary tracking replica backlog refcount - start with backlog (60939 ms) -=== (dual-channel-replication external:skip) Starting server 127.0.0.1:23025 ok -=== () Starting server 127.0.0.1:23027 ok -[ok]: Psync established after rdb load - within grace period (540 ms) -=== (dual-channel-replication external:skip) Starting server 127.0.0.1:23029 ok -=== () Starting server 127.0.0.1:23031 ok -[ok]: Psync established after RDB load - beyond grace period (2512 ms) -Waiting for process 2841803 to exit... -=== (dual-channel-replication external:skip) Starting server 127.0.0.1:23033 ok -=== () Starting server 127.0.0.1:23035 ok -[ok]: Test dual-channel-replication primary gets cob overrun before established psync (1580 ms) -[ok]: Test dual-channel-replication primary gets cob overrun during replica rdb load (1182 ms) -=== (dual-channel-replication external:skip) Starting server 127.0.0.1:23037 ok -=== () Starting server 127.0.0.1:23039 ok -=== () Starting server 127.0.0.1:23041 ok -[ok]: Sync should continue if not all slaves dropped dual-channel-replication yes (32891 ms) -[ok]: Primary abort sync if all slaves dropped dual-channel-replication yes (6108 ms) -=== (dual-channel-replication external:skip) Starting server 127.0.0.1:23043 ok -=== () Starting server 127.0.0.1:23045 ok -=== () Starting server 127.0.0.1:23047 ok -[ok]: Sync should continue if not all slaves dropped dual-channel-replication no (17196 ms) -[ok]: Primary abort sync if all slaves dropped dual-channel-replication no (6089 ms) -=== (dual-channel-replication external:skip) Starting server 127.0.0.1:23049 ok -=== () Starting server 127.0.0.1:23051 ok -[ok]: Test dual-channel-replication replica main channel disconnected (5429 ms) -[ok]: Test dual channel replication slave of no one after main conn kill (2 ms) -[ok]: Test dual-channel-replication replica rdb connection disconnected (11639 ms) -[ok]: Test dual channel replication slave of no one after rdb conn kill (3 ms) -[ok]: Test dual-channel-replication primary reject set-rdb-client after client killed (12336 ms) -=== (dual-channel-replication external:skip) Starting server 127.0.0.1:23053 ok -=== () Starting server 127.0.0.1:23055 ok -=== () Starting server 127.0.0.1:23057 ok -[ok]: Test replica unable to join dual channel replication sync after started (1121 ms) -=== (dual-channel-replication external:skip) Starting server 127.0.0.1:23059 ok -=== () Starting server 127.0.0.1:23061 ok -[ok]: Replica recover rdb-connection killed (10084 ms) -[ok]: Replica recover main-connection killed (10040 ms) -[115/124 done]: integration/dual-channel-replication (326 seconds) -Testing integration/aof-multi-part -=== () Starting server 127.0.0.1:23063 -[ok]: Multi Part AOF can't load data when some file missing (125 ms) -=== () Starting server 127.0.0.1:23065 -[ok]: Multi Part AOF can't load data when the sequence not increase monotonically (131 ms) -=== () Starting server 127.0.0.1:23067 -[ok]: Multi Part AOF can't load data when there are blank lines in the manifest file (142 ms) -=== () Starting server 127.0.0.1:23069 -[ok]: Multi Part AOF can't load data when there is a duplicate base file (199 ms) -=== () Starting server 127.0.0.1:23071 -[ok]: Multi Part AOF can't load data when the manifest format is wrong (type unknown) (130 ms) -=== () Starting server 127.0.0.1:23073 -[ok]: Multi Part AOF can't load data when the manifest format is wrong (missing key) (137 ms) -=== () Starting server 127.0.0.1:23075 -[ok]: Multi Part AOF can't load data when the manifest format is wrong (line too short) (133 ms) -=== () Starting server 127.0.0.1:23077 -[ok]: Multi Part AOF can't load data when the manifest format is wrong (line too long) (126 ms) -=== () Starting server 127.0.0.1:23079 -[ok]: Multi Part AOF can't load data when the manifest format is wrong (odd parameter) (132 ms) -=== () Starting server 127.0.0.1:23081 -[ok]: Multi Part AOF can't load data when the manifest file is empty (137 ms) -=== () Starting server 127.0.0.1:23083 ok -[ok]: Multi Part AOF can start when no aof and no manifest (793 ms) -=== () Starting server 127.0.0.1:23085 ok -[ok]: Multi Part AOF can start when we have en empty AOF dir (614 ms) -=== () Starting server 127.0.0.1:23087 ok -[ok]: Multi Part AOF can load data discontinuously increasing sequence (696 ms) -=== () Starting server 127.0.0.1:23089 ok -[ok]: Multi Part AOF can load data when manifest add new k-v (839 ms) -=== () Starting server 127.0.0.1:23091 ok -[ok]: Multi Part AOF can load data when some AOFs are empty (272 ms) -=== () Starting server 127.0.0.1:23093 ok - -Waiting for background AOF rewrite to finish... [ok]: Multi Part AOF can load data from old version valkey (rdb preamble no) (394 ms) -=== () Starting server 127.0.0.1:23095 ok - -Waiting for background AOF rewrite to finish... [ok]: Multi Part AOF can load data from old version valkey (rdb preamble yes) (385 ms) -=== () Starting server 127.0.0.1:23097 ok -[ok]: Multi Part AOF can continue the upgrade from the interrupted upgrade state (262 ms) -=== () Starting server 127.0.0.1:23099 ok -[ok]: Multi Part AOF can be loaded correctly when both server dir and aof dir contain old AOF (256 ms) -=== () Starting server 127.0.0.1:23101 -[ok]: Multi Part AOF can't load data when the manifest contains the old AOF file name but the file does not exist in server dir and aof dir (136 ms) -=== () Starting server 127.0.0.1:23103 ok -=== () Starting server 127.0.0.1:23105 ok - -Waiting for background AOF rewrite to finish... -Waiting for background AOF rewrite to finish... [ok]: Multi Part AOF can upgrade when when two servers share the same server dir (server1) (134 ms) - -Waiting for background AOF rewrite to finish... [ok]: Multi Part AOF can upgrade when when two servers share the same server dir (server2) (89 ms) -[ok]: Multi Part AOF can upgrade when when two servers share the same server dir (797 ms) -=== () Starting server 127.0.0.1:23107 ok - -Waiting for background AOF rewrite to finish... -Waiting for background AOF rewrite to finish... [ok]: Multi Part AOF can handle appendfilename contains whitespaces (375 ms) -=== () Starting server 127.0.0.1:23109 ok -[ok]: Multi Part AOF can create BASE (RDB format) when server starts from empty (252 ms) -=== () Starting server 127.0.0.1:23111 ok -[ok]: Multi Part AOF can create BASE (AOF format) when server starts from empty (268 ms) -=== () Starting server 127.0.0.1:23113 ok - -Waiting for background AOF rewrite to finish... -Waiting for background AOF rewrite to finish... -Waiting for background AOF rewrite to finish... -Waiting for background AOF rewrite to finish... -Waiting for background AOF rewrite to finish... -Waiting for background AOF rewrite to finish... -Waiting for background AOF rewrite to finish... -Waiting for background AOF rewrite to finish... -Waiting for background AOF rewrite to finish... -Waiting for background AOF rewrite to finish... -Waiting for background AOF rewrite to finish... -Waiting for background AOF rewrite to finish... -Waiting for background AOF rewrite to finish... -Waiting for background AOF rewrite to finish... -Waiting for background AOF rewrite to finish... -Waiting for background AOF rewrite to finish... -Waiting for background AOF rewrite to finish... -Waiting for background AOF rewrite to finish... -Waiting for background AOF rewrite to finish... -Waiting for background AOF rewrite to finish... -Waiting for background AOF rewrite to finish... [ok]: Changing aof-use-rdb-preamble during rewrite process (1389 ms) -=== (Multi Part AOF) Starting server 127.0.0.1:23115 ok -[ok]: Make sure aof manifest appendonly.aof.manifest not in aof directory (1 ms) - -Waiting for background AOF rewrite to finish... -Waiting for background AOF rewrite to finish... -Waiting for background AOF rewrite to finish... [ok]: AOF enable will create manifest file (337 ms) - -Waiting for background AOF rewrite to finish... -Waiting for background AOF rewrite to finish... -Waiting for background AOF rewrite to finish... -Waiting for background AOF rewrite to finish... -Waiting for background AOF rewrite to finish... -Waiting for background AOF rewrite to finish... -Waiting for background AOF rewrite to finish... -Waiting for background AOF rewrite to finish... [ok]: AOF multiple rewrite failures will open multiple INCR AOFs (1387 ms) - -Waiting for background AOF rewrite to finish... -Waiting for background AOF rewrite to finish... -Waiting for background AOF rewrite to finish... -Waiting for background AOF rewrite to finish... -Waiting for background AOF rewrite to finish... -Waiting for background AOF rewrite to finish... [ok]: AOF rewrite doesn't open new aof when AOF turn off (488 ms) - -Waiting for background AOF rewrite to finish... -Waiting for background AOF rewrite to finish... -Waiting for background AOF rewrite to finish... [ok]: AOF enable/disable auto gc (182 ms) - -Waiting for background AOF rewrite to finish... -Waiting for background AOF rewrite to finish... [ok]: AOF can produce consecutive sequence number after reload (137 ms) - -Waiting for background AOF rewrite to finish... -Waiting for background AOF rewrite to finish... [ok]: AOF enable during BGSAVE will not write data util AOFRW finish (442 ms) - -Waiting for background AOF rewrite to finish... -Waiting for background AOF rewrite to finish... -Waiting for background AOF rewrite to finish... -Waiting for background AOF rewrite to finish... -Waiting for background AOF rewrite to finish... -Waiting for background AOF rewrite to finish... -Waiting for background AOF rewrite to finish... -Waiting for background AOF rewrite to finish... [ok]: AOF will trigger limit when AOFRW fails many times (833 ms) -=== () Starting server 127.0.0.1:23117 ok - -Waiting for background AOF rewrite to finish... -Waiting for background AOF rewrite to finish... -Waiting for background AOF rewrite to finish... -Waiting for background AOF rewrite to finish... [ok]: AOF will open a temporary INCR AOF to accumulate data until the first AOFRW success when AOF is dynamically enabled (2543 ms) -[116/124 done]: integration/aof-multi-part (16 seconds) -Testing integration/psync2 -=== (psync2 external:skip) Starting server 127.0.0.1:23119 ok -=== () Starting server 127.0.0.1:23121 ok -=== () Starting server 127.0.0.1:23123 ok -=== () Starting server 127.0.0.1:23125 ok -=== () Starting server 127.0.0.1:23127 ok -[ok]: PSYNC2: --- CYCLE 1 --- (0 ms) -[ok]: PSYNC2: [NEW LAYOUT] Set #3 as master (1 ms) -[ok]: PSYNC2: Set #4 to replicate from #3 (7 ms) -[ok]: PSYNC2: Set #0 to replicate from #3 (2 ms) -[ok]: PSYNC2: Set #1 to replicate from #3 (2 ms) -[ok]: PSYNC2: Set #2 to replicate from #3 (1 ms) -[ok]: PSYNC2: cluster is consistent after failover (3 ms) -[ok]: PSYNC2: generate load while killing replication links (5001 ms) -[ok]: PSYNC2: cluster is consistent after load (x = 23176) (21 ms) -[ok]: PSYNC2: total sum of full synchronizations is exactly 4 (4 ms) -[ok]: PSYNC2: --- CYCLE 2 --- (0 ms) -[ok]: PSYNC2: [NEW LAYOUT] Set #4 as master (0 ms) -[ok]: PSYNC2: Set #0 to replicate from #4 (1 ms) -[ok]: PSYNC2: Set #1 to replicate from #4 (1 ms) -[ok]: PSYNC2: Set #2 to replicate from #4 (1 ms) -[ok]: PSYNC2: Set #3 to replicate from #1 (1 ms) -[ok]: PSYNC2: cluster is consistent after failover (4 ms) -[ok]: PSYNC2: generate load while killing replication links (5000 ms) -[ok]: PSYNC2: cluster is consistent after load (x = 40869) (1 ms) -[ok]: PSYNC2: total sum of full synchronizations is exactly 4 (2 ms) -[ok]: PSYNC2: --- CYCLE 3 --- (0 ms) -[ok]: PSYNC2: [NEW LAYOUT] Set #0 as master (0 ms) -[ok]: PSYNC2: Set #4 to replicate from #1 (2 ms) -[ok]: PSYNC2: Set #1 to replicate from #2 (3 ms) -[ok]: PSYNC2: Set #2 to replicate from #0 (4 ms) -[ok]: PSYNC2: Set #3 to replicate from #2 (3 ms) -[ok]: PSYNC2: cluster is consistent after failover (10 ms) -[ok]: PSYNC2: generate load while killing replication links (5001 ms) -[ok]: PSYNC2: cluster is consistent after load (x = 56782) (7 ms) -[ok]: PSYNC2: total sum of full synchronizations is exactly 4 (3 ms) -[ok]: PSYNC2: --- CYCLE 4 --- (0 ms) -[ok]: PSYNC2: [NEW LAYOUT] Set #3 as master (1 ms) -[ok]: PSYNC2: Set #0 to replicate from #1 (2 ms) -[ok]: PSYNC2: Set #4 to replicate from #1 (7 ms) -[ok]: PSYNC2: Set #1 to replicate from #3 (2 ms) -[ok]: PSYNC2: Set #2 to replicate from #1 (5 ms) -[ok]: PSYNC2: cluster is consistent after failover (44 ms) -[ok]: PSYNC2: generate load while killing replication links (5016 ms) -[ok]: PSYNC2: cluster is consistent after load (x = 60107) (34 ms) -[ok]: PSYNC2: total sum of full synchronizations is exactly 4 (44 ms) -[ok]: PSYNC2: --- CYCLE 5 --- (0 ms) -[ok]: PSYNC2: [NEW LAYOUT] Set #3 as master (14 ms) -[ok]: PSYNC2: Set #4 to replicate from #0 (7 ms) -[ok]: PSYNC2: Set #0 to replicate from #1 (7 ms) -[ok]: PSYNC2: Set #1 to replicate from #3 (9 ms) -[ok]: PSYNC2: Set #2 to replicate from #1 (11 ms) -[ok]: PSYNC2: cluster is consistent after failover (25 ms) -[ok]: PSYNC2: generate load while killing replication links (5002 ms) -[ok]: PSYNC2: cluster is consistent after load (x = 68020) (22 ms) -[ok]: PSYNC2: total sum of full synchronizations is exactly 4 (19 ms) -[ok]: PSYNC2: Bring the master back again for next test (37 ms) -[ok]: PSYNC2: Partial resync after restart using RDB aux fields (515 ms) -[117/124 done]: integration/psync2 (45 seconds) -Testing integration/replica-redirect -=== (needs:repl external:skip) Starting server 127.0.0.1:23129 ok -=== () Starting server 127.0.0.1:23131 ok -[ok]: write command inside MULTI is QUEUED, EXEC should be REDIRECT (179 ms) -[ok]: write command inside MULTI is REDIRECT, EXEC should be EXECABORT (68 ms) -[ok]: replica allow read command by default (7 ms) -[ok]: replica reply READONLY error for write command by default (11 ms) -[ok]: replica redirect read and write command after CLIENT CAPA REDIRECT (16 ms) -[ok]: non-data access commands are not redirected (8 ms) -[ok]: replica allow read command in READONLY mode (14 ms) -[ok]: client paused during failover-in-progress (592 ms) -[118/124 done]: integration/replica-redirect (3 seconds) -Testing integration/convert-ziplist-zset-on-load -=== () Starting server 127.0.0.1:23133 ok -[ok]: RDB load ziplist zset: converts to listpack when RDB loading (32 ms) -=== () Starting server 127.0.0.1:23135 ok -[ok]: RDB load ziplist zset: converts to skiplist when zset-max-ziplist-entries is exceeded (28 ms) -[119/124 done]: integration/convert-ziplist-zset-on-load (1 seconds) -Testing integration/dismiss-mem -=== (dismiss external:skip) Starting server 127.0.0.1:23137 ok - -Waiting for background AOF rewrite to finish... -Waiting for background AOF rewrite to finish... -Waiting for background AOF rewrite to finish... -Waiting for background AOF rewrite to finish... -Waiting for background AOF rewrite to finish... -Waiting for background AOF rewrite to finish... -Waiting for background AOF rewrite to finish... [ok]: dismiss all data types memory (2475 ms) - -Waiting for background save to finish... -Waiting for background save to finish... [ok]: dismiss client output buffer (1281 ms) - -Waiting for background save to finish... -Waiting for background save to finish... [ok]: dismiss client query buffer (286 ms) -=== () Starting server 127.0.0.1:23139 ok - -Waiting for background save to finish... -Waiting for background save to finish... [ok]: dismiss replication backlog (1857 ms) -[120/124 done]: integration/dismiss-mem (7 seconds) -Testing integration/aof -=== () Starting server 127.0.0.1:23141 ok -[ok]: Unfinished MULTI: Server should start if load-truncated is yes (11 ms) -=== () Starting server 127.0.0.1:23143 ok -[ok]: Short read: Server should start if load-truncated is yes (39 ms) -[ok]: Truncated AOF loaded: we expect foo to be equal to 5 (36 ms) -[ok]: Append a new command after loading an incomplete AOF (0 ms) -=== () Starting server 127.0.0.1:23145 ok -[ok]: Short read + command: Server should start (18 ms) -[ok]: Truncated AOF loaded: we expect foo to be equal to 6 now (39 ms) -=== () Starting server 127.0.0.1:23147 -[ok]: Bad format: Server should have logged an error (18 ms) -=== () Starting server 127.0.0.1:23149 -[ok]: Unfinished MULTI: Server should have logged an error (13 ms) -=== () Starting server 127.0.0.1:23151 -[ok]: Short read: Server should have logged an error (11 ms) -[ok]: Short read: Utility should confirm the AOF is not valid (123 ms) -[ok]: Short read: Utility should show the abnormal line num in AOF (60 ms) -[ok]: Short read: Utility should be able to fix the AOF (58 ms) -=== () Starting server 127.0.0.1:23153 ok -[ok]: Fixed AOF: Server should have been started (15 ms) -[ok]: Fixed AOF: Keyspace should contain values that were parseable (28 ms) -=== () Starting server 127.0.0.1:23155 ok -[ok]: AOF+SPOP: Server should have been started (18 ms) -[ok]: AOF+SPOP: Set should have 1 member (37 ms) -=== () Starting server 127.0.0.1:23157 ok -[ok]: AOF+SPOP: Server should have been started (6 ms) -[ok]: AOF+SPOP: Set should have 1 member (35 ms) -=== () Starting server 127.0.0.1:23159 ok -[ok]: AOF+EXPIRE: Server should have been started (23 ms) -[ok]: AOF+EXPIRE: List should be empty (35 ms) -=== () Starting server 127.0.0.1:23161 ok -[ok]: Server should not try to convert DEL into EXPIREAT for EXPIRE -1 (10 ms) -=== () Starting server 127.0.0.1:23163 ok -[ok]: AOF fsync always barrier issue (7917 ms) -=== () Starting server 127.0.0.1:23165 ok -[ok]: GETEX should not append to AOF (27 ms) -=== () Starting server 127.0.0.1:23167 -[ok]: Unknown command: Server should have logged an error (16 ms) -=== () Starting server 127.0.0.1:23169 ok -[ok]: AOF+LMPOP/BLMPOP: pop elements from the list (158 ms) -=== () Starting server 127.0.0.1:23171 ok -[ok]: AOF+LMPOP/BLMPOP: after pop elements from the list (21 ms) -=== () Starting server 127.0.0.1:23173 ok -[ok]: AOF+ZMPOP/BZMPOP: pop elements from the zset (192 ms) -=== () Starting server 127.0.0.1:23175 ok -[ok]: AOF+ZMPOP/BZMPOP: after pop elements from the zset (48 ms) -=== () Starting server 127.0.0.1:23177 ok - -Waiting for background AOF rewrite to finish... -Waiting for background AOF rewrite to finish... -Waiting for background AOF rewrite to finish... -Waiting for background AOF rewrite to finish... -Waiting for background AOF rewrite to finish... -Waiting for background AOF rewrite to finish... [ok]: Generate timestamp annotations in AOF (3156 ms) -=== () Starting server 127.0.0.1:23179 ok -[ok]: Successfully load AOF which has timestamp annotations inside (116 ms) -=== () Starting server 127.0.0.1:23181 ok -=== () Starting server 127.0.0.1:23183 ok -=== () Starting server 127.0.0.1:23185 ok -[ok]: Truncate AOF to specific timestamp (2980 ms) -=== () Starting server 127.0.0.1:23187 ok -[ok]: EVAL timeout with slow verbatim Lua script from AOF (4572 ms) -=== () Starting server 127.0.0.1:23189 ok -[ok]: EVAL can process writes from AOF in read-only replicas (905 ms) -[ok]: Test valkey-check-aof for old style resp AOF (170 ms) -[ok]: Test valkey-check-aof for old style resp AOF - has data in the same format as manifest (208 ms) -[ok]: Test valkey-check-aof for old style rdb-preamble AOF (49 ms) -[ok]: Test valkey-check-aof for Multi Part AOF with resp AOF base (95 ms) -[ok]: Test valkey-check-aof for Multi Part AOF with rdb-preamble AOF base (135 ms) -[ok]: Test valkey-check-aof for Multi Part AOF contains a format error (76 ms) -[ok]: Test valkey-check-aof only truncates the last file for Multi Part AOF in fix mode (207 ms) -[ok]: Test valkey-check-aof only truncates the last file for Multi Part AOF in truncate-to-timestamp mode (143 ms) -=== () Starting server 127.0.0.1:23191 ok -[ok]: FLUSHDB / FLUSHALL should persist in AOF (204 ms) -=== () Starting server 127.0.0.1:23193 ok -[ok]: Turning appendonly on and off within a transaction (81 ms) -=== () Starting server 127.0.0.1:23195 ok -[ok]: Test cluster slots / cluster shards in aof won't crash (612 ms) -[121/124 done]: integration/aof (33 seconds) -Testing integration/convert-zipmap-hash-on-load -=== () Starting server 127.0.0.1:23197 ok -[ok]: RDB load zipmap hash: converts to listpack (24 ms) -=== () Starting server 127.0.0.1:23199 ok -[ok]: RDB load zipmap hash: converts to hash table when hash-max-ziplist-entries is exceeded (31 ms) -=== () Starting server 127.0.0.1:23201 ok -[ok]: RDB load zipmap hash: converts to hash table when hash-max-ziplist-value is exceeded (27 ms) -[122/124 done]: integration/convert-zipmap-hash-on-load (2 seconds) -Testing integration/replication-buffer -=== (repl external:skip) Starting server 127.0.0.1:23203 ok -=== () Starting server 127.0.0.1:23205 ok -=== () Starting server 127.0.0.1:23207 ok -=== () Starting server 127.0.0.1:23209 ok -[ok]: All replicas share one global replication buffer dualchannel yes (454 ms) -[ok]: Replication buffer will become smaller when no replica uses dualchannel yes (28 ms) -=== (repl external:skip) Starting server 127.0.0.1:23211 ok -=== () Starting server 127.0.0.1:23213 ok -=== () Starting server 127.0.0.1:23215 ok -=== () Starting server 127.0.0.1:23217 ok -[ok]: All replicas share one global replication buffer dualchannel no (383 ms) -[ok]: Replication buffer will become smaller when no replica uses dualchannel no (47 ms) -=== (repl external:skip) Starting server 127.0.0.1:23219 ok -=== () Starting server 127.0.0.1:23221 ok -=== () Starting server 127.0.0.1:23223 ok -[ok]: Replication backlog size can outgrow the backlog limit config dualchannel yes (10563 ms) -[ok]: Replica could use replication buffer (beyond backlog config) for partial resynchronization dualchannel yes (9962 ms) -[ok]: Replication backlog memory will become smaller if disconnecting with replica dualchannel yes (716 ms) -=== (repl external:skip) Starting server 127.0.0.1:23225 ok -=== () Starting server 127.0.0.1:23227 ok -=== () Starting server 127.0.0.1:23229 ok -[ok]: Replication backlog size can outgrow the backlog limit config dualchannel no (10957 ms) -[ok]: Replica could use replication buffer (beyond backlog config) for partial resynchronization dualchannel no (9461 ms) -[ok]: Replication backlog memory will become smaller if disconnecting with replica dualchannel no (593 ms) -=== (repl external:skip) Starting server 127.0.0.1:23231 ok -=== () Starting server 127.0.0.1:23233 ok -[ok]: Partial resynchronization is successful even client-output-buffer-limit is less than repl-backlog-size. dualchannel yes (7000 ms) -=== (repl external:skip) Starting server 127.0.0.1:23235 ok -=== () Starting server 127.0.0.1:23237 ok -[ok]: The update of replBufBlock's repl_offset is ok - Regression test for #11666 (44 ms) -[ok]: Replica client-output-buffer size is limited to backlog_limit/16 when no replication data is pending. dualchannel yes (2780 ms) -=== (repl external:skip) Starting server 127.0.0.1:23239 ok -=== () Starting server 127.0.0.1:23241 ok -[ok]: Partial resynchronization is successful even client-output-buffer-limit is less than repl-backlog-size. dualchannel no (7135 ms) -=== (repl external:skip) Starting server 127.0.0.1:23243 ok -=== () Starting server 127.0.0.1:23245 ok -[ok]: The update of replBufBlock's repl_offset is ok - Regression test for #11666 (57 ms) -[ok]: Replica client-output-buffer size is limited to backlog_limit/16 when no replication data is pending. dualchannel no (2600 ms) -[123/124 done]: integration/replication-buffer (74 seconds) -Testing integration/psync2-master-restart -=== (psync2 external:skip) Starting server 127.0.0.1:23247 ok -=== () Starting server 127.0.0.1:23249 ok -=== () Starting server 127.0.0.1:23251 ok -[ok]: PSYNC2: Partial resync after Master restart using RDB aux fields when offset is 0 (468 ms) -[ok]: PSYNC2: Partial resync after Master restart using RDB aux fields with data (2206 ms) -[ok]: PSYNC2: Partial resync after Master restart using RDB aux fields with expire (3000 ms) -[ok]: PSYNC2: Full resync after Master restart when too many key expired (4223 ms) -[124/124 done]: integration/psync2-master-restart (22 seconds) - - The End - -Execution time of different units: - 0 seconds - unit/oom-score-adj - 11 seconds - unit/client-eviction - 11 seconds - unit/introspection-2 - 6 seconds - unit/querybuf - 34 seconds - unit/wait - 50 seconds - unit/scan - 2 seconds - unit/networking - 3 seconds - unit/shutdown - 52 seconds - unit/dump - 2 seconds - unit/bitfield - 2 seconds - unit/pause - 10 seconds - unit/latency-monitor - 1 seconds - unit/replybufsize - 142 seconds - unit/maxmemory - 4 seconds - unit/auth - 2 seconds - unit/limits - 67 seconds - unit/hyperloglog - 1 seconds - unit/info-command - 1 seconds - unit/slowlog - 4 seconds - unit/info - 2 seconds - unit/lazyfree - 1 seconds - unit/acl-v2 - 2 seconds - unit/violations - 5 seconds - unit/bitops - 3 seconds - unit/tracking - 1 seconds - unit/protocol - 24 seconds - unit/aofrw - 15 seconds - unit/other - 19 seconds - unit/expire - 0 seconds - unit/printver - 1287 seconds - unit/memefficiency - 9 seconds - unit/acl - 2 seconds - unit/tls - 11 seconds - unit/multi - 2 seconds - unit/pubsubshard - 36 seconds - unit/obuf-limits - 63 seconds - unit/scripting - 0 seconds - unit/quit - 7 seconds - unit/functions - 3 seconds - unit/pubsub - 86 seconds - unit/geo - 6 seconds - unit/keyspace - 28 seconds - unit/introspection - 47 seconds - unit/sort - 53 seconds - unit/type/string - 63 seconds - unit/type/hash - 80 seconds - unit/type/list - 174 seconds - unit/type/set - 1 seconds - unit/type/incr - 134 seconds - unit/type/list-2 - 164 seconds - unit/type/zset - 12 seconds - unit/type/stream-cgroups - 145 seconds - unit/type/list-3 - 113 seconds - unit/type/stream - 720 seconds - unit/cluster/slot-stats - 13 seconds - unit/cluster/misc - 146 seconds - unit/cluster/cluster-response-tls - 12 seconds - unit/cluster/sharded-pubsub - 26 seconds - unit/cluster/slot-migration-response - 7 seconds - unit/cluster/half-migrated-slot - 139 seconds - unit/cluster/diskless-load-swapdb - 8 seconds - unit/cluster/human-announced-nodename - 17 seconds - unit/cluster/no-failover-option - 53 seconds - unit/cluster/failover - 0 seconds - unit/cluster/cluster-multiple-meets - 30 seconds - unit/cluster/failover2 - 49 seconds - unit/cluster/slave-stop-cond - 23 seconds - unit/cluster/replica-in-sync - 309 seconds - unit/cluster/replica-migration - 32 seconds - unit/cluster/consistency-check - 0 seconds - unit/cluster/cli - 0 seconds - unit/cluster/base - 4 seconds - unit/cluster/info - 15 seconds - unit/cluster/announced-endpoints - 62 seconds - unit/cluster/slot-ownership - 31 seconds - unit/cluster/manual-takeover - 15 seconds - unit/cluster/transactions-on-replica - 1109 seconds - unit/cluster/many-slot-migration - 30 seconds - unit/cluster/failure-marking - 17 seconds - unit/cluster/hostnames - 15 seconds - unit/cluster/pubsubshard-slot-migration - 13 seconds - unit/cluster/cluster-nodes-slots - 0 seconds - unit/cluster/cluster-reliable-meet - 27 seconds - unit/cluster/links - 17 seconds - unit/cluster/cluster-shards - 354 seconds - unit/cluster/cluster-slots - 14 seconds - unit/cluster/pubsubshard - 3 seconds - unit/cluster/scripting - 40 seconds - unit/cluster/faildet - 33 seconds - unit/cluster/update-msg - 94 seconds - unit/cluster/slave-selection - 16 seconds - unit/cluster/pubsub - 32 seconds - unit/cluster/announce-client-ip - 108 seconds - unit/cluster/slot-migration - 165 seconds - unit/cluster/manual-failover - 5 seconds - unit/cluster/multi-slot-operations - 20 seconds - integration/corrupt-dump-fuzzer - 23 seconds - integration/valkey-cli - 17 seconds - integration/shutdown - 20 seconds - integration/replication-3 - 8 seconds - integration/failover - 22 seconds - integration/rdb - 1 seconds - integration/convert-ziplist-hash-on-load - 49 seconds - integration/replication-4 - 15 seconds - integration/valkey-benchmark - 24 seconds - integration/psync2-reg - 55 seconds - integration/corrupt-dump - 28 seconds - integration/replication-2 - 19 seconds - integration/psync2-pingoff - 406 seconds - integration/replication-psync - 2 seconds - integration/logging - 28 seconds - integration/block-repl - 330 seconds - integration/replication - 6 seconds - integration/aof-race - 326 seconds - integration/dual-channel-replication - 16 seconds - integration/aof-multi-part - 45 seconds - integration/psync2 - 3 seconds - integration/replica-redirect - 1 seconds - integration/convert-ziplist-zset-on-load - 7 seconds - integration/dismiss-mem - 33 seconds - integration/aof - 2 seconds - integration/convert-zipmap-hash-on-load - 74 seconds - integration/replication-buffer - 22 seconds - integration/psync2-master-restart - -!!! WARNING The following tests failed: - -*** [err]: Active defrag big list: standalone in tests/unit/memefficiency.tcl -Expected 1.39 >= 1.7 (context: type eval line 40 cmd {assert {$frag >= $expected_frag}} proc ::test) -Cleanup: may take some time... OK +[ok]: Cluster should start ok (6 ms) +[ok]: Can't read from replica without READONLY (29 ms) +[ok]: Can't read from replica after READWRITE (2 ms) +[ok]: Can read from replica after READONLY (12 ms) +[ok]: Can perform HSET primary and HGET from replica (150 ms) +[ok]: Can MULTI-EXEC transaction of HGET operations from replica (7 ms) +[ok]: MULTI-EXEC with write operations is MOVED (18 ms) +[ok]: write command is QUEUED, then EXEC should be MOVED after failover (7295 ms) +[exception]: Executing test client: CLUSTERDOWN The cluster is down. +CLUSTERDOWN The cluster is down + while executing +"$primary XADD k * foo bar" + ("uplevel" body line 13) + invoked from within +"uplevel 1 $code" + (procedure "test" line 58) + invoked from within +"test "read-only blocking operations from replica" { + set rd [valkey_deferring_client -1] + $rd readonly + $rd read + $rd XREAD BLOCK 0 STREAM..." + ("uplevel" body line 78) + invoked from within +"uplevel 1 $code" + (procedure "cluster_setup" line 35) + invoked from within +"cluster_setup 1 1 2 continuous_slot_allocation default_replica_allocation { + +test "Cluster should start ok" { + wait_for_cluster_state ok +} + +set pri..." + ("uplevel" body line 1) + invoked from within +"uplevel 1 $code " + (procedure "start_server" line 2) + invoked from within +"start_server {overrides {cluster-enabled yes cluster-ping-interval 100 cluster-node-timeout 3000} tags {external:skip cluster}} {cluster_setup 1 1 2 c..." + ("uplevel" body line 1) + invoked from within +"uplevel 1 $code " + (procedure "start_server" line 2) + invoked from within +"start_server {overrides {cluster-enabled yes cluster-ping-interval 100 cluster-node-timeout 3000} tags {external:skip cluster}} {start_server {overrid..." + ("uplevel" body line 1) + invoked from within +"uplevel 1 $code" + (procedure "start_multiple_servers" line 5) + invoked from within +"start_multiple_servers $node_count $options $code" + (procedure "start_cluster" line 16) + invoked from within +"start_cluster 1 1 {tags {external:skip cluster}} { + +test "Cluster should start ok" { + wait_for_cluster_state ok +} + +set primary [srv 0 "client"] +set..." + (file "tests/unit/cluster/transactions-on-replica.tcl" line 3) + invoked from within +"source $path" + (procedure "execute_test_file" line 4) + invoked from within +"execute_test_file $data" + (procedure "test_client_main" line 10) + invoked from within +"test_client_main $::test_server_port " timeout 30m ./runtest-cluster --tls || true Starting valkey #0 at port 30000 Starting valkey #1 at port 30002 @@ -7207,250 +7656,242 @@ Starting valkey #18 at port 30036 Starting valkey #19 at port 30038 Testing unit: 03-failover-loop.tcl -13:36:13> (init) Restart killed instances: OK -13:36:13> Cluster nodes are reachable: OK -13:36:14> Cluster nodes hard reset: OK -13:36:16> Cluster Join and auto-discovery test: OK -13:36:18> Before slots allocation, all nodes report cluster failure: OK -13:36:18> Create a 5 nodes cluster: OK -13:36:22> Cluster is up: OK +23:57:10> (init) Restart killed instances: OK +23:57:10> Cluster nodes are reachable: OK +23:57:11> Cluster nodes hard reset: OK +23:57:34> Cluster Join and auto-discovery test: OK +23:57:47> Before slots allocation, all nodes report cluster failure: OK +23:57:47> Create a 5 nodes cluster: OK +23:58:00> Cluster is up: OK --- Iteration 19 --- -13:36:23> Wait for slave of #2 to sync: OK -13:36:23> Cluster is writable before failover: OK -13:36:24> Terminating node #2: OK -13:36:26> Wait failover by #7 with old epoch 3: OK -13:36:32> Cluster should eventually be up again: OK -13:36:34> Cluster is writable again: OK -13:36:34> Restarting node #2: OK -13:36:37> Instance #2 is now a slave: OK -13:36:37> We can read back the value we set before: OK +23:58:00> Cluster is writable before failover: OK +23:58:01> Terminating node #9: OK +23:58:01> Cluster should eventually be up again: OK +23:58:01> Cluster is writable again: OK +23:58:01> Restarting node #9: OK +23:58:02> Instance #9 is now a slave: OK +23:58:02> We can read back the value we set before: OK --- Iteration 18 --- -13:36:38> Wait for slave of #4 to sync: OK -13:36:38> Cluster is writable before failover: OK -13:36:39> Terminating node #4: OK -13:36:40> Wait failover by #9 with old epoch 5: OK -13:36:45> Cluster should eventually be up again: OK -13:36:45> Cluster is writable again: OK -13:36:47> Restarting node #4: OK -13:36:47> Instance #4 is now a slave: OK -13:36:47> We can read back the value we set before: OK +23:58:03> Wait for slave of #2 to sync: OK +23:58:03> Cluster is writable before failover: OK +23:58:03> Terminating node #2: OK +23:58:04> Wait failover by #7 with old epoch 3: OK +23:58:10> Cluster should eventually be up again: OK +23:58:10> Cluster is writable again: OK +23:58:10> Restarting node #2: OK +23:58:11> Instance #2 is now a slave: OK +23:58:11> We can read back the value we set before: OK --- Iteration 17 --- -13:36:48> Cluster is writable before failover: OK -13:36:49> Terminating node #2: OK -13:36:49> Cluster should eventually be up again: OK -13:36:49> Cluster is writable again: OK -13:36:50> Restarting node #2: OK -13:36:50> Instance #2 is now a slave: OK -13:36:50> We can read back the value we set before: OK +23:58:11> Cluster is writable before failover: OK +23:58:11> Terminating node #2: OK +23:58:12> Cluster should eventually be up again: OK +23:58:12> Cluster is writable again: OK +23:58:12> Restarting node #2: OK +23:58:13> Instance #2 is now a slave: OK +23:58:13> We can read back the value we set before: OK --- Iteration 16 --- -13:36:51> Cluster is writable before failover: OK -13:36:51> Terminating node #2: OK -13:36:51> Cluster should eventually be up again: OK -13:36:51> Cluster is writable again: OK -13:36:52> Restarting node #2: OK -13:36:53> Instance #2 is now a slave: OK -13:36:53> We can read back the value we set before: OK +23:58:13> Wait for slave of #4 to sync: OK +23:58:13> Cluster is writable before failover: OK +23:58:14> Terminating node #4: OK +23:58:18> Wait failover by #9 with old epoch 5: OK +23:58:22> Cluster should eventually be up again: OK +23:58:22> Cluster is writable again: OK +23:58:23> Restarting node #4: OK +23:58:23> Instance #4 is now a slave: OK +23:58:23> We can read back the value we set before: OK --- Iteration 15 --- -13:36:54> Wait for slave of #1 to sync: OK -13:36:54> Cluster is writable before failover: OK -13:36:54> Terminating node #1: OK -13:36:55> Wait failover by #6 with old epoch 2: OK -13:37:01> Cluster should eventually be up again: OK -13:37:02> Cluster is writable again: OK -13:37:03> Restarting node #1: OK -13:37:04> Instance #1 is now a slave: OK -13:37:04> We can read back the value we set before: OK +23:58:23> Cluster is writable before failover: OK +23:58:23> Terminating node #5: OK +23:58:24> Cluster should eventually be up again: OK +23:58:24> Cluster is writable again: OK +23:58:24> Restarting node #5: OK +23:58:28> Instance #5 is now a slave: OK +23:58:28> We can read back the value we set before: OK --- Iteration 14 --- -13:37:05> Wait for slave of #7 to sync: OK -13:37:05> Cluster is writable before failover: OK -13:37:05> Terminating node #7: OK -13:37:06> Wait failover by #2 with old epoch 21: OK -13:37:10> Cluster should eventually be up again: OK -13:37:11> Cluster is writable again: OK -13:37:12> Restarting node #7: OK -13:37:12> Instance #7 is now a slave: OK -13:37:12> We can read back the value we set before: OK +23:58:28> Wait for slave of #9 to sync: OK +23:58:28> Cluster is writable before failover: OK +23:58:29> Terminating node #9: OK +23:58:30> Wait failover by #4 with old epoch 22: OK +23:58:36> Cluster should eventually be up again: OK +23:58:36> Cluster is writable again: OK +23:58:37> Restarting node #9: OK +23:58:41> Instance #9 is now a slave: OK +23:58:41> We can read back the value we set before: OK --- Iteration 13 --- -13:37:14> Wait for slave of #2 to sync: OK -13:37:14> Cluster is writable before failover: OK -13:37:14> Terminating node #2: OK -13:37:15> Wait failover by #7 with old epoch 24: OK -13:37:20> Cluster should eventually be up again: OK -13:37:20> Cluster is writable again: OK -13:37:21> Restarting node #2: OK -13:37:21> Instance #2 is now a slave: OK -13:37:21> We can read back the value we set before: OK +23:58:42> Wait for slave of #3 to sync: OK +23:58:42> Cluster is writable before failover: OK +23:58:42> Terminating node #3: OK +23:58:48> Wait failover by #8 with old epoch 4: OK +23:58:52> Cluster should eventually be up again: OK +23:58:54> Cluster is writable again: OK +23:58:55> Restarting node #3: OK +23:59:06> Instance #3 is now a slave: OK +23:59:06> We can read back the value we set before: OK --- Iteration 12 --- -13:37:22> Cluster is writable before failover: OK -13:37:23> Terminating node #2: OK -13:37:23> Cluster should eventually be up again: OK -13:37:23> Cluster is writable again: OK -13:37:23> Restarting node #2: OK -13:37:24> Instance #2 is now a slave: OK -13:37:24> We can read back the value we set before: OK +23:59:06> Cluster is writable before failover: FAILED: Expected CLUSTERDOWN The cluster is down eq {OK} (context: type eval line 4 cmd {assert {$err eq {OK}}} proc ::test) +(Jumping to next unit after error) --- Iteration 11 --- -13:37:25> Wait for slave of #9 to sync: OK -13:37:25> Cluster is writable before failover: OK -13:37:26> Terminating node #9: OK -13:37:27> Wait failover by #4 with old epoch 22: OK -13:37:31> Cluster should eventually be up again: OK -13:37:31> Cluster is writable again: OK -13:37:32> Restarting node #9: OK -13:37:33> Instance #9 is now a slave: OK -13:37:33> We can read back the value we set before: OK +23:59:24> Wait for slave of #7 to sync: OK +23:59:25> Cluster is writable before failover: OK +23:59:25> Terminating node #7: OK +23:59:26> Wait failover by #2 with old epoch 21: OK +23:59:31> Cluster should eventually be up again: OK +23:59:31> Cluster is writable again: OK +23:59:31> Restarting node #7: OK +23:59:32> Instance #7 is now a slave: OK +23:59:32> We can read back the value we set before: OK --- Iteration 10 --- -13:37:34> Cluster is writable before failover: OK -13:37:35> Terminating node #8: OK -13:37:35> Cluster should eventually be up again: OK -13:37:36> Cluster is writable again: OK -13:37:36> Restarting node #8: OK -13:37:37> Instance #8 is now a slave: OK -13:37:37> We can read back the value we set before: OK +23:59:34> Wait for slave of #3 to sync: OK +23:59:34> Cluster is writable before failover: OK +23:59:34> Terminating node #3: OK +23:59:35> Wait failover by #8 with old epoch 26: OK +23:59:39> Cluster should eventually be up again: OK +23:59:39> Cluster is writable again: OK +23:59:39> Restarting node #3: OK +23:59:39> Instance #3 is now a slave: OK +23:59:39> We can read back the value we set before: OK --- Iteration 9 --- -13:37:38> Cluster is writable before failover: OK -13:37:39> Terminating node #8: OK -13:37:39> Cluster should eventually be up again: OK -13:37:39> Cluster is writable again: OK -13:37:40> Restarting node #8: OK -13:37:40> Instance #8 is now a slave: OK -13:37:40> We can read back the value we set before: OK +23:59:39> Cluster is writable before failover: OK +23:59:39> Terminating node #7: OK +23:59:40> Cluster should eventually be up again: OK +23:59:40> Cluster is writable again: OK +23:59:40> Restarting node #7: OK +23:59:40> Instance #7 is now a slave: OK +23:59:40> We can read back the value we set before: OK --- Iteration 8 --- -13:37:42> Wait for slave of #6 to sync: OK -13:37:42> Cluster is writable before failover: OK -13:37:42> Terminating node #6: OK -13:37:44> Wait failover by #1 with old epoch 23: OK -13:37:49> Cluster should eventually be up again: OK -13:37:49> Cluster is writable again: OK -13:37:50> Restarting node #6: OK -13:37:51> Instance #6 is now a slave: OK -13:37:51> We can read back the value we set before: OK +23:59:40> Cluster is writable before failover: OK +23:59:40> Terminating node #5: OK +23:59:40> Cluster should eventually be up again: OK +23:59:40> Cluster is writable again: OK +23:59:40> Restarting node #5: OK +23:59:40> Instance #5 is now a slave: OK +23:59:40> We can read back the value we set before: OK --- Iteration 7 --- -13:37:52> Wait for slave of #1 to sync: OK -13:37:52> Cluster is writable before failover: OK -13:37:53> Terminating node #1: OK -13:37:54> Wait failover by #6 with old epoch 27: OK -13:37:59> Cluster should eventually be up again: OK -13:37:59> Cluster is writable again: OK -13:38:00> Restarting node #1: OK -13:38:01> Instance #1 is now a slave: OK -13:38:01> We can read back the value we set before: OK +23:59:40> Cluster is writable before failover: OK +23:59:40> Terminating node #5: OK +23:59:40> Cluster should eventually be up again: OK +23:59:40> Cluster is writable again: OK +23:59:40> Restarting node #5: OK +23:59:40> Instance #5 is now a slave: OK +23:59:40> We can read back the value we set before: OK --- Iteration 6 --- -13:38:03> Wait for slave of #0 to sync: OK -13:38:03> Cluster is writable before failover: OK -13:38:03> Terminating node #0: OK -13:38:04> Wait failover by #5 with old epoch 1: OK -13:38:09> Cluster should eventually be up again: OK -13:38:09> Cluster is writable again: OK -13:38:10> Restarting node #0: OK -13:38:10> Instance #0 is now a slave: OK -13:38:10> We can read back the value we set before: OK +23:59:41> Wait for slave of #1 to sync: OK +23:59:41> Cluster is writable before failover: OK +23:59:41> Terminating node #1: OK +23:59:41> Wait failover by #6 with old epoch 2: OK +23:59:46> Cluster should eventually be up again: OK +23:59:46> Cluster is writable again: OK +23:59:47> Restarting node #1: OK +23:59:47> Instance #1 is now a slave: OK +23:59:47> We can read back the value we set before: OK --- Iteration 5 --- -13:38:12> Wait for slave of #3 to sync: OK -13:38:12> Cluster is writable before failover: OK -13:38:13> Terminating node #3: OK -13:38:14> Wait failover by #8 with old epoch 4: OK -13:38:19> Cluster should eventually be up again: OK -13:38:20> Cluster is writable again: OK -13:38:21> Restarting node #3: OK -13:38:21> Instance #3 is now a slave: OK -13:38:21> We can read back the value we set before: OK +23:59:48> Wait for slave of #6 to sync: OK +23:59:48> Cluster is writable before failover: OK +23:59:48> Terminating node #6: OK +23:59:49> Wait failover by #1 with old epoch 29: OK +23:59:55> Cluster should eventually be up again: OK +23:59:55> Cluster is writable again: OK +23:59:56> Restarting node #6: OK +23:59:56> Instance #6 is now a slave: OK +23:59:56> We can read back the value we set before: OK --- Iteration 4 --- -13:38:22> Wait for slave of #4 to sync: OK -13:38:22> Cluster is writable before failover: OK -13:38:23> Terminating node #4: OK -13:38:24> Wait failover by #9 with old epoch 26: OK -13:38:29> Cluster should eventually be up again: OK -13:38:30> Cluster is writable again: OK -13:38:30> Restarting node #4: OK -13:38:31> Instance #4 is now a slave: OK -13:38:31> We can read back the value we set before: OK +23:59:56> Cluster is writable before failover: OK +23:59:56> Terminating node #3: OK +23:59:56> Cluster should eventually be up again: OK +23:59:56> Cluster is writable again: OK +23:59:56> Restarting node #3: OK +23:59:56> Instance #3 is now a slave: OK +23:59:56> We can read back the value we set before: OK --- Iteration 3 --- -13:38:33> Wait for slave of #6 to sync: OK -13:38:33> Cluster is writable before failover: OK -13:38:33> Terminating node #6: OK -13:38:35> Wait failover by #1 with old epoch 28: OK -13:38:39> Cluster should eventually be up again: OK -13:38:40> Cluster is writable again: OK -13:38:41> Restarting node #6: OK -13:38:41> Instance #6 is now a slave: OK -13:38:41> We can read back the value we set before: OK +23:59:56> Cluster is writable before failover: OK +23:59:56> Terminating node #5: OK +23:59:57> Cluster should eventually be up again: OK +23:59:57> Cluster is writable again: OK +23:59:57> Restarting node #5: OK +23:59:57> Instance #5 is now a slave: OK +23:59:57> We can read back the value we set before: OK --- Iteration 2 --- -13:38:42> Wait for slave of #5 to sync: OK -13:38:42> Cluster is writable before failover: OK -13:38:43> Terminating node #5: OK -13:38:44> Wait failover by #0 with old epoch 29: OK -13:38:49> Cluster should eventually be up again: OK -13:38:49> Cluster is writable again: OK -13:38:50> Restarting node #5: OK -13:38:51> Instance #5 is now a slave: OK -13:38:51> We can read back the value we set before: OK +23:59:57> Cluster is writable before failover: OK +23:59:57> Terminating node #6: OK +23:59:57> Cluster should eventually be up again: OK +23:59:57> Cluster is writable again: OK +23:59:57> Restarting node #6: OK +23:59:57> Instance #6 is now a slave: OK +23:59:57> We can read back the value we set before: OK --- Iteration 1 --- -13:38:52> Cluster is writable before failover: OK -13:38:52> Terminating node #3: OK -13:38:52> Cluster should eventually be up again: OK -13:38:52> Cluster is writable again: OK -13:38:53> Restarting node #3: OK -13:38:54> Instance #3 is now a slave: OK -13:38:54> We can read back the value we set before: OK -13:38:55> Post condition: current_epoch >= my_epoch everywhere: OK +23:59:57> Cluster is writable before failover: OK +23:59:58> Terminating node #6: OK +23:59:58> Cluster should eventually be up again: OK +23:59:58> Cluster is writable again: OK +23:59:58> Restarting node #6: OK +23:59:58> Instance #6 is now a slave: OK +23:59:58> We can read back the value we set before: OK +23:59:58> Post condition: current_epoch >= my_epoch everywhere: OK Testing unit: 04-resharding.tcl -13:38:55> (init) Restart killed instances: OK -13:38:55> Cluster nodes are reachable: OK -13:38:55> Cluster nodes hard reset: OK -13:39:00> Cluster Join and auto-discovery test: OK -13:39:03> Before slots allocation, all nodes report cluster failure: OK -13:39:03> Create a 5 nodes cluster: OK -13:39:07> Cluster is up: OK -13:39:07> Enable AOF in all the instances: OK +23:59:58> (init) Restart killed instances: OK +23:59:58> Cluster nodes are reachable: OK +23:59:58> Cluster nodes hard reset: OK +00:00:06> Cluster Join and auto-discovery test: OK +00:00:11> Before slots allocation, all nodes report cluster failure: OK +00:00:11> Create a 5 nodes cluster: OK +00:00:15> Cluster is up: OK +00:00:15> Enable AOF in all the instances: OK Testing TLS cluster on start node 127.0.0.1:30000, plaintext port 30001 -13:39:08> Cluster consistency during live resharding: WWWWWWWWWWWWWWWWWWWWWWWWWWWWWWWWWWWWWWWWWWWWWWWWWWWWWWWWWWWWWWWWWWWWWWWWWWWWWWWWWWWWWWWWWWWWWWWWWWWW...Starting resharding...W......................................................................................................................................................................................................................................................................................................................W..................................................................................................................................................................................................................................................................................................................W...............................................................................................................................................................................................................................................................................W.............................................................................WWWWWW...Starting resharding...W....................................................................................................................................................................................................................................................................................................................W.........................................................................................................................................................................................................................................................W........................................................................................................................................................................................................................................W....................................................................................................................................................................................................W.....................WWWWW...Starting resharding...W.............................................................................................................................................................................................................................................................................................W...................................................................................................................................................................................................................................................W...........................................................................................................................................................................................................................................................................W.............................................................................................................................................................WWWWWW...Starting resharding...W.......................................................................................................................................................................W...................................................................................................................................................................................................................................................................................................W....................................................................................................................................................................................................................................................W.........................................................................................................................................................................................W..........................................................................................................................................WWWWW...Starting resharding...W........................................................................................................................................................................................................................................................................W.................................................................................................................................................................................................................W.................................................................................................................................................................................................................................................................................W.........................................................................................................................................................................................................................................................................WWWWWW...Starting resharding...W..................................................................................................................................................................................................................................................................W.....................................................................................................................................................................................................................................................................................W.......................................................................................................................................................................................................................................................W.........................................................................................................................................................................................................................WWWWWW...Starting resharding...W............................................................................................................................................................W..............................................................................................................................................................W............................................................................................................................................................W.....................................................................................................................................................................................................................................W........................................................................................................................................................................................................................................................................................W........WWWW...Starting resharding...W................................................................................................................................................................................................................................................................................................W.................................................................................................................................................................................................................W.............................................................................................................................................................................................................................................................W............................................................................................................................................................................................................................................W................................................WWWWW...Starting resharding...W...........................................................................................................................................................W.................................................................................................................................................................................................................................W..................................................................................................................................................................................................................................................................................W..............................................................................................................................................................................................................................................................................................W................................................WWWWW...Starting resharding...W.................................................................................................................................................................................................................................................................................W................................W..............................W..................................................................................................................................................................................W.................................................................................................................................................W..................................................................................................................................................................................................................................................................................W.....................................................................................WWWOK -13:57:05> Verify 50000 keys for consistency with logical content: OK -13:58:38> Terminate and restart all the instances: OK -13:58:57> Cluster should eventually be up again: OK -13:58:59> Verify 50000 keys after the restart: OK -14:00:33> Disable AOF in all the instances: OK -14:00:34> Verify slaves consistency: OK -14:00:37> Dump sanitization was skipped for migrations: OK +00:00:15> Cluster consistency during live resharding: WWWWWWWWWWWWWWWWWWWWWWWWWWWWWWWWWWWWWWWWWWWWWWWWWWWWWWWWWWWWWWWWWWWWWWWWWWWWWWWWWWWWWWWWWWWWWWWWWWWW...Starting resharding...W.............................................W............................................................................................................................................W.................................................................................................................W........................................................................................................W..................................................................................................................................W...............................................................................................................W............................................................................W..................................................................................................W.............................................................................................................W.............................................Starting resharding...W...........................................................................W.........................................................................................................................................W.................................................................................................W............................................................................................................................W...................................................................................................W....................................................................................................................................................................W...............................................................................................................W............................................................................................................W....................................................................................W...Starting resharding...W....................................................................................W...................................................................................................................................W.................................................................................................W......................................................................................W...........................WWWWW...Starting resharding...W.................................W............*** Please fix your cluster problems before resharding +.......................................WWWWWWWW...Starting resharding...W.............................................W*** Please fix your cluster problems before resharding +.......................................WWWWWWWW...Starting resharding...W.............................................*** Please fix your cluster problems before resharding +.......................................WWWWWWWWW...Starting resharding...W.............................................*** Please fix your cluster problems before resharding +.......................................WWWWWWWWW...Starting resharding...W.............................................*** Please fix your cluster problems before resharding +.......................................WWWWWWWWW...Starting resharding...W.............................................*** Please fix your cluster problems before resharding +......W.................................WWWWWWWW...Starting resharding...W.................................W............*** Please fix your cluster problems before resharding +.......................................WWWWWWWWOK +00:01:51> Verify 50000 keys for consistency with logical content: OK +00:01:59> Terminate and restart all the instances: OK +00:02:05> Cluster should eventually be up again: OK +00:02:07> Verify 50000 keys after the restart: OK +00:02:22> Disable AOF in all the instances: OK +00:02:22> Verify slaves consistency: OK +00:02:23> Dump sanitization was skipped for migrations: OK Testing unit: 07-replica-migration.tcl -14:00:37> (init) Restart killed instances: OK -14:00:37> Cluster nodes are reachable: OK -14:00:38> Cluster nodes hard reset: OK -14:00:42> Cluster Join and auto-discovery test: OK -14:00:44> Before slots allocation, all nodes report cluster failure: OK -14:00:44> Create a 5 nodes cluster: OK -14:00:48> Cluster is up: OK -14:00:48> Each master should have two replicas attached: OK -14:00:48> Killing all the slaves of master #0 and #1: OK -14:00:55> Master #0 should have at least one replica: OK -14:01:01> Master #1 should have at least one replica: OK -14:01:01> Master #2 should have at least one replica: OK -14:01:01> Master #3 should have at least one replica: OK -14:01:01> Master #4 should have at least one replica: OK -14:01:01> (init) Restart killed instances: valkey/5 valkey/6 valkey/10 valkey/11 OK -14:01:03> Cluster nodes are reachable: OK -14:01:03> Cluster nodes hard reset: OK -14:01:09> Cluster Join and auto-discovery test: OK -14:01:11> Before slots allocation, all nodes report cluster failure: OK -14:01:11> Create a 5 nodes cluster: OK -14:01:21> Cluster is up: OK -14:01:21> Kill slave #7 of master #2. Only slave left is #12 now: OK -14:01:21> Killing master node #2, #12 should failover: OK -14:01:22> Wait for failover: OK -14:01:27> Cluster should eventually be up again: OK -14:01:27> Cluster is writable: OK -14:01:28> Instance 12 is now a master without slaves: OK -14:01:28> Master #12 should get at least one migrated replica: OK +00:02:23> (init) Restart killed instances: OK +00:02:23> Cluster nodes are reachable: OK +00:02:23> Cluster nodes hard reset: OK +00:02:24> Cluster Join and auto-discovery test: OK +00:02:27> Before slots allocation, all nodes report cluster failure: OK +00:02:27> Create a 5 nodes cluster: OK +00:02:31> Cluster is up: OK +00:02:31> Each master should have two replicas attached: OK +00:02:31> Killing all the slaves of master #0 and #1: OK +00:02:36> Master #0 should have at least one replica: OK +00:02:41> Master #1 should have at least one replica: OK +00:02:41> Master #2 should have at least one replica: OK +00:02:41> Master #3 should have at least one replica: OK +00:02:41> Master #4 should have at least one replica: OK +00:02:41> (init) Restart killed instances: valkey/5 valkey/6 valkey/10 valkey/11 OK +00:02:42> Cluster nodes are reachable: OK +00:02:42> Cluster nodes hard reset: OK +00:02:42> Cluster Join and auto-discovery test: OK +00:02:43> Before slots allocation, all nodes report cluster failure: OK +00:02:43> Create a 5 nodes cluster: OK +00:02:48> Cluster is up: OK +00:02:48> Kill slave #7 of master #2. Only slave left is #12 now: OK +00:02:48> Killing master node #2, #12 should failover: OK +00:02:48> Wait for failover: OK +00:02:53> Cluster should eventually be up again: OK +00:02:53> Cluster is writable: OK +00:02:53> Instance 12 is now a master without slaves: OK +00:02:53> Master #12 should get at least one migrated replica: OK Testing unit: 12-replica-migration-2.tcl -14:01:32> (init) Restart killed instances: valkey/2 valkey/7 OK -14:01:34> Cluster nodes are reachable: OK -14:01:34> Cluster nodes hard reset: OK -14:01:38> Cluster Join and auto-discovery test: OK -14:01:40> Before slots allocation, all nodes report cluster failure: OK -14:01:40> Create a 5 nodes cluster: OK -14:01:49> Cluster is up: OK -14:01:49> Each master should have at least two replicas attached: OK -14:01:49> Set allow-replica-migration yes: OK -14:01:49> Resharding all the master #0 slots away from it: >>> Performing Cluster Check (using node 127.0.0.1:30000) +00:02:58> (init) Restart killed instances: valkey/2 valkey/7 OK +00:02:59> Cluster nodes are reachable: OK +00:02:59> Cluster nodes hard reset: OK +00:02:59> Cluster Join and auto-discovery test: OK +00:03:01> Before slots allocation, all nodes report cluster failure: OK +00:03:01> Create a 5 nodes cluster: OK +00:03:04> Cluster is up: OK +00:03:04> Each master should have at least two replicas attached: OK +00:03:04> Set allow-replica-migration yes: OK +00:03:04> Resharding all the master #0 slots away from it: >>> Performing Cluster Check (using node 127.0.0.1:30000) [OK] All nodes agree about slots configuration. >>> Check for open slots... >>> Check slots coverage... @@ -7458,266 +7899,1807 @@ >>> Rebalancing across 5 nodes. Total weight = 4.00 Moving 820 slots from 127.0.0.1:30000 to 127.0.0.1:30002 #################################################################################################################################################################################################################################################################################################################################################################################################################################################################################################################################################################################################################################################################################################################################################################################################################################################### +Moving 820 slots from 127.0.0.1:30000 to 127.0.0.1:30006 +#################################################################################################################################################################################################################################################################################################################################################################################################################################################################################################################################################################################################################################################################################################################################################################################################################################################### Moving 820 slots from 127.0.0.1:30000 to 127.0.0.1:30004 -#########################################################################################################################################################################################################################################################################################################################################################################################################./runtest-sentinel || true -Starting sentinel #0 at port 20006 -Starting sentinel #1 at port 20007 -Starting sentinel #2 at port 20008 -Starting sentinel #3 at port 20009 -Starting sentinel #4 at port 20010 +#################################################################################################################################################################################################################################################################################################################################################################################################################################################################################################################################################################################################################################################################################################################################################################################################################################################### +Moving 816 slots from 127.0.0.1:30000 to 127.0.0.1:30008 +################################################################################################################################################################################################################################################################################################################################################################################################################################################################################################################################################################################################################################################################################################################################################################################################################################################ +OK +00:04:51> Master #0 who lost all slots should turn into a replica without replicas: OK +00:04:51> Resharding back some slot to master #0: >>> Performing Cluster Check (using node 127.0.0.1:30000) +[OK] All nodes agree about slots configuration. +>>> Check for open slots... +>>> Check slots coverage... +[OK] All 16384 slots covered. +*** No rebalancing needed! All nodes are within the 2.00% threshold. +OK +00:05:01> Master #0 should re-acquire one or more replicas: OK +Testing unit: 12.1-replica-migration-3.tcl +00:05:01> (init) Restart killed instances: OK +00:05:01> Cluster nodes are reachable: OK +00:05:01> Cluster nodes hard reset: OK +00:05:01> Cluster Join and auto-discovery test: OK +00:05:06> Before slots allocation, all nodes report cluster failure: OK +00:05:08> Create a 5 nodes cluster: OK +00:05:11> Cluster is up: OK +00:05:11> Each master should have at least two replicas attached: OK +00:05:11> Set allow-replica-migration no: OK +00:05:11> Resharding all the master #0 slots away from it: >>> Performing Cluster Check (using node 127.0.0.1:30000) +[OK] All nodes agree about slots configuration. +>>> Check for open slots... +>>> Check slots coverage... +[OK] All 16384 slots covered. +>>> Rebalancing across 5 nodes. Total weight = 4.00 +Moving 820 slots from 127.0.0.1:30000 to 127.0.0.1:30004 +#################################################################################################################################################################################################################################################################################################################################################################################################################################################################################################################################################################################################################################################################################################################################################################################################################################################### +Moving 820 slots from 127.0.0.1:30000 to 127.0.0.1:30002 +#################################################################################################################################################################################################################################################################################################################################################################################################################################################################################################################################################################################################################################################################################################################################################################################################################################################### +Moving 820 slots from 127.0.0.1:30000 to 127.0.0.1:30006 +#################################################################################################################################################################################################################################################################################################################################################################################################################################################################################################################################################################################################################################################################################################################################################################################################################################################### +Moving 816 slots from 127.0.0.1:30000 to 127.0.0.1:30008 +################################################################################################################################################################################################################################################################################################################################################################################################################################################################################################################################################################################################################################################################################################################################################################################################################################################ +OK +00:06:53> Wait cluster to be stable: OK +00:06:53> Master #0 still should have its replicas: OK +00:06:53> Each master should have at least two replicas attached: OK +Testing unit: 28-cluster-shards.tcl +00:06:53> (init) Restart killed instances: OK +00:06:53> Cluster nodes are reachable: OK +00:06:53> Cluster nodes hard reset: OK +00:06:54> Cluster Join and auto-discovery test: OK +00:07:05> Before slots allocation, all nodes report cluster failure: OK +00:07:05> Create a 8 nodes cluster with 4 shards: OK +00:07:05> Cluster should start ok: OK +00:07:10> Set cluster hostnames and verify they are propagated: OK +00:07:11> Verify information about the shards: OK +00:07:12> Verify no slot shard: OK +00:07:12> Kill a node and tell the replica to immediately takeover: OK +00:07:12> Verify health as fail for killed node: OK +00:07:17> Restarting primary node: OK +00:07:17> Instance #0 gets converted into a replica: OK +00:07:17> Test the replica reports a loading state while it's loading: OK +00:07:20> Regression test for a crash when calling SHARDS during handshake: OK +00:07:20> Cluster is up: OK +00:07:23> Shard ids are unique: OK +00:07:23> CLUSTER MYSHARDID reports same id for both primary and replica: OK +00:07:23> New replica receives primary's shard id: OK +00:07:23> CLUSTER MYSHARDID reports same shard id after shard restart: OK +00:07:26> CLUSTER MYSHARDID reports same shard id after cluster restart: OK +Cleaning up... +killing stale instance 2622575 +killing stale instance 2622851 +killing stale instance 2623828 +killing stale instance 2624075 +killing stale instance 2624429 +killing stale instance 2624707 +killing stale instance 2625031 +killing stale instance 2625372 +killing stale instance 2625698 +killing stale instance 2626012 +killing stale instance 2634544 +killing stale instance 2634550 +killing stale instance 2683582 +killing stale instance 2683592 +killing stale instance 2683603 +killing stale instance 2683610 +killing stale instance 2683616 +killing stale instance 2683622 +killing stale instance 2683631 +killing stale instance 2683642 +WARNING 1 test(s) failed. +./runtest-sentinel || true +Starting sentinel #0 at port 20000 +Starting sentinel #1 at port 20001 +Starting sentinel #2 at port 20002 +Starting sentinel #3 at port 20003 +Starting sentinel #4 at port 20004 Starting valkey #0 at port 30000 Starting valkey #1 at port 30001 Starting valkey #2 at port 30002 Starting valkey #3 at port 30003 Starting valkey #4 at port 30004 Testing unit: 00-base.tcl -14:06:10> (start-init) Flush config and compare rewrite config file lines: OK -14:06:11> (init) Restart killed instances: OK -14:06:11> (init) Remove old primary entry from sentinels: OK -14:06:11> (init) Create a primary-replicas cluster of 5 instances: OK -14:06:11> (init) Sentinels can start monitoring a primary: OK -14:06:12> (init) Sentinels can talk with the primary: OK -14:06:12> (init) Sentinels are able to auto-discover other sentinels: OK -14:06:14> (init) Sentinels are able to auto-discover replicas: OK -14:06:14> Sentinel command flag infrastructure works correctly: OK -14:06:14> SENTINEL HELP output the sentinel subcommand help: OK -14:06:14> SENTINEL MYID return the sentinel instance ID: OK -14:06:14> SENTINEL INFO CACHE returns the cached info: OK -14:06:14> SENTINEL PENDING-SCRIPTS returns the information about pending scripts: OK -14:06:14> SENTINEL PRIMARIES returns a list of monitored primaries: OK -14:06:14> SENTINEL SENTINELS returns a list of sentinel instances: OK -14:06:14> SENTINEL SLAVES returns a list of the monitored replicas: OK -14:06:14> SENTINEL SIMULATE-FAILURE HELP list supported flags: OK -14:06:14> Basic failover works if the primary is down: OK -14:06:18> New primary 127.0.0.1:30002 role matches: OK -14:06:18> All the other slaves now point to the new primary: OK -14:06:18> The old primary eventually gets reconfigured as a slave: OK -14:06:29> ODOWN is not possible without N (quorum) Sentinels reports: OK -14:06:34> Failover is not possible without majority agreement: OK -14:06:37> Failover works if we configure for absolute agreement: OK -14:06:42> New primary 127.0.0.1:30001 role matches: OK -14:06:42> SENTINEL RESET can resets the primary: OK -14:06:42> SENTINEL IS-PRIMARY-DOWN-BY-ADDR checks if the primary is down: OK +00:07:33> (start-init) Flush config and compare rewrite config file lines: OK +00:07:33> (init) Restart killed instances: OK +00:07:33> (init) Remove old primary entry from sentinels: OK +00:07:33> (init) Create a primary-replicas cluster of 5 instances: OK +00:07:33> (init) Sentinels can start monitoring a primary: OK +00:07:33> (init) Sentinels can talk with the primary: OK +00:07:33> (init) Sentinels are able to auto-discover other sentinels: OK +00:07:35> (init) Sentinels are able to auto-discover replicas: OK +00:07:35> Sentinel command flag infrastructure works correctly: OK +00:07:35> SENTINEL HELP output the sentinel subcommand help: OK +00:07:35> SENTINEL MYID return the sentinel instance ID: OK +00:07:35> SENTINEL INFO CACHE returns the cached info: OK +00:07:35> SENTINEL PENDING-SCRIPTS returns the information about pending scripts: OK +00:07:35> SENTINEL PRIMARIES returns a list of monitored primaries: OK +00:07:35> SENTINEL SENTINELS returns a list of sentinel instances: OK +00:07:35> SENTINEL SLAVES returns a list of the monitored replicas: OK +00:07:35> SENTINEL SIMULATE-FAILURE HELP list supported flags: OK +00:07:35> Basic failover works if the primary is down: OK +00:07:39> New primary 127.0.0.1:30002 role matches: OK +00:07:39> All the other slaves now point to the new primary: OK +00:07:39> The old primary eventually gets reconfigured as a slave: OK +00:07:49> ODOWN is not possible without N (quorum) Sentinels reports: OK +00:07:49> Failover is not possible without majority agreement: OK +00:07:50> Failover works if we configure for absolute agreement: OK +00:07:54> New primary 127.0.0.1:30001 role matches: OK +00:07:54> SENTINEL RESET can resets the primary: OK +00:07:54> SENTINEL IS-PRIMARY-DOWN-BY-ADDR checks if the primary is down: OK Testing unit: 01-conf-update.tcl -14:06:42> (init) Restart killed instances: OK -14:06:42> (init) Remove old primary entry from sentinels: OK -14:06:42> (init) Create a primary-replicas cluster of 5 instances: OK -14:06:42> (init) Sentinels can start monitoring a primary: OK -14:06:43> (init) Sentinels can talk with the primary: OK -14:06:43> (init) Sentinels are able to auto-discover other sentinels: OK -14:06:44> (init) Sentinels are able to auto-discover replicas: OK -14:06:44> We can failover with Sentinel 1 crashed: OK -14:06:48> After Sentinel 1 is restarted, its config gets updated: OK -14:06:49> New primary 127.0.0.1:30002 role matches: OK -14:06:49> Update log level: OK +00:07:54> (init) Restart killed instances: OK +00:07:54> (init) Remove old primary entry from sentinels: OK +00:07:54> (init) Create a primary-replicas cluster of 5 instances: OK +00:07:54> (init) Sentinels can start monitoring a primary: OK +00:07:54> (init) Sentinels can talk with the primary: OK +00:07:54> (init) Sentinels are able to auto-discover other sentinels: OK +00:07:56> (init) Sentinels are able to auto-discover replicas: OK +00:07:56> We can failover with Sentinel 1 crashed: OK +00:08:00> After Sentinel 1 is restarted, its config gets updated: OK +00:08:00> New primary 127.0.0.1:30002 role matches: OK +00:08:00> Update log level: OK Testing unit: 02-replicas-reconf.tcl -14:06:49> (init) Restart killed instances: OK -14:06:49> (init) Remove old primary entry from sentinels: OK -14:06:49> (init) Create a primary-replicas cluster of 5 instances: OK -14:06:49> (init) Sentinels can start monitoring a primary: OK -14:06:50> (init) Sentinels can talk with the primary: OK -14:06:50> (init) Sentinels are able to auto-discover other sentinels: OK -14:06:51> (init) Sentinels are able to auto-discover replicas: OK -14:06:51> Check that slaves replicate from current primary: OK -14:06:51> Crash the primary and force a failover: OK -14:06:58> Check that slaves replicate from current primary: OK -14:07:10> Kill a slave instance: OK -14:07:10> Crash the primary and force a failover: OK -14:07:15> Check that slaves replicate from current primary: OK -14:07:16> Wait for failover to end: OK -14:07:16> Restart killed slave and test replication of slaves again...: OK -14:07:16> Check that slaves replicate from current primary: OK +00:08:00> (init) Restart killed instances: OK +00:08:00> (init) Remove old primary entry from sentinels: OK +00:08:00> (init) Create a primary-replicas cluster of 5 instances: OK +00:08:00> (init) Sentinels can start monitoring a primary: OK +00:08:01> (init) Sentinels can talk with the primary: OK +00:08:01> (init) Sentinels are able to auto-discover other sentinels: OK +00:08:03> (init) Sentinels are able to auto-discover replicas: OK +00:08:03> Check that slaves replicate from current primary: OK +00:08:03> Crash the primary and force a failover: OK +00:08:06> Check that slaves replicate from current primary: OK +00:08:16> Kill a slave instance: OK +00:08:17> Crash the primary and force a failover: OK +00:08:21> Check that slaves replicate from current primary: OK +00:08:22> Wait for failover to end: OK +00:08:22> Restart killed slave and test replication of slaves again...: OK +00:08:22> Check that slaves replicate from current primary: OK Testing unit: 03-runtime-reconf.tcl -14:07:28> (init) Restart killed instances: OK -14:07:28> (init) Remove old primary entry from sentinels: OK -14:07:28> (init) Create a primary-replicas cluster of 5 instances: OK -14:07:28> (init) Sentinels can start monitoring a primary: OK -14:07:29> (init) Sentinels can talk with the primary: OK -14:07:29> (init) Sentinels are able to auto-discover other sentinels: OK -14:07:29> (init) Sentinels are able to auto-discover replicas: OK -14:07:29> Sentinels (re)connection following SENTINEL SET myprimary auth-pass: OK -14:07:31> Sentinels (re)connection following primary ACL change: OK -14:07:35> Set parameters in normal case: OK -14:07:38> Set parameters in normal case with bad format: OK -14:07:38> Sentinel Set with other error situations: OK +00:08:32> (init) Restart killed instances: OK +00:08:32> (init) Remove old primary entry from sentinels: OK +00:08:32> (init) Create a primary-replicas cluster of 5 instances: OK +00:08:32> (init) Sentinels can start monitoring a primary: OK +00:08:32> (init) Sentinels can talk with the primary: OK +00:08:32> (init) Sentinels are able to auto-discover other sentinels: OK +00:08:32> (init) Sentinels are able to auto-discover replicas: OK +00:08:32> Sentinels (re)connection following SENTINEL SET myprimary auth-pass: OK +00:08:34> Sentinels (re)connection following primary ACL change: OK +00:08:36> Set parameters in normal case: OK +00:08:36> Set parameters in normal case with bad format: OK +00:08:36> Sentinel Set with other error situations: OK Testing unit: 04-slave-selection.tcl Testing unit: 05-manual.tcl -14:07:38> (init) Restart killed instances: OK -14:07:38> (init) Remove old primary entry from sentinels: OK -14:07:38> (init) Create a primary-replicas cluster of 5 instances: OK -14:07:38> (init) Sentinels can start monitoring a primary: OK -14:07:39> (init) Sentinels can talk with the primary: OK -14:07:39> (init) Sentinels are able to auto-discover other sentinels: OK -14:07:40> (init) Sentinels are able to auto-discover replicas: OK -14:07:40> Manual failover works: OK -14:07:41> New primary 127.0.0.1:30002 role matches: OK -14:07:41> All the other slaves now point to the new primary: OK -14:07:41> The old primary eventually gets reconfigured as a slave: OK -14:07:45> (init) Restart killed instances: OK -14:07:45> (init) Remove old primary entry from sentinels: OK -14:07:45> (init) Create a primary-replicas cluster of 5 instances: OK -14:07:46> (init) Sentinels can start monitoring a primary: OK -14:07:47> (init) Sentinels can talk with the primary: OK -14:07:47> (init) Sentinels are able to auto-discover other sentinels: OK -14:07:47> (init) Sentinels are able to auto-discover replicas: OK -14:07:47> SENTINEL SIMULATE-FAILURE crash-after-election works: OK -14:07:47> (init) Restart killed instances: OK -14:07:47> (init) Remove old primary entry from sentinels: OK -14:07:47> (init) Create a primary-replicas cluster of 5 instances: OK -14:07:47> (init) Sentinels can start monitoring a primary: OK -14:07:48> (init) Sentinels can talk with the primary: OK -14:07:48> (init) Sentinels are able to auto-discover other sentinels: OK -14:07:49> (init) Sentinels are able to auto-discover replicas: OK -14:07:49> SENTINEL SIMULATE-FAILURE crash-after-promotion works: OK +00:08:36> (init) Restart killed instances: OK +00:08:36> (init) Remove old primary entry from sentinels: OK +00:08:36> (init) Create a primary-replicas cluster of 5 instances: OK +00:08:36> (init) Sentinels can start monitoring a primary: OK +00:08:37> (init) Sentinels can talk with the primary: OK +00:08:37> (init) Sentinels are able to auto-discover other sentinels: OK +00:08:39> (init) Sentinels are able to auto-discover replicas: OK +00:08:39> Manual failover works: OK +00:08:40> New primary 127.0.0.1:30002 role matches: OK +00:08:40> All the other slaves now point to the new primary: OK +00:08:40> The old primary eventually gets reconfigured as a slave: OK +00:08:44> (init) Restart killed instances: OK +00:08:44> (init) Remove old primary entry from sentinels: OK +00:08:44> (init) Create a primary-replicas cluster of 5 instances: OK +00:08:44> (init) Sentinels can start monitoring a primary: OK +00:08:44> (init) Sentinels can talk with the primary: OK +00:08:44> (init) Sentinels are able to auto-discover other sentinels: OK +00:08:45> (init) Sentinels are able to auto-discover replicas: OK +00:08:45> SENTINEL SIMULATE-FAILURE crash-after-election works: OK +00:08:45> (init) Restart killed instances: OK +00:08:45> (init) Remove old primary entry from sentinels: OK +00:08:45> (init) Create a primary-replicas cluster of 5 instances: OK +00:08:45> (init) Sentinels can start monitoring a primary: OK +00:08:45> (init) Sentinels can talk with the primary: OK +00:08:45> (init) Sentinels are able to auto-discover other sentinels: OK +00:08:47> (init) Sentinels are able to auto-discover replicas: OK +00:08:47> SENTINEL SIMULATE-FAILURE crash-after-promotion works: OK Testing unit: 06-ckquorum.tcl -14:07:51> (init) Restart killed instances: OK -14:07:51> (init) Remove old primary entry from sentinels: OK -14:07:53> (init) Create a primary-replicas cluster of 5 instances: OK -14:07:53> (init) Sentinels can start monitoring a primary: OK -14:08:04> (init) Sentinels can talk with the primary: OK -14:08:04> (init) Sentinels are able to auto-discover other sentinels: OK -14:08:04> (init) Sentinels are able to auto-discover replicas: OK -14:08:04> CKQUORUM reports OK and the right amount of Sentinels: OK -14:08:04> CKQUORUM detects quorum cannot be reached: OK -14:08:04> CKQUORUM detects failover authorization cannot be reached: OK +00:08:49> (init) Restart killed instances: OK +00:08:49> (init) Remove old primary entry from sentinels: OK +00:08:49> (init) Create a primary-replicas cluster of 5 instances: OK +00:08:49> (init) Sentinels can start monitoring a primary: OK +00:08:49> (init) Sentinels can talk with the primary: OK +00:08:49> (init) Sentinels are able to auto-discover other sentinels: OK +00:08:51> (init) Sentinels are able to auto-discover replicas: OK +00:08:51> CKQUORUM reports OK and the right amount of Sentinels: OK +00:08:51> CKQUORUM detects quorum cannot be reached: OK +00:08:51> CKQUORUM detects failover authorization cannot be reached: OK Testing unit: 07-down-conditions.tcl -14:08:07> (init) Restart killed instances: OK -14:08:07> (init) Remove old primary entry from sentinels: OK -14:08:07> (init) Create a primary-replicas cluster of 5 instances: OK -14:08:07> (init) Sentinels can start monitoring a primary: OK -14:08:08> (init) Sentinels can talk with the primary: OK -14:08:08> (init) Sentinels are able to auto-discover other sentinels: OK -14:08:13> (init) Sentinels are able to auto-discover replicas: OK -14:08:14> Crash the majority of Sentinels to prevent failovers for this unit: OK -14:08:15> SDOWN is triggered by non-responding but not crashed instance: OK -14:08:17> SDOWN is triggered by crashed instance: OK -14:08:20> SDOWN is triggered by primaries advertising as slaves: OK -14:08:24> SDOWN is triggered by misconfigured instance replying with errors: OK -14:08:35> SDOWN is triggered if we rename PING to PONG: OK +00:08:53> (init) Restart killed instances: OK +00:08:53> (init) Remove old primary entry from sentinels: OK +00:08:53> (init) Create a primary-replicas cluster of 5 instances: OK +00:08:53> (init) Sentinels can start monitoring a primary: OK +00:08:53> (init) Sentinels can talk with the primary: OK +00:08:53> (init) Sentinels are able to auto-discover other sentinels: OK +00:08:55> (init) Sentinels are able to auto-discover replicas: OK +00:08:55> Crash the majority of Sentinels to prevent failovers for this unit: OK +00:08:55> SDOWN is triggered by non-responding but not crashed instance: OK +00:08:58> SDOWN is triggered by crashed instance: OK +00:09:00> SDOWN is triggered by primaries advertising as slaves: OK +00:09:04> SDOWN is triggered by misconfigured instance replying with errors: OK +00:09:15> SDOWN is triggered if we rename PING to PONG: OK Testing unit: 08-hostname-conf.tcl -14:08:37> (pre-init) Configure instances and sentinel for hostname use: sentinel/0 sentinel/1 sentinel/2 OK -14:08:38> (init) Restart killed instances: OK -14:08:38> (init) Remove old primary entry from sentinels: OK -14:08:38> (init) Create a primary-replicas cluster of 5 instances: OK -14:08:38> (init) Sentinels can start monitoring a primary: OK -14:08:39> (init) Sentinels can talk with the primary: OK -14:08:39> (init) Sentinels are able to auto-discover other sentinels: OK -14:08:41> (init) Sentinels are able to auto-discover replicas: OK -14:08:41> Sentinel announces hostnames: OK -14:08:43> (post-cleanup) Configure instances and sentinel for IPs: OK +00:09:17> (pre-init) Configure instances and sentinel for hostname use: sentinel/0 sentinel/1 sentinel/2 OK +00:09:18> (init) Restart killed instances: OK +00:09:18> (init) Remove old primary entry from sentinels: OK +00:09:18> (init) Create a primary-replicas cluster of 5 instances: OK +00:09:18> (init) Sentinels can start monitoring a primary: OK +00:09:18> (init) Sentinels can talk with the primary: OK +00:09:18> (init) Sentinels are able to auto-discover other sentinels: OK +00:09:20> (init) Sentinels are able to auto-discover replicas: OK +00:09:20> Sentinel announces hostnames: OK +00:09:20> (post-cleanup) Configure instances and sentinel for IPs: OK Testing unit: 09-acl-support.tcl -14:08:44> (init) Restart killed instances: OK -14:08:44> (init) Remove old primary entry from sentinels: OK -14:08:44> (init) Create a primary-replicas cluster of 5 instances: OK -14:08:44> (init) Sentinels can start monitoring a primary: OK -14:08:45> (init) Sentinels can talk with the primary: OK -14:08:45> (init) Sentinels are able to auto-discover other sentinels: OK -14:08:46> (init) Sentinels are able to auto-discover replicas: OK -14:08:46> (post-init) Set up ACL configuration: OK -14:08:46> SENTINEL CONFIG SET handles on-the-fly credentials reconfiguration: OK -14:08:50> (post-cleanup) Tear down ACL configuration: OK +00:09:20> (init) Restart killed instances: OK +00:09:20> (init) Remove old primary entry from sentinels: OK +00:09:20> (init) Create a primary-replicas cluster of 5 instances: OK +00:09:20> (init) Sentinels can start monitoring a primary: OK +00:09:20> (init) Sentinels can talk with the primary: OK +00:09:20> (init) Sentinels are able to auto-discover other sentinels: OK +00:09:22> (init) Sentinels are able to auto-discover replicas: OK +00:09:22> (post-init) Set up ACL configuration: OK +00:09:22> SENTINEL CONFIG SET handles on-the-fly credentials reconfiguration: OK +00:09:29> (post-cleanup) Tear down ACL configuration: OK Testing unit: 10-replica-priority.tcl -14:08:54> (init) Restart killed instances: OK -14:08:54> (init) Remove old primary entry from sentinels: OK -14:08:54> (init) Create a primary-replicas cluster of 5 instances: OK -14:08:54> (init) Sentinels can start monitoring a primary: OK -14:08:55> (init) Sentinels can talk with the primary: OK -14:08:55> (init) Sentinels are able to auto-discover other sentinels: OK -14:08:56> (init) Sentinels are able to auto-discover replicas: OK -14:08:56> Check acceptable replica-priority values: OK -14:08:56> Set replica-announced=yes on all replicas: OK -14:08:56> Check sentinel replies with 4 replicas: OK -14:08:56> Set replica-announced=no on 2 replicas: OK -14:08:56> Check sentinel replies with 2 replicas: OK -14:08:57> Set replica-announced=yes on all replicas: OK -14:08:57> Check sentinel replies with 4 replicas: OK +00:09:29> (init) Restart killed instances: OK +00:09:29> (init) Remove old primary entry from sentinels: OK +00:09:29> (init) Create a primary-replicas cluster of 5 instances: OK +00:09:29> (init) Sentinels can start monitoring a primary: OK +00:09:29> (init) Sentinels can talk with the primary: OK +00:09:29> (init) Sentinels are able to auto-discover other sentinels: OK +00:09:31> (init) Sentinels are able to auto-discover replicas: OK +00:09:31> Check acceptable replica-priority values: OK +00:09:31> Set replica-announced=yes on all replicas: OK +00:09:31> Check sentinel replies with 4 replicas: OK +00:09:31> Set replica-announced=no on 2 replicas: OK +00:09:31> Check sentinel replies with 2 replicas: OK +00:09:31> Set replica-announced=yes on all replicas: OK +00:09:31> Check sentinel replies with 4 replicas: OK Testing unit: 11-port-0.tcl -14:08:57> (init) Restart killed instances: OK -14:08:57> (init) Remove old primary entry from sentinels: OK -14:08:57> (init) Create a primary-replicas cluster of 5 instances: OK -14:08:57> (init) Sentinels can start monitoring a primary: OK -14:08:59> (init) Sentinels can talk with the primary: OK -14:08:59> (init) Sentinels are able to auto-discover other sentinels: OK -14:08:59> (init) Sentinels are able to auto-discover replicas: OK -14:08:59> Start/Stop sentinel on same port with a different runID should not change the total number of sentinels: OK +00:09:31> (init) Restart killed instances: OK +00:09:31> (init) Remove old primary entry from sentinels: OK +00:09:31> (init) Create a primary-replicas cluster of 5 instances: OK +00:09:31> (init) Sentinels can start monitoring a primary: OK +00:09:31> (init) Sentinels can talk with the primary: OK +00:09:31> (init) Sentinels are able to auto-discover other sentinels: OK +00:09:31> (init) Sentinels are able to auto-discover replicas: OK +00:09:31> Start/Stop sentinel on same port with a different runID should not change the total number of sentinels: OK Testing unit: 12-primary-reboot.tcl -14:09:00> (init) Restart killed instances: sentinel/4 OK -14:09:00> (init) Remove old primary entry from sentinels: OK -14:09:04> (init) Create a primary-replicas cluster of 5 instances: OK -14:09:04> (init) Sentinels can start monitoring a primary: OK -14:09:09> (init) Sentinels can talk with the primary: OK -14:09:09> (init) Sentinels are able to auto-discover other sentinels: OK -14:09:09> (init) Sentinels are able to auto-discover replicas: OK -14:09:09> Primary reboot in very short time: OK -14:09:24> New primary 127.0.0.1:30002 role matches: OK -14:09:24> All the other slaves now point to the new primary: OK -14:09:24> The old primary eventually gets reconfigured as a slave: OK +00:09:33> (init) Restart killed instances: sentinel/4 OK +00:09:33> (init) Remove old primary entry from sentinels: OK +00:09:33> (init) Create a primary-replicas cluster of 5 instances: OK +00:09:33> (init) Sentinels can start monitoring a primary: OK +00:09:33> (init) Sentinels can talk with the primary: OK +00:09:33> (init) Sentinels are able to auto-discover other sentinels: OK +00:09:35> (init) Sentinels are able to auto-discover replicas: OK +00:09:35> Primary reboot in very short time: OK +00:09:43> New primary 127.0.0.1:30002 role matches: OK +00:09:43> All the other slaves now point to the new primary: OK +00:09:43> The old primary eventually gets reconfigured as a slave: OK Testing unit: 13-info-command.tcl -14:09:59> (init) Restart killed instances: OK -14:09:59> (init) Remove old primary entry from sentinels: OK -14:09:59> (init) Create a primary-replicas cluster of 5 instances: OK -14:09:59> (init) Sentinels can start monitoring a primary: OK -14:10:00> (init) Sentinels can talk with the primary: OK -14:10:00> (init) Sentinels are able to auto-discover other sentinels: OK -14:10:02> (init) Sentinels are able to auto-discover replicas: OK -14:10:02> info command with at most one argument: OK -14:10:02> info command with one sub-section: OK -14:10:02> info command with multiple sub-sections: OK +00:09:53> (init) Restart killed instances: OK +00:09:53> (init) Remove old primary entry from sentinels: OK +00:09:53> (init) Create a primary-replicas cluster of 5 instances: OK +00:09:53> (init) Sentinels can start monitoring a primary: OK +00:09:53> (init) Sentinels can talk with the primary: OK +00:09:53> (init) Sentinels are able to auto-discover other sentinels: OK +00:09:55> (init) Sentinels are able to auto-discover replicas: OK +00:09:55> info command with at most one argument: OK +00:09:55> info command with one sub-section: OK +00:09:55> info command with multiple sub-sections: OK Testing unit: 14-debug-command.tcl -14:10:02> (init) Restart killed instances: OK -14:10:02> (init) Remove old primary entry from sentinels: OK -14:10:02> (init) Create a primary-replicas cluster of 5 instances: OK -14:10:02> (init) Sentinels can start monitoring a primary: OK -14:10:03> (init) Sentinels can talk with the primary: OK -14:10:03> (init) Sentinels are able to auto-discover other sentinels: OK -14:10:04> (init) Sentinels are able to auto-discover replicas: OK -14:10:04> Sentinel debug test with arguments and without argument: OK +00:09:55> (init) Restart killed instances: OK +00:09:55> (init) Remove old primary entry from sentinels: OK +00:09:55> (init) Create a primary-replicas cluster of 5 instances: OK +00:09:55> (init) Sentinels can start monitoring a primary: OK +00:09:55> (init) Sentinels can talk with the primary: OK +00:09:55> (init) Sentinels are able to auto-discover other sentinels: OK +00:09:57> (init) Sentinels are able to auto-discover replicas: OK +00:09:57> Sentinel debug test with arguments and without argument: OK Testing unit: 15-sentinel-deprecated-commands.tcl -14:10:04> (init) Restart killed instances: OK -14:10:04> (init) Remove old primary entry from sentinels: OK -14:10:05> (init) Create a primary-replicas cluster of 5 instances: OK -14:10:05> (init) Sentinels can start monitoring a primary: OK -14:10:06> (init) Sentinels can talk with the primary: OK -14:10:06> (init) Sentinels are able to auto-discover other sentinels: OK -14:10:07> (init) Sentinels are able to auto-discover replicas: OK -14:10:07> SENTINEL MASTERS returns a list of monitored masters (SENTINEL MASTERS as a deprecated command): OK -14:10:07> SENTINEL SLAVES returns a list of the monitored slaves (SENTINEL SLAVES as a deprecated command): OK -14:10:07> SENTINEL MASTER returns the information list of the monitored master (SENTINEL MASTER as a deprecated command): OK -14:10:07> SENTINEL IS-MASTER-DOWN-BY-ADDR checks if the primary is down (SENTINEL IS-MASTER-DOWN-BY-ADDR as a deprecated command): OK +00:09:57> (init) Restart killed instances: OK +00:09:57> (init) Remove old primary entry from sentinels: OK +00:09:57> (init) Create a primary-replicas cluster of 5 instances: OK +00:09:57> (init) Sentinels can start monitoring a primary: OK +00:09:57> (init) Sentinels can talk with the primary: OK +00:09:57> (init) Sentinels are able to auto-discover other sentinels: OK +00:10:00> (init) Sentinels are able to auto-discover replicas: OK +00:10:00> SENTINEL MASTERS returns a list of monitored masters (SENTINEL MASTERS as a deprecated command): OK +00:10:00> SENTINEL SLAVES returns a list of the monitored slaves (SENTINEL SLAVES as a deprecated command): OK +00:10:00> SENTINEL MASTER returns the information list of the monitored master (SENTINEL MASTER as a deprecated command): OK +00:10:00> SENTINEL IS-MASTER-DOWN-BY-ADDR checks if the primary is down (SENTINEL IS-MASTER-DOWN-BY-ADDR as a deprecated command): OK Testing unit: 16-config-set-config-get.tcl -14:10:07> (init) Restart killed instances: OK -14:10:07> (init) Remove old primary entry from sentinels: OK -14:10:08> (init) Create a primary-replicas cluster of 5 instances: OK -14:10:08> (init) Sentinels can start monitoring a primary: OK -14:10:17> (init) Sentinels can talk with the primary: OK -14:10:17> (init) Sentinels are able to auto-discover other sentinels: OK -14:10:17> (init) Sentinels are able to auto-discover replicas: OK -14:10:17> SENTINEL CONFIG SET and SENTINEL CONFIG GET handles multiple variables: OK -14:10:18> SENTINEL CONFIG GET for duplicate and unknown variables: OK -14:10:18> SENTINEL CONFIG GET for patterns: OK -14:10:18> SENTINEL CONFIG SET duplicate variables: OK -14:10:18> SENTINEL CONFIG SET, one option does not exist: OK -14:10:18> SENTINEL CONFIG SET, one option with wrong value: OK -14:10:18> SENTINEL CONFIG SET, wrong number of arguments: OK +00:10:02> (init) Restart killed instances: OK +00:10:02> (init) Remove old primary entry from sentinels: OK +00:10:05> (init) Create a primary-replicas cluster of 5 instances: OK +00:10:05> (init) Sentinels can start monitoring a primary: OK +00:10:05> (init) Sentinels can talk with the primary: OK +00:10:05> (init) Sentinels are able to auto-discover other sentinels: OK +00:10:07> (init) Sentinels are able to auto-discover replicas: OK +00:10:07> SENTINEL CONFIG SET and SENTINEL CONFIG GET handles multiple variables: OK +00:10:07> SENTINEL CONFIG GET for duplicate and unknown variables: OK +00:10:07> SENTINEL CONFIG GET for patterns: OK +00:10:07> SENTINEL CONFIG SET duplicate variables: OK +00:10:07> SENTINEL CONFIG SET, one option does not exist: OK +00:10:07> SENTINEL CONFIG SET, one option with wrong value: OK +00:10:07> SENTINEL CONFIG SET, wrong number of arguments: OK Cleaning up... -killing stale instance 3195588 -killing stale instance 3195625 -killing stale instance 3195644 -killing stale instance 3203843 -killing stale instance 3205905 -killing stale instance 3207873 -killing stale instance 3210193 -killing stale instance 3213606 -killing stale instance 3213632 -killing stale instance 3213662 -killing stale instance 3216749 -killing stale instance 3216807 -killing stale instance 3218236 +killing stale instance 2684105 +killing stale instance 2684117 +killing stale instance 2684123 +killing stale instance 2690947 +killing stale instance 2691408 +killing stale instance 2691928 +killing stale instance 2693500 +killing stale instance 2696309 +killing stale instance 2696315 +killing stale instance 2696326 +killing stale instance 2697878 +killing stale instance 2697909 +killing stale instance 2698341 GOOD! No errors. # Clean up after gen-test-certs.sh rm -rf tests/tls || true # Other cleanup find tests/tmp ! -name .gitignore -type f -exec rm -rfv {} + +removed 'tests/tmp/server.30294.11/stderr' +removed 'tests/tmp/server.30294.11/stdout' +removed 'tests/tmp/server.30294.470/nodes.conf' +removed 'tests/tmp/server.30294.470/stderr' +removed 'tests/tmp/server.30294.470/stdout' +removed 'tests/tmp/server.30294.218/stderr' +removed 'tests/tmp/server.30294.218/stdout' +removed 'tests/tmp/valkey.conf.30294.227' +removed 'tests/tmp/server.30294.177/stderr' +removed 'tests/tmp/server.30294.177/stdout' +removed 'tests/tmp/valkey.conf.30294.577' +removed 'tests/tmp/server.30294.648/nodes.conf' +removed 'tests/tmp/server.30294.648/stderr' +removed 'tests/tmp/server.30294.648/stdout' +removed 'tests/tmp/server.30294.796/nodes.conf' +removed 'tests/tmp/server.30294.796/stderr' +removed 'tests/tmp/server.30294.796/stdout' +removed 'tests/tmp/valkey.conf.30294.557' +removed 'tests/tmp/valkey.conf.30294.353' +removed 'tests/tmp/server.30294.306/stderr' +removed 'tests/tmp/server.30294.306/stdout' +removed 'tests/tmp/valkey.conf.30294.46' +removed 'tests/tmp/valkey.conf.30294.555' +removed 'tests/tmp/server.30294.135/stderr' +removed 'tests/tmp/server.30294.135/stdout' +removed 'tests/tmp/valkey.conf.30294.205' +removed 'tests/tmp/server.30294.93/stderr' +removed 'tests/tmp/server.30294.93/stdout' +removed 'tests/tmp/server.30294.494/nodes.conf' +removed 'tests/tmp/server.30294.494/stderr' +removed 'tests/tmp/server.30294.494/stdout' +removed 'tests/tmp/valkey.conf.30294.50' +removed 'tests/tmp/valkey.conf.30294.629' +removed 'tests/tmp/server.30294.762/nodes.conf' +removed 'tests/tmp/server.30294.762/stderr' +removed 'tests/tmp/server.30294.762/stdout' +removed 'tests/tmp/server.30294.91/stderr' +removed 'tests/tmp/server.30294.91/stdout' +removed 'tests/tmp/valkey.conf.30294.455' +removed 'tests/tmp/valkey.conf.30294.647' +removed 'tests/tmp/valkey.conf.30294.291' +removed 'tests/tmp/valkey.conf.30294.645' +removed 'tests/tmp/valkey.conf.30294.68' +removed 'tests/tmp/server.30294.27/stderr' +removed 'tests/tmp/server.30294.27/stdout' +removed 'tests/tmp/valkey.conf.30294.295' +removed 'tests/tmp/server.30294.226/stderr' +removed 'tests/tmp/server.30294.226/stdout' +removed 'tests/tmp/valkey.conf.30294.509' +removed 'tests/tmp/server.30294.51/nodes.conf' +removed 'tests/tmp/server.30294.51/stderr' +removed 'tests/tmp/server.30294.51/stdout' +removed 'tests/tmp/server.30294.57/stderr' +removed 'tests/tmp/server.30294.57/stdout' +removed 'tests/tmp/server.30294.720/nodes.conf' +removed 'tests/tmp/server.30294.720/stderr' +removed 'tests/tmp/server.30294.720/stdout' +removed 'tests/tmp/server.30294.438/stderr' +removed 'tests/tmp/server.30294.438/stdout' +removed 'tests/tmp/server.30294.480/nodes.conf' +removed 'tests/tmp/server.30294.480/stderr' +removed 'tests/tmp/server.30294.480/stdout' +removed 'tests/tmp/server.30294.778/nodes.conf' +removed 'tests/tmp/server.30294.778/stderr' +removed 'tests/tmp/server.30294.778/stdout' +removed 'tests/tmp/valkey.conf.30294.331' +removed 'tests/tmp/server.30294.238/stderr' +removed 'tests/tmp/server.30294.238/stdout' +removed 'tests/tmp/valkey.conf.30294.130' +removed 'tests/tmp/server.30294.33/stderr' +removed 'tests/tmp/server.30294.33/stdout' +removed 'tests/tmp/server.30294.290/stderr' +removed 'tests/tmp/server.30294.290/stdout' +removed 'tests/tmp/server.30294.123/stderr' +removed 'tests/tmp/server.30294.123/stdout' +removed 'tests/tmp/server.30294.404/stderr' +removed 'tests/tmp/server.30294.404/stdout' +removed 'tests/tmp/valkey.conf.30294.429' +removed 'tests/tmp/server.30294.7/stderr' +removed 'tests/tmp/server.30294.7/stdout' +removed 'tests/tmp/server.30294.310/stderr' +removed 'tests/tmp/server.30294.310/stdout' +removed 'tests/tmp/valkey.conf.30294.201' +removed 'tests/tmp/server.30294.522/nodes.conf' +removed 'tests/tmp/server.30294.522/stderr' +removed 'tests/tmp/server.30294.522/stdout' +removed 'tests/tmp/server.30294.506/nodes.conf' +removed 'tests/tmp/server.30294.506/stderr' +removed 'tests/tmp/server.30294.506/stdout' +removed 'tests/tmp/valkey.conf.30294.150' +removed 'tests/tmp/valkey.conf.30294.715' +removed 'tests/tmp/valkey.conf.30294.132' +removed 'tests/tmp/valkey.conf.30294.285' +removed 'tests/tmp/valkey.conf.30294.605' +removed 'tests/tmp/server.30294.802/nodes.conf' +removed 'tests/tmp/server.30294.802/stderr' +removed 'tests/tmp/server.30294.802/stdout' +removed 'tests/tmp/server.30294.424/stderr' +removed 'tests/tmp/server.30294.424/stdout' +removed 'tests/tmp/valkey.conf.30294.551' +removed 'tests/tmp/valkey.conf.30294.401' +removed 'tests/tmp/server.30294.540/nodes.conf' +removed 'tests/tmp/server.30294.540/stderr' +removed 'tests/tmp/server.30294.540/stdout' +removed 'tests/tmp/valkey.conf.30294.180' +removed 'tests/tmp/server.30294.826/nodes.conf' +removed 'tests/tmp/server.30294.826/stderr' +removed 'tests/tmp/server.30294.826/stdout' +removed 'tests/tmp/server.30294.630/nodes.conf' +removed 'tests/tmp/server.30294.630/stderr' +removed 'tests/tmp/server.30294.630/stdout' +removed 'tests/tmp/valkey.conf.30294.18' +removed 'tests/tmp/valkey.conf.30294.657' +removed 'tests/tmp/server.30294.688/nodes.conf' +removed 'tests/tmp/server.30294.688/stderr' +removed 'tests/tmp/server.30294.688/stdout' +removed 'tests/tmp/valkey.conf.30294.247' +removed 'tests/tmp/server.30294.832/nodes.conf' +removed 'tests/tmp/server.30294.832/stderr' +removed 'tests/tmp/server.30294.832/stdout' +removed 'tests/tmp/server.30294.129/stderr' +removed 'tests/tmp/server.30294.129/stdout' +removed 'tests/tmp/server.30294.354/stderr' +removed 'tests/tmp/server.30294.354/stdout' +removed 'tests/tmp/valkey.conf.30294.241' +removed 'tests/tmp/valkey.conf.30294.86' +removed 'tests/tmp/server.30294.268/stderr' +removed 'tests/tmp/server.30294.268/stdout' +removed 'tests/tmp/valkey.conf.30294.309' +removed 'tests/tmp/valkey.conf.30294.495' +removed 'tests/tmp/valkey.conf.30294.391' +removed 'tests/tmp/valkey.conf.30294.823' +removed 'tests/tmp/valkey.conf.30294.521' +removed 'tests/tmp/server.30294.596/nodes.conf' +removed 'tests/tmp/server.30294.596/stderr' +removed 'tests/tmp/server.30294.596/stdout' +removed 'tests/tmp/valkey.conf.30294.749' +removed 'tests/tmp/server.30294.212/stderr' +removed 'tests/tmp/server.30294.212/stdout' +removed 'tests/tmp/valkey.conf.30294.627' +removed 'tests/tmp/valkey.conf.30294.573' +removed 'tests/tmp/server.30294.788/nodes.conf' +removed 'tests/tmp/server.30294.788/stderr' +removed 'tests/tmp/server.30294.788/stdout' +removed 'tests/tmp/valkey.conf.30294.70' +removed 'tests/tmp/valkey.conf.30294.207' +removed 'tests/tmp/valkey.conf.30294.593' +removed 'tests/tmp/server.30294.464/stderr' +removed 'tests/tmp/server.30294.464/stdout' +removed 'tests/tmp/server.30294.496/nodes.conf' +removed 'tests/tmp/server.30294.496/stderr' +removed 'tests/tmp/server.30294.496/stdout' +removed 'tests/tmp/server.30294.508/nodes.conf' +removed 'tests/tmp/server.30294.508/stderr' +removed 'tests/tmp/server.30294.508/stdout' +removed 'tests/tmp/valkey.conf.30294.337' +removed 'tests/tmp/valkey.conf.30294.299' +removed 'tests/tmp/valkey.conf.30294.413' +removed 'tests/tmp/server.30294.782/nodes.conf' +removed 'tests/tmp/server.30294.782/stderr' +removed 'tests/tmp/server.30294.782/stdout' +removed 'tests/tmp/valkey.conf.30294.471' +removed 'tests/tmp/server.30294.416/stderr' +removed 'tests/tmp/server.30294.416/stdout' +removed 'tests/tmp/server.30294.350/stderr' +removed 'tests/tmp/server.30294.350/stdout' +removed 'tests/tmp/valkey.conf.30294.253' +removed 'tests/tmp/server.30294.3/stderr' +removed 'tests/tmp/server.30294.3/stdout' +removed 'tests/tmp/valkey.conf.30294.595' +removed 'tests/tmp/server.30294.220/stderr' +removed 'tests/tmp/server.30294.220/stdout' +removed 'tests/tmp/valkey.conf.30294.691' +removed 'tests/tmp/server.30294.262/stderr' +removed 'tests/tmp/server.30294.262/stdout' +removed 'tests/tmp/valkey.conf.30294.363' +removed 'tests/tmp/server.30294.652/nodes.conf' +removed 'tests/tmp/server.30294.652/stderr' +removed 'tests/tmp/server.30294.652/stdout' +removed 'tests/tmp/valkey.conf.30294.134' +removed 'tests/tmp/server.30294.560/nodes.conf' +removed 'tests/tmp/server.30294.560/stderr' +removed 'tests/tmp/server.30294.560/stdout' +removed 'tests/tmp/server.30294.318/stderr' +removed 'tests/tmp/server.30294.318/stdout' +removed 'tests/tmp/valkey.conf.30294.769' +removed 'tests/tmp/server.30294.504/nodes.conf' +removed 'tests/tmp/server.30294.504/stderr' +removed 'tests/tmp/server.30294.504/stdout' +removed 'tests/tmp/valkey.conf.30294.591' +removed 'tests/tmp/server.30294.478/nodes.conf' +removed 'tests/tmp/server.30294.478/stderr' +removed 'tests/tmp/server.30294.478/stdout' +removed 'tests/tmp/server.30294.246/stderr' +removed 'tests/tmp/server.30294.246/stdout' +removed 'tests/tmp/valkey.conf.30294.799' +removed 'tests/tmp/valkey.conf.30294.803' +removed 'tests/tmp/valkey.conf.30294.617' +removed 'tests/tmp/valkey.conf.30294.245' +removed 'tests/tmp/valkey.conf.30294.697' +removed 'tests/tmp/server.30294.366/stderr' +removed 'tests/tmp/server.30294.366/stdout' +removed 'tests/tmp/server.30294.468/stderr' +removed 'tests/tmp/server.30294.468/stdout' +removed 'tests/tmp/server.30294.718/nodes.conf' +removed 'tests/tmp/server.30294.718/stderr' +removed 'tests/tmp/server.30294.718/stdout' +removed 'tests/tmp/server.30294.37/stderr' +removed 'tests/tmp/server.30294.37/stdout' +removed 'tests/tmp/server.30294.260/stderr' +removed 'tests/tmp/server.30294.260/stdout' +removed 'tests/tmp/valkey.conf.30294.62' +removed 'tests/tmp/valkey.conf.30294.581' +removed 'tests/tmp/server.30294.584/nodes.conf' +removed 'tests/tmp/server.30294.584/stderr' +removed 'tests/tmp/server.30294.584/stdout' +removed 'tests/tmp/valkey.conf.30294.623' +removed 'tests/tmp/valkey.conf.30294.611' +removed 'tests/tmp/valkey.conf.30294.585' +removed 'tests/tmp/valkey.conf.30294.24' +removed 'tests/tmp/valkey.conf.30294.26' +removed 'tests/tmp/server.30294.638/nodes.conf' +removed 'tests/tmp/server.30294.638/stderr' +removed 'tests/tmp/server.30294.638/stdout' +removed 'tests/tmp/valkey.conf.30294.625' +removed 'tests/tmp/valkey.conf.30294.74' +removed 'tests/tmp/server.30294.698/nodes.conf' +removed 'tests/tmp/server.30294.698/stderr' +removed 'tests/tmp/server.30294.698/stdout' +removed 'tests/tmp/server.30294.798/nodes.conf' +removed 'tests/tmp/server.30294.798/stderr' +removed 'tests/tmp/server.30294.798/stdout' +removed 'tests/tmp/server.30294.358/stderr' +removed 'tests/tmp/server.30294.358/stdout' +removed 'tests/tmp/server.30294.674/nodes.conf' +removed 'tests/tmp/server.30294.674/stderr' +removed 'tests/tmp/server.30294.674/stdout' +removed 'tests/tmp/server.30294.708/nodes.conf' +removed 'tests/tmp/server.30294.708/stderr' +removed 'tests/tmp/server.30294.708/stdout' +removed 'tests/tmp/valkey.conf.30294.487' +removed 'tests/tmp/valkey.conf.30294.64' +removed 'tests/tmp/valkey.conf.30294.735' +removed 'tests/tmp/server.30294.184/stderr' +removed 'tests/tmp/server.30294.184/stdout' +removed 'tests/tmp/valkey.conf.30294.377' +removed 'tests/tmp/server.30294.208/stderr' +removed 'tests/tmp/server.30294.208/stdout' +removed 'tests/tmp/server.30294.234/stderr' +removed 'tests/tmp/server.30294.234/stdout' +removed 'tests/tmp/valkey.conf.30294.637' +removed 'tests/tmp/valkey.conf.30294.176' +removed 'tests/tmp/valkey.conf.30294.589' +removed 'tests/tmp/server.30294.63/stderr' +removed 'tests/tmp/server.30294.63/stdout' +removed 'tests/tmp/server.30294.542/nodes.conf' +removed 'tests/tmp/server.30294.542/stderr' +removed 'tests/tmp/server.30294.542/stdout' +removed 'tests/tmp/valkey.conf.30294.313' +removed 'tests/tmp/valkey.conf.30294.599' +removed 'tests/tmp/server.30294.678/nodes.conf' +removed 'tests/tmp/server.30294.678/stderr' +removed 'tests/tmp/server.30294.678/stdout' +removed 'tests/tmp/valkey.conf.30294.188' +removed 'tests/tmp/server.30294.786/nodes.conf' +removed 'tests/tmp/server.30294.786/stderr' +removed 'tests/tmp/server.30294.786/stdout' +removed 'tests/tmp/valkey.conf.30294.651' +removed 'tests/tmp/server.30294.686/nodes.conf' +removed 'tests/tmp/server.30294.686/stderr' +removed 'tests/tmp/server.30294.686/stdout' +removed 'tests/tmp/server.30294.137/stderr' +removed 'tests/tmp/server.30294.137/stdout' +removed 'tests/tmp/server.30294.484/nodes.conf' +removed 'tests/tmp/server.30294.484/stderr' +removed 'tests/tmp/server.30294.484/stdout' +removed 'tests/tmp/valkey.conf.30294.174' +removed 'tests/tmp/valkey.conf.30294.373' +removed 'tests/tmp/server.30294.626/nodes.conf' +removed 'tests/tmp/server.30294.626/stderr' +removed 'tests/tmp/server.30294.626/stdout' +removed 'tests/tmp/valkey.conf.30294.499' +removed 'tests/tmp/server.30294.340/stderr' +removed 'tests/tmp/server.30294.340/stdout' +removed 'tests/tmp/valkey.conf.30294.701' +removed 'tests/tmp/valkey.conf.30294.569' +removed 'tests/tmp/valkey.conf.30294.839' +removed 'tests/tmp/server.30294.486/nodes.conf' +removed 'tests/tmp/server.30294.486/stderr' +removed 'tests/tmp/server.30294.486/stdout' +removed 'tests/tmp/server.30294.157/stderr' +removed 'tests/tmp/server.30294.157/stdout' +removed 'tests/tmp/valkey.conf.30294.148' +removed 'tests/tmp/valkey.conf.30294.387' +removed 'tests/tmp/server.30294.476/nodes.conf' +removed 'tests/tmp/server.30294.476/stderr' +removed 'tests/tmp/server.30294.476/stdout' +removed 'tests/tmp/valkey.conf.30294.621' +removed 'tests/tmp/valkey.conf.30294.507' +removed 'tests/tmp/server.30294.121/stderr' +removed 'tests/tmp/server.30294.121/stdout' +removed 'tests/tmp/resetchannels.acl.30294.192/nodefaultuser.acl' +removed 'tests/tmp/resetchannels.acl.30294.192/stderr' +removed 'tests/tmp/resetchannels.acl.30294.192/stdout' +removed 'tests/tmp/resetchannels.acl.30294.192/default.conf' +removed 'tests/tmp/valkey.conf.30294.335' +removed 'tests/tmp/valkey.conf.30294.52' +removed 'tests/tmp/valkey.conf.30294.677' +removed 'tests/tmp/valkey.conf.30294.797' +removed 'tests/tmp/valkey.conf.30294.371' +removed 'tests/tmp/valkey.conf.30294.475' +removed 'tests/tmp/server.30294.109/stderr' +removed 'tests/tmp/server.30294.109/stdout' +removed 'tests/tmp/valkey.conf.30294.435' +removed 'tests/tmp/valkey.conf.30294.267' +removed 'tests/tmp/server.30294.55/stderr' +removed 'tests/tmp/server.30294.55/stdout' +removed 'tests/tmp/valkey.conf.30294.445' +removed 'tests/tmp/valkey.conf.30294.545' +removed 'tests/tmp/server.30294.734/nodes.conf' +removed 'tests/tmp/server.30294.734/stderr' +removed 'tests/tmp/server.30294.734/stdout' +removed 'tests/tmp/valkey.conf.30294.505' +removed 'tests/tmp/server.30294.728/nodes.conf' +removed 'tests/tmp/server.30294.728/stderr' +removed 'tests/tmp/server.30294.728/stdout' +removed 'tests/tmp/valkey.conf.30294.297' +removed 'tests/tmp/server.30294.530/nodes.conf' +removed 'tests/tmp/server.30294.530/stderr' +removed 'tests/tmp/server.30294.530/stdout' +removed 'tests/tmp/server.30294.730/nodes.conf' +removed 'tests/tmp/server.30294.730/stderr' +removed 'tests/tmp/server.30294.730/stdout' +removed 'tests/tmp/server.30294.566/nodes.conf' +removed 'tests/tmp/server.30294.566/stderr' +removed 'tests/tmp/server.30294.566/stdout' +removed 'tests/tmp/server.30294.45/stderr' +removed 'tests/tmp/server.30294.45/stdout' +removed 'tests/tmp/server.30294.742/nodes.conf' +removed 'tests/tmp/server.30294.742/stderr' +removed 'tests/tmp/server.30294.742/stdout' +removed 'tests/tmp/valkey.conf.30294.419' +removed 'tests/tmp/server.30294.1/stderr' +removed 'tests/tmp/server.30294.1/stdout' +removed 'tests/tmp/server.30294.232/stderr' +removed 'tests/tmp/server.30294.232/stdout' +removed 'tests/tmp/valkey.conf.30294.281' +removed 'tests/tmp/valkey.conf.30294.607' +removed 'tests/tmp/valkey.conf.30294.178' +removed 'tests/tmp/server.30294.23/stderr' +removed 'tests/tmp/server.30294.23/stdout' +removed 'tests/tmp/server.30294.714/nodes.conf' +removed 'tests/tmp/server.30294.714/stderr' +removed 'tests/tmp/server.30294.714/stdout' +removed 'tests/tmp/valkey.conf.30294.601' +removed 'tests/tmp/valkey.conf.30294.301' +removed 'tests/tmp/server.30294.694/nodes.conf' +removed 'tests/tmp/server.30294.694/stderr' +removed 'tests/tmp/server.30294.694/stdout' +removed 'tests/tmp/server.30294.412/stderr' +removed 'tests/tmp/server.30294.412/stdout' +removed 'tests/tmp/server.30294.328/stderr' +removed 'tests/tmp/server.30294.328/stdout' +removed 'tests/tmp/valkey.conf.30294.211' +removed 'tests/tmp/server.30294.704/nodes.conf' +removed 'tests/tmp/server.30294.704/stderr' +removed 'tests/tmp/server.30294.704/stdout' +removed 'tests/tmp/server.30294.204/stderr' +removed 'tests/tmp/server.30294.204/stdout' +removed 'tests/tmp/server.30294.266/stderr' +removed 'tests/tmp/server.30294.266/stdout' +removed 'tests/tmp/server.30294.580/nodes.conf' +removed 'tests/tmp/server.30294.580/stderr' +removed 'tests/tmp/server.30294.580/stdout' +removed 'tests/tmp/valkey.conf.30294.575' +removed 'tests/tmp/valkey.conf.30294.347' +removed 'tests/tmp/server.30294.538/nodes.conf' +removed 'tests/tmp/server.30294.538/stderr' +removed 'tests/tmp/server.30294.538/stdout' +removed 'tests/tmp/valkey.conf.30294.116' +removed 'tests/tmp/valkey.conf.30294.515' +removed 'tests/tmp/valkey.conf.30294.221' +removed 'tests/tmp/server.30294.346/stderr' +removed 'tests/tmp/server.30294.346/stdout' +removed 'tests/tmp/valkey.conf.30294.459' +removed 'tests/tmp/server.30294.101/stderr' +removed 'tests/tmp/server.30294.101/stdout' +removed 'tests/tmp/valkey.conf.30294.170' +removed 'tests/tmp/server.30294.546/nodes.conf' +removed 'tests/tmp/server.30294.546/stderr' +removed 'tests/tmp/server.30294.546/stdout' +removed 'tests/tmp/valkey.conf.30294.703' +removed 'tests/tmp/server.30294.482/nodes.conf' +removed 'tests/tmp/server.30294.482/stderr' +removed 'tests/tmp/server.30294.482/stdout' +removed 'tests/tmp/server.30294.822/nodes.conf' +removed 'tests/tmp/server.30294.822/stderr' +removed 'tests/tmp/server.30294.822/stdout' +removed 'tests/tmp/valkey.conf.30294.675' +removed 'tests/tmp/server.30294.512/nodes.conf' +removed 'tests/tmp/server.30294.512/stderr' +removed 'tests/tmp/server.30294.512/stdout' +removed 'tests/tmp/valkey.conf.30294.114' +removed 'tests/tmp/server.30294.724/nodes.conf' +removed 'tests/tmp/server.30294.724/stderr' +removed 'tests/tmp/server.30294.724/stdout' +removed 'tests/tmp/server.30294.732/nodes.conf' +removed 'tests/tmp/server.30294.732/stderr' +removed 'tests/tmp/server.30294.732/stdout' +removed 'tests/tmp/server.30294.248/stderr' +removed 'tests/tmp/server.30294.248/stdout' +removed 'tests/tmp/server.30294.646/nodes.conf' +removed 'tests/tmp/server.30294.646/stderr' +removed 'tests/tmp/server.30294.646/stdout' +removed 'tests/tmp/valkey.conf.30294.757' +removed 'tests/tmp/server.30294.228/stderr' +removed 'tests/tmp/server.30294.228/stdout' +removed 'tests/tmp/valkey.conf.30294.255' +removed 'tests/tmp/server.30294.498/nodes.conf' +removed 'tests/tmp/server.30294.498/stderr' +removed 'tests/tmp/server.30294.498/stdout' +removed 'tests/tmp/server.30294.338/stderr' +removed 'tests/tmp/server.30294.338/stdout' +removed 'tests/tmp/server.30294.632/nodes.conf' +removed 'tests/tmp/server.30294.632/stderr' +removed 'tests/tmp/server.30294.632/stdout' +removed 'tests/tmp/server.30294.155/stderr' +removed 'tests/tmp/server.30294.155/stdout' +removed 'tests/tmp/valkey.conf.30294.44' +removed 'tests/tmp/valkey.conf.30294.118' +removed 'tests/tmp/server.30294.780/nodes.conf' +removed 'tests/tmp/server.30294.780/stderr' +removed 'tests/tmp/server.30294.780/stdout' +removed 'tests/tmp/server.30294.163/stderr' +removed 'tests/tmp/server.30294.163/stdout' +removed 'tests/tmp/valkey.conf.30294.641' +removed 'tests/tmp/server.30294.558/nodes.conf' +removed 'tests/tmp/server.30294.558/stderr' +removed 'tests/tmp/server.30294.558/stdout' +removed 'tests/tmp/server.30294.716/nodes.conf' +removed 'tests/tmp/server.30294.716/stderr' +removed 'tests/tmp/server.30294.716/stdout' +removed 'tests/tmp/server.30294.230/stderr' +removed 'tests/tmp/server.30294.230/stdout' +removed 'tests/tmp/server.30294.738/nodes.conf' +removed 'tests/tmp/server.30294.738/stderr' +removed 'tests/tmp/server.30294.738/stdout' +removed 'tests/tmp/server.30294.816/nodes.conf' +removed 'tests/tmp/server.30294.816/stderr' +removed 'tests/tmp/server.30294.816/stdout' +removed 'tests/tmp/valkey.conf.30294.160' +removed 'tests/tmp/valkey.conf.30294.80' +removed 'tests/tmp/server.30294.462/stderr' +removed 'tests/tmp/server.30294.462/stdout' +removed 'tests/tmp/valkey.conf.30294.747' +removed 'tests/tmp/server.30294.107/stderr' +removed 'tests/tmp/server.30294.107/stdout' +removed 'tests/tmp/valkey.conf.30294.343' +removed 'tests/tmp/valkey.conf.30294.461' +removed 'tests/tmp/valkey.conf.30294.563' +removed 'tests/tmp/server.30294.89/stderr' +removed 'tests/tmp/server.30294.89/stdout' +removed 'tests/tmp/valkey.conf.30294.449' +removed 'tests/tmp/server.30294.682/nodes.conf' +removed 'tests/tmp/server.30294.682/stderr' +removed 'tests/tmp/server.30294.682/stdout' +removed 'tests/tmp/valkey.conf.30294.765' +removed 'tests/tmp/server.30294.622/nodes.conf' +removed 'tests/tmp/server.30294.622/stderr' +removed 'tests/tmp/server.30294.622/stdout' +removed 'tests/tmp/selectors.acl.30294.181/stderr' +removed 'tests/tmp/selectors.acl.30294.181/stdout' +removed 'tests/tmp/selectors.acl.30294.181/default.conf' +removed 'tests/tmp/selectors.acl.30294.181/userwithselectors.acl' +removed 'tests/tmp/server.30294.746/nodes.conf' +removed 'tests/tmp/server.30294.746/stderr' +removed 'tests/tmp/server.30294.746/stdout' +removed 'tests/tmp/valkey.conf.30294.329' +removed 'tests/tmp/valkey.conf.30294.144' +removed 'tests/tmp/server.30294.696/nodes.conf' +removed 'tests/tmp/server.30294.696/stderr' +removed 'tests/tmp/server.30294.696/stdout' +removed 'tests/tmp/valkey.conf.30294.489' +removed 'tests/tmp/server.30294.492/nodes.conf' +removed 'tests/tmp/server.30294.492/stderr' +removed 'tests/tmp/server.30294.492/stdout' +removed 'tests/tmp/valkey.conf.30294.323' +removed 'tests/tmp/valkey.conf.30294.321' +removed 'tests/tmp/server.30294.406/stderr' +removed 'tests/tmp/server.30294.406/stdout' +removed 'tests/tmp/valkey.conf.30294.745' +removed 'tests/tmp/valkey.conf.30294.807' +removed 'tests/tmp/server.30294.514/nodes.conf' +removed 'tests/tmp/server.30294.514/stderr' +removed 'tests/tmp/server.30294.514/stdout' +removed 'tests/tmp/valkey.conf.30294.146' +removed 'tests/tmp/valkey.conf.30294.421' +removed 'tests/tmp/server.30294.270/stderr' +removed 'tests/tmp/server.30294.270/stdout' +removed 'tests/tmp/valkey.conf.30294.411' +removed 'tests/tmp/valkey.conf.30294.142' +removed 'tests/tmp/valkey.conf.30294.753' +removed 'tests/tmp/valkey.conf.30294.729' +removed 'tests/tmp/server.30294.582/nodes.conf' +removed 'tests/tmp/server.30294.582/stderr' +removed 'tests/tmp/server.30294.582/stdout' +removed 'tests/tmp/valkey.conf.30294.687' +removed 'tests/tmp/server.30294.552/nodes.conf' +removed 'tests/tmp/server.30294.552/stderr' +removed 'tests/tmp/server.30294.552/stdout' +removed 'tests/tmp/valkey.conf.30294.553' +removed 'tests/tmp/server.30294.87/stderr' +removed 'tests/tmp/server.30294.87/stdout' +removed 'tests/tmp/valkey.conf.30294.28' +removed 'tests/tmp/server.30294.175/stderr' +removed 'tests/tmp/server.30294.175/stdout' +removed 'tests/tmp/valkey.conf.30294.76' +removed 'tests/tmp/server.30294.426/stderr' +removed 'tests/tmp/server.30294.426/stdout' +removed 'tests/tmp/server.30294.85/stderr' +removed 'tests/tmp/server.30294.85/stdout' +removed 'tests/tmp/server.30294.432/stderr' +removed 'tests/tmp/server.30294.432/stdout' +removed 'tests/tmp/valkey.conf.30294.269' +removed 'tests/tmp/server.30294.69/stderr' +removed 'tests/tmp/server.30294.69/stdout' +removed 'tests/tmp/valkey.conf.30294.6' +removed 'tests/tmp/valkey.conf.30294.731' +removed 'tests/tmp/valkey.conf.30294.463' +removed 'tests/tmp/server.30294.224/stderr' +removed 'tests/tmp/server.30294.224/stdout' +removed 'tests/tmp/valkey.conf.30294.8' +removed 'tests/tmp/server.30294.99/stderr' +removed 'tests/tmp/server.30294.99/stdout' +removed 'tests/tmp/server.30294.286/stderr' +removed 'tests/tmp/server.30294.286/stdout' +removed 'tests/tmp/server.30294.222/stderr' +removed 'tests/tmp/server.30294.222/stdout' +removed 'tests/tmp/valkey.conf.30294.719' +removed 'tests/tmp/server.30294.127/stderr' +removed 'tests/tmp/server.30294.127/stdout' +removed 'tests/tmp/server.30294.252/stderr' +removed 'tests/tmp/server.30294.252/stdout' +removed 'tests/tmp/server.30294.564/nodes.conf' +removed 'tests/tmp/server.30294.564/stderr' +removed 'tests/tmp/server.30294.564/stdout' +removed 'tests/tmp/server.30294.618/nodes.conf' +removed 'tests/tmp/server.30294.618/stderr' +removed 'tests/tmp/server.30294.618/stdout' +removed 'tests/tmp/server.30294.500/nodes.conf' +removed 'tests/tmp/server.30294.500/stderr' +removed 'tests/tmp/server.30294.500/stdout' +removed 'tests/tmp/server.30294.356/stderr' +removed 'tests/tmp/server.30294.356/stdout' +removed 'tests/tmp/valkey.conf.30294.517' +removed 'tests/tmp/server.30294.758/nodes.conf' +removed 'tests/tmp/server.30294.758/stderr' +removed 'tests/tmp/server.30294.758/stdout' +removed 'tests/tmp/server.30294.31/stderr' +removed 'tests/tmp/server.30294.31/stdout' +removed 'tests/tmp/valkey.conf.30294.243' +removed 'tests/tmp/server.30294.572/nodes.conf' +removed 'tests/tmp/server.30294.572/stderr' +removed 'tests/tmp/server.30294.572/stdout' +removed 'tests/tmp/valkey.conf.30294.663' +removed 'tests/tmp/server.30294.812/nodes.conf' +removed 'tests/tmp/server.30294.812/stderr' +removed 'tests/tmp/server.30294.812/stdout' +removed 'tests/tmp/valkey.conf.30294.317' +removed 'tests/tmp/server.30294.650/nodes.conf' +removed 'tests/tmp/server.30294.650/stderr' +removed 'tests/tmp/server.30294.650/stdout' +removed 'tests/tmp/valkey.conf.30294.439' +removed 'tests/tmp/valkey.conf.30294.20' +removed 'tests/tmp/valkey.conf.30294.385' +removed 'tests/tmp/server.30294.131/stderr' +removed 'tests/tmp/server.30294.131/stdout' +removed 'tests/tmp/server.30294.818/nodes.conf' +removed 'tests/tmp/server.30294.818/stderr' +removed 'tests/tmp/server.30294.818/stdout' +removed 'tests/tmp/valkey.conf.30294.36' +removed 'tests/tmp/valkey.conf.30294.381' +removed 'tests/tmp/valkey.conf.30294.565' +removed 'tests/tmp/server.30294.578/nodes.conf' +removed 'tests/tmp/server.30294.578/stderr' +removed 'tests/tmp/server.30294.578/stdout' +removed 'tests/tmp/server.30294.520/nodes.conf' +removed 'tests/tmp/server.30294.520/stderr' +removed 'tests/tmp/server.30294.520/stdout' +removed 'tests/tmp/server.30294.113/stderr' +removed 'tests/tmp/server.30294.113/stdout' +removed 'tests/tmp/valkey.conf.30294.203' +removed 'tests/tmp/server.30294.676/nodes.conf' +removed 'tests/tmp/server.30294.676/stderr' +removed 'tests/tmp/server.30294.676/stdout' +removed 'tests/tmp/valkey.conf.30294.325' +removed 'tests/tmp/valkey.conf.30294.136' +removed 'tests/tmp/valkey.conf.30294.829' +removed 'tests/tmp/valkey.conf.30294.104' +removed 'tests/tmp/valkey.conf.30294.793' +removed 'tests/tmp/valkey.conf.30294.102' +removed 'tests/tmp/server.30294.634/nodes.conf' +removed 'tests/tmp/server.30294.634/stderr' +removed 'tests/tmp/server.30294.634/stdout' +removed 'tests/tmp/valkey.conf.30294.251' +removed 'tests/tmp/valkey.conf.30294.561' +removed 'tests/tmp/server.30294.536/nodes.conf' +removed 'tests/tmp/server.30294.536/stderr' +removed 'tests/tmp/server.30294.536/stdout' +removed 'tests/tmp/valkey.conf.30294.727' +removed 'tests/tmp/valkey.conf.30294.138' +removed 'tests/tmp/server.30294.466/stderr' +removed 'tests/tmp/server.30294.466/stdout' +removed 'tests/tmp/server.30294.756/nodes.conf' +removed 'tests/tmp/server.30294.756/stderr' +removed 'tests/tmp/server.30294.756/stdout' +removed 'tests/tmp/valkey.conf.30294.14' +removed 'tests/tmp/server.30294.364/stderr' +removed 'tests/tmp/server.30294.364/stdout' +removed 'tests/tmp/server.30294.604/nodes.conf' +removed 'tests/tmp/server.30294.604/stderr' +removed 'tests/tmp/server.30294.604/stdout' +removed 'tests/tmp/valkey.conf.30294.497' +removed 'tests/tmp/server.30294.446/stderr' +removed 'tests/tmp/server.30294.446/stdout' +removed 'tests/tmp/valkey.conf.30294.215' +removed 'tests/tmp/valkey.conf.30294.531' +removed 'tests/tmp/server.30294.700/nodes.conf' +removed 'tests/tmp/server.30294.700/stderr' +removed 'tests/tmp/server.30294.700/stdout' +removed 'tests/tmp/server.30294.362/stderr' +removed 'tests/tmp/server.30294.362/stdout' +removed 'tests/tmp/valkey.conf.30294.503' +removed 'tests/tmp/valkey.conf.30294.669' +removed 'tests/tmp/server.30294.450/stderr' +removed 'tests/tmp/server.30294.450/stdout' +removed 'tests/tmp/server.30294.49/stderr' +removed 'tests/tmp/server.30294.49/stdout' +removed 'tests/tmp/valkey.conf.30294.763' +removed 'tests/tmp/server.30294.556/nodes.conf' +removed 'tests/tmp/server.30294.556/stderr' +removed 'tests/tmp/server.30294.556/stdout' +removed 'tests/tmp/server.30294.606/nodes.conf' +removed 'tests/tmp/server.30294.606/stderr' +removed 'tests/tmp/server.30294.606/stdout' +removed 'tests/tmp/server.30294.236/stderr' +removed 'tests/tmp/server.30294.236/stdout' +removed 'tests/tmp/valkey.conf.30294.233' +removed 'tests/tmp/valkey.conf.30294.126' +removed 'tests/tmp/valkey.conf.30294.90' +removed 'tests/tmp/server.30294.368/stderr' +removed 'tests/tmp/server.30294.368/stdout' +removed 'tests/tmp/server.30294.161/stderr' +removed 'tests/tmp/server.30294.161/stdout' +removed 'tests/tmp/server.30294.664/nodes.conf' +removed 'tests/tmp/server.30294.664/stderr' +removed 'tests/tmp/server.30294.664/stdout' +removed 'tests/tmp/server.30294.141/stderr' +removed 'tests/tmp/server.30294.141/stdout' +removed 'tests/tmp/valkey.conf.30294.535' +removed 'tests/tmp/valkey.conf.30294.196' +removed 'tests/tmp/valkey.conf.30294.441' +removed 'tests/tmp/server.30294.47/stderr' +removed 'tests/tmp/server.30294.47/stdout' +removed 'tests/tmp/valkey.conf.30294.755' +removed 'tests/tmp/valkey.conf.30294.293' +removed 'tests/tmp/server.30294.510/nodes.conf' +removed 'tests/tmp/server.30294.510/stderr' +removed 'tests/tmp/server.30294.510/stdout' +removed 'tests/tmp/valkey.conf.30294.349' +removed 'tests/tmp/server.30294.754/nodes.conf' +removed 'tests/tmp/server.30294.754/stderr' +removed 'tests/tmp/server.30294.754/stdout' +removed 'tests/tmp/valkey.conf.30294.787' +removed 'tests/tmp/valkey.conf.30294.223' +removed 'tests/tmp/valkey.conf.30294.771' +removed 'tests/tmp/server.30294.59/stderr' +removed 'tests/tmp/server.30294.59/stdout' +removed 'tests/tmp/server.30294.147/stderr' +removed 'tests/tmp/server.30294.147/stdout' +removed 'tests/tmp/server.30294.422/stderr' +removed 'tests/tmp/server.30294.422/stdout' +removed 'tests/tmp/server.30294.274/stderr' +removed 'tests/tmp/server.30294.274/stdout' +removed 'tests/tmp/server.30294.524/nodes.conf' +removed 'tests/tmp/server.30294.524/stderr' +removed 'tests/tmp/server.30294.524/stdout' +removed 'tests/tmp/server.30294.768/nodes.conf' +removed 'tests/tmp/server.30294.768/stderr' +removed 'tests/tmp/server.30294.768/stdout' +removed 'tests/tmp/valkey.conf.30294.777' +removed 'tests/tmp/server.30294.458/stderr' +removed 'tests/tmp/server.30294.458/stdout' +removed 'tests/tmp/valkey.conf.30294.112' +removed 'tests/tmp/valkey.conf.30294.825' +removed 'tests/tmp/valkey.conf.30294.643' +removed 'tests/tmp/valkey.conf.30294.721' +removed 'tests/tmp/server.30294.444/stderr' +removed 'tests/tmp/server.30294.444/stdout' +removed 'tests/tmp/server.30294.264/stderr' +removed 'tests/tmp/server.30294.264/stdout' +removed 'tests/tmp/valkey.conf.30294.783' +removed 'tests/tmp/server.30294.442/stderr' +removed 'tests/tmp/server.30294.442/stdout' +removed 'tests/tmp/valkey.conf.30294.56' +removed 'tests/tmp/valkey.conf.30294.263' +removed 'tests/tmp/server.30294.636/nodes.conf' +removed 'tests/tmp/server.30294.636/stderr' +removed 'tests/tmp/server.30294.636/stdout' +removed 'tests/tmp/valkey.conf.30294.409' +removed 'tests/tmp/valkey.conf.30294.603' +removed 'tests/tmp/server.30294.320/stderr' +removed 'tests/tmp/server.30294.320/stdout' +removed 'tests/tmp/server.30294.408/stderr' +removed 'tests/tmp/server.30294.408/stdout' +removed 'tests/tmp/valkey.conf.30294.679' +removed 'tests/tmp/valkey.conf.30294.527' +removed 'tests/tmp/valkey.conf.30294.723' +removed 'tests/tmp/valkey.conf.30294.367' +removed 'tests/tmp/server.30294.772/nodes.conf' +removed 'tests/tmp/server.30294.772/stderr' +removed 'tests/tmp/server.30294.772/stdout' +removed 'tests/tmp/valkey.conf.30294.22' +removed 'tests/tmp/valkey.conf.30294.229' +removed 'tests/tmp/valkey.conf.30294.717' +removed 'tests/tmp/valkey.conf.30294.477' +removed 'tests/tmp/valkey.conf.30294.405' +removed 'tests/tmp/server.30294.67/stderr' +removed 'tests/tmp/server.30294.67/stdout' +removed 'tests/tmp/valkey.conf.30294.781' +removed 'tests/tmp/valkey.conf.30294.537' +removed 'tests/tmp/valkey.conf.30294.821' +removed 'tests/tmp/valkey.conf.30294.365' +removed 'tests/tmp/valkey.conf.30294.16' +removed 'tests/tmp/server.30294.312/stderr' +removed 'tests/tmp/server.30294.312/stdout' +removed 'tests/tmp/server.30294.574/nodes.conf' +removed 'tests/tmp/server.30294.574/stderr' +removed 'tests/tmp/server.30294.574/stdout' +removed 'tests/tmp/valkey.conf.30294.10' +removed 'tests/tmp/valkey.conf.30294.225' +removed 'tests/tmp/valkey.conf.30294.733' +removed 'tests/tmp/server.30294.322/stderr' +removed 'tests/tmp/server.30294.322/stdout' +removed 'tests/tmp/server.30294.388/stderr' +removed 'tests/tmp/server.30294.388/stdout' +removed 'tests/tmp/valkey.conf.30294.811' +removed 'tests/tmp/valkey.conf.30294.837' +removed 'tests/tmp/valkey.conf.30294.659' +removed 'tests/tmp/valkey.conf.30294.695' +removed 'tests/tmp/server.30294.774/nodes.conf' +removed 'tests/tmp/server.30294.774/stderr' +removed 'tests/tmp/server.30294.774/stdout' +removed 'tests/tmp/valkey.conf.30294.743' +removed 'tests/tmp/valkey.conf.30294.249' +removed 'tests/tmp/server.30294.119/stderr' +removed 'tests/tmp/server.30294.119/stdout' +removed 'tests/tmp/server.30294.644/nodes.conf' +removed 'tests/tmp/server.30294.644/stderr' +removed 'tests/tmp/server.30294.644/stdout' +removed 'tests/tmp/valkey.conf.30294.739' +removed 'tests/tmp/valkey.conf.30294.359' +removed 'tests/tmp/valkey.conf.30294.665' +removed 'tests/tmp/server.30294.608/nodes.conf' +removed 'tests/tmp/server.30294.608/stderr' +removed 'tests/tmp/server.30294.608/stdout' +removed 'tests/tmp/valkey.conf.30294.541' +removed 'tests/tmp/server.30294.95/stderr' +removed 'tests/tmp/server.30294.95/stdout' +removed 'tests/tmp/valkey.conf.30294.543' +removed 'tests/tmp/server.30294.278/stderr' +removed 'tests/tmp/server.30294.278/stdout' +removed 'tests/tmp/valkey.conf.30294.120' +removed 'tests/tmp/server.30294.169/nodes.conf' +removed 'tests/tmp/server.30294.169/stderr' +removed 'tests/tmp/server.30294.169/stdout' +removed 'tests/tmp/valkey.conf.30294.397' +removed 'tests/tmp/valkey.conf.30294.108' +removed 'tests/tmp/server.30294.726/nodes.conf' +removed 'tests/tmp/server.30294.726/stderr' +removed 'tests/tmp/server.30294.726/stdout' +removed 'tests/tmp/server.30294.145/stderr' +removed 'tests/tmp/server.30294.145/stdout' +removed 'tests/tmp/server.30294.302/stderr' +removed 'tests/tmp/server.30294.302/stdout' +removed 'tests/tmp/valkey.conf.30294.393' +removed 'tests/tmp/valkey.conf.30294.473' +removed 'tests/tmp/server.30294.612/nodes.conf' +removed 'tests/tmp/server.30294.612/stderr' +removed 'tests/tmp/server.30294.612/stdout' +removed 'tests/tmp/server.30294.654/nodes.conf' +removed 'tests/tmp/server.30294.654/stderr' +removed 'tests/tmp/server.30294.654/stdout' +removed 'tests/tmp/server.30294.526/nodes.conf' +removed 'tests/tmp/server.30294.526/stderr' +removed 'tests/tmp/server.30294.526/stdout' +removed 'tests/tmp/valkey.conf.30294.801' +removed 'tests/tmp/server.30294.692/nodes.conf' +removed 'tests/tmp/server.30294.692/stderr' +removed 'tests/tmp/server.30294.692/stdout' +removed 'tests/tmp/server.30294.820/nodes.conf' +removed 'tests/tmp/server.30294.820/stderr' +removed 'tests/tmp/server.30294.820/stdout' +removed 'tests/tmp/server.30294.418/stderr' +removed 'tests/tmp/server.30294.418/stdout' +removed 'tests/tmp/server.30294.206/stderr' +removed 'tests/tmp/server.30294.206/stdout' +removed 'tests/tmp/valkey.conf.30294.713' +removed 'tests/tmp/server.30294.242/stderr' +removed 'tests/tmp/server.30294.242/stdout' +removed 'tests/tmp/server.30294.610/nodes.conf' +removed 'tests/tmp/server.30294.610/stderr' +removed 'tests/tmp/server.30294.610/stdout' +removed 'tests/tmp/server.30294.794/nodes.conf' +removed 'tests/tmp/server.30294.794/stderr' +removed 'tests/tmp/server.30294.794/stdout' +removed 'tests/tmp/server.30294.804/nodes.conf' +removed 'tests/tmp/server.30294.804/stderr' +removed 'tests/tmp/server.30294.804/stdout' +removed 'tests/tmp/server.30294.214/stderr' +removed 'tests/tmp/server.30294.214/stdout' +removed 'tests/tmp/server.30294.810/nodes.conf' +removed 'tests/tmp/server.30294.810/stderr' +removed 'tests/tmp/server.30294.810/stdout' +removed 'tests/tmp/server.30294.344/stderr' +removed 'tests/tmp/server.30294.344/stdout' +removed 'tests/tmp/server.30294.210/nodes.conf' +removed 'tests/tmp/server.30294.210/stderr' +removed 'tests/tmp/server.30294.210/stdout' +removed 'tests/tmp/valkey.conf.30294.333' +removed 'tests/tmp/valkey.conf.30294.191' +removed 'tests/tmp/server.30294.396/stderr' +removed 'tests/tmp/server.30294.396/stdout' +removed 'tests/tmp/server.30294.288/stderr' +removed 'tests/tmp/server.30294.288/stdout' +removed 'tests/tmp/valkey.conf.30294.82' +removed 'tests/tmp/server.30294.706/nodes.conf' +removed 'tests/tmp/server.30294.706/stderr' +removed 'tests/tmp/server.30294.706/stdout' +removed 'tests/tmp/server.30294.490/nodes.conf' +removed 'tests/tmp/server.30294.490/stderr' +removed 'tests/tmp/server.30294.490/stdout' +removed 'tests/tmp/valkey.conf.30294.699' +removed 'tests/tmp/server.30294.250/stderr' +removed 'tests/tmp/server.30294.250/stdout' +removed 'tests/tmp/server.30294.250/somename' +removed 'tests/tmp/valkey.conf.30294.523' +removed 'tests/tmp/valkey.conf.30294.635' +removed 'tests/tmp/server.30294.17/stderr' +removed 'tests/tmp/server.30294.17/stdout' +removed 'tests/tmp/server.30294.256/stderr' +removed 'tests/tmp/server.30294.256/stdout' +removed 'tests/tmp/valkey.conf.30294.655' +removed 'tests/tmp/server.30294.830/nodes.conf' +removed 'tests/tmp/server.30294.830/stderr' +removed 'tests/tmp/server.30294.830/stdout' +removed 'tests/tmp/valkey.conf.30294.84' +removed 'tests/tmp/valkey.conf.30294.511' +removed 'tests/tmp/valkey.conf.30294.383' +removed 'tests/tmp/valkey.conf.30294.305' +removed 'tests/tmp/valkey.conf.30294.341' +removed 'tests/tmp/valkey.conf.30294.485' +removed 'tests/tmp/server.30294.384/stderr' +removed 'tests/tmp/server.30294.384/stdout' +removed 'tests/tmp/server.30294.488/nodes.conf' +removed 'tests/tmp/server.30294.488/stderr' +removed 'tests/tmp/server.30294.488/stdout' +removed 'tests/tmp/valkey.conf.30294.319' +removed 'tests/tmp/valkey.conf.30294.427' +removed 'tests/tmp/server.30294.656/nodes.conf' +removed 'tests/tmp/server.30294.656/stderr' +removed 'tests/tmp/server.30294.656/stdout' +removed 'tests/tmp/valkey.conf.30294.519' +removed 'tests/tmp/valkey.conf.30294.361' +removed 'tests/tmp/resetchannels.acl.30294.189/nodefaultuser.acl' +removed 'tests/tmp/resetchannels.acl.30294.189/stderr' +removed 'tests/tmp/resetchannels.acl.30294.189/stdout' +removed 'tests/tmp/resetchannels.acl.30294.189/default.conf' +removed 'tests/tmp/server.30294.740/nodes.conf' +removed 'tests/tmp/server.30294.740/stderr' +removed 'tests/tmp/server.30294.740/stdout' +removed 'tests/tmp/server.30294.666/nodes.conf' +removed 'tests/tmp/server.30294.666/stderr' +removed 'tests/tmp/server.30294.666/stdout' +removed 'tests/tmp/valkey.conf.30294.327' +removed 'tests/tmp/server.30294.41/stderr' +removed 'tests/tmp/server.30294.41/stdout' +removed 'tests/tmp/server.30294.216/stderr' +removed 'tests/tmp/server.30294.216/stdout' +removed 'tests/tmp/valkey.conf.30294.156' +removed 'tests/tmp/valkey.conf.30294.38' +removed 'tests/tmp/server.30294.662/nodes.conf' +removed 'tests/tmp/server.30294.662/stderr' +removed 'tests/tmp/server.30294.662/stdout' +removed 'tests/tmp/server.30294.502/nodes.conf' +removed 'tests/tmp/server.30294.502/stderr' +removed 'tests/tmp/server.30294.502/stdout' +removed 'tests/tmp/valkey.conf.30294.369' +removed 'tests/tmp/server.30294.61/nodes.conf' +removed 'tests/tmp/server.30294.61/stderr' +removed 'tests/tmp/server.30294.61/stdout' +removed 'tests/tmp/server.30294.376/stderr' +removed 'tests/tmp/server.30294.376/stdout' +removed 'tests/tmp/valkey.conf.30294.351' +removed 'tests/tmp/server.30294.400/stderr' +removed 'tests/tmp/server.30294.400/stdout' +removed 'tests/tmp/valkey.conf.30294.447' +removed 'tests/tmp/valkey.conf.30294.805' +removed 'tests/tmp/valkey.conf.30294.96' +removed 'tests/tmp/valkey.conf.30294.773' +removed 'tests/tmp/server.30294.394/stderr' +removed 'tests/tmp/server.30294.394/stdout' +removed 'tests/tmp/valkey.conf.30294.815' +removed 'tests/tmp/server.30294.660/nodes.conf' +removed 'tests/tmp/server.30294.660/stderr' +removed 'tests/tmp/server.30294.660/stdout' +removed 'tests/tmp/valkey.conf.30294.779' +removed 'tests/tmp/server.30294.300/stderr' +removed 'tests/tmp/server.30294.300/stdout' +removed 'tests/tmp/valkey.conf.30294.152' +removed 'tests/tmp/valkey.conf.30294.34' +removed 'tests/tmp/valkey.conf.30294.106' +removed 'tests/tmp/server.30294.806/nodes.conf' +removed 'tests/tmp/server.30294.806/stderr' +removed 'tests/tmp/server.30294.806/stdout' +removed 'tests/tmp/valkey.conf.30294.423' +removed 'tests/tmp/server.30294.680/nodes.conf' +removed 'tests/tmp/server.30294.680/stderr' +removed 'tests/tmp/server.30294.680/stdout' +removed 'tests/tmp/valkey.conf.30294.128' +removed 'tests/tmp/valkey.conf.30294.813' +removed 'tests/tmp/valkey.conf.30294.513' +removed 'tests/tmp/server.30294.620/nodes.conf' +removed 'tests/tmp/server.30294.620/stderr' +removed 'tests/tmp/server.30294.620/stdout' +removed 'tests/tmp/valkey.conf.30294.465' +removed 'tests/tmp/server.30294.452/stderr' +removed 'tests/tmp/server.30294.452/stdout' +removed 'tests/tmp/valkey.conf.30294.185' +removed 'tests/tmp/valkey.conf.30294.667' +removed 'tests/tmp/server.30294.616/nodes.conf' +removed 'tests/tmp/server.30294.616/stderr' +removed 'tests/tmp/server.30294.616/stdout' +removed 'tests/tmp/valkey.conf.30294.265' +removed 'tests/tmp/server.30294.528/nodes.conf' +removed 'tests/tmp/server.30294.528/stderr' +removed 'tests/tmp/server.30294.528/stdout' +removed 'tests/tmp/valkey.conf.30294.315' +removed 'tests/tmp/valkey.conf.30294.525' +removed 'tests/tmp/server.30294.474/nodes.conf' +removed 'tests/tmp/server.30294.474/stderr' +removed 'tests/tmp/server.30294.474/stdout' +removed 'tests/tmp/server.30294.374/stderr' +removed 'tests/tmp/server.30294.374/stdout' +removed 'tests/tmp/valkey.conf.30294.493' +removed 'tests/tmp/server.30294.167/stderr' +removed 'tests/tmp/server.30294.167/stdout' +removed 'tests/tmp/server.30294.548/nodes.conf' +removed 'tests/tmp/server.30294.548/stderr' +removed 'tests/tmp/server.30294.548/stdout' +removed 'tests/tmp/server.30294.702/nodes.conf' +removed 'tests/tmp/server.30294.702/stderr' +removed 'tests/tmp/server.30294.702/stdout' +removed 'tests/tmp/server.30294.800/nodes.conf' +removed 'tests/tmp/server.30294.800/stderr' +removed 'tests/tmp/server.30294.800/stdout' +removed 'tests/tmp/valkey.conf.30294.124' +removed 'tests/tmp/valkey.conf.30294.567' +removed 'tests/tmp/server.30294.159/stderr' +removed 'tests/tmp/server.30294.159/stdout' +removed 'tests/tmp/server.30294.15/stderr' +removed 'tests/tmp/server.30294.15/stdout' +removed 'tests/tmp/valkey.conf.30294.283' +removed 'tests/tmp/valkey.conf.30294.257' +removed 'tests/tmp/valkey.conf.30294.78' +removed 'tests/tmp/server.30294.628/nodes.conf' +removed 'tests/tmp/server.30294.628/stderr' +removed 'tests/tmp/server.30294.628/stdout' +removed 'tests/tmp/valkey.conf.30294.375' +removed 'tests/tmp/valkey.conf.30294.209' +removed 'tests/tmp/server.30294.598/nodes.conf' +removed 'tests/tmp/server.30294.598/stderr' +removed 'tests/tmp/server.30294.598/stdout' +removed 'tests/tmp/server.30294.592/nodes.conf' +removed 'tests/tmp/server.30294.592/stderr' +removed 'tests/tmp/server.30294.592/stdout' +removed 'tests/tmp/server.30294.668/nodes.conf' +removed 'tests/tmp/server.30294.668/stderr' +removed 'tests/tmp/server.30294.668/stdout' +removed 'tests/tmp/server.30294.324/stderr' +removed 'tests/tmp/server.30294.324/stdout' +removed 'tests/tmp/valkey.conf.30294.213' +removed 'tests/tmp/server.30294.420/stderr' +removed 'tests/tmp/server.30294.420/stdout' +removed 'tests/tmp/valkey.conf.30294.54' +removed 'tests/tmp/valkey.conf.30294.40' +removed 'tests/tmp/server.30294.684/nodes.conf' +removed 'tests/tmp/server.30294.684/stderr' +removed 'tests/tmp/server.30294.684/stdout' +removed 'tests/tmp/valkey.conf.30294.237' +removed 'tests/tmp/valkey.conf.30294.289' +removed 'tests/tmp/valkey.conf.30294.661' +removed 'tests/tmp/server.30294.75/stderr' +removed 'tests/tmp/server.30294.75/stdout' +removed 'tests/tmp/server.30294.53/stderr' +removed 'tests/tmp/server.30294.53/stdout' +removed 'tests/tmp/valkey.conf.30294.122' +removed 'tests/tmp/server.30294.81/stderr' +removed 'tests/tmp/server.30294.81/stdout' +removed 'tests/tmp/server.30294.13/stderr' +removed 'tests/tmp/server.30294.13/stdout' +removed 'tests/tmp/valkey.conf.30294.791' +removed 'tests/tmp/server.30294.448/stderr' +removed 'tests/tmp/server.30294.448/stdout' +removed 'tests/tmp/server.30294.670/nodes.conf' +removed 'tests/tmp/server.30294.670/stderr' +removed 'tests/tmp/server.30294.670/stdout' +removed 'tests/tmp/server.30294.784/nodes.conf' +removed 'tests/tmp/server.30294.784/stderr' +removed 'tests/tmp/server.30294.784/stdout' +removed 'tests/tmp/server.30294.770/nodes.conf' +removed 'tests/tmp/server.30294.770/stderr' +removed 'tests/tmp/server.30294.770/stdout' +removed 'tests/tmp/server.30294.382/stderr' +removed 'tests/tmp/server.30294.382/stdout' +removed 'tests/tmp/server.30294.398/stderr' +removed 'tests/tmp/server.30294.398/stdout' +removed 'tests/tmp/valkey.conf.30294.261' +removed 'tests/tmp/valkey.conf.30294.759' +removed 'tests/tmp/valkey.conf.30294.168' +removed 'tests/tmp/valkey.conf.30294.425' +removed 'tests/tmp/server.30294.380/stderr' +removed 'tests/tmp/server.30294.380/stdout' +removed 'tests/tmp/valkey.conf.30294.399' +removed 'tests/tmp/valkey.conf.30294.110' +removed 'tests/tmp/server.30294.414/stderr' +removed 'tests/tmp/server.30294.414/stdout' +removed 'tests/tmp/valkey.conf.30294.307' +removed 'tests/tmp/server.30294.77/stderr' +removed 'tests/tmp/server.30294.77/stdout' +removed 'tests/tmp/valkey.conf.30294.162' +removed 'tests/tmp/server.30294.29/nodes.conf' +removed 'tests/tmp/server.30294.29/stderr' +removed 'tests/tmp/server.30294.29/stdout' +removed 'tests/tmp/valkey.conf.30294.457' +removed 'tests/tmp/server.30294.614/nodes.conf' +removed 'tests/tmp/server.30294.614/stderr' +removed 'tests/tmp/server.30294.614/stdout' +removed 'tests/tmp/valkey.conf.30294.833' +removed 'tests/tmp/valkey.conf.30294.403' +removed 'tests/tmp/server.30294.153/stderr' +removed 'tests/tmp/server.30294.153/stdout' +removed 'tests/tmp/server.30294.179/stderr' +removed 'tests/tmp/server.30294.179/stdout' +removed 'tests/tmp/server.30294.760/nodes.conf' +removed 'tests/tmp/server.30294.760/stderr' +removed 'tests/tmp/server.30294.760/stdout' +removed 'tests/tmp/server.30294.254/stderr' +removed 'tests/tmp/server.30294.254/stdout' +removed 'tests/tmp/valkey.conf.30294.275' +removed 'tests/tmp/valkey.conf.30294.453' +removed 'tests/tmp/valkey.conf.30294.217' +removed 'tests/tmp/valkey.conf.30294.164' +removed 'tests/tmp/server.30294.103/stderr' +removed 'tests/tmp/server.30294.103/stdout' +removed 'tests/tmp/server.30294.171/stderr' +removed 'tests/tmp/server.30294.171/stdout' +removed 'tests/tmp/server.30294.97/stderr' +removed 'tests/tmp/server.30294.97/stdout' +removed 'tests/tmp/valkey.conf.30294.741' +removed 'tests/tmp/server.30294.586/nodes.conf' +removed 'tests/tmp/server.30294.586/stderr' +removed 'tests/tmp/server.30294.586/stdout' +removed 'tests/tmp/server.30294.202/stderr' +removed 'tests/tmp/server.30294.202/stdout' +removed 'tests/tmp/server.30294.352/stderr' +removed 'tests/tmp/server.30294.352/stdout' +removed 'tests/tmp/valkey.conf.30294.737' +removed 'tests/tmp/server.30294.35/stderr' +removed 'tests/tmp/server.30294.35/stdout' +removed 'tests/tmp/valkey.conf.30294.615' +removed 'tests/tmp/server.30294.814/nodes.conf' +removed 'tests/tmp/server.30294.814/stderr' +removed 'tests/tmp/server.30294.814/stdout' +removed 'tests/tmp/server.30294.308/stderr' +removed 'tests/tmp/server.30294.308/stdout' +removed 'tests/tmp/valkey.conf.30294.579' +removed 'tests/tmp/valkey.conf.30294.395' +removed 'tests/tmp/valkey.conf.30294.100' +removed 'tests/tmp/valkey.conf.30294.501' +removed 'tests/tmp/valkey.conf.30294.795' +removed 'tests/tmp/server.30294.314/stderr' +removed 'tests/tmp/server.30294.314/stdout' +removed 'tests/tmp/valkey.conf.30294.819' +removed 'tests/tmp/server.30294.19/stderr' +removed 'tests/tmp/server.30294.19/stdout' +removed 'tests/tmp/valkey.conf.30294.785' +removed 'tests/tmp/server.30294.115/stderr' +removed 'tests/tmp/server.30294.115/stdout' +removed 'tests/tmp/server.30294.600/nodes.conf' +removed 'tests/tmp/server.30294.600/stderr' +removed 'tests/tmp/server.30294.600/stdout' +removed 'tests/tmp/valkey.conf.30294.479' +removed 'tests/tmp/valkey.conf.30294.609' +removed 'tests/tmp/valkey.conf.30294.72' +removed 'tests/tmp/server.30294.672/nodes.conf' +removed 'tests/tmp/server.30294.672/stderr' +removed 'tests/tmp/server.30294.672/stdout' +removed 'tests/tmp/server.30294.390/stderr' +removed 'tests/tmp/server.30294.390/stdout' +removed 'tests/tmp/valkey.conf.30294.789' +removed 'tests/tmp/valkey.conf.30294.66' +removed 'tests/tmp/valkey.conf.30294.431' +removed 'tests/tmp/valkey.conf.30294.437' +removed 'tests/tmp/server.30294.280/stderr' +removed 'tests/tmp/server.30294.280/stdout' +removed 'tests/tmp/valkey.conf.30294.183' +removed 'tests/tmp/server.30294.402/stderr' +removed 'tests/tmp/server.30294.402/stdout' +removed 'tests/tmp/server.30294.544/nodes.conf' +removed 'tests/tmp/server.30294.544/stderr' +removed 'tests/tmp/server.30294.544/stdout' +removed 'tests/tmp/valkey.conf.30294.529' +removed 'tests/tmp/server.30294.336/stderr' +removed 'tests/tmp/server.30294.336/stdout' +removed 'tests/tmp/server.30294.642/nodes.conf' +removed 'tests/tmp/server.30294.642/stderr' +removed 'tests/tmp/server.30294.642/stdout' +removed 'tests/tmp/valkey.conf.30294.158' +removed 'tests/tmp/server.30294.410/stderr' +removed 'tests/tmp/server.30294.410/stdout' +removed 'tests/tmp/valkey.conf.30294.12' +removed 'tests/tmp/server.30294.828/nodes.conf' +removed 'tests/tmp/server.30294.828/stderr' +removed 'tests/tmp/server.30294.828/stdout' +removed 'tests/tmp/server.30294.195/stderr' +removed 'tests/tmp/server.30294.195/stdout' +removed 'tests/tmp/valkey.conf.30294.277' +removed 'tests/tmp/server.30294.430/stderr' +removed 'tests/tmp/server.30294.430/stdout' +removed 'tests/tmp/valkey.conf.30294.707' +removed 'tests/tmp/server.30294.39/stderr' +removed 'tests/tmp/server.30294.39/stdout' +removed 'tests/tmp/valkey.conf.30294.379' +removed 'tests/tmp/valkey.conf.30294.705' +removed 'tests/tmp/valkey.conf.30294.2' +removed 'tests/tmp/server.30294.722/nodes.conf' +removed 'tests/tmp/server.30294.722/stderr' +removed 'tests/tmp/server.30294.722/stdout' +removed 'tests/tmp/valkey.conf.30294.571' +removed 'tests/tmp/valkey.conf.30294.415' +removed 'tests/tmp/server.30294.436/stderr' +removed 'tests/tmp/server.30294.436/stdout' +removed 'tests/tmp/valkey.conf.30294.639' +removed 'tests/tmp/valkey.conf.30294.172' +removed 'tests/tmp/server.30294.360/stderr' +removed 'tests/tmp/server.30294.360/stdout' +removed 'tests/tmp/valkey.conf.30294.48' +removed 'tests/tmp/server.30294.456/stderr' +removed 'tests/tmp/server.30294.456/stdout' +removed 'tests/tmp/valkey.conf.30294.287' +removed 'tests/tmp/server.30294.554/nodes.conf' +removed 'tests/tmp/server.30294.554/stderr' +removed 'tests/tmp/server.30294.554/stdout' +removed 'tests/tmp/valkey.conf.30294.407' +removed 'tests/tmp/server.30294.111/stderr' +removed 'tests/tmp/server.30294.111/stdout' +removed 'tests/tmp/valkey.conf.30294.166' +removed 'tests/tmp/valkey.conf.30294.817' +removed 'tests/tmp/valkey.conf.30294.547' +removed 'tests/tmp/valkey.conf.30294.683' +removed 'tests/tmp/server.30294.9/stderr' +removed 'tests/tmp/server.30294.9/stdout' +removed 'tests/tmp/server.30294.117/stderr' +removed 'tests/tmp/server.30294.117/stdout' +removed 'tests/tmp/server.30294.316/stderr' +removed 'tests/tmp/server.30294.316/stdout' +removed 'tests/tmp/valkey.conf.30294.42' +removed 'tests/tmp/valkey.conf.30294.194' +removed 'tests/tmp/server.30294.334/stderr' +removed 'tests/tmp/server.30294.334/stdout' +removed 'tests/tmp/server.30294.764/nodes.conf' +removed 'tests/tmp/server.30294.764/stderr' +removed 'tests/tmp/server.30294.764/stdout' +removed 'tests/tmp/server.30294.440/stderr' +removed 'tests/tmp/server.30294.440/stdout' +removed 'tests/tmp/server.30294.133/stderr' +removed 'tests/tmp/server.30294.133/stdout' +removed 'tests/tmp/server.30294.105/stderr' +removed 'tests/tmp/server.30294.105/stdout' +removed 'tests/tmp/valkey.conf.30294.32' +removed 'tests/tmp/valkey.conf.30294.761' +removed 'tests/tmp/valkey.conf.30294.835' +removed 'tests/tmp/server.30294.139/stderr' +removed 'tests/tmp/server.30294.139/stdout' +removed 'tests/tmp/server.30294.21/nodes.conf' +removed 'tests/tmp/server.30294.21/stderr' +removed 'tests/tmp/server.30294.21/stdout' +removed 'tests/tmp/valkey.conf.30294.355' +removed 'tests/tmp/server.30294.428/stderr' +removed 'tests/tmp/server.30294.428/stdout' +removed 'tests/tmp/server.30294.258/stderr' +removed 'tests/tmp/server.30294.258/stdout' +removed 'tests/tmp/server.30294.836/nodes.conf' +removed 'tests/tmp/server.30294.836/dump.rdb' +removed 'tests/tmp/server.30294.836/stderr' +removed 'tests/tmp/server.30294.836/stdout' +removed 'tests/tmp/valkey.conf.30294.681' +removed 'tests/tmp/valkey.conf.30294.94' +removed 'tests/tmp/valkey.conf.30294.389' +removed 'tests/tmp/valkey.conf.30294.633' +removed 'tests/tmp/valkey.conf.30294.311' +removed 'tests/tmp/valkey.conf.30294.533' +removed 'tests/tmp/valkey.conf.30294.767' +removed 'tests/tmp/valkey.conf.30294.154' +removed 'tests/tmp/server.30294.292/stderr' +removed 'tests/tmp/server.30294.292/stdout' +removed 'tests/tmp/server.30294.298/stderr' +removed 'tests/tmp/server.30294.298/stdout' +removed 'tests/tmp/valkey.conf.30294.467' +removed 'tests/tmp/server.30294.378/stderr' +removed 'tests/tmp/server.30294.378/stdout' +removed 'tests/tmp/server.30294.710/nodes.conf' +removed 'tests/tmp/server.30294.710/stderr' +removed 'tests/tmp/server.30294.710/stdout' +removed 'tests/tmp/server.30294.690/nodes.conf' +removed 'tests/tmp/server.30294.690/stderr' +removed 'tests/tmp/server.30294.690/stdout' +removed 'tests/tmp/server.30294.125/stderr' +removed 'tests/tmp/server.30294.125/stdout' +removed 'tests/tmp/valkey.conf.30294.92' +removed 'tests/tmp/valkey.conf.30294.303' +removed 'tests/tmp/server.30294.472/nodes.conf' +removed 'tests/tmp/server.30294.472/stderr' +removed 'tests/tmp/server.30294.472/stdout' +removed 'tests/tmp/valkey.conf.30294.831' +removed 'tests/tmp/server.30294.838/nodes.conf' +removed 'tests/tmp/server.30294.838/stderr' +removed 'tests/tmp/server.30294.838/stdout' +removed 'tests/tmp/server.30294.624/nodes.conf' +removed 'tests/tmp/server.30294.624/stderr' +removed 'tests/tmp/server.30294.624/stdout' +removed 'tests/tmp/valkey.conf.30294.583' +removed 'tests/tmp/valkey.conf.30294.88' +removed 'tests/tmp/server.30294.576/nodes.conf' +removed 'tests/tmp/server.30294.576/stderr' +removed 'tests/tmp/server.30294.576/stdout' +removed 'tests/tmp/server.30294.460/stderr' +removed 'tests/tmp/server.30294.460/stdout' +removed 'tests/tmp/server.30294.752/nodes.conf' +removed 'tests/tmp/server.30294.752/stderr' +removed 'tests/tmp/server.30294.752/stdout' +removed 'tests/tmp/server.30294.79/stderr' +removed 'tests/tmp/server.30294.79/stdout' +removed 'tests/tmp/server.30294.792/nodes.conf' +removed 'tests/tmp/server.30294.792/stderr' +removed 'tests/tmp/server.30294.792/stdout' +removed 'tests/tmp/server.30294.736/nodes.conf' +removed 'tests/tmp/server.30294.736/stderr' +removed 'tests/tmp/server.30294.736/stdout' +removed 'tests/tmp/server.30294.348/stderr' +removed 'tests/tmp/server.30294.348/stdout' +removed 'tests/tmp/server.30294.594/nodes.conf' +removed 'tests/tmp/server.30294.594/stderr' +removed 'tests/tmp/server.30294.594/stdout' +removed 'tests/tmp/server.30294.658/nodes.conf' +removed 'tests/tmp/server.30294.658/stderr' +removed 'tests/tmp/server.30294.658/stdout' +removed 'tests/tmp/server.30294.272/stderr' +removed 'tests/tmp/server.30294.272/stdout' +removed 'tests/tmp/server.30294.532/nodes.conf' +removed 'tests/tmp/server.30294.532/stderr' +removed 'tests/tmp/server.30294.532/stdout' +removed 'tests/tmp/valkey.conf.30294.711' +removed 'tests/tmp/valkey.conf.30294.725' +removed 'tests/tmp/server.30294.454/stderr' +removed 'tests/tmp/server.30294.454/stdout' +removed 'tests/tmp/valkey.conf.30294.491' +removed 'tests/tmp/server.acl.30294.186/stderr' +removed 'tests/tmp/server.acl.30294.186/stdout' +removed 'tests/tmp/server.acl.30294.186/user.acl' +removed 'tests/tmp/server.30294.284/stderr' +removed 'tests/tmp/server.30294.284/stdout' +removed 'tests/tmp/server.30294.43/stderr' +removed 'tests/tmp/server.30294.43/stdout' +removed 'tests/tmp/server.30294.151/stderr' +removed 'tests/tmp/server.30294.151/stdout' +removed 'tests/tmp/valkey.conf.30294.98' +removed 'tests/tmp/valkey.conf.30294.339' +removed 'tests/tmp/valkey.conf.30294.469' +removed 'tests/tmp/server.30294.25/stderr' +removed 'tests/tmp/server.30294.25/stdout' +removed 'tests/tmp/server.30294.386/stderr' +removed 'tests/tmp/server.30294.386/stdout' +removed 'tests/tmp/valkey.conf.30294.587' +removed 'tests/tmp/server.30294.294/stderr' +removed 'tests/tmp/server.30294.294/stdout' +removed 'tests/tmp/valkey.conf.30294.58' +removed 'tests/tmp/valkey.conf.30294.631' +removed 'tests/tmp/server.30294.434/stderr' +removed 'tests/tmp/server.30294.434/stdout' +removed 'tests/tmp/server.30294.73/stderr' +removed 'tests/tmp/server.30294.73/stdout' +removed 'tests/tmp/server.30294.143/stderr' +removed 'tests/tmp/server.30294.143/stdout' +removed 'tests/tmp/valkey.conf.30294.30' +removed 'tests/tmp/server.30294.534/nodes.conf' +removed 'tests/tmp/server.30294.534/stderr' +removed 'tests/tmp/server.30294.534/stdout' +removed 'tests/tmp/valkey.conf.30294.279' +removed 'tests/tmp/valkey.conf.30294.259' +removed 'tests/tmp/valkey.conf.30294.271' +removed 'tests/tmp/valkey.conf.30294.685' +removed 'tests/tmp/valkey.conf.30294.4' +removed 'tests/tmp/valkey.conf.30294.809' +removed 'tests/tmp/valkey.conf.30294.239' +removed 'tests/tmp/server.30294.750/nodes.conf' +removed 'tests/tmp/server.30294.750/stderr' +removed 'tests/tmp/server.30294.750/stdout' +removed 'tests/tmp/valkey.conf.30294.775' +removed 'tests/tmp/server.30294.550/nodes.conf' +removed 'tests/tmp/server.30294.550/stderr' +removed 'tests/tmp/server.30294.550/stdout' +removed 'tests/tmp/valkey.conf.30294.357' +removed 'tests/tmp/valkey.conf.30294.549' +removed 'tests/tmp/server.30294.65/stderr' +removed 'tests/tmp/server.30294.65/stdout' +removed 'tests/tmp/valkey.conf.30294.649' +removed 'tests/tmp/valkey.conf.30294.60' +removed 'tests/tmp/valkey.conf.30294.671' +removed 'tests/tmp/server.30294.776/nodes.conf' +removed 'tests/tmp/server.30294.776/stderr' +removed 'tests/tmp/server.30294.776/stdout' +removed 'tests/tmp/server.30294.824/nodes.conf' +removed 'tests/tmp/server.30294.824/stderr' +removed 'tests/tmp/server.30294.824/stdout' +removed 'tests/tmp/server.30294.588/nodes.conf' +removed 'tests/tmp/server.30294.588/stderr' +removed 'tests/tmp/server.30294.588/stdout' +removed 'tests/tmp/server.30294.370/stderr' +removed 'tests/tmp/server.30294.370/stdout' +removed 'tests/tmp/server.30294.304/stderr' +removed 'tests/tmp/server.30294.304/stdout' +removed 'tests/tmp/server.30294.282/stderr' +removed 'tests/tmp/server.30294.282/stdout' +removed 'tests/tmp/server.30294.834/nodes.conf' +removed 'tests/tmp/server.30294.834/stderr' +removed 'tests/tmp/server.30294.834/stdout' +removed 'tests/tmp/valkey.conf.30294.345' +removed 'tests/tmp/server.30294.240/stderr' +removed 'tests/tmp/server.30294.240/stdout' +removed 'tests/tmp/valkey.conf.30294.613' +removed 'tests/tmp/server.30294.372/stderr' +removed 'tests/tmp/server.30294.372/stdout' +removed 'tests/tmp/server.30294.165/stderr' +removed 'tests/tmp/server.30294.165/stdout' +removed 'tests/tmp/valkey.conf.30294.235' +removed 'tests/tmp/valkey.conf.30294.597' +removed 'tests/tmp/server.30294.640/nodes.conf' +removed 'tests/tmp/server.30294.640/stderr' +removed 'tests/tmp/server.30294.640/stdout' +removed 'tests/tmp/valkey.conf.30294.219' +removed 'tests/tmp/valkey.conf.30294.539' +removed 'tests/tmp/server.30294.326/stderr' +removed 'tests/tmp/server.30294.326/stdout' +removed 'tests/tmp/valkey.conf.30294.443' +removed 'tests/tmp/valkey.conf.30294.653' +removed 'tests/tmp/server.30294.766/nodes.conf' +removed 'tests/tmp/server.30294.766/stderr' +removed 'tests/tmp/server.30294.766/stdout' +removed 'tests/tmp/server.30294.518/nodes.conf' +removed 'tests/tmp/server.30294.518/stderr' +removed 'tests/tmp/server.30294.518/stdout' +removed 'tests/tmp/valkey.conf.30294.709' +removed 'tests/tmp/valkey.conf.30294.693' +removed 'tests/tmp/server.30294.602/nodes.conf' +removed 'tests/tmp/server.30294.602/stderr' +removed 'tests/tmp/server.30294.602/stdout' +removed 'tests/tmp/server.30294.296/stderr' +removed 'tests/tmp/server.30294.296/stdout' +removed 'tests/tmp/valkey.conf.30294.481' +removed 'tests/tmp/server.30294.744/nodes.conf' +removed 'tests/tmp/server.30294.744/stderr' +removed 'tests/tmp/server.30294.744/stdout' +removed 'tests/tmp/server.30294.568/nodes.conf' +removed 'tests/tmp/server.30294.568/stderr' +removed 'tests/tmp/server.30294.568/stdout' +removed 'tests/tmp/server.30294.332/stderr' +removed 'tests/tmp/server.30294.332/stdout' +removed 'tests/tmp/valkey.conf.30294.751' +removed 'tests/tmp/server.30294.808/nodes.conf' +removed 'tests/tmp/server.30294.808/stderr' +removed 'tests/tmp/server.30294.808/stdout' +removed 'tests/tmp/server.30294.173/stderr' +removed 'tests/tmp/server.30294.173/stdout' +removed 'tests/tmp/server.30294.392/stderr' +removed 'tests/tmp/server.30294.392/stdout' +removed 'tests/tmp/server.30294.197/stderr' +removed 'tests/tmp/server.30294.197/stdout' +removed 'tests/tmp/valkey.conf.30294.689' +removed 'tests/tmp/server.30294.244/stderr' +removed 'tests/tmp/server.30294.244/stdout' +removed 'tests/tmp/server.30294.748/nodes.conf' +removed 'tests/tmp/server.30294.748/stderr' +removed 'tests/tmp/server.30294.748/stdout' +removed 'tests/tmp/valkey.conf.30294.673' +removed 'tests/tmp/valkey.conf.30294.417' +removed 'tests/tmp/server.30294.342/stderr' +removed 'tests/tmp/server.30294.342/stdout' +removed 'tests/tmp/valkey.conf.30294.140' +removed 'tests/tmp/server.30294.71/stderr' +removed 'tests/tmp/server.30294.71/stdout' +removed 'tests/tmp/server.30294.516/nodes.conf' +removed 'tests/tmp/server.30294.516/stderr' +removed 'tests/tmp/server.30294.516/stdout' +removed 'tests/tmp/valkey.conf.30294.483' +removed 'tests/tmp/valkey.conf.30294.559' +removed 'tests/tmp/server.30294.330/stderr' +removed 'tests/tmp/server.30294.330/stdout' +removed 'tests/tmp/valkey.conf.30294.231' +removed 'tests/tmp/server.30294.590/nodes.conf' +removed 'tests/tmp/server.30294.590/stderr' +removed 'tests/tmp/server.30294.590/stdout' +removed 'tests/tmp/server.30294.562/nodes.conf' +removed 'tests/tmp/server.30294.562/stderr' +removed 'tests/tmp/server.30294.562/stdout' +removed 'tests/tmp/duplicate.acl.30294.199/stderr' +removed 'tests/tmp/duplicate.acl.30294.199/stdout' +removed 'tests/tmp/duplicate.acl.30294.199/default.conf' +removed 'tests/tmp/duplicate.acl.30294.199/user.acl' +removed 'tests/tmp/valkey.conf.30294.619' +removed 'tests/tmp/server.30294.712/nodes.conf' +removed 'tests/tmp/server.30294.712/stderr' +removed 'tests/tmp/server.30294.712/stdout' +removed 'tests/tmp/server.30294.83/stderr' +removed 'tests/tmp/server.30294.83/stdout' +removed 'tests/tmp/server.30294.790/nodes.conf' +removed 'tests/tmp/server.30294.790/stderr' +removed 'tests/tmp/server.30294.790/stdout' +removed 'tests/tmp/server.30294.5/stderr' +removed 'tests/tmp/server.30294.5/stdout' +removed 'tests/tmp/valkey.conf.30294.433' +removed 'tests/tmp/server.30294.276/stderr' +removed 'tests/tmp/server.30294.276/stdout' +removed 'tests/tmp/valkey.conf.30294.827' +removed 'tests/tmp/valkey.conf.30294.451' +removed 'tests/tmp/valkey.conf.30294.198' +removed 'tests/tmp/server.30294.149/stderr' +removed 'tests/tmp/server.30294.149/stdout' +removed 'tests/tmp/server.30294.570/nodes.conf' +removed 'tests/tmp/server.30294.570/stderr' +removed 'tests/tmp/server.30294.570/stdout' +removed 'tests/tmp/valkey.conf.30294.273' make[1]: Leaving directory '/build/reproducible-path/valkey-8.0.2+dfsg1' create-stamp debian/debhelper-build-stamp dh_prep @@ -7749,15 +9731,15 @@ dh_strip -a dh_makeshlibs -a dh_shlibdeps -a -dpkg-shlibdeps: warning: package could avoid a useless dependency if debian/valkey-tools/usr/bin/valkey-cli debian/valkey-tools/usr/bin/valkey-check-rdb debian/valkey-tools/usr/bin/valkey-check-aof debian/valkey-tools/usr/bin/valkey-benchmark were not linked against libatomic.so.1 (they use none of the library's symbols) -dpkg-shlibdeps: warning: package could avoid a useless dependency if debian/valkey-tools/usr/bin/valkey-cli debian/valkey-tools/usr/bin/valkey-check-rdb debian/valkey-tools/usr/bin/valkey-check-aof debian/valkey-tools/usr/bin/valkey-benchmark were not linked against liblzf.so.1 (they use none of the library's symbols) +dpkg-shlibdeps: warning: package could avoid a useless dependency if debian/valkey-tools/usr/bin/valkey-benchmark debian/valkey-tools/usr/bin/valkey-cli debian/valkey-tools/usr/bin/valkey-check-rdb debian/valkey-tools/usr/bin/valkey-check-aof were not linked against libatomic.so.1 (they use none of the library's symbols) +dpkg-shlibdeps: warning: package could avoid a useless dependency if debian/valkey-tools/usr/bin/valkey-benchmark debian/valkey-tools/usr/bin/valkey-cli debian/valkey-tools/usr/bin/valkey-check-rdb debian/valkey-tools/usr/bin/valkey-check-aof were not linked against liblzf.so.1 (they use none of the library's symbols) dh_installdeb dh_gencontrol dh_md5sums dh_builddeb -dpkg-deb: building package 'valkey-sentinel' in '../valkey-sentinel_8.0.2+dfsg1-1_amd64.deb'. dpkg-deb: building package 'valkey-tools-dbgsym' in '../valkey-tools-dbgsym_8.0.2+dfsg1-1_amd64.deb'. dpkg-deb: building package 'valkey-tools' in '../valkey-tools_8.0.2+dfsg1-1_amd64.deb'. +dpkg-deb: building package 'valkey-sentinel' in '../valkey-sentinel_8.0.2+dfsg1-1_amd64.deb'. dpkg-deb: building package 'valkey-server' in '../valkey-server_8.0.2+dfsg1-1_amd64.deb'. dpkg-genbuildinfo --build=binary -O../valkey_8.0.2+dfsg1-1_amd64.buildinfo dpkg-genchanges --build=binary -O../valkey_8.0.2+dfsg1-1_amd64.changes @@ -7766,12 +9748,14 @@ dpkg-buildpackage: info: binary-only upload (no source included) dpkg-genchanges: info: including full source code in upload I: copying local configuration +I: user script /srv/workspace/pbuilder/4180030/tmp/hooks/B01_cleanup starting +I: user script /srv/workspace/pbuilder/4180030/tmp/hooks/B01_cleanup finished I: unmounting dev/ptmx filesystem I: unmounting dev/pts filesystem I: unmounting dev/shm filesystem I: unmounting proc filesystem I: unmounting sys filesystem I: cleaning the build env -I: removing directory /srv/workspace/pbuilder/1755253 and its subdirectories -I: Current time: Sun Mar 9 14:11:11 -12 2025 -I: pbuilder-time-stamp: 1741572671 +I: removing directory /srv/workspace/pbuilder/4180030 and its subdirectories +I: Current time: Mon Apr 13 00:10:24 +14 2026 +I: pbuilder-time-stamp: 1775988624