Tarantool development patches archive
 help / color / mirror / Atom feed
From: Sergey Bronnikov <sergeyb@tarantool.org>
To: "Alexander V. Tikhonov" <avtikhon@tarantool.org>
Cc: tarantool-patches@dev.tarantool.org
Subject: Re: [Tarantool-patches] [sergepetrenko@tarantool.org: Re: [PATCH v1] Divide replication/mist.test.lua]
Date: Thu, 3 Sep 2020 12:32:10 +0300	[thread overview]
Message-ID: <13c35d18-e188-a68a-417e-8ae605599979@tarantool.org> (raw)
In-Reply-To: <20200903075417.GA24428@hpalx>

Hi,

I remember that we already discussed naming format for regression tests when

Alexander sent patches for box/net.box [1] and decided to add gh issue 
number as a postfix to the name.

BTW, ddl.test.lua was splitted to the tests with gh issue number as a 
postfix - [2].


1. 
https://lists.tarantool.org/pipermail/tarantool-patches/2020-April/015355.html

2. 
https://github.com/tarantool/tarantool/commit/4a8d1ebda3e0eb7b7d6d14f66b77a9c6d8d0815e

On 03.09.2020 10:54, Alexander V. Tikhonov wrote:
> Hi Sergey, could you please confirm the the new tests style used
> in dividing box/net.box*.test.lua test.
>
>
> ----- Forwarded message from Serge Petrenko <sergepetrenko@tarantool.org> -----
>
> Date: Thu, 3 Sep 2020 10:45:17 +0300
> From: Serge Petrenko <sergepetrenko@tarantool.org>
> To: "Alexander V. Tikhonov" <avtikhon@tarantool.org>
> Cc: Mergen Imeev <imeevma@gmail.com>, tarantool-patches@dev.tarantool.org, Alexander Turenko <alexander.turenko@tarantool.org>, Yukhin Kirill <k.yukhin@corp.mail.ru>
> Subject: Re: [Tarantool-patches] [PATCH v1] Divide replication/mist.test.lua
> User-Agent: Mozilla/5.0 (Macintosh; Intel Mac OS X 10.15; rv:68.0) Gecko/20100101 Thunderbird/68.12.0
>
>
> 02.09.2020 13:08, Alexander V. Tikhonov пишет:
>> Hi Sergey, thanks for the review. The major comment on tests naming
>> that it was discussed during fixes in issue [1], and decided to use
>> this naming format, check as an example of it with:
>>
>>     ls -al test/box/net.box*.test.lua
>>
>> [1] - https://github.com/tarantool/tarantool/issues/4880
> Hi, Alex!
>
> Strange. I was guided by our earlier decision and replication tests:
>
> ls test/replication/gh-*.test.lua
> test/replication/gh-4114-local-space-replication.test.lua
> test/replication/gh-4402-info-errno.test.lua
> test/replication/gh-4605-empty-password.test.lua
> test/replication/gh-4606-admin-creds.test.lua
> test/replication/gh-4729-netbox-group-id.test.lua
> test/replication/gh-4730-applier-rollback.test.lua
> test/replication/gh-4739-vclock-assert.test.lua
> test/replication/gh-4928-tx-boundaries.test.lua
> test/replication/gh-5140-qsync-casc-rollback.test.lua
> test/replication/gh-5144-qsync-dup-confirm.test.lua
> test/replication/gh-5163-qsync-restart-crash.test.lua
> test/replication/gh-5167-qsync-rollback-snap.test.lua
> test/replication/gh-5195-qsync-replica-write.test.lua
>
> Anyway, if someone can confirm that we're using another naming
> scheme now, I'm ok with it.
>
>> On Wed, Sep 02, 2020 at 11:09:30AM +0300, Serge Petrenko wrote:
>>> Hi! Thanks for the patch!
>>>
>>> Please see my comments below.
>>>
>>> 17.07.2020 12:25, Alexander V. Tikhonov пишет:
>>>> To fix flaky issues of replication/misc.test.lua the test had to be
>>>> divided into smaller tests to be able to localize the flaky results:
>>>>
>>>>      misc_assert_connecting_master_twice_gh-3610.test.lua
>>>>      misc_assert_on_server_die_gh-2991.test.lua
>>>>      misc_assert_replica_on_applier_disconnect_gh-3510.test.lua
>>>>      misc_crash_on_box_concurrent_update_gh-3606.test.lua
>>>>      misc_heartbeats_on_master_changes_gh-3160.test.lua
>>>>      misc_no_failure_on_error_reading_wal_gh-4399.test.lua
>>>>      misc_no_panic_on_connected_gh-3637.test.lua
>>>>      misc_no_restart_on_same_configuration_gh-3711.test.lua
>>>>      misc_no_socket_leak_on_replica_disconnect_gh-3642.test.lua
>>>>      misc_orphan_on_reconfiguration_error_gh-4424.test.lua
>>>>      misc_rebootstrap_from_ro_master_gh-3111.test.lua
>>>>      misc_replica_checks_cluster_id_gh-3704.test.lua
>>>>      misc_return_on_quorum_0_gh-3760.test.lua
>>>>      misc_value_not_replicated_on_iproto_request_gh-3247.test.lua
>>> We use a different naming scheme for regression tests.
>>> It should be `gh-xxxx-bug-description`, also notice the dashes
>>> between words instead of the underscores.
>>>
>>>> Needed for #4940
>>>> ---
>>>>
>>>> Github: https://github.com/tarantool/tarantool/tree/avtikhon/gh-4940-replication-misc
>>>> Issue: https://github.com/tarantool/tarantool/issues/4940
>>>>
>>>>     test/replication/misc.result                  | 866 ------------------
>>>>     test/replication/misc.test.lua                | 356 -------
>>>>     ...ert_connecting_master_twice_gh-3610.result |  86 ++
>>>>     ...t_connecting_master_twice_gh-3610.test.lua |  34 +
>>>>     .../misc_assert_on_server_die_gh-2991.result  |  31 +
>>>>     ...misc_assert_on_server_die_gh-2991.test.lua |  12 +
>>>>     ...plica_on_applier_disconnect_gh-3510.result |  49 +
>>>>     ...ica_on_applier_disconnect_gh-3510.test.lua |  17 +
>>>>     ...sh_on_box_concurrent_update_gh-3606.result |  50 +
>>>>     ..._on_box_concurrent_update_gh-3606.test.lua |  19 +
>>>>     ...eartbeats_on_master_changes_gh-3160.result |  76 ++
>>>>     ...rtbeats_on_master_changes_gh-3160.test.lua |  40 +
>>>>     ...ailure_on_error_reading_wal_gh-4399.result |  97 ++
>>>>     ...lure_on_error_reading_wal_gh-4399.test.lua |  39 +
>>>>     .../misc_no_panic_on_connected_gh-3637.result |  72 ++
>>>>     ...isc_no_panic_on_connected_gh-3637.test.lua |  33 +
>>>>     ...start_on_same_configuration_gh-3711.result | 107 +++
>>>>     ...art_on_same_configuration_gh-3711.test.lua |  41 +
>>>>     ..._leak_on_replica_disconnect_gh-3642.result |  98 ++
>>>>     ...eak_on_replica_disconnect_gh-3642.test.lua |  44 +
>>>>     ...an_on_reconfiguration_error_gh-4424.result |  88 ++
>>>>     ..._on_reconfiguration_error_gh-4424.test.lua |  37 +
>>>>     ..._rebootstrap_from_ro_master_gh-3111.result |  58 ++
>>>>     ...ebootstrap_from_ro_master_gh-3111.test.lua |  20 +
>>>>     ...c_replica_checks_cluster_id_gh-3704.result |  71 ++
>>>>     ...replica_checks_cluster_id_gh-3704.test.lua |  26 +
>>>>     .../misc_return_on_quorum_0_gh-3760.result    |  48 +
>>>>     .../misc_return_on_quorum_0_gh-3760.test.lua  |  27 +
>>>>     ...eplicated_on_iproto_request_gh-3247.result |  90 ++
>>>>     ...licated_on_iproto_request_gh-3247.test.lua |  33 +
>>>>     test/replication/suite.cfg                    |  15 +-
>>>>     31 files changed, 1457 insertions(+), 1223 deletions(-)
>>>>     delete mode 100644 test/replication/misc.result
>>>>     delete mode 100644 test/replication/misc.test.lua
>>>>     create mode 100644 test/replication/misc_assert_connecting_master_twice_gh-3610.result
>>>>     create mode 100644 test/replication/misc_assert_connecting_master_twice_gh-3610.test.lua
>>>>     create mode 100644 test/replication/misc_assert_on_server_die_gh-2991.result
>>>>     create mode 100644 test/replication/misc_assert_on_server_die_gh-2991.test.lua
>>>>     create mode 100644 test/replication/misc_assert_replica_on_applier_disconnect_gh-3510.result
>>>>     create mode 100644 test/replication/misc_assert_replica_on_applier_disconnect_gh-3510.test.lua
>>>>     create mode 100644 test/replication/misc_crash_on_box_concurrent_update_gh-3606.result
>>>>     create mode 100644 test/replication/misc_crash_on_box_concurrent_update_gh-3606.test.lua
>>>>     create mode 100644 test/replication/misc_heartbeats_on_master_changes_gh-3160.result
>>>>     create mode 100644 test/replication/misc_heartbeats_on_master_changes_gh-3160.test.lua
>>>>     create mode 100644 test/replication/misc_no_failure_on_error_reading_wal_gh-4399.result
>>>>     create mode 100644 test/replication/misc_no_failure_on_error_reading_wal_gh-4399.test.lua
>>>>     create mode 100644 test/replication/misc_no_panic_on_connected_gh-3637.result
>>>>     create mode 100644 test/replication/misc_no_panic_on_connected_gh-3637.test.lua
>>>>     create mode 100644 test/replication/misc_no_restart_on_same_configuration_gh-3711.result
>>>>     create mode 100644 test/replication/misc_no_restart_on_same_configuration_gh-3711.test.lua
>>>>     create mode 100644 test/replication/misc_no_socket_leak_on_replica_disconnect_gh-3642.result
>>>>     create mode 100644 test/replication/misc_no_socket_leak_on_replica_disconnect_gh-3642.test.lua
>>>>     create mode 100644 test/replication/misc_orphan_on_reconfiguration_error_gh-4424.result
>>>>     create mode 100644 test/replication/misc_orphan_on_reconfiguration_error_gh-4424.test.lua
>>>>     create mode 100644 test/replication/misc_rebootstrap_from_ro_master_gh-3111.result
>>>>     create mode 100644 test/replication/misc_rebootstrap_from_ro_master_gh-3111.test.lua
>>>>     create mode 100644 test/replication/misc_replica_checks_cluster_id_gh-3704.result
>>>>     create mode 100644 test/replication/misc_replica_checks_cluster_id_gh-3704.test.lua
>>>>     create mode 100644 test/replication/misc_return_on_quorum_0_gh-3760.result
>>>>     create mode 100644 test/replication/misc_return_on_quorum_0_gh-3760.test.lua
>>>>     create mode 100644 test/replication/misc_value_not_replicated_on_iproto_request_gh-3247.result
>>>>     create mode 100644 test/replication/misc_value_not_replicated_on_iproto_request_gh-3247.test.lua
>>>>
>>>> diff --git a/test/replication/misc.result b/test/replication/misc.result
>>>> deleted file mode 100644
>>>> index e5d1f560e..000000000
>>>> --- a/test/replication/misc.result
>>>> +++ /dev/null
>>>> @@ -1,866 +0,0 @@
>>>> -uuid = require('uuid')
>>>> ----
>>>> -...
>>>> -test_run = require('test_run').new()
>>>> ----
>>>> -...
>>>> -box.schema.user.grant('guest', 'replication')
>>>> ----
>>>> -...
>>>> --- gh-2991 - Tarantool asserts on box.cfg.replication update if one of
>>>> --- servers is dead
>>>> -replication_timeout = box.cfg.replication_timeout
>>>> ----
>>>> -...
>>>> -replication_connect_timeout = box.cfg.replication_connect_timeout
>>>> ----
>>>> -...
>>>> -box.cfg{replication_timeout=0.05, replication_connect_timeout=0.05, replication={}}
>>>> ----
>>>> -...
>>>> -box.cfg{replication_connect_quorum=2}
>>>> ----
>>>> -...
>>>> -box.cfg{replication = {'127.0.0.1:12345', box.cfg.listen}}
>>>> ----
>>>> -...
>>>> -box.info.status
>>>> ----
>>>> -- orphan
>>>> -...
>>>> -box.info.ro
>>>> ----
>>>> -- true
>>>> -...
>>>> --- gh-3606 - Tarantool crashes if box.cfg.replication is updated concurrently
>>>> -fiber = require('fiber')
>>>> ----
>>>> -...
>>>> -c = fiber.channel(2)
>>>> ----
>>>> -...
>>>> -f = function() fiber.create(function() pcall(box.cfg, {replication = {12345}}) c:put(true) end) end
>>>> ----
>>>> -...
>>>> -f()
>>>> ----
>>>> -...
>>>> -f()
>>>> ----
>>>> -...
>>>> -c:get()
>>>> ----
>>>> -- true
>>>> -...
>>>> -c:get()
>>>> ----
>>>> -- true
>>>> -...
>>>> -box.cfg{replication = "", replication_timeout = replication_timeout, replication_connect_timeout = replication_connect_timeout}
>>>> ----
>>>> -...
>>>> -box.info.status
>>>> ----
>>>> -- running
>>>> -...
>>>> -box.info.ro
>>>> ----
>>>> -- false
>>>> -...
>>>> --- gh-3111 - Allow to rebootstrap a replica from a read-only master
>>>> -replica_uuid = uuid.new()
>>>> ----
>>>> -...
>>>> -test_run:cmd('create server test with rpl_master=default, script="replication/replica_uuid.lua"')
>>>> ----
>>>> -- true
>>>> -...
>>>> -test_run:cmd(string.format('start server test with args="%s"', replica_uuid))
>>>> ----
>>>> -- true
>>>> -...
>>>> -test_run:cmd('stop server test')
>>>> ----
>>>> -- true
>>>> -...
>>>> -test_run:cmd('cleanup server test')
>>>> ----
>>>> -- true
>>>> -...
>>>> -box.cfg{read_only = true}
>>>> ----
>>>> -...
>>>> -test_run:cmd(string.format('start server test with args="%s"', replica_uuid))
>>>> ----
>>>> -- true
>>>> -...
>>>> -test_run:cmd('stop server test')
>>>> ----
>>>> -- true
>>>> -...
>>>> -test_run:cmd('cleanup server test')
>>>> ----
>>>> -- true
>>>> -...
>>>> -box.cfg{read_only = false}
>>>> ----
>>>> -...
>>>> -test_run:cmd('delete server test')
>>>> ----
>>>> -- true
>>>> -...
>>>> -test_run:cleanup_cluster()
>>>> ----
>>>> -...
>>>> --- gh-3160 - Send heartbeats if there are changes from a remote master only
>>>> -SERVERS = { 'autobootstrap1', 'autobootstrap2', 'autobootstrap3' }
>>>> ----
>>>> -...
>>>> --- Deploy a cluster.
>>>> -test_run:create_cluster(SERVERS, "replication", {args="0.03"})
>>>> ----
>>>> -...
>>>> -test_run:wait_fullmesh(SERVERS)
>>>> ----
>>>> -...
>>>> -test_run:cmd("switch autobootstrap3")
>>>> ----
>>>> -- true
>>>> -...
>>>> -test_run = require('test_run').new()
>>>> ----
>>>> -...
>>>> -fiber = require('fiber')
>>>> ----
>>>> -...
>>>> -_ = box.schema.space.create('test_timeout'):create_index('pk')
>>>> ----
>>>> -...
>>>> -test_run:cmd("setopt delimiter ';'")
>>>> ----
>>>> -- true
>>>> -...
>>>> -function wait_not_follow(replicaA, replicaB)
>>>> -    return test_run:wait_cond(function()
>>>> -        return replicaA.status ~= 'follow' or replicaB.status ~= 'follow'
>>>> -    end, box.cfg.replication_timeout)
>>>> -end;
>>>> ----
>>>> -...
>>>> -function test_timeout()
>>>> -    local replicaA = box.info.replication[1].upstream or box.info.replication[2].upstream
>>>> -    local replicaB = box.info.replication[3].upstream or box.info.replication[2].upstream
>>>> -    local follows = test_run:wait_cond(function()
>>>> -        return replicaA.status == 'follow' or replicaB.status == 'follow'
>>>> -    end)
>>>> -    if not follows then error('replicas are not in the follow status') end
>>>> -    for i = 0, 99 do
>>>> -        box.space.test_timeout:replace({1})
>>>> -        if wait_not_follow(replicaA, replicaB) then
>>>> -            return error(box.info.replication)
>>>> -        end
>>>> -    end
>>>> -    return true
>>>> -end;
>>>> ----
>>>> -...
>>>> -test_run:cmd("setopt delimiter ''");
>>>> ----
>>>> -- true
>>>> -...
>>>> -test_timeout()
>>>> ----
>>>> -- true
>>>> -...
>>>> --- gh-3247 - Sequence-generated value is not replicated in case
>>>> --- the request was sent via iproto.
>>>> -test_run:cmd("switch autobootstrap1")
>>>> ----
>>>> -- true
>>>> -...
>>>> -net_box = require('net.box')
>>>> ----
>>>> -...
>>>> -_ = box.schema.space.create('space1')
>>>> ----
>>>> -...
>>>> -_ = box.schema.sequence.create('seq')
>>>> ----
>>>> -...
>>>> -_ = box.space.space1:create_index('primary', {sequence = true} )
>>>> ----
>>>> -...
>>>> -_ = box.space.space1:create_index('secondary', {parts = {2, 'unsigned'}})
>>>> ----
>>>> -...
>>>> -box.schema.user.grant('guest', 'read,write', 'space', 'space1')
>>>> ----
>>>> -...
>>>> -c = net_box.connect(box.cfg.listen)
>>>> ----
>>>> -...
>>>> -c.space.space1:insert{box.NULL, "data"} -- fails, but bumps sequence value
>>>> ----
>>>> -- error: 'Tuple field 2 type does not match one required by operation: expected unsigned'
>>>> -...
>>>> -c.space.space1:insert{box.NULL, 1, "data"}
>>>> ----
>>>> -- [2, 1, 'data']
>>>> -...
>>>> -box.space.space1:select{}
>>>> ----
>>>> -- - [2, 1, 'data']
>>>> -...
>>>> -vclock = test_run:get_vclock("autobootstrap1")
>>>> ----
>>>> -...
>>>> -vclock[0] = nil
>>>> ----
>>>> -...
>>>> -_ = test_run:wait_vclock("autobootstrap2", vclock)
>>>> ----
>>>> -...
>>>> -test_run:cmd("switch autobootstrap2")
>>>> ----
>>>> -- true
>>>> -...
>>>> -box.space.space1:select{}
>>>> ----
>>>> -- - [2, 1, 'data']
>>>> -...
>>>> -test_run:cmd("switch autobootstrap1")
>>>> ----
>>>> -- true
>>>> -...
>>>> -box.space.space1:drop()
>>>> ----
>>>> -...
>>>> -test_run:cmd("switch default")
>>>> ----
>>>> -- true
>>>> -...
>>>> -test_run:drop_cluster(SERVERS)
>>>> ----
>>>> -...
>>>> -test_run:cleanup_cluster()
>>>> ----
>>>> -...
>>>> --- gh-3642 - Check that socket file descriptor doesn't leak
>>>> --- when a replica is disconnected.
>>>> -rlimit = require('rlimit')
>>>> ----
>>>> -...
>>>> -lim = rlimit.limit()
>>>> ----
>>>> -...
>>>> -rlimit.getrlimit(rlimit.RLIMIT_NOFILE, lim)
>>>> ----
>>>> -...
>>>> -old_fno = lim.rlim_cur
>>>> ----
>>>> -...
>>>> -lim.rlim_cur = 64
>>>> ----
>>>> -...
>>>> -rlimit.setrlimit(rlimit.RLIMIT_NOFILE, lim)
>>>> ----
>>>> -...
>>>> -test_run:cmd('create server sock with rpl_master=default, script="replication/replica.lua"')
>>>> ----
>>>> -- true
>>>> -...
>>>> -test_run:cmd('start server sock')
>>>> ----
>>>> -- true
>>>> -...
>>>> -test_run:cmd('switch sock')
>>>> ----
>>>> -- true
>>>> -...
>>>> -test_run = require('test_run').new()
>>>> ----
>>>> -...
>>>> -fiber = require('fiber')
>>>> ----
>>>> -...
>>>> -test_run:cmd("setopt delimiter ';'")
>>>> ----
>>>> -- true
>>>> -...
>>>> -for i = 1, 64 do
>>>> -    local replication = box.cfg.replication
>>>> -    box.cfg{replication = {}}
>>>> -    box.cfg{replication = replication}
>>>> -    while box.info.replication[1].upstream.status ~= 'follow' do
>>>> -        fiber.sleep(0.001)
>>>> -    end
>>>> -end;
>>>> ----
>>>> -...
>>>> -test_run:cmd("setopt delimiter ''");
>>>> ----
>>>> -- true
>>>> -...
>>>> -box.info.replication[1].upstream.status
>>>> ----
>>>> -- follow
>>>> -...
>>>> -test_run:cmd('switch default')
>>>> ----
>>>> -- true
>>>> -...
>>>> -lim.rlim_cur = old_fno
>>>> ----
>>>> -...
>>>> -rlimit.setrlimit(rlimit.RLIMIT_NOFILE, lim)
>>>> ----
>>>> -...
>>>> -test_run:cmd("stop server sock")
>>>> ----
>>>> -- true
>>>> -...
>>>> -test_run:cmd("cleanup server sock")
>>>> ----
>>>> -- true
>>>> -...
>>>> -test_run:cmd("delete server sock")
>>>> ----
>>>> -- true
>>>> -...
>>>> -test_run:cleanup_cluster()
>>>> ----
>>>> -...
>>>> -box.schema.user.revoke('guest', 'replication')
>>>> ----
>>>> -...
>>>> --- gh-3510 assertion failure in replica_on_applier_disconnect()
>>>> -test_run:cmd('create server er_load1 with script="replication/er_load1.lua"')
>>>> ----
>>>> -- true
>>>> -...
>>>> -test_run:cmd('create server er_load2 with script="replication/er_load2.lua"')
>>>> ----
>>>> -- true
>>>> -...
>>>> -test_run:cmd('start server er_load1 with wait=False, wait_load=False')
>>>> ----
>>>> -- true
>>>> -...
>>>> --- Instance er_load2 will fail with error ER_REPLICASET_UUID_MISMATCH.
>>>> --- This is OK since we only test here that er_load1 doesn't assert.
>>>> -test_run:cmd('start server er_load2 with wait=True, wait_load=True, crash_expected = True')
>>>> ----
>>>> -- false
>>>> -...
>>>> -test_run:cmd('stop server er_load1')
>>>> ----
>>>> -- true
>>>> -...
>>>> --- er_load2 exits automatically.
>>>> -test_run:cmd('cleanup server er_load1')
>>>> ----
>>>> -- true
>>>> -...
>>>> -test_run:cmd('cleanup server er_load2')
>>>> ----
>>>> -- true
>>>> -...
>>>> -test_run:cmd('delete server er_load1')
>>>> ----
>>>> -- true
>>>> -...
>>>> -test_run:cmd('delete server er_load2')
>>>> ----
>>>> -- true
>>>> -...
>>>> -test_run:cleanup_cluster()
>>>> ----
>>>> -...
>>>> ---
>>>> --- Test case for gh-3637, gh-4550. Before the fix replica would
>>>> --- exit with an error if a user does not exist or a password is
>>>> --- incorrect. Now check that we don't hang/panic and successfully
>>>> --- connect.
>>>> ---
>>>> -fiber = require('fiber')
>>>> ----
>>>> -...
>>>> -test_run:cmd("create server replica_auth with rpl_master=default, script='replication/replica_auth.lua'")
>>>> ----
>>>> -- true
>>>> -...
>>>> -test_run:cmd("start server replica_auth with wait=False, wait_load=False, args='cluster:pass 0.05'")
>>>> ----
>>>> -- true
>>>> -...
>>>> --- Wait a bit to make sure replica waits till user is created.
>>>> -fiber.sleep(0.1)
>>>> ----
>>>> -...
>>>> -box.schema.user.create('cluster')
>>>> ----
>>>> -...
>>>> --- The user is created. Let the replica fail auth request due to
>>>> --- a wrong password.
>>>> -fiber.sleep(0.1)
>>>> ----
>>>> -...
>>>> -box.schema.user.passwd('cluster', 'pass')
>>>> ----
>>>> -...
>>>> -box.schema.user.grant('cluster', 'replication')
>>>> ----
>>>> -...
>>>> -while box.info.replication[2] == nil do fiber.sleep(0.01) end
>>>> ----
>>>> -...
>>>> -vclock = test_run:get_vclock('default')
>>>> ----
>>>> -...
>>>> -vclock[0] = nil
>>>> ----
>>>> -...
>>>> -_ = test_run:wait_vclock('replica_auth', vclock)
>>>> ----
>>>> -...
>>>> -test_run:cmd("stop server replica_auth")
>>>> ----
>>>> -- true
>>>> -...
>>>> -test_run:cmd("cleanup server replica_auth")
>>>> ----
>>>> -- true
>>>> -...
>>>> -test_run:cmd("delete server replica_auth")
>>>> ----
>>>> -- true
>>>> -...
>>>> -test_run:cleanup_cluster()
>>>> ----
>>>> -...
>>>> -box.schema.user.drop('cluster')
>>>> ----
>>>> -...
>>>> ---
>>>> --- Test case for gh-3610. Before the fix replica would fail with the assertion
>>>> --- when trying to connect to the same master twice.
>>>> ---
>>>> -box.schema.user.grant('guest', 'replication')
>>>> ----
>>>> -...
>>>> -test_run:cmd("create server replica with rpl_master=default, script='replication/replica.lua'")
>>>> ----
>>>> -- true
>>>> -...
>>>> -test_run:cmd("start server replica")
>>>> ----
>>>> -- true
>>>> -...
>>>> -test_run:cmd("switch replica")
>>>> ----
>>>> -- true
>>>> -...
>>>> -replication = box.cfg.replication[1]
>>>> ----
>>>> -...
>>>> -box.cfg{replication = {replication, replication}}
>>>> ----
>>>> -- error: 'Incorrect value for option ''replication'': duplicate connection to the
>>>> -    same replica'
>>>> -...
>>>> --- Check the case when duplicate connection is detected in the background.
>>>> -test_run:cmd("switch default")
>>>> ----
>>>> -- true
>>>> -...
>>>> -listen = box.cfg.listen
>>>> ----
>>>> -...
>>>> -box.cfg{listen = ''}
>>>> ----
>>>> -...
>>>> -test_run:cmd("switch replica")
>>>> ----
>>>> -- true
>>>> -...
>>>> -box.cfg{replication_connect_quorum = 0, replication_connect_timeout = 0.01}
>>>> ----
>>>> -...
>>>> -box.cfg{replication = {replication, replication}}
>>>> ----
>>>> -...
>>>> -test_run:cmd("switch default")
>>>> ----
>>>> -- true
>>>> -...
>>>> -box.cfg{listen = listen}
>>>> ----
>>>> -...
>>>> -while test_run:grep_log('replica', 'duplicate connection') == nil do fiber.sleep(0.01) end
>>>> ----
>>>> -...
>>>> -test_run:cmd("stop server replica")
>>>> ----
>>>> -- true
>>>> -...
>>>> -test_run:cmd("cleanup server replica")
>>>> ----
>>>> -- true
>>>> -...
>>>> -test_run:cmd("delete server replica")
>>>> ----
>>>> -- true
>>>> -...
>>>> -test_run:cleanup_cluster()
>>>> ----
>>>> -...
>>>> -box.schema.user.revoke('guest', 'replication')
>>>> ----
>>>> -...
>>>> ---
>>>> --- gh-3711 Do not restart replication on box.cfg in case the
>>>> --- configuration didn't change.
>>>> ---
>>>> -box.schema.user.grant('guest', 'replication')
>>>> ----
>>>> -...
>>>> -test_run:cmd("create server replica with rpl_master=default, script='replication/replica.lua'")
>>>> ----
>>>> -- true
>>>> -...
>>>> -test_run:cmd("start server replica")
>>>> ----
>>>> -- true
>>>> -...
>>>> --- Access rights are checked only during reconnect. If the new
>>>> --- and old configurations are equivalent, no reconnect will be
>>>> --- issued and replication should continue working.
>>>> -box.schema.user.revoke('guest', 'replication')
>>>> ----
>>>> -...
>>>> -test_run:cmd("switch replica")
>>>> ----
>>>> -- true
>>>> -...
>>>> -replication = box.cfg.replication[1]
>>>> ----
>>>> -...
>>>> -box.cfg{replication = {replication}}
>>>> ----
>>>> -...
>>>> -box.info.status == 'running'
>>>> ----
>>>> -- true
>>>> -...
>>>> -box.cfg{replication = replication}
>>>> ----
>>>> -...
>>>> -box.info.status == 'running'
>>>> ----
>>>> -- true
>>>> -...
>>>> --- Check that comparison of tables works as expected as well.
>>>> -test_run:cmd("switch default")
>>>> ----
>>>> -- true
>>>> -...
>>>> -box.schema.user.grant('guest', 'replication')
>>>> ----
>>>> -...
>>>> -test_run:cmd("switch replica")
>>>> ----
>>>> -- true
>>>> -...
>>>> -replication = box.cfg.replication
>>>> ----
>>>> -...
>>>> -table.insert(replication, box.cfg.listen)
>>>> ----
>>>> -...
>>>> -test_run:cmd("switch default")
>>>> ----
>>>> -- true
>>>> -...
>>>> -box.schema.user.revoke('guest', 'replication')
>>>> ----
>>>> -...
>>>> -test_run:cmd("switch replica")
>>>> ----
>>>> -- true
>>>> -...
>>>> -box.cfg{replication = replication}
>>>> ----
>>>> -...
>>>> -box.info.status == 'running'
>>>> ----
>>>> -- true
>>>> -...
>>>> -test_run:cmd("switch default")
>>>> ----
>>>> -- true
>>>> -...
>>>> -test_run:cmd("stop server replica")
>>>> ----
>>>> -- true
>>>> -...
>>>> -test_run:cmd("cleanup server replica")
>>>> ----
>>>> -- true
>>>> -...
>>>> -test_run:cmd("delete server replica")
>>>> ----
>>>> -- true
>>>> -...
>>>> -test_run:cleanup_cluster()
>>>> ----
>>>> -...
>>>> ---
>>>> --- gh-3704 move cluster id check to replica
>>>> ---
>>>> -test_run:cmd("create server replica with rpl_master=default, script='replication/replica.lua'")
>>>> ----
>>>> -- true
>>>> -...
>>>> -box.schema.user.grant("guest", "replication")
>>>> ----
>>>> -...
>>>> -test_run:cmd("start server replica")
>>>> ----
>>>> -- true
>>>> -...
>>>> -test_run:grep_log("replica", "REPLICASET_UUID_MISMATCH")
>>>> ----
>>>> -- null
>>>> -...
>>>> -box.info.replication[2].downstream.status
>>>> ----
>>>> -- follow
>>>> -...
>>>> --- change master's cluster uuid and check that replica doesn't connect.
>>>> -test_run:cmd("stop server replica")
>>>> ----
>>>> -- true
>>>> -...
>>>> -_ = box.space._schema:replace{'cluster', tostring(uuid.new())}
>>>> ----
>>>> -...
>>>> --- master believes replica is in cluster, but their cluster UUIDs differ.
>>>> -test_run:cmd("start server replica")
>>>> ----
>>>> -- true
>>>> -...
>>>> -test_run:wait_log("replica", "REPLICASET_UUID_MISMATCH", nil, 1.0)
>>>> ----
>>>> -- REPLICASET_UUID_MISMATCH
>>>> -...
>>>> -test_run:wait_downstream(2, {status = 'stopped'})
>>>> ----
>>>> -- true
>>>> -...
>>>> -test_run:cmd("stop server replica")
>>>> ----
>>>> -- true
>>>> -...
>>>> -test_run:cmd("cleanup server replica")
>>>> ----
>>>> -- true
>>>> -...
>>>> -test_run:cmd("delete server replica")
>>>> ----
>>>> -- true
>>>> -...
>>>> -test_run:cleanup_cluster()
>>>> ----
>>>> -...
>>>> -box.schema.user.revoke('guest', 'replication')
>>>> ----
>>>> -...
>>>> ---
>>>> --- gh-4399 Check that an error reading WAL directory on subscribe
>>>> --- doesn't lead to a permanent replication failure.
>>>> ---
>>>> -box.schema.user.grant("guest", "replication")
>>>> ----
>>>> -...
>>>> -test_run:cmd("create server replica with rpl_master=default, script='replication/replica.lua'")
>>>> ----
>>>> -- true
>>>> -...
>>>> -test_run:cmd("start server replica")
>>>> ----
>>>> -- true
>>>> -...
>>>> --- Make the WAL directory inaccessible.
>>>> -fio = require('fio')
>>>> ----
>>>> -...
>>>> -path = fio.abspath(box.cfg.wal_dir)
>>>> ----
>>>> -...
>>>> -fio.chmod(path, 0)
>>>> ----
>>>> -- true
>>>> -...
>>>> --- Break replication on timeout.
>>>> -replication_timeout = box.cfg.replication_timeout
>>>> ----
>>>> -...
>>>> -box.cfg{replication_timeout = 9000}
>>>> ----
>>>> -...
>>>> -test_run:cmd("switch replica")
>>>> ----
>>>> -- true
>>>> -...
>>>> -test_run:wait_cond(function() return box.info.replication[1].upstream.status ~= 'follow' end)
>>>> ----
>>>> -- true
>>>> -...
>>>> -require('fiber').sleep(box.cfg.replication_timeout)
>>>> ----
>>>> -...
>>>> -test_run:cmd("switch default")
>>>> ----
>>>> -- true
>>>> -...
>>>> -box.cfg{replication_timeout = replication_timeout}
>>>> ----
>>>> -...
>>>> --- Restore access to the WAL directory.
>>>> --- Wait for replication to be reestablished.
>>>> -fio.chmod(path, tonumber('777', 8))
>>>> ----
>>>> -- true
>>>> -...
>>>> -test_run:cmd("switch replica")
>>>> ----
>>>> -- true
>>>> -...
>>>> -test_run:wait_cond(function() return box.info.replication[1].upstream.status == 'follow' end)
>>>> ----
>>>> -- true
>>>> -...
>>>> -test_run:cmd("switch default")
>>>> ----
>>>> -- true
>>>> -...
>>>> -test_run:cmd("stop server replica")
>>>> ----
>>>> -- true
>>>> -...
>>>> -test_run:cmd("cleanup server replica")
>>>> ----
>>>> -- true
>>>> -...
>>>> -test_run:cmd("delete server replica")
>>>> ----
>>>> -- true
>>>> -...
>>>> -test_run:cleanup_cluster()
>>>> ----
>>>> -...
>>>> -box.schema.user.revoke('guest', 'replication')
>>>> ----
>>>> -...
>>>> ---
>>>> --- gh-4424 Always enter orphan mode on error in replication
>>>> --- configuration change.
>>>> ---
>>>> -replication_connect_timeout = box.cfg.replication_connect_timeout
>>>> ----
>>>> -...
>>>> -replication_connect_quorum = box.cfg.replication_connect_quorum
>>>> ----
>>>> -...
>>>> -box.cfg{replication="12345", replication_connect_timeout=0.1, replication_connect_quorum=1}
>>>> ----
>>>> -...
>>>> -box.info.status
>>>> ----
>>>> -- orphan
>>>> -...
>>>> -box.info.ro
>>>> ----
>>>> -- true
>>>> -...
>>>> --- reset replication => leave orphan mode
>>>> -box.cfg{replication=""}
>>>> ----
>>>> -...
>>>> -box.info.status
>>>> ----
>>>> -- running
>>>> -...
>>>> -box.info.ro
>>>> ----
>>>> -- false
>>>> -...
>>>> --- no switch to orphan when quorum == 0
>>>> -box.cfg{replication="12345", replication_connect_quorum=0}
>>>> ----
>>>> -...
>>>> -box.info.status
>>>> ----
>>>> -- running
>>>> -...
>>>> -box.info.ro
>>>> ----
>>>> -- false
>>>> -...
>>>> --- we could connect to one out of two replicas. Set orphan.
>>>> -box.cfg{replication_connect_quorum=2}
>>>> ----
>>>> -...
>>>> -box.cfg{replication={box.cfg.listen, "12345"}}
>>>> ----
>>>> -...
>>>> -box.info.status
>>>> ----
>>>> -- orphan
>>>> -...
>>>> -box.info.ro
>>>> ----
>>>> -- true
>>>> -...
>>>> --- lower quorum => leave orphan mode
>>>> -box.cfg{replication_connect_quorum=1}
>>>> ----
>>>> -...
>>>> -box.info.status
>>>> ----
>>>> -- running
>>>> -...
>>>> -box.info.ro
>>>> ----
>>>> -- false
>>>> -...
>>>> ---
>>>> --- gh-3760: replication quorum 0 on reconfiguration should return
>>>> --- from box.cfg immediately.
>>>> ---
>>>> -replication = box.cfg.replication
>>>> ----
>>>> -...
>>>> -box.cfg{                                                        \
>>>> -    replication = {},                                           \
>>>> -    replication_connect_quorum = 0,                             \
>>>> -    replication_connect_timeout = 1000000                       \
>>>> -}
>>>> ----
>>>> -...
>>>> --- The call below would hang, if quorum 0 is ignored, or checked
>>>> --- too late.
>>>> -box.cfg{replication = {'localhost:12345'}}
>>>> ----
>>>> -...
>>>> -box.info.status
>>>> ----
>>>> -- running
>>>> -...
>>>> -box.cfg{                                                        \
>>>> -    replication = {},                                           \
>>>> -    replication_connect_quorum = replication_connect_quorum,    \
>>>> -    replication_connect_timeout = replication_connect_timeout   \
>>>> -}
>>>> ----
>>>> -...
>>>> diff --git a/test/replication/misc.test.lua b/test/replication/misc.test.lua
>>>> deleted file mode 100644
>>>> index d285b014a..000000000
>>>> --- a/test/replication/misc.test.lua
>>>> +++ /dev/null
>>>> @@ -1,356 +0,0 @@
>>>> -uuid = require('uuid')
>>>> -test_run = require('test_run').new()
>>>> -
>>>> -box.schema.user.grant('guest', 'replication')
>>>> -
>>>> --- gh-2991 - Tarantool asserts on box.cfg.replication update if one of
>>>> --- servers is dead
>>>> -replication_timeout = box.cfg.replication_timeout
>>>> -replication_connect_timeout = box.cfg.replication_connect_timeout
>>>> -box.cfg{replication_timeout=0.05, replication_connect_timeout=0.05, replication={}}
>>>> -box.cfg{replication_connect_quorum=2}
>>>> -box.cfg{replication = {'127.0.0.1:12345', box.cfg.listen}}
>>>> -box.info.status
>>>> -box.info.ro
>>>> -
>>>> --- gh-3606 - Tarantool crashes if box.cfg.replication is updated concurrently
>>>> -fiber = require('fiber')
>>>> -c = fiber.channel(2)
>>>> -f = function() fiber.create(function() pcall(box.cfg, {replication = {12345}}) c:put(true) end) end
>>>> -f()
>>>> -f()
>>>> -c:get()
>>>> -c:get()
>>>> -
>>>> -box.cfg{replication = "", replication_timeout = replication_timeout, replication_connect_timeout = replication_connect_timeout}
>>>> -box.info.status
>>>> -box.info.ro
>>>> -
>>>> --- gh-3111 - Allow to rebootstrap a replica from a read-only master
>>>> -replica_uuid = uuid.new()
>>>> -test_run:cmd('create server test with rpl_master=default, script="replication/replica_uuid.lua"')
>>>> -test_run:cmd(string.format('start server test with args="%s"', replica_uuid))
>>>> -test_run:cmd('stop server test')
>>>> -test_run:cmd('cleanup server test')
>>>> -box.cfg{read_only = true}
>>>> -test_run:cmd(string.format('start server test with args="%s"', replica_uuid))
>>>> -test_run:cmd('stop server test')
>>>> -test_run:cmd('cleanup server test')
>>>> -box.cfg{read_only = false}
>>>> -test_run:cmd('delete server test')
>>>> -test_run:cleanup_cluster()
>>>> -
>>>> --- gh-3160 - Send heartbeats if there are changes from a remote master only
>>>> -SERVERS = { 'autobootstrap1', 'autobootstrap2', 'autobootstrap3' }
>>>> -
>>>> --- Deploy a cluster.
>>>> -test_run:create_cluster(SERVERS, "replication", {args="0.03"})
>>>> -test_run:wait_fullmesh(SERVERS)
>>>> -test_run:cmd("switch autobootstrap3")
>>>> -test_run = require('test_run').new()
>>>> -fiber = require('fiber')
>>>> -_ = box.schema.space.create('test_timeout'):create_index('pk')
>>>> -test_run:cmd("setopt delimiter ';'")
>>>> -function wait_not_follow(replicaA, replicaB)
>>>> -    return test_run:wait_cond(function()
>>>> -        return replicaA.status ~= 'follow' or replicaB.status ~= 'follow'
>>>> -    end, box.cfg.replication_timeout)
>>>> -end;
>>>> -function test_timeout()
>>>> -    local replicaA = box.info.replication[1].upstream or box.info.replication[2].upstream
>>>> -    local replicaB = box.info.replication[3].upstream or box.info.replication[2].upstream
>>>> -    local follows = test_run:wait_cond(function()
>>>> -        return replicaA.status == 'follow' or replicaB.status == 'follow'
>>>> -    end)
>>>> -    if not follows then error('replicas are not in the follow status') end
>>>> -    for i = 0, 99 do
>>>> -        box.space.test_timeout:replace({1})
>>>> -        if wait_not_follow(replicaA, replicaB) then
>>>> -            return error(box.info.replication)
>>>> -        end
>>>> -    end
>>>> -    return true
>>>> -end;
>>>> -test_run:cmd("setopt delimiter ''");
>>>> -test_timeout()
>>>> -
>>>> --- gh-3247 - Sequence-generated value is not replicated in case
>>>> --- the request was sent via iproto.
>>>> -test_run:cmd("switch autobootstrap1")
>>>> -net_box = require('net.box')
>>>> -_ = box.schema.space.create('space1')
>>>> -_ = box.schema.sequence.create('seq')
>>>> -_ = box.space.space1:create_index('primary', {sequence = true} )
>>>> -_ = box.space.space1:create_index('secondary', {parts = {2, 'unsigned'}})
>>>> -box.schema.user.grant('guest', 'read,write', 'space', 'space1')
>>>> -c = net_box.connect(box.cfg.listen)
>>>> -c.space.space1:insert{box.NULL, "data"} -- fails, but bumps sequence value
>>>> -c.space.space1:insert{box.NULL, 1, "data"}
>>>> -box.space.space1:select{}
>>>> -vclock = test_run:get_vclock("autobootstrap1")
>>>> -vclock[0] = nil
>>>> -_ = test_run:wait_vclock("autobootstrap2", vclock)
>>>> -test_run:cmd("switch autobootstrap2")
>>>> -box.space.space1:select{}
>>>> -test_run:cmd("switch autobootstrap1")
>>>> -box.space.space1:drop()
>>>> -
>>>> -test_run:cmd("switch default")
>>>> -test_run:drop_cluster(SERVERS)
>>>> -test_run:cleanup_cluster()
>>>> -
>>>> --- gh-3642 - Check that socket file descriptor doesn't leak
>>>> --- when a replica is disconnected.
>>>> -rlimit = require('rlimit')
>>>> -lim = rlimit.limit()
>>>> -rlimit.getrlimit(rlimit.RLIMIT_NOFILE, lim)
>>>> -old_fno = lim.rlim_cur
>>>> -lim.rlim_cur = 64
>>>> -rlimit.setrlimit(rlimit.RLIMIT_NOFILE, lim)
>>>> -
>>>> -test_run:cmd('create server sock with rpl_master=default, script="replication/replica.lua"')
>>>> -test_run:cmd('start server sock')
>>>> -test_run:cmd('switch sock')
>>>> -test_run = require('test_run').new()
>>>> -fiber = require('fiber')
>>>> -test_run:cmd("setopt delimiter ';'")
>>>> -for i = 1, 64 do
>>>> -    local replication = box.cfg.replication
>>>> -    box.cfg{replication = {}}
>>>> -    box.cfg{replication = replication}
>>>> -    while box.info.replication[1].upstream.status ~= 'follow' do
>>>> -        fiber.sleep(0.001)
>>>> -    end
>>>> -end;
>>>> -test_run:cmd("setopt delimiter ''");
>>>> -
>>>> -box.info.replication[1].upstream.status
>>>> -
>>>> -test_run:cmd('switch default')
>>>> -
>>>> -lim.rlim_cur = old_fno
>>>> -rlimit.setrlimit(rlimit.RLIMIT_NOFILE, lim)
>>>> -
>>>> -test_run:cmd("stop server sock")
>>>> -test_run:cmd("cleanup server sock")
>>>> -test_run:cmd("delete server sock")
>>>> -test_run:cleanup_cluster()
>>>> -
>>>> -box.schema.user.revoke('guest', 'replication')
>>>> -
>>>> --- gh-3510 assertion failure in replica_on_applier_disconnect()
>>>> -test_run:cmd('create server er_load1 with script="replication/er_load1.lua"')
>>>> -test_run:cmd('create server er_load2 with script="replication/er_load2.lua"')
>>>> -test_run:cmd('start server er_load1 with wait=False, wait_load=False')
>>>> --- Instance er_load2 will fail with error ER_REPLICASET_UUID_MISMATCH.
>>>> --- This is OK since we only test here that er_load1 doesn't assert.
>>>> -test_run:cmd('start server er_load2 with wait=True, wait_load=True, crash_expected = True')
>>>> -test_run:cmd('stop server er_load1')
>>>> --- er_load2 exits automatically.
>>>> -test_run:cmd('cleanup server er_load1')
>>>> -test_run:cmd('cleanup server er_load2')
>>>> -test_run:cmd('delete server er_load1')
>>>> -test_run:cmd('delete server er_load2')
>>>> -test_run:cleanup_cluster()
>>>> -
>>>> ---
>>>> --- Test case for gh-3637, gh-4550. Before the fix replica would
>>>> --- exit with an error if a user does not exist or a password is
>>>> --- incorrect. Now check that we don't hang/panic and successfully
>>>> --- connect.
>>>> ---
>>>> -fiber = require('fiber')
>>>> -test_run:cmd("create server replica_auth with rpl_master=default, script='replication/replica_auth.lua'")
>>>> -test_run:cmd("start server replica_auth with wait=False, wait_load=False, args='cluster:pass 0.05'")
>>>> --- Wait a bit to make sure replica waits till user is created.
>>>> -fiber.sleep(0.1)
>>>> -box.schema.user.create('cluster')
>>>> --- The user is created. Let the replica fail auth request due to
>>>> --- a wrong password.
>>>> -fiber.sleep(0.1)
>>>> -box.schema.user.passwd('cluster', 'pass')
>>>> -box.schema.user.grant('cluster', 'replication')
>>>> -
>>>> -while box.info.replication[2] == nil do fiber.sleep(0.01) end
>>>> -vclock = test_run:get_vclock('default')
>>>> -vclock[0] = nil
>>>> -_ = test_run:wait_vclock('replica_auth', vclock)
>>>> -
>>>> -test_run:cmd("stop server replica_auth")
>>>> -test_run:cmd("cleanup server replica_auth")
>>>> -test_run:cmd("delete server replica_auth")
>>>> -test_run:cleanup_cluster()
>>>> -
>>>> -box.schema.user.drop('cluster')
>>>> -
>>>> ---
>>>> --- Test case for gh-3610. Before the fix replica would fail with the assertion
>>>> --- when trying to connect to the same master twice.
>>>> ---
>>>> -box.schema.user.grant('guest', 'replication')
>>>> -test_run:cmd("create server replica with rpl_master=default, script='replication/replica.lua'")
>>>> -test_run:cmd("start server replica")
>>>> -test_run:cmd("switch replica")
>>>> -replication = box.cfg.replication[1]
>>>> -box.cfg{replication = {replication, replication}}
>>>> -
>>>> --- Check the case when duplicate connection is detected in the background.
>>>> -test_run:cmd("switch default")
>>>> -listen = box.cfg.listen
>>>> -box.cfg{listen = ''}
>>>> -
>>>> -test_run:cmd("switch replica")
>>>> -box.cfg{replication_connect_quorum = 0, replication_connect_timeout = 0.01}
>>>> -box.cfg{replication = {replication, replication}}
>>>> -
>>>> -test_run:cmd("switch default")
>>>> -box.cfg{listen = listen}
>>>> -while test_run:grep_log('replica', 'duplicate connection') == nil do fiber.sleep(0.01) end
>>>> -
>>>> -test_run:cmd("stop server replica")
>>>> -test_run:cmd("cleanup server replica")
>>>> -test_run:cmd("delete server replica")
>>>> -test_run:cleanup_cluster()
>>>> -box.schema.user.revoke('guest', 'replication')
>>>> -
>>>> ---
>>>> --- gh-3711 Do not restart replication on box.cfg in case the
>>>> --- configuration didn't change.
>>>> ---
>>>> -box.schema.user.grant('guest', 'replication')
>>>> -test_run:cmd("create server replica with rpl_master=default, script='replication/replica.lua'")
>>>> -test_run:cmd("start server replica")
>>>> -
>>>> --- Access rights are checked only during reconnect. If the new
>>>> --- and old configurations are equivalent, no reconnect will be
>>>> --- issued and replication should continue working.
>>>> -box.schema.user.revoke('guest', 'replication')
>>>> -test_run:cmd("switch replica")
>>>> -replication = box.cfg.replication[1]
>>>> -box.cfg{replication = {replication}}
>>>> -box.info.status == 'running'
>>>> -box.cfg{replication = replication}
>>>> -box.info.status == 'running'
>>>> -
>>>> --- Check that comparison of tables works as expected as well.
>>>> -test_run:cmd("switch default")
>>>> -box.schema.user.grant('guest', 'replication')
>>>> -test_run:cmd("switch replica")
>>>> -replication = box.cfg.replication
>>>> -table.insert(replication, box.cfg.listen)
>>>> -test_run:cmd("switch default")
>>>> -box.schema.user.revoke('guest', 'replication')
>>>> -test_run:cmd("switch replica")
>>>> -box.cfg{replication = replication}
>>>> -box.info.status == 'running'
>>>> -
>>>> -test_run:cmd("switch default")
>>>> -test_run:cmd("stop server replica")
>>>> -test_run:cmd("cleanup server replica")
>>>> -test_run:cmd("delete server replica")
>>>> -test_run:cleanup_cluster()
>>>> -
>>>> ---
>>>> --- gh-3704 move cluster id check to replica
>>>> ---
>>>> -test_run:cmd("create server replica with rpl_master=default, script='replication/replica.lua'")
>>>> -box.schema.user.grant("guest", "replication")
>>>> -test_run:cmd("start server replica")
>>>> -test_run:grep_log("replica", "REPLICASET_UUID_MISMATCH")
>>>> -box.info.replication[2].downstream.status
>>>> --- change master's cluster uuid and check that replica doesn't connect.
>>>> -test_run:cmd("stop server replica")
>>>> -_ = box.space._schema:replace{'cluster', tostring(uuid.new())}
>>>> --- master believes replica is in cluster, but their cluster UUIDs differ.
>>>> -test_run:cmd("start server replica")
>>>> -test_run:wait_log("replica", "REPLICASET_UUID_MISMATCH", nil, 1.0)
>>>> -test_run:wait_downstream(2, {status = 'stopped'})
>>>> -
>>>> -test_run:cmd("stop server replica")
>>>> -test_run:cmd("cleanup server replica")
>>>> -test_run:cmd("delete server replica")
>>>> -test_run:cleanup_cluster()
>>>> -box.schema.user.revoke('guest', 'replication')
>>>> -
>>>> ---
>>>> --- gh-4399 Check that an error reading WAL directory on subscribe
>>>> --- doesn't lead to a permanent replication failure.
>>>> ---
>>>> -box.schema.user.grant("guest", "replication")
>>>> -test_run:cmd("create server replica with rpl_master=default, script='replication/replica.lua'")
>>>> -test_run:cmd("start server replica")
>>>> -
>>>> --- Make the WAL directory inaccessible.
>>>> -fio = require('fio')
>>>> -path = fio.abspath(box.cfg.wal_dir)
>>>> -fio.chmod(path, 0)
>>>> -
>>>> --- Break replication on timeout.
>>>> -replication_timeout = box.cfg.replication_timeout
>>>> -box.cfg{replication_timeout = 9000}
>>>> -test_run:cmd("switch replica")
>>>> -test_run:wait_cond(function() return box.info.replication[1].upstream.status ~= 'follow' end)
>>>> -require('fiber').sleep(box.cfg.replication_timeout)
>>>> -test_run:cmd("switch default")
>>>> -box.cfg{replication_timeout = replication_timeout}
>>>> -
>>>> --- Restore access to the WAL directory.
>>>> --- Wait for replication to be reestablished.
>>>> -fio.chmod(path, tonumber('777', 8))
>>>> -test_run:cmd("switch replica")
>>>> -test_run:wait_cond(function() return box.info.replication[1].upstream.status == 'follow' end)
>>>> -test_run:cmd("switch default")
>>>> -
>>>> -test_run:cmd("stop server replica")
>>>> -test_run:cmd("cleanup server replica")
>>>> -test_run:cmd("delete server replica")
>>>> -test_run:cleanup_cluster()
>>>> -box.schema.user.revoke('guest', 'replication')
>>>> -
>>>> ---
>>>> --- gh-4424 Always enter orphan mode on error in replication
>>>> --- configuration change.
>>>> ---
>>>> -replication_connect_timeout = box.cfg.replication_connect_timeout
>>>> -replication_connect_quorum = box.cfg.replication_connect_quorum
>>>> -box.cfg{replication="12345", replication_connect_timeout=0.1, replication_connect_quorum=1}
>>>> -box.info.status
>>>> -box.info.ro
>>>> --- reset replication => leave orphan mode
>>>> -box.cfg{replication=""}
>>>> -box.info.status
>>>> -box.info.ro
>>>> --- no switch to orphan when quorum == 0
>>>> -box.cfg{replication="12345", replication_connect_quorum=0}
>>>> -box.info.status
>>>> -box.info.ro
>>>> -
>>>> --- we could connect to one out of two replicas. Set orphan.
>>>> -box.cfg{replication_connect_quorum=2}
>>>> -box.cfg{replication={box.cfg.listen, "12345"}}
>>>> -box.info.status
>>>> -box.info.ro
>>>> --- lower quorum => leave orphan mode
>>>> -box.cfg{replication_connect_quorum=1}
>>>> -box.info.status
>>>> -box.info.ro
>>>> -
>>>> ---
>>>> --- gh-3760: replication quorum 0 on reconfiguration should return
>>>> --- from box.cfg immediately.
>>>> ---
>>>> -replication = box.cfg.replication
>>>> -box.cfg{                                                        \
>>>> -    replication = {},                                           \
>>>> -    replication_connect_quorum = 0,                             \
>>>> -    replication_connect_timeout = 1000000                       \
>>>> -}
>>>> --- The call below would hang, if quorum 0 is ignored, or checked
>>>> --- too late.
>>>> -box.cfg{replication = {'localhost:12345'}}
>>>> -box.info.status
>>>> -box.cfg{                                                        \
>>>> -    replication = {},                                           \
>>>> -    replication_connect_quorum = replication_connect_quorum,    \
>>>> -    replication_connect_timeout = replication_connect_timeout   \
>>>> -}
>>>> diff --git a/test/replication/misc_assert_connecting_master_twice_gh-3610.result b/test/replication/misc_assert_connecting_master_twice_gh-3610.result
>>>> new file mode 100644
>>>> index 000000000..d7b7cc25b
>>>> --- /dev/null
>>>> +++ b/test/replication/misc_assert_connecting_master_twice_gh-3610.result
>>>> @@ -0,0 +1,86 @@
>>>> +test_run = require('test_run').new()
>>>> +---
>>>> +...
>>>> +test_run:cmd("restart server default")
>>> I don't  think you should restart the server here.
>>> The test seems fine without it. Same about almost all the
>>> testcases below.
>>>> +uuid = require('uuid')
>>>> +---
>>>> +...
>>>> +fiber = require('fiber')
>>>> +---
>>>> +...
>>> UUID and fiber modules aren't used in this test, so please remove
>>> these two extra lines. Same about almost all the testcases below.
>>>
>>>
>>> The branch contains this extraneous change:
>>>
>>> +replication_connect_quorum = box.cfg.replication_connect_quorum
>>>
>>> +---
>>> +...
>>> +replication_connect_timeout = box.cfg.replication_connect_timeout
>>> +---
>>>
>>> +...
>>>
>>> You don't need to save these options, since you only change them on a
>>> replica.
>>>
>>>> +--
>>>> +-- Test case for gh-3610. Before the fix replica would fail with the assertion
>>>> +-- when trying to connect to the same master twice.
>>>> +--
>>>> +box.schema.user.grant('guest', 'replication')
>>>> +---
>>>> +...
>>>> +test_run:cmd("create server replica with rpl_master=default, script='replication/replica.lua'")
>>>> +---
>>>> +- true
>>>> +...
>>>> +test_run:cmd("start server replica")
>>>> +---
>>>> +- true
>>>> +...
>>>> +test_run:cmd("switch replica")
>>>> +---
>>>> +- true
>>>> +...
>>>> +replication = box.cfg.replication[1]
>>>> +---
>>>> +...
>>>> +box.cfg{replication = {replication, replication}}
>>>> +---
>>>> +- error: 'Incorrect value for option ''replication'': duplicate connection to the
>>>> +    same replica'
>>>> +...
>>>> +-- Check the case when duplicate connection is detected in the background.
>>>> +test_run:cmd("switch default")
>>>> +---
>>>> +- true
>>>> +...
>>>> +listen = box.cfg.listen
>>>> +---
>>>> +...
>>>> +box.cfg{listen = ''}
>>>> +---
>>>> +...
>>>> +test_run:cmd("switch replica")
>>>> +---
>>>> +- true
>>>> +...
>>>> +box.cfg{replication_connect_quorum = 0, replication_connect_timeout = 0.01}
>>>> +---
>>>> +...
>>>> +box.cfg{replication = {replication, replication}}
>>>> +---
>>>> +...
>>>> +test_run:cmd("switch default")
>>>> +---
>>>> +- true
>>>> +...
>>>> +box.cfg{listen = listen}
>>>> +---
>>>> +...
>>>> +while test_run:grep_log('replica', 'duplicate connection') == nil do fiber.sleep(0.01) end
>>>> +---
>>>> +...
>>>> +test_run:cmd("stop server replica")
>>>> +---
>>>> +- true
>>>> +...
>>>> +test_run:cmd("cleanup server replica")
>>>> +---
>>>> +- true
>>>> +...
>>>> +test_run:cmd("delete server replica")
>>>> +---
>>>> +- true
>>>> +...
>>>> +test_run:cleanup_cluster()
>>>> +---
>>>> +...
>>>> +box.schema.user.revoke('guest', 'replication')
>>>> +---
>>>> +...
>>> +box.cfg{replication = "", \
>>> +        replication_connect_quorum = replication_connect_quorum, \
>>> +        replication_connect_timeout = replication_connect_timeout}
>>> +---
>>> +...
>>>
>>> The change I spoke of above. This piece is also unneeded.
>>>
>>>> diff --git a/test/replication/misc_assert_connecting_master_twice_gh-3610.test.lua b/test/replication/misc_assert_connecting_master_twice_gh-3610.test.lua
>>>> new file mode 100644
>>>> index 000000000..9b12e623b
>>>> --- /dev/null
>>>> +++ b/test/replication/misc_assert_connecting_master_twice_gh-3610.test.lua
>>>> @@ -0,0 +1,34 @@
>>>> +test_run = require('test_run').new()
>>>> +test_run:cmd("restart server default")
>>>> +uuid = require('uuid')
>>>> +fiber = require('fiber')
>>>> +
>>>> +--
>>>> +-- Test case for gh-3610. Before the fix replica would fail with the assertion
>>>> +-- when trying to connect to the same master twice.
>>>> +--
>>>> +box.schema.user.grant('guest', 'replication')
>>>> +test_run:cmd("create server replica with rpl_master=default, script='replication/replica.lua'")
>>>> +test_run:cmd("start server replica")
>>>> +test_run:cmd("switch replica")
>>>> +replication = box.cfg.replication[1]
>>>> +box.cfg{replication = {replication, replication}}
>>>> +
>>>> +-- Check the case when duplicate connection is detected in the background.
>>>> +test_run:cmd("switch default")
>>>> +listen = box.cfg.listen
>>>> +box.cfg{listen = ''}
>>>> +
>>>> +test_run:cmd("switch replica")
>>>> +box.cfg{replication_connect_quorum = 0, replication_connect_timeout = 0.01}
>>>> +box.cfg{replication = {replication, replication}}
>>>> +
>>>> +test_run:cmd("switch default")
>>>> +box.cfg{listen = listen}
>>>> +while test_run:grep_log('replica', 'duplicate connection') == nil do fiber.sleep(0.01) end
>>>> +
>>>> +test_run:cmd("stop server replica")
>>>> +test_run:cmd("cleanup server replica")
>>>> +test_run:cmd("delete server replica")
>>>> +test_run:cleanup_cluster()
>>>> +box.schema.user.revoke('guest', 'replication')
>>>> diff --git a/test/replication/misc_assert_on_server_die_gh-2991.result b/test/replication/misc_assert_on_server_die_gh-2991.result
>>>> new file mode 100644
>>>> index 000000000..ea9e80f6b
>>>> --- /dev/null
>>>> +++ b/test/replication/misc_assert_on_server_die_gh-2991.result
>>>> @@ -0,0 +1,31 @@
>>>> +uuid = require('uuid')
>>>> +---
>>>> +...
>>> UUID module isn't used in this testcase, please remove the extra line.
>>>> +test_run = require('test_run').new()
>>>> +---
>>>> +...
>>>> +-- gh-2991 - Tarantool asserts on box.cfg.replication update if one of
>>>> +-- servers is dead
>>>> +replication_timeout = box.cfg.replication_timeout
>>>> +---
>>>> +...
>>>> +replication_connect_timeout = box.cfg.replication_connect_timeout
>>>> +---
>>>> +...
>>>> +box.cfg{replication_timeout=0.05, replication_connect_timeout=0.05, replication={}}
>>>> +---
>>>> +...
>>>> +box.cfg{replication_connect_quorum=2}
>>>> +---
>>>> +...
>>>> +box.cfg{replication = {'127.0.0.1:12345', box.cfg.listen}}
>>>> +---
>>>> +...
>>>> +box.info.status
>>>> +---
>>>> +- orphan
>>>> +...
>>>> +box.info.ro
>>>> +---
>>>> +- true
>>>> +...
>>>> diff --git a/test/replication/misc_assert_on_server_die_gh-2991.test.lua b/test/replication/misc_assert_on_server_die_gh-2991.test.lua
>>>> new file mode 100644
>>>> index 000000000..adda839f7
>>>> --- /dev/null
>>>> +++ b/test/replication/misc_assert_on_server_die_gh-2991.test.lua
>>>> @@ -0,0 +1,12 @@
>>>> +uuid = require('uuid')
>>>> +test_run = require('test_run').new()
>>>> +
>>>> +-- gh-2991 - Tarantool asserts on box.cfg.replication update if one of
>>>> +-- servers is dead
>>>> +replication_timeout = box.cfg.replication_timeout
>>>> +replication_connect_timeout = box.cfg.replication_connect_timeout
>>>> +box.cfg{replication_timeout=0.05, replication_connect_timeout=0.05, replication={}}
>>>> +box.cfg{replication_connect_quorum=2}
>>>> +box.cfg{replication = {'127.0.0.1:12345', box.cfg.listen}}
>>>> +box.info.status
>>>> +box.info.ro
>>>> diff --git a/test/replication/misc_assert_replica_on_applier_disconnect_gh-3510.result b/test/replication/misc_assert_replica_on_applier_disconnect_gh-3510.result
>>>> new file mode 100644
>>>> index 000000000..82f5d9a23
>>>> --- /dev/null
>>>> +++ b/test/replication/misc_assert_replica_on_applier_disconnect_gh-3510.result
>>>> @@ -0,0 +1,49 @@
>>>> +uuid = require('uuid')
>>>> +---
>>>> +...
>>> Same about the uuid.
>>>> +test_run = require('test_run').new()
>>>> +---
>>>> +...
>>>> +-- gh-3510 assertion failure in replica_on_applier_disconnect()
>>>> +test_run:cmd('create server er_load1 with script="replication/er_load1.lua"')
>>>> +---
>>>> +- true
>>>> +...
>>>> +test_run:cmd('create server er_load2 with script="replication/er_load2.lua"')
>>>> +---
>>>> +- true
>>>> +...
>>>> +test_run:cmd('start server er_load1 with wait=False, wait_load=False')
>>>> +---
>>>> +- true
>>>> +...
>>>> +-- Instance er_load2 will fail with error ER_REPLICASET_UUID_MISMATCH.
>>>> +-- This is OK since we only test here that er_load1 doesn't assert.
>>>> +test_run:cmd('start server er_load2 with wait=True, wait_load=True, crash_expected = True')
>>>> +---
>>>> +- false
>>>> +...
>>>> +test_run:cmd('stop server er_load1')
>>>> +---
>>>> +- true
>>>> +...
>>>> +-- er_load2 exits automatically.
>>>> +test_run:cmd('cleanup server er_load1')
>>>> +---
>>>> +- true
>>>> +...
>>>> +test_run:cmd('cleanup server er_load2')
>>>> +---
>>>> +- true
>>>> +...
>>>> +test_run:cmd('delete server er_load1')
>>>> +---
>>>> +- true
>>>> +...
>>>> +test_run:cmd('delete server er_load2')
>>>> +---
>>>> +- true
>>>> +...
>>>> +test_run:cleanup_cluster()
>>>> +---
>>>> +...
>>>> diff --git a/test/replication/misc_assert_replica_on_applier_disconnect_gh-3510.test.lua b/test/replication/misc_assert_replica_on_applier_disconnect_gh-3510.test.lua
>>>> new file mode 100644
>>>> index 000000000..4e1d2a41e
>>>> --- /dev/null
>>>> +++ b/test/replication/misc_assert_replica_on_applier_disconnect_gh-3510.test.lua
>>>> @@ -0,0 +1,17 @@
>>>> +uuid = require('uuid')
>>>> +test_run = require('test_run').new()
>>>> +
>>>> +-- gh-3510 assertion failure in replica_on_applier_disconnect()
>>>> +test_run:cmd('create server er_load1 with script="replication/er_load1.lua"')
>>>> +test_run:cmd('create server er_load2 with script="replication/er_load2.lua"')
>>>> +test_run:cmd('start server er_load1 with wait=False, wait_load=False')
>>>> +-- Instance er_load2 will fail with error ER_REPLICASET_UUID_MISMATCH.
>>>> +-- This is OK since we only test here that er_load1 doesn't assert.
>>>> +test_run:cmd('start server er_load2 with wait=True, wait_load=True, crash_expected = True')
>>>> +test_run:cmd('stop server er_load1')
>>>> +-- er_load2 exits automatically.
>>>> +test_run:cmd('cleanup server er_load1')
>>>> +test_run:cmd('cleanup server er_load2')
>>>> +test_run:cmd('delete server er_load1')
>>>> +test_run:cmd('delete server er_load2')
>>>> +test_run:cleanup_cluster()
>>>> diff --git a/test/replication/misc_crash_on_box_concurrent_update_gh-3606.result b/test/replication/misc_crash_on_box_concurrent_update_gh-3606.result
>>>> new file mode 100644
>>>> index 000000000..b43b00849
>>>> --- /dev/null
>>>> +++ b/test/replication/misc_crash_on_box_concurrent_update_gh-3606.result
>>>> @@ -0,0 +1,50 @@
>>>> +uuid = require('uuid')
>>>> +---
>>>> +...
>>> Same about UUID.
>>>> +test_run = require('test_run').new()
>>>> +---
>>>> +...
>>>> +replication_timeout = box.cfg.replication_timeout
>>>> +---
>>>> +...
>>>> +replication_connect_timeout = box.cfg.replication_connect_timeout
>>>> +---
>>>> +...
>>>> +box.cfg{replication_timeout=0.05, replication_connect_timeout=0.05, replication={}}
>>>> +---
>>>> +...
>>>> +-- gh-3606 - Tarantool crashes if box.cfg.replication is updated concurrently
>>>> +fiber = require('fiber')
>>>> +---
>>>> +...
>>>> +c = fiber.channel(2)
>>>> +---
>>>> +...
>>>> +f = function() fiber.create(function() pcall(box.cfg, {replication = {12345}}) c:put(true) end) end
>>>> +---
>>>> +...
>>>> +f()
>>>> +---
>>>> +...
>>>> +f()
>>>> +---
>>>> +...
>>>> +c:get()
>>>> +---
>>>> +- true
>>>> +...
>>>> +c:get()
>>>> +---
>>>> +- true
>>>> +...
>>>> +box.cfg{replication = "", replication_timeout = replication_timeout, replication_connect_timeout = replication_connect_timeout}
>>>> +---
>>>> +...
>>>> +box.info.status
>>>> +---
>>>> +- running
>>>> +...
>>>> +box.info.ro
>>>> +---
>>>> +- false
>>>> +...
>>>> diff --git a/test/replication/misc_crash_on_box_concurrent_update_gh-3606.test.lua b/test/replication/misc_crash_on_box_concurrent_update_gh-3606.test.lua
>>>> new file mode 100644
>>>> index 000000000..17f9c6bc6
>>>> --- /dev/null
>>>> +++ b/test/replication/misc_crash_on_box_concurrent_update_gh-3606.test.lua
>>>> @@ -0,0 +1,19 @@
>>>> +uuid = require('uuid')
>>>> +test_run = require('test_run').new()
>>>> +
>>>> +replication_timeout = box.cfg.replication_timeout
>>>> +replication_connect_timeout = box.cfg.replication_connect_timeout
>>>> +box.cfg{replication_timeout=0.05, replication_connect_timeout=0.05, replication={}}
>>>> +
>>>> +-- gh-3606 - Tarantool crashes if box.cfg.replication is updated concurrently
>>>> +fiber = require('fiber')
>>>> +c = fiber.channel(2)
>>>> +f = function() fiber.create(function() pcall(box.cfg, {replication = {12345}}) c:put(true) end) end
>>>> +f()
>>>> +f()
>>>> +c:get()
>>>> +c:get()
>>>> +
>>>> +box.cfg{replication = "", replication_timeout = replication_timeout, replication_connect_timeout = replication_connect_timeout}
>>>> +box.info.status
>>>> +box.info.ro
>>>> diff --git a/test/replication/misc_heartbeats_on_master_changes_gh-3160.result b/test/replication/misc_heartbeats_on_master_changes_gh-3160.result
>>>> new file mode 100644
>>>> index 000000000..cdb463614
>>>> --- /dev/null
>>>> +++ b/test/replication/misc_heartbeats_on_master_changes_gh-3160.result
>>>> @@ -0,0 +1,76 @@
>>>> +uuid = require('uuid')
>>>> +---
>>>> +...
>>> Same about UUID.
>>>
>>>> +test_run = require('test_run').new()
>>>> +---
>>>> +...
>>> The branch contains this change:
>>>
>>> +...
>>> +replication_timeout = box.cfg.replication_timeout
>>> +---
>>> +...
>>> +replication_connect_timeout = box.cfg.replication_connect_timeout
>>> +---
>>> +...
>>> +box.cfg{replication_timeout=0.05, replication_connect_timeout=0.05,
>>> replication={}}
>>> +---
>>> +...
>>>
>>> Which is extraneous in this test. The original testcase uses default timeout
>>> values,
>>> and I don't think we should change them.
>>>
>>>> +-- gh-3160 - Send heartbeats if there are changes from a remote master only
>>>> +SERVERS = { 'autobootstrap1', 'autobootstrap2', 'autobootstrap3' }
>>>> +---
>>>> +...
>>>> +-- Deploy a cluster.
>>>> +test_run:create_cluster(SERVERS, "replication", {args="0.03"})
>>>> +---
>>>> +...
>>>> +test_run:wait_fullmesh(SERVERS)
>>>> +---
>>>> +...
>>>> +test_run:cmd("switch autobootstrap3")
>>>> +---
>>>> +- true
>>>> +...
>>>> +test_run = require('test_run').new()
>>>> +---
>>>> +...
>>>> +fiber = require('fiber')
>>>> +---
>>>> +...
>>>> +_ = box.schema.space.create('test_timeout'):create_index('pk')
>>>> +---
>>>> +...
>>>> +test_run:cmd("setopt delimiter ';'")
>>>> +---
>>>> +- true
>>>> +...
>>>> +function wait_not_follow(replicaA, replicaB)
>>>> +    return test_run:wait_cond(function()
>>>> +        return replicaA.status ~= 'follow' or replicaB.status ~= 'follow'
>>>> +    end, box.cfg.replication_timeout)
>>>> +end;
>>>> +---
>>>> +...
>>>> +function test_timeout()
>>>> +    local replicaA = box.info.replication[1].upstream or box.info.replication[2].upstream
>>>> +    local replicaB = box.info.replication[3].upstream or box.info.replication[2].upstream
>>>> +    local follows = test_run:wait_cond(function()
>>>> +        return replicaA.status == 'follow' or replicaB.status == 'follow'
>>>> +    end)
>>>> +    if not follows then error('replicas are not in the follow status') end
>>>> +    for i = 0, 99 do
>>>> +        box.space.test_timeout:replace({1})
>>>> +        if wait_not_follow(replicaA, replicaB) then
>>>> +            return error(box.info.replication)
>>>> +        end
>>>> +    end
>>>> +    return true
>>>> +end;
>>>> +---
>>>> +...
>>>> +test_run:cmd("setopt delimiter ''");
>>>> +---
>>>> +- true
>>>> +...
>>>> +test_timeout()
>>>> +---
>>>> +- true
>>>> +...
>>>> +test_run:cmd("switch default")
>>>> +---
>>>> +- true
>>>> +...
>>>> +test_run:drop_cluster(SERVERS)
>>>> +---
>>>> +...
>>>> +test_run:cleanup_cluster()
>>> No need to call cleanup_cluster(). The 'default' instance wasn't part of the
>>> cluster
>>> you ran in this test.
>>>
>>>
>>> +...
>>> +box.cfg{replication = "", replication_timeout = replication_timeout, \
>>> +        replication_connect_timeout = replication_connect_timeout}
>>> +---
>>> +...
>>>
>>> One more extraneous change, related to the one above.
>>>
>>>
>>>> +---
>>>> +...
>>>> diff --git a/test/replication/misc_heartbeats_on_master_changes_gh-3160.test.lua b/test/replication/misc_heartbeats_on_master_changes_gh-3160.test.lua
>>>> new file mode 100644
>>>> index 000000000..eba8a7725
>>>> --- /dev/null
>>>> +++ b/test/replication/misc_heartbeats_on_master_changes_gh-3160.test.lua
>>>> @@ -0,0 +1,40 @@
>>>> +uuid = require('uuid')
>>>> +test_run = require('test_run').new()
>>>> +
>>>> +-- gh-3160 - Send heartbeats if there are changes from a remote master only
>>>> +SERVERS = { 'autobootstrap1', 'autobootstrap2', 'autobootstrap3' }
>>>> +
>>>> +-- Deploy a cluster.
>>>> +test_run:create_cluster(SERVERS, "replication", {args="0.03"})
>>>> +test_run:wait_fullmesh(SERVERS)
>>>> +test_run:cmd("switch autobootstrap3")
>>>> +test_run = require('test_run').new()
>>>> +fiber = require('fiber')
>>>> +_ = box.schema.space.create('test_timeout'):create_index('pk')
>>>> +test_run:cmd("setopt delimiter ';'")
>>>> +function wait_not_follow(replicaA, replicaB)
>>>> +    return test_run:wait_cond(function()
>>>> +        return replicaA.status ~= 'follow' or replicaB.status ~= 'follow'
>>>> +    end, box.cfg.replication_timeout)
>>>> +end;
>>>> +function test_timeout()
>>>> +    local replicaA = box.info.replication[1].upstream or box.info.replication[2].upstream
>>>> +    local replicaB = box.info.replication[3].upstream or box.info.replication[2].upstream
>>>> +    local follows = test_run:wait_cond(function()
>>>> +        return replicaA.status == 'follow' or replicaB.status == 'follow'
>>>> +    end)
>>>> +    if not follows then error('replicas are not in the follow status') end
>>>> +    for i = 0, 99 do
>>>> +        box.space.test_timeout:replace({1})
>>>> +        if wait_not_follow(replicaA, replicaB) then
>>>> +            return error(box.info.replication)
>>>> +        end
>>>> +    end
>>>> +    return true
>>>> +end;
>>>> +test_run:cmd("setopt delimiter ''");
>>>> +test_timeout()
>>>> +
>>>> +test_run:cmd("switch default")
>>>> +test_run:drop_cluster(SERVERS)
>>>> +test_run:cleanup_cluster()
>>>> diff --git a/test/replication/misc_no_failure_on_error_reading_wal_gh-4399.result b/test/replication/misc_no_failure_on_error_reading_wal_gh-4399.result
>>>> new file mode 100644
>>>> index 000000000..64a4de26c
>>>> --- /dev/null
>>>> +++ b/test/replication/misc_no_failure_on_error_reading_wal_gh-4399.result
>>>> @@ -0,0 +1,97 @@
>>>> +test_run = require('test_run').new()
>>>> +---
>>>> +...
>>>> +test_run:cmd("restart server default")
>>>> +uuid = require('uuid')
>>>> +---
>>>> +...
>>>> +fiber = require('fiber')
>>> Same about restarting the server, requiring uuid and fiber. Not needed.
>>>> +---
>>>> +...
>>>> +--
>>>> +-- gh-4399 Check that an error reading WAL directory on subscribe
>>>> +-- doesn't lead to a permanent replication failure.
>>>> +--
>>>> +box.schema.user.grant("guest", "replication")
>>>> +---
>>>> +...
>>>> +test_run:cmd("create server replica with rpl_master=default, script='replication/replica.lua'")
>>>> +---
>>>> +- true
>>>> +...
>>>> +test_run:cmd("start server replica")
>>>> +---
>>>> +- true
>>>> +...
>>>> +-- Make the WAL directory inaccessible.
>>>> +fio = require('fio')
>>>> +---
>>>> +...
>>>> +path = fio.abspath(box.cfg.wal_dir)
>>>> +---
>>>> +...
>>>> +fio.chmod(path, 0)
>>>> +---
>>>> +- true
>>>> +...
>>>> +-- Break replication on timeout.
>>>> +replication_timeout = box.cfg.replication_timeout
>>>> +---
>>>> +...
>>>> +box.cfg{replication_timeout = 9000}
>>>> +---
>>>> +...
>>>> +test_run:cmd("switch replica")
>>>> +---
>>>> +- true
>>>> +...
>>>> +test_run:wait_cond(function() return box.info.replication[1].upstream.status ~= 'follow' end)
>>>> +---
>>>> +- true
>>>> +...
>>>> +require('fiber').sleep(box.cfg.replication_timeout)
>>>> +---
>>>> +...
>>>> +test_run:cmd("switch default")
>>>> +---
>>>> +- true
>>>> +...
>>>> +box.cfg{replication_timeout = replication_timeout}
>>>> +---
>>>> +...
>>>> +-- Restore access to the WAL directory.
>>>> +-- Wait for replication to be reestablished.
>>>> +fio.chmod(path, tonumber('777', 8))
>>>> +---
>>>> +- true
>>>> +...
>>>> +test_run:cmd("switch replica")
>>>> +---
>>>> +- true
>>>> +...
>>>> +test_run:wait_cond(function() return box.info.replication[1].upstream.status == 'follow' end)
>>>> +---
>>>> +- true
>>>> +...
>>>> +test_run:cmd("switch default")
>>>> +---
>>>> +- true
>>>> +...
>>>> +test_run:cmd("stop server replica")
>>>> +---
>>>> +- true
>>>> +...
>>>> +test_run:cmd("cleanup server replica")
>>>> +---
>>>> +- true
>>>> +...
>>>> +test_run:cmd("delete server replica")
>>>> +---
>>>> +- true
>>>> +...
>>>> +test_run:cleanup_cluster()
>>>> +---
>>>> +...
>>>> +box.schema.user.revoke('guest', 'replication')
>>>> +---
>>>> +...
>>>> diff --git a/test/replication/misc_no_failure_on_error_reading_wal_gh-4399.test.lua b/test/replication/misc_no_failure_on_error_reading_wal_gh-4399.test.lua
>>>> new file mode 100644
>>>> index 000000000..15e19a211
>>>> --- /dev/null
>>>> +++ b/test/replication/misc_no_failure_on_error_reading_wal_gh-4399.test.lua
>>>> @@ -0,0 +1,39 @@
>>>> +test_run = require('test_run').new()
>>>> +test_run:cmd("restart server default")
>>>> +uuid = require('uuid')
>>>> +fiber = require('fiber')
>>>> +
>>>> +--
>>>> +-- gh-4399 Check that an error reading WAL directory on subscribe
>>>> +-- doesn't lead to a permanent replication failure.
>>>> +--
>>>> +box.schema.user.grant("guest", "replication")
>>>> +test_run:cmd("create server replica with rpl_master=default, script='replication/replica.lua'")
>>>> +test_run:cmd("start server replica")
>>>> +
>>>> +-- Make the WAL directory inaccessible.
>>>> +fio = require('fio')
>>>> +path = fio.abspath(box.cfg.wal_dir)
>>>> +fio.chmod(path, 0)
>>>> +
>>>> +-- Break replication on timeout.
>>>> +replication_timeout = box.cfg.replication_timeout
>>>> +box.cfg{replication_timeout = 9000}
>>>> +test_run:cmd("switch replica")
>>>> +test_run:wait_cond(function() return box.info.replication[1].upstream.status ~= 'follow' end)
>>>> +require('fiber').sleep(box.cfg.replication_timeout)
>>>> +test_run:cmd("switch default")
>>>> +box.cfg{replication_timeout = replication_timeout}
>>>> +
>>>> +-- Restore access to the WAL directory.
>>>> +-- Wait for replication to be reestablished.
>>>> +fio.chmod(path, tonumber('777', 8))
>>>> +test_run:cmd("switch replica")
>>>> +test_run:wait_cond(function() return box.info.replication[1].upstream.status == 'follow' end)
>>>> +test_run:cmd("switch default")
>>>> +
>>>> +test_run:cmd("stop server replica")
>>>> +test_run:cmd("cleanup server replica")
>>>> +test_run:cmd("delete server replica")
>>>> +test_run:cleanup_cluster()
>>>> +box.schema.user.revoke('guest', 'replication')
>>>> diff --git a/test/replication/misc_no_panic_on_connected_gh-3637.result b/test/replication/misc_no_panic_on_connected_gh-3637.result
>>>> new file mode 100644
>>>> index 000000000..693b18c42
>>>> --- /dev/null
>>>> +++ b/test/replication/misc_no_panic_on_connected_gh-3637.result
>>>> @@ -0,0 +1,72 @@
>>>> +test_run = require('test_run').new()
>>>> +---
>>>> +...
>>>> +test_run:cmd("restart server default")
>>>> +uuid = require('uuid')
>>>> +---
>>>> +...
>>> Again, no need for uuid here. And no need for restart.
>>>> +--
>>>> +-- Test case for gh-3637, gh-4550. Before the fix replica would
>>>> +-- exit with an error if a user does not exist or a password is
>>>> +-- incorrect. Now check that we don't hang/panic and successfully
>>>> +-- connect.
>>>> +--
>>>> +fiber = require('fiber')
>>>> +---
>>>> +...
>>>> +test_run:cmd("create server replica_auth with rpl_master=default, script='replication/replica_auth.lua'")
>>>> +---
>>>> +- true
>>>> +...
>>>> +test_run:cmd("start server replica_auth with wait=False, wait_load=False, args='cluster:pass 0.05'")
>>>> +---
>>>> +- true
>>>> +...
>>>> +-- Wait a bit to make sure replica waits till user is created.
>>>> +fiber.sleep(0.1)
>>>> +---
>>>> +...
>>>> +box.schema.user.create('cluster')
>>>> +---
>>>> +...
>>>> +-- The user is created. Let the replica fail auth request due to
>>>> +-- a wrong password.
>>>> +fiber.sleep(0.1)
>>>> +---
>>>> +...
>>>> +box.schema.user.passwd('cluster', 'pass')
>>>> +---
>>>> +...
>>>> +box.schema.user.grant('cluster', 'replication')
>>>> +---
>>>> +...
>>>> +while box.info.replication[2] == nil do fiber.sleep(0.01) end
>>>> +---
>>>> +...
>>>> +vclock = test_run:get_vclock('default')
>>>> +---
>>>> +...
>>>> +vclock[0] = nil
>>>> +---
>>>> +...
>>>> +_ = test_run:wait_vclock('replica_auth', vclock)
>>>> +---
>>>> +...
>>>> +test_run:cmd("stop server replica_auth")
>>>> +---
>>>> +- true
>>>> +...
>>>> +test_run:cmd("cleanup server replica_auth")
>>>> +---
>>>> +- true
>>>> +...
>>>> +test_run:cmd("delete server replica_auth")
>>>> +---
>>>> +- true
>>>> +...
>>>> +test_run:cleanup_cluster()
>>>> +---
>>>> +...
>>>> +box.schema.user.drop('cluster')
>>>> +---
>>>> +...
>>>> diff --git a/test/replication/misc_no_panic_on_connected_gh-3637.test.lua b/test/replication/misc_no_panic_on_connected_gh-3637.test.lua
>>>> new file mode 100644
>>>> index 000000000..a1e51198b
>>>> --- /dev/null
>>>> +++ b/test/replication/misc_no_panic_on_connected_gh-3637.test.lua
>>>> @@ -0,0 +1,33 @@
>>>> +test_run = require('test_run').new()
>>>> +test_run:cmd("restart server default")
>>>> +uuid = require('uuid')
>>>> +
>>>> +--
>>>> +-- Test case for gh-3637, gh-4550. Before the fix replica would
>>>> +-- exit with an error if a user does not exist or a password is
>>>> +-- incorrect. Now check that we don't hang/panic and successfully
>>>> +-- connect.
>>>> +--
>>>> +fiber = require('fiber')
>>>> +test_run:cmd("create server replica_auth with rpl_master=default, script='replication/replica_auth.lua'")
>>>> +test_run:cmd("start server replica_auth with wait=False, wait_load=False, args='cluster:pass 0.05'")
>>>> +-- Wait a bit to make sure replica waits till user is created.
>>>> +fiber.sleep(0.1)
>>>> +box.schema.user.create('cluster')
>>>> +-- The user is created. Let the replica fail auth request due to
>>>> +-- a wrong password.
>>>> +fiber.sleep(0.1)
>>>> +box.schema.user.passwd('cluster', 'pass')
>>>> +box.schema.user.grant('cluster', 'replication')
>>>> +
>>>> +while box.info.replication[2] == nil do fiber.sleep(0.01) end
>>>> +vclock = test_run:get_vclock('default')
>>>> +vclock[0] = nil
>>>> +_ = test_run:wait_vclock('replica_auth', vclock)
>>>> +
>>>> +test_run:cmd("stop server replica_auth")
>>>> +test_run:cmd("cleanup server replica_auth")
>>>> +test_run:cmd("delete server replica_auth")
>>>> +test_run:cleanup_cluster()
>>>> +
>>>> +box.schema.user.drop('cluster')
>>>> diff --git a/test/replication/misc_no_restart_on_same_configuration_gh-3711.result b/test/replication/misc_no_restart_on_same_configuration_gh-3711.result
>>>> new file mode 100644
>>>> index 000000000..dd9ef848c
>>>> --- /dev/null
>>>> +++ b/test/replication/misc_no_restart_on_same_configuration_gh-3711.result
>>>> @@ -0,0 +1,107 @@
>>>> +test_run = require('test_run').new()
>>>> +---
>>>> +...
>>>> +test_run:cmd("restart server default")
>>>> +uuid = require('uuid')
>>>> +---
>>>> +...
>>>> +fiber = require('fiber')
>>> Same about restart, uuid and fiber.
>>>> +---
>>>> +...
>>>> +--
>>>> +-- gh-3711 Do not restart replication on box.cfg in case the
>>>> +-- configuration didn't change.
>>>> +--
>>>> +box.schema.user.grant('guest', 'replication')
>>>> +---
>>>> +...
>>>> +test_run:cmd("create server replica with rpl_master=default, script='replication/replica.lua'")
>>>> +---
>>>> +- true
>>>> +...
>>>> +test_run:cmd("start server replica")
>>>> +---
>>>> +- true
>>>> +...
>>>> +-- Access rights are checked only during reconnect. If the new
>>>> +-- and old configurations are equivalent, no reconnect will be
>>>> +-- issued and replication should continue working.
>>>> +box.schema.user.revoke('guest', 'replication')
>>>> +---
>>>> +...
>>>> +test_run:cmd("switch replica")
>>>> +---
>>>> +- true
>>>> +...
>>>> +replication = box.cfg.replication[1]
>>>> +---
>>>> +...
>>>> +box.cfg{replication = {replication}}
>>>> +---
>>>> +...
>>>> +box.info.status == 'running'
>>>> +---
>>>> +- true
>>>> +...
>>>> +box.cfg{replication = replication}
>>>> +---
>>>> +...
>>>> +box.info.status == 'running'
>>>> +---
>>>> +- true
>>>> +...
>>>> +-- Check that comparison of tables works as expected as well.
>>>> +test_run:cmd("switch default")
>>>> +---
>>>> +- true
>>>> +...
>>>> +box.schema.user.grant('guest', 'replication')
>>>> +---
>>>> +...
>>>> +test_run:cmd("switch replica")
>>>> +---
>>>> +- true
>>>> +...
>>>> +replication = box.cfg.replication
>>>> +---
>>>> +...
>>>> +table.insert(replication, box.cfg.listen)
>>>> +---
>>>> +...
>>>> +test_run:cmd("switch default")
>>>> +---
>>>> +- true
>>>> +...
>>>> +box.schema.user.revoke('guest', 'replication')
>>>> +---
>>>> +...
>>>> +test_run:cmd("switch replica")
>>>> +---
>>>> +- true
>>>> +...
>>>> +box.cfg{replication = replication}
>>>> +---
>>>> +...
>>>> +box.info.status == 'running'
>>>> +---
>>>> +- true
>>>> +...
>>>> +test_run:cmd("switch default")
>>>> +---
>>>> +- true
>>>> +...
>>>> +test_run:cmd("stop server replica")
>>>> +---
>>>> +- true
>>>> +...
>>>> +test_run:cmd("cleanup server replica")
>>>> +---
>>>> +- true
>>>> +...
>>>> +test_run:cmd("delete server replica")
>>>> +---
>>>> +- true
>>>> +...
>>>> +test_run:cleanup_cluster()
>>>> +---
>>>> +...
>>>> diff --git a/test/replication/misc_no_restart_on_same_configuration_gh-3711.test.lua b/test/replication/misc_no_restart_on_same_configuration_gh-3711.test.lua
>>>> new file mode 100644
>>>> index 000000000..14b522ead
>>>> --- /dev/null
>>>> +++ b/test/replication/misc_no_restart_on_same_configuration_gh-3711.test.lua
>>>> @@ -0,0 +1,41 @@
>>>> +test_run = require('test_run').new()
>>>> +test_run:cmd("restart server default")
>>>> +uuid = require('uuid')
>>>> +fiber = require('fiber')
>>>> +
>>>> +--
>>>> +-- gh-3711 Do not restart replication on box.cfg in case the
>>>> +-- configuration didn't change.
>>>> +--
>>>> +box.schema.user.grant('guest', 'replication')
>>>> +test_run:cmd("create server replica with rpl_master=default, script='replication/replica.lua'")
>>>> +test_run:cmd("start server replica")
>>>> +
>>>> +-- Access rights are checked only during reconnect. If the new
>>>> +-- and old configurations are equivalent, no reconnect will be
>>>> +-- issued and replication should continue working.
>>>> +box.schema.user.revoke('guest', 'replication')
>>>> +test_run:cmd("switch replica")
>>>> +replication = box.cfg.replication[1]
>>>> +box.cfg{replication = {replication}}
>>>> +box.info.status == 'running'
>>>> +box.cfg{replication = replication}
>>>> +box.info.status == 'running'
>>>> +
>>>> +-- Check that comparison of tables works as expected as well.
>>>> +test_run:cmd("switch default")
>>>> +box.schema.user.grant('guest', 'replication')
>>>> +test_run:cmd("switch replica")
>>>> +replication = box.cfg.replication
>>>> +table.insert(replication, box.cfg.listen)
>>>> +test_run:cmd("switch default")
>>>> +box.schema.user.revoke('guest', 'replication')
>>>> +test_run:cmd("switch replica")
>>>> +box.cfg{replication = replication}
>>>> +box.info.status == 'running'
>>>> +
>>>> +test_run:cmd("switch default")
>>>> +test_run:cmd("stop server replica")
>>>> +test_run:cmd("cleanup server replica")
>>>> +test_run:cmd("delete server replica")
>>>> +test_run:cleanup_cluster()
>>>> diff --git a/test/replication/misc_no_socket_leak_on_replica_disconnect_gh-3642.result b/test/replication/misc_no_socket_leak_on_replica_disconnect_gh-3642.result
>>>> new file mode 100644
>>>> index 000000000..3680bcebb
>>>> --- /dev/null
>>>> +++ b/test/replication/misc_no_socket_leak_on_replica_disconnect_gh-3642.result
>>>> @@ -0,0 +1,98 @@
>>>> +test_run = require('test_run').new()
>>>> +---
>>>> +...
>>>> +test_run:cmd("restart server default")
>>> Unneeded restart. Correct me if I'm wrong.
>>>
>>>> +uuid = require('uuid')
>>> You don't need the uuid module in this testcase.
>>>
>>>> +---
>>>> +...
>>>> +box.schema.user.grant('guest', 'replication')
>>>> +---
>>>> +...
>>>> +-- gh-3642 - Check that socket file descriptor doesn't leak
>>>> +-- when a replica is disconnected.
>>>> +rlimit = require('rlimit')
>>>> +---
>>>> +...
>>>> +lim = rlimit.limit()
>>>> +---
>>>> +...
>>>> +rlimit.getrlimit(rlimit.RLIMIT_NOFILE, lim)
>>>> +---
>>>> +...
>>>> +old_fno = lim.rlim_cur
>>>> +---
>>>> +...
>>>> +lim.rlim_cur = 64
>>>> +---
>>>> +...
>>>> +rlimit.setrlimit(rlimit.RLIMIT_NOFILE, lim)
>>>> +---
>>>> +...
>>>> +test_run:cmd('create server sock with rpl_master=default, script="replication/replica.lua"')
>>>> +---
>>>> +- true
>>>> +...
>>>> +test_run:cmd('start server sock')
>>>> +---
>>>> +- true
>>>> +...
>>>> +test_run:cmd('switch sock')
>>>> +---
>>>> +- true
>>>> +...
>>>> +test_run = require('test_run').new()
>>>> +---
>>>> +...
>>>> +fiber = require('fiber')
>>>> +---
>>>> +...
>>>> +test_run:cmd("setopt delimiter ';'")
>>>> +---
>>>> +- true
>>>> +...
>>>> +for i = 1, 64 do
>>>> +    local replication = box.cfg.replication
>>>> +    box.cfg{replication = {}}
>>>> +    box.cfg{replication = replication}
>>>> +    while box.info.replication[1].upstream.status ~= 'follow' do
>>>> +        fiber.sleep(0.001)
>>>> +    end
>>>> +end;
>>>> +---
>>>> +...
>>>> +test_run:cmd("setopt delimiter ''");
>>>> +---
>>>> +- true
>>>> +...
>>>> +box.info.replication[1].upstream.status
>>>> +---
>>>> +- follow
>>>> +...
>>>> +test_run:cmd('switch default')
>>>> +---
>>>> +- true
>>>> +...
>>>> +lim.rlim_cur = old_fno
>>>> +---
>>>> +...
>>>> +rlimit.setrlimit(rlimit.RLIMIT_NOFILE, lim)
>>>> +---
>>>> +...
>>>> +test_run:cmd("stop server sock")
>>>> +---
>>>> +- true
>>>> +...
>>>> +test_run:cmd("cleanup server sock")
>>>> +---
>>>> +- true
>>>> +...
>>>> +test_run:cmd("delete server sock")
>>>> +---
>>>> +- true
>>>> +...
>>>> +test_run:cleanup_cluster()
>>>> +---
>>>> +...
>>>> +box.schema.user.revoke('guest', 'replication')
>>>> +---
>>>> +...
>>>> diff --git a/test/replication/misc_no_socket_leak_on_replica_disconnect_gh-3642.test.lua b/test/replication/misc_no_socket_leak_on_replica_disconnect_gh-3642.test.lua
>>>> new file mode 100644
>>>> index 000000000..08ef9ec0d
>>>> --- /dev/null
>>>> +++ b/test/replication/misc_no_socket_leak_on_replica_disconnect_gh-3642.test.lua
>>>> @@ -0,0 +1,44 @@
>>>> +test_run = require('test_run').new()
>>>> +test_run:cmd("restart server default")
>>>> +uuid = require('uuid')
>>>> +
>>>> +box.schema.user.grant('guest', 'replication')
>>>> +
>>>> +-- gh-3642 - Check that socket file descriptor doesn't leak
>>>> +-- when a replica is disconnected.
>>>> +rlimit = require('rlimit')
>>>> +lim = rlimit.limit()
>>>> +rlimit.getrlimit(rlimit.RLIMIT_NOFILE, lim)
>>>> +old_fno = lim.rlim_cur
>>>> +lim.rlim_cur = 64
>>>> +rlimit.setrlimit(rlimit.RLIMIT_NOFILE, lim)
>>>> +
>>>> +test_run:cmd('create server sock with rpl_master=default, script="replication/replica.lua"')
>>>> +test_run:cmd('start server sock')
>>>> +test_run:cmd('switch sock')
>>>> +test_run = require('test_run').new()
>>>> +fiber = require('fiber')
>>>> +test_run:cmd("setopt delimiter ';'")
>>>> +for i = 1, 64 do
>>>> +    local replication = box.cfg.replication
>>>> +    box.cfg{replication = {}}
>>>> +    box.cfg{replication = replication}
>>>> +    while box.info.replication[1].upstream.status ~= 'follow' do
>>>> +        fiber.sleep(0.001)
>>>> +    end
>>>> +end;
>>>> +test_run:cmd("setopt delimiter ''");
>>>> +
>>>> +box.info.replication[1].upstream.status
>>>> +
>>>> +test_run:cmd('switch default')
>>>> +
>>>> +lim.rlim_cur = old_fno
>>>> +rlimit.setrlimit(rlimit.RLIMIT_NOFILE, lim)
>>>> +
>>>> +test_run:cmd("stop server sock")
>>>> +test_run:cmd("cleanup server sock")
>>>> +test_run:cmd("delete server sock")
>>>> +test_run:cleanup_cluster()
>>>> +
>>>> +box.schema.user.revoke('guest', 'replication')
>>>> diff --git a/test/replication/misc_orphan_on_reconfiguration_error_gh-4424.result b/test/replication/misc_orphan_on_reconfiguration_error_gh-4424.result
>>>> new file mode 100644
>>>> index 000000000..6a9d1148d
>>>> --- /dev/null
>>>> +++ b/test/replication/misc_orphan_on_reconfiguration_error_gh-4424.result
>>>> @@ -0,0 +1,88 @@
>>>> +uuid = require('uuid')
>>>> +---
>>>> +...
>>>> +test_run = require('test_run').new()
>>>> +---
>>>> +...
>>>> +fiber = require('fiber')
>>> UUID and fiber aren't used in this testcase
>>>
>>>> +---
>>>> +...
>>>> +--
>>>> +-- gh-4424 Always enter orphan mode on error in replication
>>>> +-- configuration change.
>>>> +--
>>>> +replication_connect_timeout = box.cfg.replication_connect_timeout
>>>> +---
>>>> +...
>>>> +replication_connect_quorum = box.cfg.replication_connect_quorum
>>>> +---
>>>> +...
>>>> +box.cfg{replication="12345", replication_connect_timeout=0.1, replication_connect_quorum=1}
>>>> +---
>>>> +...
>>>> +box.info.status
>>>> +---
>>>> +- orphan
>>>> +...
>>>> +box.info.ro
>>>> +---
>>>> +- true
>>>> +...
>>>> +-- reset replication => leave orphan mode
>>>> +box.cfg{replication=""}
>>>> +---
>>>> +...
>>>> +box.info.status
>>>> +---
>>>> +- running
>>>> +...
>>>> +box.info.ro
>>>> +---
>>>> +- false
>>>> +...
>>>> +-- no switch to orphan when quorum == 0
>>>> +box.cfg{replication="12345", replication_connect_quorum=0}
>>>> +---
>>>> +...
>>>> +box.info.status
>>>> +---
>>>> +- running
>>>> +...
>>>> +box.info.ro
>>>> +---
>>>> +- false
>>>> +...
>>>> +-- we could connect to one out of two replicas. Set orphan.
>>>> +box.cfg{replication_connect_quorum=2}
>>>> +---
>>>> +...
>>>> +box.cfg{replication={box.cfg.listen, "12345"}}
>>>> +---
>>>> +...
>>>> +box.info.status
>>>> +---
>>>> +- orphan
>>>> +...
>>>> +box.info.ro
>>>> +---
>>>> +- true
>>>> +...
>>>> +-- lower quorum => leave orphan mode
>>>> +box.cfg{replication_connect_quorum=1}
>>>> +---
>>>> +...
>>>> +box.info.status
>>>> +---
>>>> +- running
>>>> +...
>>>> +box.info.ro
>>>> +---
>>>> +- false
>>>> +...
>>>> +box.cfg{                                                        \
>>>> +    replication = {},                                           \
>>>> +    replication_connect_quorum = replication_connect_quorum,    \
>>>> +    replication_connect_timeout = replication_connect_timeout   \
>>>> +}
>>>> +---
>>>> +...
>>>> diff --git a/test/replication/misc_orphan_on_reconfiguration_error_gh-4424.test.lua b/test/replication/misc_orphan_on_reconfiguration_error_gh-4424.test.lua
>>>> new file mode 100644
>>>> index 000000000..7d1a70a36
>>>> --- /dev/null
>>>> +++ b/test/replication/misc_orphan_on_reconfiguration_error_gh-4424.test.lua
>>>> @@ -0,0 +1,37 @@
>>>> +uuid = require('uuid')
>>>> +test_run = require('test_run').new()
>>>> +fiber = require('fiber')
>>>> +
>>>> +--
>>>> +-- gh-4424 Always enter orphan mode on error in replication
>>>> +-- configuration change.
>>>> +--
>>>> +replication_connect_timeout = box.cfg.replication_connect_timeout
>>>> +replication_connect_quorum = box.cfg.replication_connect_quorum
>>>> +box.cfg{replication="12345", replication_connect_timeout=0.1, replication_connect_quorum=1}
>>>> +box.info.status
>>>> +box.info.ro
>>>> +-- reset replication => leave orphan mode
>>>> +box.cfg{replication=""}
>>>> +box.info.status
>>>> +box.info.ro
>>>> +-- no switch to orphan when quorum == 0
>>>> +box.cfg{replication="12345", replication_connect_quorum=0}
>>>> +box.info.status
>>>> +box.info.ro
>>>> +
>>>> +-- we could connect to one out of two replicas. Set orphan.
>>>> +box.cfg{replication_connect_quorum=2}
>>>> +box.cfg{replication={box.cfg.listen, "12345"}}
>>>> +box.info.status
>>>> +box.info.ro
>>>> +-- lower quorum => leave orphan mode
>>>> +box.cfg{replication_connect_quorum=1}
>>>> +box.info.status
>>>> +box.info.ro
>>>> +
>>>> +box.cfg{                                                        \
>>>> +    replication = {},                                           \
>>>> +    replication_connect_quorum = replication_connect_quorum,    \
>>>> +    replication_connect_timeout = replication_connect_timeout   \
>>>> +}
>>>> diff --git a/test/replication/misc_rebootstrap_from_ro_master_gh-3111.result b/test/replication/misc_rebootstrap_from_ro_master_gh-3111.result
>>>> new file mode 100644
>>>> index 000000000..7ffca1585
>>>> --- /dev/null
>>>> +++ b/test/replication/misc_rebootstrap_from_ro_master_gh-3111.result
>>>> @@ -0,0 +1,58 @@
>>>> +test_run = require('test_run').new()
>>>> +---
>>>> +...
>>>> +test_run:cmd("restart server default")
>>> Why do you need it? Seems like the test passes just fine without the
>>> restart.
>>> Restarting the server will take some time if this is executed after some
>>> test that
>>> performs lots of DML operations.So You better remove the change
>>>
>>>> +uuid = require('uuid')
>>>> +---
>>>> +...
>>>> +box.schema.user.grant('guest', 'replication')
>>>> +---
>>>> +...
>>>> +-- gh-3111 - Allow to rebootstrap a replica from a read-only master
>>>> +replica_uuid = uuid.new()
>>>> +---
>>>> +...
>>>> +test_run:cmd('create server test with rpl_master=default, script="replication/replica_uuid.lua"')
>>>> +---
>>>> +- true
>>>> +...
>>>> +test_run:cmd(string.format('start server test with args="%s"', replica_uuid))
>>>> +---
>>>> +- true
>>>> +...
>>>> +test_run:cmd('stop server test')
>>>> +---
>>>> +- true
>>>> +...
>>>> +test_run:cmd('cleanup server test')
>>>> +---
>>>> +- true
>>>> +...
>>>> +box.cfg{read_only = true}
>>>> +---
>>>> +...
>>>> +test_run:cmd(string.format('start server test with args="%s"', replica_uuid))
>>>> +---
>>>> +- true
>>>> +...
>>>> +test_run:cmd('stop server test')
>>>> +---
>>>> +- true
>>>> +...
>>>> +test_run:cmd('cleanup server test')
>>>> +---
>>>> +- true
>>>> +...
>>>> +box.cfg{read_only = false}
>>>> +---
>>>> +...
>>>> +test_run:cmd('delete server test')
>>>> +---
>>>> +- true
>>>> +...
>>>> +test_run:cleanup_cluster()
>>>> +---
>>>> +...
>>>> +box.schema.user.revoke('guest', 'replication')
>>>> +---
>>>> +...
>>>> diff --git a/test/replication/misc_rebootstrap_from_ro_master_gh-3111.test.lua b/test/replication/misc_rebootstrap_from_ro_master_gh-3111.test.lua
>>>> new file mode 100644
>>>> index 000000000..bb9b4a80f
>>>> --- /dev/null
>>>> +++ b/test/replication/misc_rebootstrap_from_ro_master_gh-3111.test.lua
>>>> @@ -0,0 +1,20 @@
>>>> +test_run = require('test_run').new()
>>>> +test_run:cmd("restart server default")
>>>> +uuid = require('uuid')
>>>> +
>>>> +box.schema.user.grant('guest', 'replication')
>>>> +
>>>> +-- gh-3111 - Allow to rebootstrap a replica from a read-only master
>>>> +replica_uuid = uuid.new()
>>>> +test_run:cmd('create server test with rpl_master=default, script="replication/replica_uuid.lua"')
>>>> +test_run:cmd(string.format('start server test with args="%s"', replica_uuid))
>>>> +test_run:cmd('stop server test')
>>>> +test_run:cmd('cleanup server test')
>>>> +box.cfg{read_only = true}
>>>> +test_run:cmd(string.format('start server test with args="%s"', replica_uuid))
>>>> +test_run:cmd('stop server test')
>>>> +test_run:cmd('cleanup server test')
>>>> +box.cfg{read_only = false}
>>>> +test_run:cmd('delete server test')
>>>> +test_run:cleanup_cluster()
>>>> +box.schema.user.revoke('guest', 'replication')
>>>> diff --git a/test/replication/misc_replica_checks_cluster_id_gh-3704.result b/test/replication/misc_replica_checks_cluster_id_gh-3704.result
>>>> new file mode 100644
>>>> index 000000000..e6bc8b4d8
>>>> --- /dev/null
>>>> +++ b/test/replication/misc_replica_checks_cluster_id_gh-3704.result
>>>> @@ -0,0 +1,71 @@
>>>> +test_run = require('test_run').new()
>>>> +---
>>>> +...
>>>> +test_run:cmd("restart server default")
>>>> +uuid = require('uuid')
>>>> +---
>>>> +...
>>>> +fiber = require('fiber')
>>> No need to restart the instance, and require fiber module.
>>>
>>> UUID is needed this time.
>>>
>>>> +---
>>>> +...
>>>> +--
>>>> +-- gh-3704 move cluster id check to replica
>>>> +--
>>>> +test_run:cmd("create server replica with rpl_master=default, script='replication/replica.lua'")
>>>> +---
>>>> +- true
>>>> +...
>>>> +box.schema.user.grant("guest", "replication")
>>>> +---
>>>> +...
>>>> +test_run:cmd("start server replica")
>>>> +---
>>>> +- true
>>>> +...
>>>> +test_run:grep_log("replica", "REPLICASET_UUID_MISMATCH")
>>>> +---
>>>> +- null
>>>> +...
>>>> +box.info.replication[2].downstream.status
>>>> +---
>>>> +- follow
>>>> +...
>>>> +-- change master's cluster uuid and check that replica doesn't connect.
>>>> +test_run:cmd("stop server replica")
>>>> +---
>>>> +- true
>>>> +...
>>>> +_ = box.space._schema:replace{'cluster', tostring(uuid.new())}
>>>> +---
>>>> +...
>>>> +-- master believes replica is in cluster, but their cluster UUIDs differ.
>>>> +test_run:cmd("start server replica")
>>>> +---
>>>> +- true
>>>> +...
>>>> +test_run:wait_log("replica", "REPLICASET_UUID_MISMATCH", nil, 1.0)
>>>> +---
>>>> +- REPLICASET_UUID_MISMATCH
>>>> +...
>>>> +test_run:wait_downstream(2, {status = 'stopped'})
>>>> +---
>>>> +- true
>>>> +...
>>>> +test_run:cmd("stop server replica")
>>>> +---
>>>> +- true
>>>> +...
>>>> +test_run:cmd("cleanup server replica")
>>>> +---
>>>> +- true
>>>> +...
>>>> +test_run:cmd("delete server replica")
>>>> +---
>>>> +- true
>>>> +...
>>>> +test_run:cleanup_cluster()
>>>> +---
>>>> +...
>>>> +box.schema.user.revoke('guest', 'replication')
>>>> +---
>>>> +...
>>>> diff --git a/test/replication/misc_replica_checks_cluster_id_gh-3704.test.lua b/test/replication/misc_replica_checks_cluster_id_gh-3704.test.lua
>>>> new file mode 100644
>>>> index 000000000..8a23cc1fe
>>>> --- /dev/null
>>>> +++ b/test/replication/misc_replica_checks_cluster_id_gh-3704.test.lua
>>>> @@ -0,0 +1,26 @@
>>>> +test_run = require('test_run').new()
>>>> +test_run:cmd("restart server default")
>>>> +uuid = require('uuid')
>>>> +fiber = require('fiber')
>>>> +
>>>> +--
>>>> +-- gh-3704 move cluster id check to replica
>>>> +--
>>>> +test_run:cmd("create server replica with rpl_master=default, script='replication/replica.lua'")
>>>> +box.schema.user.grant("guest", "replication")
>>>> +test_run:cmd("start server replica")
>>>> +test_run:grep_log("replica", "REPLICASET_UUID_MISMATCH")
>>>> +box.info.replication[2].downstream.status
>>>> +-- change master's cluster uuid and check that replica doesn't connect.
>>>> +test_run:cmd("stop server replica")
>>>> +_ = box.space._schema:replace{'cluster', tostring(uuid.new())}
>>>> +-- master believes replica is in cluster, but their cluster UUIDs differ.
>>>> +test_run:cmd("start server replica")
>>>> +test_run:wait_log("replica", "REPLICASET_UUID_MISMATCH", nil, 1.0)
>>>> +test_run:wait_downstream(2, {status = 'stopped'})
>>>> +
>>>> +test_run:cmd("stop server replica")
>>>> +test_run:cmd("cleanup server replica")
>>>> +test_run:cmd("delete server replica")
>>>> +test_run:cleanup_cluster()
>>>> +box.schema.user.revoke('guest', 'replication')
>>>> diff --git a/test/replication/misc_return_on_quorum_0_gh-3760.result b/test/replication/misc_return_on_quorum_0_gh-3760.result
>>>> new file mode 100644
>>>> index 000000000..2eb622896
>>>> --- /dev/null
>>>> +++ b/test/replication/misc_return_on_quorum_0_gh-3760.result
>>>> @@ -0,0 +1,48 @@
>>>> +uuid = require('uuid')
>>>> +---
>>>> +...
>>>> +test_run = require('test_run').new()
>>>> +---
>>>> +...
>>>> +fiber = require('fiber')
>>> Unneeded uuid and fiber
>>>
>>>> +---
>>>> +...
>>>> +replication_connect_timeout = box.cfg.replication_connect_timeout
>>>> +---
>>>> +...
>>>> +replication_connect_quorum = box.cfg.replication_connect_quorum
>>>> +---
>>>> +...
>>>> +box.cfg{replication="12345", replication_connect_timeout=0.1, replication_connect_quorum=1}
>>> This box.cfg call is extraneous. The test'll be just fine without it.
>>>
>>>> +---
>>>> +...
>>>> +--
>>>> +-- gh-3760: replication quorum 0 on reconfiguration should return
>>>> +-- from box.cfg immediately.
>>>> +--
>>>> +replication = box.cfg.replication
>>>> +---
>>>> +...
>>>> +box.cfg{                                                        \
>>>> +    replication = {},                                           \
>>>> +    replication_connect_quorum = 0,                             \
>>>> +    replication_connect_timeout = 1000000                       \
>>>> +}
>>>> +---
>>>> +...
>>>> +-- The call below would hang, if quorum 0 is ignored, or checked
>>>> +-- too late.
>>>> +box.cfg{replication = {'localhost:12345'}}
>>>> +---
>>>> +...
>>>> +box.info.status
>>>> +---
>>>> +- running
>>>> +...
>>>> +box.cfg{                                                        \
>>>> +    replication = {},                                           \
>>>> +    replication_connect_quorum = replication_connect_quorum,    \
>>>> +    replication_connect_timeout = replication_connect_timeout   \
>>>> +}
>>>> +---
>>>> +...
>>>> diff --git a/test/replication/misc_return_on_quorum_0_gh-3760.test.lua b/test/replication/misc_return_on_quorum_0_gh-3760.test.lua
>>>> new file mode 100644
>>>> index 000000000..9e0651032
>>>> --- /dev/null
>>>> +++ b/test/replication/misc_return_on_quorum_0_gh-3760.test.lua
>>>> @@ -0,0 +1,27 @@
>>>> +uuid = require('uuid')
>>>> +test_run = require('test_run').new()
>>>> +fiber = require('fiber')
>>>> +
>>>> +replication_connect_timeout = box.cfg.replication_connect_timeout
>>>> +replication_connect_quorum = box.cfg.replication_connect_quorum
>>>> +box.cfg{replication="12345", replication_connect_timeout=0.1, replication_connect_quorum=1}
>>>> +
>>>> +--
>>>> +-- gh-3760: replication quorum 0 on reconfiguration should return
>>>> +-- from box.cfg immediately.
>>>> +--
>>>> +replication = box.cfg.replication
>>>> +box.cfg{                                                        \
>>>> +    replication = {},                                           \
>>>> +    replication_connect_quorum = 0,                             \
>>>> +    replication_connect_timeout = 1000000                       \
>>>> +}
>>>> +-- The call below would hang, if quorum 0 is ignored, or checked
>>>> +-- too late.
>>>> +box.cfg{replication = {'localhost:12345'}}
>>>> +box.info.status
>>>> +box.cfg{                                                        \
>>>> +    replication = {},                                           \
>>>> +    replication_connect_quorum = replication_connect_quorum,    \
>>>> +    replication_connect_timeout = replication_connect_timeout   \
>>>> +}
>>>> diff --git a/test/replication/misc_value_not_replicated_on_iproto_request_gh-3247.result b/test/replication/misc_value_not_replicated_on_iproto_request_gh-3247.result
>>>> new file mode 100644
>>>> index 000000000..d416bd9a6
>>>> --- /dev/null
>>>> +++ b/test/replication/misc_value_not_replicated_on_iproto_request_gh-3247.result
>>>> @@ -0,0 +1,90 @@
>>>> +test_run = require('test_run').new()
>>>> +---
>>>> +...
>>>> +test_run:cmd("restart server default")
>>>> +uuid = require('uuid')
>>>> +---
>>>> +...
>>> You don't need  the uuid module in this testcase.
>>>
>>>
>>>> +-- Deploy a cluster.
>>>> +SERVERS = { 'autobootstrap1', 'autobootstrap2', 'autobootstrap3' }
>>>> +---
>>>> +...
>>>> +test_run:create_cluster(SERVERS, "replication", {args="0.03"})
>>>> +---
>>>> +...
>>>> +test_run:wait_fullmesh(SERVERS)
>>>> +---
>>>> +...
>>>> +-- gh-3247 - Sequence-generated value is not replicated in case
>>>> +-- the request was sent via iproto.
>>>> +test_run:cmd("switch autobootstrap1")
>>>> +---
>>>> +- true
>>>> +...
>>>> +net_box = require('net.box')
>>>> +---
>>>> +...
>>>> +_ = box.schema.space.create('space1')
>>>> +---
>>>> +...
>>>> +_ = box.schema.sequence.create('seq')
>>>> +---
>>>> +...
>>>> +_ = box.space.space1:create_index('primary', {sequence = true} )
>>>> +---
>>>> +...
>>>> +_ = box.space.space1:create_index('secondary', {parts = {2, 'unsigned'}})
>>>> +---
>>>> +...
>>>> +box.schema.user.grant('guest', 'read,write', 'space', 'space1')
>>>> +---
>>>> +...
>>>> +c = net_box.connect(box.cfg.listen)
>>>> +---
>>>> +...
>>>> +c.space.space1:insert{box.NULL, "data"} -- fails, but bumps sequence value
>>>> +---
>>>> +- error: 'Tuple field 2 type does not match one required by operation: expected unsigned'
>>>> +...
>>>> +c.space.space1:insert{box.NULL, 1, "data"}
>>>> +---
>>>> +- [2, 1, 'data']
>>>> +...
>>>> +box.space.space1:select{}
>>>> +---
>>>> +- - [2, 1, 'data']
>>>> +...
>>>> +vclock = test_run:get_vclock("autobootstrap1")
>>>> +---
>>>> +...
>>>> +vclock[0] = nil
>>>> +---
>>>> +...
>>>> +_ = test_run:wait_vclock("autobootstrap2", vclock)
>>>> +---
>>>> +...
>>>> +test_run:cmd("switch autobootstrap2")
>>>> +---
>>>> +- true
>>>> +...
>>>> +box.space.space1:select{}
>>>> +---
>>>> +- - [2, 1, 'data']
>>>> +...
>>>> +test_run:cmd("switch autobootstrap1")
>>>> +---
>>>> +- true
>>>> +...
>>>> +box.space.space1:drop()
>>>> +---
>>>> +...
>>>> +test_run:cmd("switch default")
>>>> +---
>>>> +- true
>>>> +...
>>>> +test_run:drop_cluster(SERVERS)
>>>> +---
>>>> +...
>>>> +test_run:cleanup_cluster()
>>>> +---
>>>> +...
>>>> diff --git a/test/replication/misc_value_not_replicated_on_iproto_request_gh-3247.test.lua b/test/replication/misc_value_not_replicated_on_iproto_request_gh-3247.test.lua
>>>> new file mode 100644
>>>> index 000000000..586e8b997
>>>> --- /dev/null
>>>> +++ b/test/replication/misc_value_not_replicated_on_iproto_request_gh-3247.test.lua
>>>> @@ -0,0 +1,33 @@
>>>> +test_run = require('test_run').new()
>>>> +test_run:cmd("restart server default")
>>>> +uuid = require('uuid')
>>>> +
>>>> +-- Deploy a cluster.
>>>> +SERVERS = { 'autobootstrap1', 'autobootstrap2', 'autobootstrap3' }
>>>> +test_run:create_cluster(SERVERS, "replication", {args="0.03"})
>>>> +test_run:wait_fullmesh(SERVERS)
>>>> +
>>>> +-- gh-3247 - Sequence-generated value is not replicated in case
>>>> +-- the request was sent via iproto.
>>>> +test_run:cmd("switch autobootstrap1")
>>>> +net_box = require('net.box')
>>>> +_ = box.schema.space.create('space1')
>>>> +_ = box.schema.sequence.create('seq')
>>>> +_ = box.space.space1:create_index('primary', {sequence = true} )
>>>> +_ = box.space.space1:create_index('secondary', {parts = {2, 'unsigned'}})
>>>> +box.schema.user.grant('guest', 'read,write', 'space', 'space1')
>>>> +c = net_box.connect(box.cfg.listen)
>>>> +c.space.space1:insert{box.NULL, "data"} -- fails, but bumps sequence value
>>>> +c.space.space1:insert{box.NULL, 1, "data"}
>>>> +box.space.space1:select{}
>>>> +vclock = test_run:get_vclock("autobootstrap1")
>>>> +vclock[0] = nil
>>>> +_ = test_run:wait_vclock("autobootstrap2", vclock)
>>>> +test_run:cmd("switch autobootstrap2")
>>>> +box.space.space1:select{}
>>>> +test_run:cmd("switch autobootstrap1")
>>>> +box.space.space1:drop()
>>>> +
>>>> +test_run:cmd("switch default")
>>>> +test_run:drop_cluster(SERVERS)
>>>> +test_run:cleanup_cluster()
>>>> diff --git a/test/replication/suite.cfg b/test/replication/suite.cfg
>>>> index f357b07da..e21daa5ad 100644
>>>> --- a/test/replication/suite.cfg
>>>> +++ b/test/replication/suite.cfg
>>>> @@ -1,6 +1,19 @@
>>>>     {
>>>>         "anon.test.lua": {},
>>>> -    "misc.test.lua": {},
>>>> +    "misc_assert_connecting_master_twice_gh-3610.test.lua": {},
>>>> +    "misc_assert_on_server_die_gh-2991.test.lua": {},
>>>> +    "misc_assert_replica_on_applier_disconnect_gh-3510.test.lua": {},
>>>> +    "misc_crash_on_box_concurrent_update_gh-3606.test.lua": {},
>>>> +    "misc_heartbeats_on_master_changes_gh-3160.test.lua": {},
>>>> +    "misc_no_failure_on_error_reading_wal_gh-4399.test.lua": {},
>>>> +    "misc_no_panic_on_connected_gh-3637.test.lua": {},
>>>> +    "misc_no_restart_on_same_configuration_gh-3711.test.lua": {},
>>>> +    "misc_no_socket_leak_on_replica_disconnect_gh-3642.test.lua": {},
>>>> +    "misc_orphan_on_reconfiguration_error_gh-4424.test.lua": {},
>>>> +    "misc_rebootstrap_from_ro_master_gh-3111.test.lua": {},
>>>> +    "misc_replica_checks_cluster_id_gh-3704.test.lua": {},
>>>> +    "misc_return_on_quorum_0_gh-3760.test.lua": {},
>>>> +    "misc_value_not_replicated_on_iproto_request_gh-3247.test.lua": {},
>>>>         "once.test.lua": {},
>>>>         "on_replace.test.lua": {},
>>>>         "status.test.lua": {},
>>> -- 
>>> Serge Petrenko
>>>

       reply	other threads:[~2020-09-03  9:32 UTC|newest]

Thread overview: 2+ messages / expand[flat|nested]  mbox.gz  Atom feed  top
     [not found] <20200903075417.GA24428@hpalx>
2020-09-03  9:32 ` Sergey Bronnikov [this message]
2020-09-03 11:30   ` Serge Petrenko

Reply instructions:

You may reply publicly to this message via plain-text email
using any one of the following methods:

* Save the following mbox file, import it into your mail client,
  and reply-to-all from there: mbox

  Avoid top-posting and favor interleaved quoting:
  https://en.wikipedia.org/wiki/Posting_style#Interleaved_style

* Reply using the --to, --cc, and --in-reply-to
  switches of git-send-email(1):

  git send-email \
    --in-reply-to=13c35d18-e188-a68a-417e-8ae605599979@tarantool.org \
    --to=sergeyb@tarantool.org \
    --cc=avtikhon@tarantool.org \
    --cc=tarantool-patches@dev.tarantool.org \
    --subject='Re: [Tarantool-patches] [sergepetrenko@tarantool.org: Re: [PATCH v1] Divide replication/mist.test.lua]' \
    /path/to/YOUR_REPLY

  https://kernel.org/pub/software/scm/git/docs/git-send-email.html

* If your mail client supports setting the In-Reply-To header
  via mailto: links, try the mailto: link

This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox