Tarantool development patches archive
 help / color / mirror / Atom feed
From: "Alexander V. Tikhonov" <avtikhon@tarantool.org>
To: Serge Petrenko <sergepetrenko@tarantool.org>
Cc: Mergen Imeev <imeevma@gmail.com>, tarantool-patches@dev.tarantool.org
Subject: Re: [Tarantool-patches] [PATCH v1] Divide replication/mist.test.lua
Date: Wed, 2 Sep 2020 13:08:37 +0300	[thread overview]
Message-ID: <20200902100837.GA22588@hpalx> (raw)
In-Reply-To: <b9329814-507a-6242-4384-40cfb38b70c5@tarantool.org>

Hi Sergey, thanks for the review. The major comment on tests naming
that it was discussed during fixes in issue [1], and decided to use
this naming format, check as an example of it with:

  ls -al test/box/net.box*.test.lua

[1] - https://github.com/tarantool/tarantool/issues/4880

On Wed, Sep 02, 2020 at 11:09:30AM +0300, Serge Petrenko wrote:
> Hi! Thanks for the patch!
> 
> Please see my comments below.
> 
> 17.07.2020 12:25, Alexander V. Tikhonov пишет:
> > To fix flaky issues of replication/misc.test.lua the test had to be
> > divided into smaller tests to be able to localize the flaky results:
> > 
> >    misc_assert_connecting_master_twice_gh-3610.test.lua
> >    misc_assert_on_server_die_gh-2991.test.lua
> >    misc_assert_replica_on_applier_disconnect_gh-3510.test.lua
> >    misc_crash_on_box_concurrent_update_gh-3606.test.lua
> >    misc_heartbeats_on_master_changes_gh-3160.test.lua
> >    misc_no_failure_on_error_reading_wal_gh-4399.test.lua
> >    misc_no_panic_on_connected_gh-3637.test.lua
> >    misc_no_restart_on_same_configuration_gh-3711.test.lua
> >    misc_no_socket_leak_on_replica_disconnect_gh-3642.test.lua
> >    misc_orphan_on_reconfiguration_error_gh-4424.test.lua
> >    misc_rebootstrap_from_ro_master_gh-3111.test.lua
> >    misc_replica_checks_cluster_id_gh-3704.test.lua
> >    misc_return_on_quorum_0_gh-3760.test.lua
> >    misc_value_not_replicated_on_iproto_request_gh-3247.test.lua
> 
> We use a different naming scheme for regression tests.
> It should be `gh-xxxx-bug-description`, also notice the dashes
> between words instead of the underscores.
> 
> > 
> > Needed for #4940
> > ---
> > 
> > Github: https://github.com/tarantool/tarantool/tree/avtikhon/gh-4940-replication-misc
> > Issue: https://github.com/tarantool/tarantool/issues/4940
> > 
> >   test/replication/misc.result                  | 866 ------------------
> >   test/replication/misc.test.lua                | 356 -------
> >   ...ert_connecting_master_twice_gh-3610.result |  86 ++
> >   ...t_connecting_master_twice_gh-3610.test.lua |  34 +
> >   .../misc_assert_on_server_die_gh-2991.result  |  31 +
> >   ...misc_assert_on_server_die_gh-2991.test.lua |  12 +
> >   ...plica_on_applier_disconnect_gh-3510.result |  49 +
> >   ...ica_on_applier_disconnect_gh-3510.test.lua |  17 +
> >   ...sh_on_box_concurrent_update_gh-3606.result |  50 +
> >   ..._on_box_concurrent_update_gh-3606.test.lua |  19 +
> >   ...eartbeats_on_master_changes_gh-3160.result |  76 ++
> >   ...rtbeats_on_master_changes_gh-3160.test.lua |  40 +
> >   ...ailure_on_error_reading_wal_gh-4399.result |  97 ++
> >   ...lure_on_error_reading_wal_gh-4399.test.lua |  39 +
> >   .../misc_no_panic_on_connected_gh-3637.result |  72 ++
> >   ...isc_no_panic_on_connected_gh-3637.test.lua |  33 +
> >   ...start_on_same_configuration_gh-3711.result | 107 +++
> >   ...art_on_same_configuration_gh-3711.test.lua |  41 +
> >   ..._leak_on_replica_disconnect_gh-3642.result |  98 ++
> >   ...eak_on_replica_disconnect_gh-3642.test.lua |  44 +
> >   ...an_on_reconfiguration_error_gh-4424.result |  88 ++
> >   ..._on_reconfiguration_error_gh-4424.test.lua |  37 +
> >   ..._rebootstrap_from_ro_master_gh-3111.result |  58 ++
> >   ...ebootstrap_from_ro_master_gh-3111.test.lua |  20 +
> >   ...c_replica_checks_cluster_id_gh-3704.result |  71 ++
> >   ...replica_checks_cluster_id_gh-3704.test.lua |  26 +
> >   .../misc_return_on_quorum_0_gh-3760.result    |  48 +
> >   .../misc_return_on_quorum_0_gh-3760.test.lua  |  27 +
> >   ...eplicated_on_iproto_request_gh-3247.result |  90 ++
> >   ...licated_on_iproto_request_gh-3247.test.lua |  33 +
> >   test/replication/suite.cfg                    |  15 +-
> >   31 files changed, 1457 insertions(+), 1223 deletions(-)
> >   delete mode 100644 test/replication/misc.result
> >   delete mode 100644 test/replication/misc.test.lua
> >   create mode 100644 test/replication/misc_assert_connecting_master_twice_gh-3610.result
> >   create mode 100644 test/replication/misc_assert_connecting_master_twice_gh-3610.test.lua
> >   create mode 100644 test/replication/misc_assert_on_server_die_gh-2991.result
> >   create mode 100644 test/replication/misc_assert_on_server_die_gh-2991.test.lua
> >   create mode 100644 test/replication/misc_assert_replica_on_applier_disconnect_gh-3510.result
> >   create mode 100644 test/replication/misc_assert_replica_on_applier_disconnect_gh-3510.test.lua
> >   create mode 100644 test/replication/misc_crash_on_box_concurrent_update_gh-3606.result
> >   create mode 100644 test/replication/misc_crash_on_box_concurrent_update_gh-3606.test.lua
> >   create mode 100644 test/replication/misc_heartbeats_on_master_changes_gh-3160.result
> >   create mode 100644 test/replication/misc_heartbeats_on_master_changes_gh-3160.test.lua
> >   create mode 100644 test/replication/misc_no_failure_on_error_reading_wal_gh-4399.result
> >   create mode 100644 test/replication/misc_no_failure_on_error_reading_wal_gh-4399.test.lua
> >   create mode 100644 test/replication/misc_no_panic_on_connected_gh-3637.result
> >   create mode 100644 test/replication/misc_no_panic_on_connected_gh-3637.test.lua
> >   create mode 100644 test/replication/misc_no_restart_on_same_configuration_gh-3711.result
> >   create mode 100644 test/replication/misc_no_restart_on_same_configuration_gh-3711.test.lua
> >   create mode 100644 test/replication/misc_no_socket_leak_on_replica_disconnect_gh-3642.result
> >   create mode 100644 test/replication/misc_no_socket_leak_on_replica_disconnect_gh-3642.test.lua
> >   create mode 100644 test/replication/misc_orphan_on_reconfiguration_error_gh-4424.result
> >   create mode 100644 test/replication/misc_orphan_on_reconfiguration_error_gh-4424.test.lua
> >   create mode 100644 test/replication/misc_rebootstrap_from_ro_master_gh-3111.result
> >   create mode 100644 test/replication/misc_rebootstrap_from_ro_master_gh-3111.test.lua
> >   create mode 100644 test/replication/misc_replica_checks_cluster_id_gh-3704.result
> >   create mode 100644 test/replication/misc_replica_checks_cluster_id_gh-3704.test.lua
> >   create mode 100644 test/replication/misc_return_on_quorum_0_gh-3760.result
> >   create mode 100644 test/replication/misc_return_on_quorum_0_gh-3760.test.lua
> >   create mode 100644 test/replication/misc_value_not_replicated_on_iproto_request_gh-3247.result
> >   create mode 100644 test/replication/misc_value_not_replicated_on_iproto_request_gh-3247.test.lua
> > 
> > diff --git a/test/replication/misc.result b/test/replication/misc.result
> > deleted file mode 100644
> > index e5d1f560e..000000000
> > --- a/test/replication/misc.result
> > +++ /dev/null
> > @@ -1,866 +0,0 @@
> > -uuid = require('uuid')
> > ----
> > -...
> > -test_run = require('test_run').new()
> > ----
> > -...
> > -box.schema.user.grant('guest', 'replication')
> > ----
> > -...
> > --- gh-2991 - Tarantool asserts on box.cfg.replication update if one of
> > --- servers is dead
> > -replication_timeout = box.cfg.replication_timeout
> > ----
> > -...
> > -replication_connect_timeout = box.cfg.replication_connect_timeout
> > ----
> > -...
> > -box.cfg{replication_timeout=0.05, replication_connect_timeout=0.05, replication={}}
> > ----
> > -...
> > -box.cfg{replication_connect_quorum=2}
> > ----
> > -...
> > -box.cfg{replication = {'127.0.0.1:12345', box.cfg.listen}}
> > ----
> > -...
> > -box.info.status
> > ----
> > -- orphan
> > -...
> > -box.info.ro
> > ----
> > -- true
> > -...
> > --- gh-3606 - Tarantool crashes if box.cfg.replication is updated concurrently
> > -fiber = require('fiber')
> > ----
> > -...
> > -c = fiber.channel(2)
> > ----
> > -...
> > -f = function() fiber.create(function() pcall(box.cfg, {replication = {12345}}) c:put(true) end) end
> > ----
> > -...
> > -f()
> > ----
> > -...
> > -f()
> > ----
> > -...
> > -c:get()
> > ----
> > -- true
> > -...
> > -c:get()
> > ----
> > -- true
> > -...
> > -box.cfg{replication = "", replication_timeout = replication_timeout, replication_connect_timeout = replication_connect_timeout}
> > ----
> > -...
> > -box.info.status
> > ----
> > -- running
> > -...
> > -box.info.ro
> > ----
> > -- false
> > -...
> > --- gh-3111 - Allow to rebootstrap a replica from a read-only master
> > -replica_uuid = uuid.new()
> > ----
> > -...
> > -test_run:cmd('create server test with rpl_master=default, script="replication/replica_uuid.lua"')
> > ----
> > -- true
> > -...
> > -test_run:cmd(string.format('start server test with args="%s"', replica_uuid))
> > ----
> > -- true
> > -...
> > -test_run:cmd('stop server test')
> > ----
> > -- true
> > -...
> > -test_run:cmd('cleanup server test')
> > ----
> > -- true
> > -...
> > -box.cfg{read_only = true}
> > ----
> > -...
> > -test_run:cmd(string.format('start server test with args="%s"', replica_uuid))
> > ----
> > -- true
> > -...
> > -test_run:cmd('stop server test')
> > ----
> > -- true
> > -...
> > -test_run:cmd('cleanup server test')
> > ----
> > -- true
> > -...
> > -box.cfg{read_only = false}
> > ----
> > -...
> > -test_run:cmd('delete server test')
> > ----
> > -- true
> > -...
> > -test_run:cleanup_cluster()
> > ----
> > -...
> > --- gh-3160 - Send heartbeats if there are changes from a remote master only
> > -SERVERS = { 'autobootstrap1', 'autobootstrap2', 'autobootstrap3' }
> > ----
> > -...
> > --- Deploy a cluster.
> > -test_run:create_cluster(SERVERS, "replication", {args="0.03"})
> > ----
> > -...
> > -test_run:wait_fullmesh(SERVERS)
> > ----
> > -...
> > -test_run:cmd("switch autobootstrap3")
> > ----
> > -- true
> > -...
> > -test_run = require('test_run').new()
> > ----
> > -...
> > -fiber = require('fiber')
> > ----
> > -...
> > -_ = box.schema.space.create('test_timeout'):create_index('pk')
> > ----
> > -...
> > -test_run:cmd("setopt delimiter ';'")
> > ----
> > -- true
> > -...
> > -function wait_not_follow(replicaA, replicaB)
> > -    return test_run:wait_cond(function()
> > -        return replicaA.status ~= 'follow' or replicaB.status ~= 'follow'
> > -    end, box.cfg.replication_timeout)
> > -end;
> > ----
> > -...
> > -function test_timeout()
> > -    local replicaA = box.info.replication[1].upstream or box.info.replication[2].upstream
> > -    local replicaB = box.info.replication[3].upstream or box.info.replication[2].upstream
> > -    local follows = test_run:wait_cond(function()
> > -        return replicaA.status == 'follow' or replicaB.status == 'follow'
> > -    end)
> > -    if not follows then error('replicas are not in the follow status') end
> > -    for i = 0, 99 do
> > -        box.space.test_timeout:replace({1})
> > -        if wait_not_follow(replicaA, replicaB) then
> > -            return error(box.info.replication)
> > -        end
> > -    end
> > -    return true
> > -end;
> > ----
> > -...
> > -test_run:cmd("setopt delimiter ''");
> > ----
> > -- true
> > -...
> > -test_timeout()
> > ----
> > -- true
> > -...
> > --- gh-3247 - Sequence-generated value is not replicated in case
> > --- the request was sent via iproto.
> > -test_run:cmd("switch autobootstrap1")
> > ----
> > -- true
> > -...
> > -net_box = require('net.box')
> > ----
> > -...
> > -_ = box.schema.space.create('space1')
> > ----
> > -...
> > -_ = box.schema.sequence.create('seq')
> > ----
> > -...
> > -_ = box.space.space1:create_index('primary', {sequence = true} )
> > ----
> > -...
> > -_ = box.space.space1:create_index('secondary', {parts = {2, 'unsigned'}})
> > ----
> > -...
> > -box.schema.user.grant('guest', 'read,write', 'space', 'space1')
> > ----
> > -...
> > -c = net_box.connect(box.cfg.listen)
> > ----
> > -...
> > -c.space.space1:insert{box.NULL, "data"} -- fails, but bumps sequence value
> > ----
> > -- error: 'Tuple field 2 type does not match one required by operation: expected unsigned'
> > -...
> > -c.space.space1:insert{box.NULL, 1, "data"}
> > ----
> > -- [2, 1, 'data']
> > -...
> > -box.space.space1:select{}
> > ----
> > -- - [2, 1, 'data']
> > -...
> > -vclock = test_run:get_vclock("autobootstrap1")
> > ----
> > -...
> > -vclock[0] = nil
> > ----
> > -...
> > -_ = test_run:wait_vclock("autobootstrap2", vclock)
> > ----
> > -...
> > -test_run:cmd("switch autobootstrap2")
> > ----
> > -- true
> > -...
> > -box.space.space1:select{}
> > ----
> > -- - [2, 1, 'data']
> > -...
> > -test_run:cmd("switch autobootstrap1")
> > ----
> > -- true
> > -...
> > -box.space.space1:drop()
> > ----
> > -...
> > -test_run:cmd("switch default")
> > ----
> > -- true
> > -...
> > -test_run:drop_cluster(SERVERS)
> > ----
> > -...
> > -test_run:cleanup_cluster()
> > ----
> > -...
> > --- gh-3642 - Check that socket file descriptor doesn't leak
> > --- when a replica is disconnected.
> > -rlimit = require('rlimit')
> > ----
> > -...
> > -lim = rlimit.limit()
> > ----
> > -...
> > -rlimit.getrlimit(rlimit.RLIMIT_NOFILE, lim)
> > ----
> > -...
> > -old_fno = lim.rlim_cur
> > ----
> > -...
> > -lim.rlim_cur = 64
> > ----
> > -...
> > -rlimit.setrlimit(rlimit.RLIMIT_NOFILE, lim)
> > ----
> > -...
> > -test_run:cmd('create server sock with rpl_master=default, script="replication/replica.lua"')
> > ----
> > -- true
> > -...
> > -test_run:cmd('start server sock')
> > ----
> > -- true
> > -...
> > -test_run:cmd('switch sock')
> > ----
> > -- true
> > -...
> > -test_run = require('test_run').new()
> > ----
> > -...
> > -fiber = require('fiber')
> > ----
> > -...
> > -test_run:cmd("setopt delimiter ';'")
> > ----
> > -- true
> > -...
> > -for i = 1, 64 do
> > -    local replication = box.cfg.replication
> > -    box.cfg{replication = {}}
> > -    box.cfg{replication = replication}
> > -    while box.info.replication[1].upstream.status ~= 'follow' do
> > -        fiber.sleep(0.001)
> > -    end
> > -end;
> > ----
> > -...
> > -test_run:cmd("setopt delimiter ''");
> > ----
> > -- true
> > -...
> > -box.info.replication[1].upstream.status
> > ----
> > -- follow
> > -...
> > -test_run:cmd('switch default')
> > ----
> > -- true
> > -...
> > -lim.rlim_cur = old_fno
> > ----
> > -...
> > -rlimit.setrlimit(rlimit.RLIMIT_NOFILE, lim)
> > ----
> > -...
> > -test_run:cmd("stop server sock")
> > ----
> > -- true
> > -...
> > -test_run:cmd("cleanup server sock")
> > ----
> > -- true
> > -...
> > -test_run:cmd("delete server sock")
> > ----
> > -- true
> > -...
> > -test_run:cleanup_cluster()
> > ----
> > -...
> > -box.schema.user.revoke('guest', 'replication')
> > ----
> > -...
> > --- gh-3510 assertion failure in replica_on_applier_disconnect()
> > -test_run:cmd('create server er_load1 with script="replication/er_load1.lua"')
> > ----
> > -- true
> > -...
> > -test_run:cmd('create server er_load2 with script="replication/er_load2.lua"')
> > ----
> > -- true
> > -...
> > -test_run:cmd('start server er_load1 with wait=False, wait_load=False')
> > ----
> > -- true
> > -...
> > --- Instance er_load2 will fail with error ER_REPLICASET_UUID_MISMATCH.
> > --- This is OK since we only test here that er_load1 doesn't assert.
> > -test_run:cmd('start server er_load2 with wait=True, wait_load=True, crash_expected = True')
> > ----
> > -- false
> > -...
> > -test_run:cmd('stop server er_load1')
> > ----
> > -- true
> > -...
> > --- er_load2 exits automatically.
> > -test_run:cmd('cleanup server er_load1')
> > ----
> > -- true
> > -...
> > -test_run:cmd('cleanup server er_load2')
> > ----
> > -- true
> > -...
> > -test_run:cmd('delete server er_load1')
> > ----
> > -- true
> > -...
> > -test_run:cmd('delete server er_load2')
> > ----
> > -- true
> > -...
> > -test_run:cleanup_cluster()
> > ----
> > -...
> > ---
> > --- Test case for gh-3637, gh-4550. Before the fix replica would
> > --- exit with an error if a user does not exist or a password is
> > --- incorrect. Now check that we don't hang/panic and successfully
> > --- connect.
> > ---
> > -fiber = require('fiber')
> > ----
> > -...
> > -test_run:cmd("create server replica_auth with rpl_master=default, script='replication/replica_auth.lua'")
> > ----
> > -- true
> > -...
> > -test_run:cmd("start server replica_auth with wait=False, wait_load=False, args='cluster:pass 0.05'")
> > ----
> > -- true
> > -...
> > --- Wait a bit to make sure replica waits till user is created.
> > -fiber.sleep(0.1)
> > ----
> > -...
> > -box.schema.user.create('cluster')
> > ----
> > -...
> > --- The user is created. Let the replica fail auth request due to
> > --- a wrong password.
> > -fiber.sleep(0.1)
> > ----
> > -...
> > -box.schema.user.passwd('cluster', 'pass')
> > ----
> > -...
> > -box.schema.user.grant('cluster', 'replication')
> > ----
> > -...
> > -while box.info.replication[2] == nil do fiber.sleep(0.01) end
> > ----
> > -...
> > -vclock = test_run:get_vclock('default')
> > ----
> > -...
> > -vclock[0] = nil
> > ----
> > -...
> > -_ = test_run:wait_vclock('replica_auth', vclock)
> > ----
> > -...
> > -test_run:cmd("stop server replica_auth")
> > ----
> > -- true
> > -...
> > -test_run:cmd("cleanup server replica_auth")
> > ----
> > -- true
> > -...
> > -test_run:cmd("delete server replica_auth")
> > ----
> > -- true
> > -...
> > -test_run:cleanup_cluster()
> > ----
> > -...
> > -box.schema.user.drop('cluster')
> > ----
> > -...
> > ---
> > --- Test case for gh-3610. Before the fix replica would fail with the assertion
> > --- when trying to connect to the same master twice.
> > ---
> > -box.schema.user.grant('guest', 'replication')
> > ----
> > -...
> > -test_run:cmd("create server replica with rpl_master=default, script='replication/replica.lua'")
> > ----
> > -- true
> > -...
> > -test_run:cmd("start server replica")
> > ----
> > -- true
> > -...
> > -test_run:cmd("switch replica")
> > ----
> > -- true
> > -...
> > -replication = box.cfg.replication[1]
> > ----
> > -...
> > -box.cfg{replication = {replication, replication}}
> > ----
> > -- error: 'Incorrect value for option ''replication'': duplicate connection to the
> > -    same replica'
> > -...
> > --- Check the case when duplicate connection is detected in the background.
> > -test_run:cmd("switch default")
> > ----
> > -- true
> > -...
> > -listen = box.cfg.listen
> > ----
> > -...
> > -box.cfg{listen = ''}
> > ----
> > -...
> > -test_run:cmd("switch replica")
> > ----
> > -- true
> > -...
> > -box.cfg{replication_connect_quorum = 0, replication_connect_timeout = 0.01}
> > ----
> > -...
> > -box.cfg{replication = {replication, replication}}
> > ----
> > -...
> > -test_run:cmd("switch default")
> > ----
> > -- true
> > -...
> > -box.cfg{listen = listen}
> > ----
> > -...
> > -while test_run:grep_log('replica', 'duplicate connection') == nil do fiber.sleep(0.01) end
> > ----
> > -...
> > -test_run:cmd("stop server replica")
> > ----
> > -- true
> > -...
> > -test_run:cmd("cleanup server replica")
> > ----
> > -- true
> > -...
> > -test_run:cmd("delete server replica")
> > ----
> > -- true
> > -...
> > -test_run:cleanup_cluster()
> > ----
> > -...
> > -box.schema.user.revoke('guest', 'replication')
> > ----
> > -...
> > ---
> > --- gh-3711 Do not restart replication on box.cfg in case the
> > --- configuration didn't change.
> > ---
> > -box.schema.user.grant('guest', 'replication')
> > ----
> > -...
> > -test_run:cmd("create server replica with rpl_master=default, script='replication/replica.lua'")
> > ----
> > -- true
> > -...
> > -test_run:cmd("start server replica")
> > ----
> > -- true
> > -...
> > --- Access rights are checked only during reconnect. If the new
> > --- and old configurations are equivalent, no reconnect will be
> > --- issued and replication should continue working.
> > -box.schema.user.revoke('guest', 'replication')
> > ----
> > -...
> > -test_run:cmd("switch replica")
> > ----
> > -- true
> > -...
> > -replication = box.cfg.replication[1]
> > ----
> > -...
> > -box.cfg{replication = {replication}}
> > ----
> > -...
> > -box.info.status == 'running'
> > ----
> > -- true
> > -...
> > -box.cfg{replication = replication}
> > ----
> > -...
> > -box.info.status == 'running'
> > ----
> > -- true
> > -...
> > --- Check that comparison of tables works as expected as well.
> > -test_run:cmd("switch default")
> > ----
> > -- true
> > -...
> > -box.schema.user.grant('guest', 'replication')
> > ----
> > -...
> > -test_run:cmd("switch replica")
> > ----
> > -- true
> > -...
> > -replication = box.cfg.replication
> > ----
> > -...
> > -table.insert(replication, box.cfg.listen)
> > ----
> > -...
> > -test_run:cmd("switch default")
> > ----
> > -- true
> > -...
> > -box.schema.user.revoke('guest', 'replication')
> > ----
> > -...
> > -test_run:cmd("switch replica")
> > ----
> > -- true
> > -...
> > -box.cfg{replication = replication}
> > ----
> > -...
> > -box.info.status == 'running'
> > ----
> > -- true
> > -...
> > -test_run:cmd("switch default")
> > ----
> > -- true
> > -...
> > -test_run:cmd("stop server replica")
> > ----
> > -- true
> > -...
> > -test_run:cmd("cleanup server replica")
> > ----
> > -- true
> > -...
> > -test_run:cmd("delete server replica")
> > ----
> > -- true
> > -...
> > -test_run:cleanup_cluster()
> > ----
> > -...
> > ---
> > --- gh-3704 move cluster id check to replica
> > ---
> > -test_run:cmd("create server replica with rpl_master=default, script='replication/replica.lua'")
> > ----
> > -- true
> > -...
> > -box.schema.user.grant("guest", "replication")
> > ----
> > -...
> > -test_run:cmd("start server replica")
> > ----
> > -- true
> > -...
> > -test_run:grep_log("replica", "REPLICASET_UUID_MISMATCH")
> > ----
> > -- null
> > -...
> > -box.info.replication[2].downstream.status
> > ----
> > -- follow
> > -...
> > --- change master's cluster uuid and check that replica doesn't connect.
> > -test_run:cmd("stop server replica")
> > ----
> > -- true
> > -...
> > -_ = box.space._schema:replace{'cluster', tostring(uuid.new())}
> > ----
> > -...
> > --- master believes replica is in cluster, but their cluster UUIDs differ.
> > -test_run:cmd("start server replica")
> > ----
> > -- true
> > -...
> > -test_run:wait_log("replica", "REPLICASET_UUID_MISMATCH", nil, 1.0)
> > ----
> > -- REPLICASET_UUID_MISMATCH
> > -...
> > -test_run:wait_downstream(2, {status = 'stopped'})
> > ----
> > -- true
> > -...
> > -test_run:cmd("stop server replica")
> > ----
> > -- true
> > -...
> > -test_run:cmd("cleanup server replica")
> > ----
> > -- true
> > -...
> > -test_run:cmd("delete server replica")
> > ----
> > -- true
> > -...
> > -test_run:cleanup_cluster()
> > ----
> > -...
> > -box.schema.user.revoke('guest', 'replication')
> > ----
> > -...
> > ---
> > --- gh-4399 Check that an error reading WAL directory on subscribe
> > --- doesn't lead to a permanent replication failure.
> > ---
> > -box.schema.user.grant("guest", "replication")
> > ----
> > -...
> > -test_run:cmd("create server replica with rpl_master=default, script='replication/replica.lua'")
> > ----
> > -- true
> > -...
> > -test_run:cmd("start server replica")
> > ----
> > -- true
> > -...
> > --- Make the WAL directory inaccessible.
> > -fio = require('fio')
> > ----
> > -...
> > -path = fio.abspath(box.cfg.wal_dir)
> > ----
> > -...
> > -fio.chmod(path, 0)
> > ----
> > -- true
> > -...
> > --- Break replication on timeout.
> > -replication_timeout = box.cfg.replication_timeout
> > ----
> > -...
> > -box.cfg{replication_timeout = 9000}
> > ----
> > -...
> > -test_run:cmd("switch replica")
> > ----
> > -- true
> > -...
> > -test_run:wait_cond(function() return box.info.replication[1].upstream.status ~= 'follow' end)
> > ----
> > -- true
> > -...
> > -require('fiber').sleep(box.cfg.replication_timeout)
> > ----
> > -...
> > -test_run:cmd("switch default")
> > ----
> > -- true
> > -...
> > -box.cfg{replication_timeout = replication_timeout}
> > ----
> > -...
> > --- Restore access to the WAL directory.
> > --- Wait for replication to be reestablished.
> > -fio.chmod(path, tonumber('777', 8))
> > ----
> > -- true
> > -...
> > -test_run:cmd("switch replica")
> > ----
> > -- true
> > -...
> > -test_run:wait_cond(function() return box.info.replication[1].upstream.status == 'follow' end)
> > ----
> > -- true
> > -...
> > -test_run:cmd("switch default")
> > ----
> > -- true
> > -...
> > -test_run:cmd("stop server replica")
> > ----
> > -- true
> > -...
> > -test_run:cmd("cleanup server replica")
> > ----
> > -- true
> > -...
> > -test_run:cmd("delete server replica")
> > ----
> > -- true
> > -...
> > -test_run:cleanup_cluster()
> > ----
> > -...
> > -box.schema.user.revoke('guest', 'replication')
> > ----
> > -...
> > ---
> > --- gh-4424 Always enter orphan mode on error in replication
> > --- configuration change.
> > ---
> > -replication_connect_timeout = box.cfg.replication_connect_timeout
> > ----
> > -...
> > -replication_connect_quorum = box.cfg.replication_connect_quorum
> > ----
> > -...
> > -box.cfg{replication="12345", replication_connect_timeout=0.1, replication_connect_quorum=1}
> > ----
> > -...
> > -box.info.status
> > ----
> > -- orphan
> > -...
> > -box.info.ro
> > ----
> > -- true
> > -...
> > --- reset replication => leave orphan mode
> > -box.cfg{replication=""}
> > ----
> > -...
> > -box.info.status
> > ----
> > -- running
> > -...
> > -box.info.ro
> > ----
> > -- false
> > -...
> > --- no switch to orphan when quorum == 0
> > -box.cfg{replication="12345", replication_connect_quorum=0}
> > ----
> > -...
> > -box.info.status
> > ----
> > -- running
> > -...
> > -box.info.ro
> > ----
> > -- false
> > -...
> > --- we could connect to one out of two replicas. Set orphan.
> > -box.cfg{replication_connect_quorum=2}
> > ----
> > -...
> > -box.cfg{replication={box.cfg.listen, "12345"}}
> > ----
> > -...
> > -box.info.status
> > ----
> > -- orphan
> > -...
> > -box.info.ro
> > ----
> > -- true
> > -...
> > --- lower quorum => leave orphan mode
> > -box.cfg{replication_connect_quorum=1}
> > ----
> > -...
> > -box.info.status
> > ----
> > -- running
> > -...
> > -box.info.ro
> > ----
> > -- false
> > -...
> > ---
> > --- gh-3760: replication quorum 0 on reconfiguration should return
> > --- from box.cfg immediately.
> > ---
> > -replication = box.cfg.replication
> > ----
> > -...
> > -box.cfg{                                                        \
> > -    replication = {},                                           \
> > -    replication_connect_quorum = 0,                             \
> > -    replication_connect_timeout = 1000000                       \
> > -}
> > ----
> > -...
> > --- The call below would hang, if quorum 0 is ignored, or checked
> > --- too late.
> > -box.cfg{replication = {'localhost:12345'}}
> > ----
> > -...
> > -box.info.status
> > ----
> > -- running
> > -...
> > -box.cfg{                                                        \
> > -    replication = {},                                           \
> > -    replication_connect_quorum = replication_connect_quorum,    \
> > -    replication_connect_timeout = replication_connect_timeout   \
> > -}
> > ----
> > -...
> > diff --git a/test/replication/misc.test.lua b/test/replication/misc.test.lua
> > deleted file mode 100644
> > index d285b014a..000000000
> > --- a/test/replication/misc.test.lua
> > +++ /dev/null
> > @@ -1,356 +0,0 @@
> > -uuid = require('uuid')
> > -test_run = require('test_run').new()
> > -
> > -box.schema.user.grant('guest', 'replication')
> > -
> > --- gh-2991 - Tarantool asserts on box.cfg.replication update if one of
> > --- servers is dead
> > -replication_timeout = box.cfg.replication_timeout
> > -replication_connect_timeout = box.cfg.replication_connect_timeout
> > -box.cfg{replication_timeout=0.05, replication_connect_timeout=0.05, replication={}}
> > -box.cfg{replication_connect_quorum=2}
> > -box.cfg{replication = {'127.0.0.1:12345', box.cfg.listen}}
> > -box.info.status
> > -box.info.ro
> > -
> > --- gh-3606 - Tarantool crashes if box.cfg.replication is updated concurrently
> > -fiber = require('fiber')
> > -c = fiber.channel(2)
> > -f = function() fiber.create(function() pcall(box.cfg, {replication = {12345}}) c:put(true) end) end
> > -f()
> > -f()
> > -c:get()
> > -c:get()
> > -
> > -box.cfg{replication = "", replication_timeout = replication_timeout, replication_connect_timeout = replication_connect_timeout}
> > -box.info.status
> > -box.info.ro
> > -
> > --- gh-3111 - Allow to rebootstrap a replica from a read-only master
> > -replica_uuid = uuid.new()
> > -test_run:cmd('create server test with rpl_master=default, script="replication/replica_uuid.lua"')
> > -test_run:cmd(string.format('start server test with args="%s"', replica_uuid))
> > -test_run:cmd('stop server test')
> > -test_run:cmd('cleanup server test')
> > -box.cfg{read_only = true}
> > -test_run:cmd(string.format('start server test with args="%s"', replica_uuid))
> > -test_run:cmd('stop server test')
> > -test_run:cmd('cleanup server test')
> > -box.cfg{read_only = false}
> > -test_run:cmd('delete server test')
> > -test_run:cleanup_cluster()
> > -
> > --- gh-3160 - Send heartbeats if there are changes from a remote master only
> > -SERVERS = { 'autobootstrap1', 'autobootstrap2', 'autobootstrap3' }
> > -
> > --- Deploy a cluster.
> > -test_run:create_cluster(SERVERS, "replication", {args="0.03"})
> > -test_run:wait_fullmesh(SERVERS)
> > -test_run:cmd("switch autobootstrap3")
> > -test_run = require('test_run').new()
> > -fiber = require('fiber')
> > -_ = box.schema.space.create('test_timeout'):create_index('pk')
> > -test_run:cmd("setopt delimiter ';'")
> > -function wait_not_follow(replicaA, replicaB)
> > -    return test_run:wait_cond(function()
> > -        return replicaA.status ~= 'follow' or replicaB.status ~= 'follow'
> > -    end, box.cfg.replication_timeout)
> > -end;
> > -function test_timeout()
> > -    local replicaA = box.info.replication[1].upstream or box.info.replication[2].upstream
> > -    local replicaB = box.info.replication[3].upstream or box.info.replication[2].upstream
> > -    local follows = test_run:wait_cond(function()
> > -        return replicaA.status == 'follow' or replicaB.status == 'follow'
> > -    end)
> > -    if not follows then error('replicas are not in the follow status') end
> > -    for i = 0, 99 do
> > -        box.space.test_timeout:replace({1})
> > -        if wait_not_follow(replicaA, replicaB) then
> > -            return error(box.info.replication)
> > -        end
> > -    end
> > -    return true
> > -end;
> > -test_run:cmd("setopt delimiter ''");
> > -test_timeout()
> > -
> > --- gh-3247 - Sequence-generated value is not replicated in case
> > --- the request was sent via iproto.
> > -test_run:cmd("switch autobootstrap1")
> > -net_box = require('net.box')
> > -_ = box.schema.space.create('space1')
> > -_ = box.schema.sequence.create('seq')
> > -_ = box.space.space1:create_index('primary', {sequence = true} )
> > -_ = box.space.space1:create_index('secondary', {parts = {2, 'unsigned'}})
> > -box.schema.user.grant('guest', 'read,write', 'space', 'space1')
> > -c = net_box.connect(box.cfg.listen)
> > -c.space.space1:insert{box.NULL, "data"} -- fails, but bumps sequence value
> > -c.space.space1:insert{box.NULL, 1, "data"}
> > -box.space.space1:select{}
> > -vclock = test_run:get_vclock("autobootstrap1")
> > -vclock[0] = nil
> > -_ = test_run:wait_vclock("autobootstrap2", vclock)
> > -test_run:cmd("switch autobootstrap2")
> > -box.space.space1:select{}
> > -test_run:cmd("switch autobootstrap1")
> > -box.space.space1:drop()
> > -
> > -test_run:cmd("switch default")
> > -test_run:drop_cluster(SERVERS)
> > -test_run:cleanup_cluster()
> > -
> > --- gh-3642 - Check that socket file descriptor doesn't leak
> > --- when a replica is disconnected.
> > -rlimit = require('rlimit')
> > -lim = rlimit.limit()
> > -rlimit.getrlimit(rlimit.RLIMIT_NOFILE, lim)
> > -old_fno = lim.rlim_cur
> > -lim.rlim_cur = 64
> > -rlimit.setrlimit(rlimit.RLIMIT_NOFILE, lim)
> > -
> > -test_run:cmd('create server sock with rpl_master=default, script="replication/replica.lua"')
> > -test_run:cmd('start server sock')
> > -test_run:cmd('switch sock')
> > -test_run = require('test_run').new()
> > -fiber = require('fiber')
> > -test_run:cmd("setopt delimiter ';'")
> > -for i = 1, 64 do
> > -    local replication = box.cfg.replication
> > -    box.cfg{replication = {}}
> > -    box.cfg{replication = replication}
> > -    while box.info.replication[1].upstream.status ~= 'follow' do
> > -        fiber.sleep(0.001)
> > -    end
> > -end;
> > -test_run:cmd("setopt delimiter ''");
> > -
> > -box.info.replication[1].upstream.status
> > -
> > -test_run:cmd('switch default')
> > -
> > -lim.rlim_cur = old_fno
> > -rlimit.setrlimit(rlimit.RLIMIT_NOFILE, lim)
> > -
> > -test_run:cmd("stop server sock")
> > -test_run:cmd("cleanup server sock")
> > -test_run:cmd("delete server sock")
> > -test_run:cleanup_cluster()
> > -
> > -box.schema.user.revoke('guest', 'replication')
> > -
> > --- gh-3510 assertion failure in replica_on_applier_disconnect()
> > -test_run:cmd('create server er_load1 with script="replication/er_load1.lua"')
> > -test_run:cmd('create server er_load2 with script="replication/er_load2.lua"')
> > -test_run:cmd('start server er_load1 with wait=False, wait_load=False')
> > --- Instance er_load2 will fail with error ER_REPLICASET_UUID_MISMATCH.
> > --- This is OK since we only test here that er_load1 doesn't assert.
> > -test_run:cmd('start server er_load2 with wait=True, wait_load=True, crash_expected = True')
> > -test_run:cmd('stop server er_load1')
> > --- er_load2 exits automatically.
> > -test_run:cmd('cleanup server er_load1')
> > -test_run:cmd('cleanup server er_load2')
> > -test_run:cmd('delete server er_load1')
> > -test_run:cmd('delete server er_load2')
> > -test_run:cleanup_cluster()
> > -
> > ---
> > --- Test case for gh-3637, gh-4550. Before the fix replica would
> > --- exit with an error if a user does not exist or a password is
> > --- incorrect. Now check that we don't hang/panic and successfully
> > --- connect.
> > ---
> > -fiber = require('fiber')
> > -test_run:cmd("create server replica_auth with rpl_master=default, script='replication/replica_auth.lua'")
> > -test_run:cmd("start server replica_auth with wait=False, wait_load=False, args='cluster:pass 0.05'")
> > --- Wait a bit to make sure replica waits till user is created.
> > -fiber.sleep(0.1)
> > -box.schema.user.create('cluster')
> > --- The user is created. Let the replica fail auth request due to
> > --- a wrong password.
> > -fiber.sleep(0.1)
> > -box.schema.user.passwd('cluster', 'pass')
> > -box.schema.user.grant('cluster', 'replication')
> > -
> > -while box.info.replication[2] == nil do fiber.sleep(0.01) end
> > -vclock = test_run:get_vclock('default')
> > -vclock[0] = nil
> > -_ = test_run:wait_vclock('replica_auth', vclock)
> > -
> > -test_run:cmd("stop server replica_auth")
> > -test_run:cmd("cleanup server replica_auth")
> > -test_run:cmd("delete server replica_auth")
> > -test_run:cleanup_cluster()
> > -
> > -box.schema.user.drop('cluster')
> > -
> > ---
> > --- Test case for gh-3610. Before the fix replica would fail with the assertion
> > --- when trying to connect to the same master twice.
> > ---
> > -box.schema.user.grant('guest', 'replication')
> > -test_run:cmd("create server replica with rpl_master=default, script='replication/replica.lua'")
> > -test_run:cmd("start server replica")
> > -test_run:cmd("switch replica")
> > -replication = box.cfg.replication[1]
> > -box.cfg{replication = {replication, replication}}
> > -
> > --- Check the case when duplicate connection is detected in the background.
> > -test_run:cmd("switch default")
> > -listen = box.cfg.listen
> > -box.cfg{listen = ''}
> > -
> > -test_run:cmd("switch replica")
> > -box.cfg{replication_connect_quorum = 0, replication_connect_timeout = 0.01}
> > -box.cfg{replication = {replication, replication}}
> > -
> > -test_run:cmd("switch default")
> > -box.cfg{listen = listen}
> > -while test_run:grep_log('replica', 'duplicate connection') == nil do fiber.sleep(0.01) end
> > -
> > -test_run:cmd("stop server replica")
> > -test_run:cmd("cleanup server replica")
> > -test_run:cmd("delete server replica")
> > -test_run:cleanup_cluster()
> > -box.schema.user.revoke('guest', 'replication')
> > -
> > ---
> > --- gh-3711 Do not restart replication on box.cfg in case the
> > --- configuration didn't change.
> > ---
> > -box.schema.user.grant('guest', 'replication')
> > -test_run:cmd("create server replica with rpl_master=default, script='replication/replica.lua'")
> > -test_run:cmd("start server replica")
> > -
> > --- Access rights are checked only during reconnect. If the new
> > --- and old configurations are equivalent, no reconnect will be
> > --- issued and replication should continue working.
> > -box.schema.user.revoke('guest', 'replication')
> > -test_run:cmd("switch replica")
> > -replication = box.cfg.replication[1]
> > -box.cfg{replication = {replication}}
> > -box.info.status == 'running'
> > -box.cfg{replication = replication}
> > -box.info.status == 'running'
> > -
> > --- Check that comparison of tables works as expected as well.
> > -test_run:cmd("switch default")
> > -box.schema.user.grant('guest', 'replication')
> > -test_run:cmd("switch replica")
> > -replication = box.cfg.replication
> > -table.insert(replication, box.cfg.listen)
> > -test_run:cmd("switch default")
> > -box.schema.user.revoke('guest', 'replication')
> > -test_run:cmd("switch replica")
> > -box.cfg{replication = replication}
> > -box.info.status == 'running'
> > -
> > -test_run:cmd("switch default")
> > -test_run:cmd("stop server replica")
> > -test_run:cmd("cleanup server replica")
> > -test_run:cmd("delete server replica")
> > -test_run:cleanup_cluster()
> > -
> > ---
> > --- gh-3704 move cluster id check to replica
> > ---
> > -test_run:cmd("create server replica with rpl_master=default, script='replication/replica.lua'")
> > -box.schema.user.grant("guest", "replication")
> > -test_run:cmd("start server replica")
> > -test_run:grep_log("replica", "REPLICASET_UUID_MISMATCH")
> > -box.info.replication[2].downstream.status
> > --- change master's cluster uuid and check that replica doesn't connect.
> > -test_run:cmd("stop server replica")
> > -_ = box.space._schema:replace{'cluster', tostring(uuid.new())}
> > --- master believes replica is in cluster, but their cluster UUIDs differ.
> > -test_run:cmd("start server replica")
> > -test_run:wait_log("replica", "REPLICASET_UUID_MISMATCH", nil, 1.0)
> > -test_run:wait_downstream(2, {status = 'stopped'})
> > -
> > -test_run:cmd("stop server replica")
> > -test_run:cmd("cleanup server replica")
> > -test_run:cmd("delete server replica")
> > -test_run:cleanup_cluster()
> > -box.schema.user.revoke('guest', 'replication')
> > -
> > ---
> > --- gh-4399 Check that an error reading WAL directory on subscribe
> > --- doesn't lead to a permanent replication failure.
> > ---
> > -box.schema.user.grant("guest", "replication")
> > -test_run:cmd("create server replica with rpl_master=default, script='replication/replica.lua'")
> > -test_run:cmd("start server replica")
> > -
> > --- Make the WAL directory inaccessible.
> > -fio = require('fio')
> > -path = fio.abspath(box.cfg.wal_dir)
> > -fio.chmod(path, 0)
> > -
> > --- Break replication on timeout.
> > -replication_timeout = box.cfg.replication_timeout
> > -box.cfg{replication_timeout = 9000}
> > -test_run:cmd("switch replica")
> > -test_run:wait_cond(function() return box.info.replication[1].upstream.status ~= 'follow' end)
> > -require('fiber').sleep(box.cfg.replication_timeout)
> > -test_run:cmd("switch default")
> > -box.cfg{replication_timeout = replication_timeout}
> > -
> > --- Restore access to the WAL directory.
> > --- Wait for replication to be reestablished.
> > -fio.chmod(path, tonumber('777', 8))
> > -test_run:cmd("switch replica")
> > -test_run:wait_cond(function() return box.info.replication[1].upstream.status == 'follow' end)
> > -test_run:cmd("switch default")
> > -
> > -test_run:cmd("stop server replica")
> > -test_run:cmd("cleanup server replica")
> > -test_run:cmd("delete server replica")
> > -test_run:cleanup_cluster()
> > -box.schema.user.revoke('guest', 'replication')
> > -
> > ---
> > --- gh-4424 Always enter orphan mode on error in replication
> > --- configuration change.
> > ---
> > -replication_connect_timeout = box.cfg.replication_connect_timeout
> > -replication_connect_quorum = box.cfg.replication_connect_quorum
> > -box.cfg{replication="12345", replication_connect_timeout=0.1, replication_connect_quorum=1}
> > -box.info.status
> > -box.info.ro
> > --- reset replication => leave orphan mode
> > -box.cfg{replication=""}
> > -box.info.status
> > -box.info.ro
> > --- no switch to orphan when quorum == 0
> > -box.cfg{replication="12345", replication_connect_quorum=0}
> > -box.info.status
> > -box.info.ro
> > -
> > --- we could connect to one out of two replicas. Set orphan.
> > -box.cfg{replication_connect_quorum=2}
> > -box.cfg{replication={box.cfg.listen, "12345"}}
> > -box.info.status
> > -box.info.ro
> > --- lower quorum => leave orphan mode
> > -box.cfg{replication_connect_quorum=1}
> > -box.info.status
> > -box.info.ro
> > -
> > ---
> > --- gh-3760: replication quorum 0 on reconfiguration should return
> > --- from box.cfg immediately.
> > ---
> > -replication = box.cfg.replication
> > -box.cfg{                                                        \
> > -    replication = {},                                           \
> > -    replication_connect_quorum = 0,                             \
> > -    replication_connect_timeout = 1000000                       \
> > -}
> > --- The call below would hang, if quorum 0 is ignored, or checked
> > --- too late.
> > -box.cfg{replication = {'localhost:12345'}}
> > -box.info.status
> > -box.cfg{                                                        \
> > -    replication = {},                                           \
> > -    replication_connect_quorum = replication_connect_quorum,    \
> > -    replication_connect_timeout = replication_connect_timeout   \
> > -}
> > diff --git a/test/replication/misc_assert_connecting_master_twice_gh-3610.result b/test/replication/misc_assert_connecting_master_twice_gh-3610.result
> > new file mode 100644
> > index 000000000..d7b7cc25b
> > --- /dev/null
> > +++ b/test/replication/misc_assert_connecting_master_twice_gh-3610.result
> > @@ -0,0 +1,86 @@
> > +test_run = require('test_run').new()
> > +---
> > +...
> > +test_run:cmd("restart server default")
> I don't  think you should restart the server here.
> The test seems fine without it. Same about almost all the
> testcases below.
> > +uuid = require('uuid')
> > +---
> > +...
> > +fiber = require('fiber')
> > +---
> > +...
> 
> UUID and fiber modules aren't used in this test, so please remove
> these two extra lines. Same about almost all the testcases below.
> 
> 
> The branch contains this extraneous change:
> 
> +replication_connect_quorum = box.cfg.replication_connect_quorum
> 
> +---
> +...
> +replication_connect_timeout = box.cfg.replication_connect_timeout
> +---
> 
> +...
> 
> You don't need to save these options, since you only change them on a
> replica.
> 
> > +--
> > +-- Test case for gh-3610. Before the fix replica would fail with the assertion
> > +-- when trying to connect to the same master twice.
> > +--
> > +box.schema.user.grant('guest', 'replication')
> > +---
> > +...
> > +test_run:cmd("create server replica with rpl_master=default, script='replication/replica.lua'")
> > +---
> > +- true
> > +...
> > +test_run:cmd("start server replica")
> > +---
> > +- true
> > +...
> > +test_run:cmd("switch replica")
> > +---
> > +- true
> > +...
> > +replication = box.cfg.replication[1]
> > +---
> > +...
> > +box.cfg{replication = {replication, replication}}
> > +---
> > +- error: 'Incorrect value for option ''replication'': duplicate connection to the
> > +    same replica'
> > +...
> > +-- Check the case when duplicate connection is detected in the background.
> > +test_run:cmd("switch default")
> > +---
> > +- true
> > +...
> > +listen = box.cfg.listen
> > +---
> > +...
> > +box.cfg{listen = ''}
> > +---
> > +...
> > +test_run:cmd("switch replica")
> > +---
> > +- true
> > +...
> > +box.cfg{replication_connect_quorum = 0, replication_connect_timeout = 0.01}
> > +---
> > +...
> > +box.cfg{replication = {replication, replication}}
> > +---
> > +...
> > +test_run:cmd("switch default")
> > +---
> > +- true
> > +...
> > +box.cfg{listen = listen}
> > +---
> > +...
> > +while test_run:grep_log('replica', 'duplicate connection') == nil do fiber.sleep(0.01) end
> > +---
> > +...
> > +test_run:cmd("stop server replica")
> > +---
> > +- true
> > +...
> > +test_run:cmd("cleanup server replica")
> > +---
> > +- true
> > +...
> > +test_run:cmd("delete server replica")
> > +---
> > +- true
> > +...
> > +test_run:cleanup_cluster()
> > +---
> > +...
> > +box.schema.user.revoke('guest', 'replication')
> > +---
> > +...
> +box.cfg{replication = "", \
> +        replication_connect_quorum = replication_connect_quorum, \
> +        replication_connect_timeout = replication_connect_timeout}
> +---
> +...
> 
> The change I spoke of above. This piece is also unneeded.
> 
> > diff --git a/test/replication/misc_assert_connecting_master_twice_gh-3610.test.lua b/test/replication/misc_assert_connecting_master_twice_gh-3610.test.lua
> > new file mode 100644
> > index 000000000..9b12e623b
> > --- /dev/null
> > +++ b/test/replication/misc_assert_connecting_master_twice_gh-3610.test.lua
> > @@ -0,0 +1,34 @@
> > +test_run = require('test_run').new()
> > +test_run:cmd("restart server default")
> > +uuid = require('uuid')
> > +fiber = require('fiber')
> > +
> > +--
> > +-- Test case for gh-3610. Before the fix replica would fail with the assertion
> > +-- when trying to connect to the same master twice.
> > +--
> > +box.schema.user.grant('guest', 'replication')
> > +test_run:cmd("create server replica with rpl_master=default, script='replication/replica.lua'")
> > +test_run:cmd("start server replica")
> > +test_run:cmd("switch replica")
> > +replication = box.cfg.replication[1]
> > +box.cfg{replication = {replication, replication}}
> > +
> > +-- Check the case when duplicate connection is detected in the background.
> > +test_run:cmd("switch default")
> > +listen = box.cfg.listen
> > +box.cfg{listen = ''}
> > +
> > +test_run:cmd("switch replica")
> > +box.cfg{replication_connect_quorum = 0, replication_connect_timeout = 0.01}
> > +box.cfg{replication = {replication, replication}}
> > +
> > +test_run:cmd("switch default")
> > +box.cfg{listen = listen}
> > +while test_run:grep_log('replica', 'duplicate connection') == nil do fiber.sleep(0.01) end
> > +
> > +test_run:cmd("stop server replica")
> > +test_run:cmd("cleanup server replica")
> > +test_run:cmd("delete server replica")
> > +test_run:cleanup_cluster()
> > +box.schema.user.revoke('guest', 'replication')
> > diff --git a/test/replication/misc_assert_on_server_die_gh-2991.result b/test/replication/misc_assert_on_server_die_gh-2991.result
> > new file mode 100644
> > index 000000000..ea9e80f6b
> > --- /dev/null
> > +++ b/test/replication/misc_assert_on_server_die_gh-2991.result
> > @@ -0,0 +1,31 @@
> > +uuid = require('uuid')
> > +---
> > +...
> UUID module isn't used in this testcase, please remove the extra line.
> > +test_run = require('test_run').new()
> > +---
> > +...
> > +-- gh-2991 - Tarantool asserts on box.cfg.replication update if one of
> > +-- servers is dead
> > +replication_timeout = box.cfg.replication_timeout
> > +---
> > +...
> > +replication_connect_timeout = box.cfg.replication_connect_timeout
> > +---
> > +...
> > +box.cfg{replication_timeout=0.05, replication_connect_timeout=0.05, replication={}}
> > +---
> > +...
> > +box.cfg{replication_connect_quorum=2}
> > +---
> > +...
> > +box.cfg{replication = {'127.0.0.1:12345', box.cfg.listen}}
> > +---
> > +...
> > +box.info.status
> > +---
> > +- orphan
> > +...
> > +box.info.ro
> > +---
> > +- true
> > +...
> > diff --git a/test/replication/misc_assert_on_server_die_gh-2991.test.lua b/test/replication/misc_assert_on_server_die_gh-2991.test.lua
> > new file mode 100644
> > index 000000000..adda839f7
> > --- /dev/null
> > +++ b/test/replication/misc_assert_on_server_die_gh-2991.test.lua
> > @@ -0,0 +1,12 @@
> > +uuid = require('uuid')
> > +test_run = require('test_run').new()
> > +
> > +-- gh-2991 - Tarantool asserts on box.cfg.replication update if one of
> > +-- servers is dead
> > +replication_timeout = box.cfg.replication_timeout
> > +replication_connect_timeout = box.cfg.replication_connect_timeout
> > +box.cfg{replication_timeout=0.05, replication_connect_timeout=0.05, replication={}}
> > +box.cfg{replication_connect_quorum=2}
> > +box.cfg{replication = {'127.0.0.1:12345', box.cfg.listen}}
> > +box.info.status
> > +box.info.ro
> > diff --git a/test/replication/misc_assert_replica_on_applier_disconnect_gh-3510.result b/test/replication/misc_assert_replica_on_applier_disconnect_gh-3510.result
> > new file mode 100644
> > index 000000000..82f5d9a23
> > --- /dev/null
> > +++ b/test/replication/misc_assert_replica_on_applier_disconnect_gh-3510.result
> > @@ -0,0 +1,49 @@
> > +uuid = require('uuid')
> > +---
> > +...
> Same about the uuid.
> > +test_run = require('test_run').new()
> > +---
> > +...
> > +-- gh-3510 assertion failure in replica_on_applier_disconnect()
> > +test_run:cmd('create server er_load1 with script="replication/er_load1.lua"')
> > +---
> > +- true
> > +...
> > +test_run:cmd('create server er_load2 with script="replication/er_load2.lua"')
> > +---
> > +- true
> > +...
> > +test_run:cmd('start server er_load1 with wait=False, wait_load=False')
> > +---
> > +- true
> > +...
> > +-- Instance er_load2 will fail with error ER_REPLICASET_UUID_MISMATCH.
> > +-- This is OK since we only test here that er_load1 doesn't assert.
> > +test_run:cmd('start server er_load2 with wait=True, wait_load=True, crash_expected = True')
> > +---
> > +- false
> > +...
> > +test_run:cmd('stop server er_load1')
> > +---
> > +- true
> > +...
> > +-- er_load2 exits automatically.
> > +test_run:cmd('cleanup server er_load1')
> > +---
> > +- true
> > +...
> > +test_run:cmd('cleanup server er_load2')
> > +---
> > +- true
> > +...
> > +test_run:cmd('delete server er_load1')
> > +---
> > +- true
> > +...
> > +test_run:cmd('delete server er_load2')
> > +---
> > +- true
> > +...
> > +test_run:cleanup_cluster()
> > +---
> > +...
> > diff --git a/test/replication/misc_assert_replica_on_applier_disconnect_gh-3510.test.lua b/test/replication/misc_assert_replica_on_applier_disconnect_gh-3510.test.lua
> > new file mode 100644
> > index 000000000..4e1d2a41e
> > --- /dev/null
> > +++ b/test/replication/misc_assert_replica_on_applier_disconnect_gh-3510.test.lua
> > @@ -0,0 +1,17 @@
> > +uuid = require('uuid')
> > +test_run = require('test_run').new()
> > +
> > +-- gh-3510 assertion failure in replica_on_applier_disconnect()
> > +test_run:cmd('create server er_load1 with script="replication/er_load1.lua"')
> > +test_run:cmd('create server er_load2 with script="replication/er_load2.lua"')
> > +test_run:cmd('start server er_load1 with wait=False, wait_load=False')
> > +-- Instance er_load2 will fail with error ER_REPLICASET_UUID_MISMATCH.
> > +-- This is OK since we only test here that er_load1 doesn't assert.
> > +test_run:cmd('start server er_load2 with wait=True, wait_load=True, crash_expected = True')
> > +test_run:cmd('stop server er_load1')
> > +-- er_load2 exits automatically.
> > +test_run:cmd('cleanup server er_load1')
> > +test_run:cmd('cleanup server er_load2')
> > +test_run:cmd('delete server er_load1')
> > +test_run:cmd('delete server er_load2')
> > +test_run:cleanup_cluster()
> > diff --git a/test/replication/misc_crash_on_box_concurrent_update_gh-3606.result b/test/replication/misc_crash_on_box_concurrent_update_gh-3606.result
> > new file mode 100644
> > index 000000000..b43b00849
> > --- /dev/null
> > +++ b/test/replication/misc_crash_on_box_concurrent_update_gh-3606.result
> > @@ -0,0 +1,50 @@
> > +uuid = require('uuid')
> > +---
> > +...
> Same about UUID.
> > +test_run = require('test_run').new()
> > +---
> > +...
> > +replication_timeout = box.cfg.replication_timeout
> > +---
> > +...
> > +replication_connect_timeout = box.cfg.replication_connect_timeout
> > +---
> > +...
> > +box.cfg{replication_timeout=0.05, replication_connect_timeout=0.05, replication={}}
> > +---
> > +...
> > +-- gh-3606 - Tarantool crashes if box.cfg.replication is updated concurrently
> > +fiber = require('fiber')
> > +---
> > +...
> > +c = fiber.channel(2)
> > +---
> > +...
> > +f = function() fiber.create(function() pcall(box.cfg, {replication = {12345}}) c:put(true) end) end
> > +---
> > +...
> > +f()
> > +---
> > +...
> > +f()
> > +---
> > +...
> > +c:get()
> > +---
> > +- true
> > +...
> > +c:get()
> > +---
> > +- true
> > +...
> > +box.cfg{replication = "", replication_timeout = replication_timeout, replication_connect_timeout = replication_connect_timeout}
> > +---
> > +...
> > +box.info.status
> > +---
> > +- running
> > +...
> > +box.info.ro
> > +---
> > +- false
> > +...
> > diff --git a/test/replication/misc_crash_on_box_concurrent_update_gh-3606.test.lua b/test/replication/misc_crash_on_box_concurrent_update_gh-3606.test.lua
> > new file mode 100644
> > index 000000000..17f9c6bc6
> > --- /dev/null
> > +++ b/test/replication/misc_crash_on_box_concurrent_update_gh-3606.test.lua
> > @@ -0,0 +1,19 @@
> > +uuid = require('uuid')
> > +test_run = require('test_run').new()
> > +
> > +replication_timeout = box.cfg.replication_timeout
> > +replication_connect_timeout = box.cfg.replication_connect_timeout
> > +box.cfg{replication_timeout=0.05, replication_connect_timeout=0.05, replication={}}
> > +
> > +-- gh-3606 - Tarantool crashes if box.cfg.replication is updated concurrently
> > +fiber = require('fiber')
> > +c = fiber.channel(2)
> > +f = function() fiber.create(function() pcall(box.cfg, {replication = {12345}}) c:put(true) end) end
> > +f()
> > +f()
> > +c:get()
> > +c:get()
> > +
> > +box.cfg{replication = "", replication_timeout = replication_timeout, replication_connect_timeout = replication_connect_timeout}
> > +box.info.status
> > +box.info.ro
> > diff --git a/test/replication/misc_heartbeats_on_master_changes_gh-3160.result b/test/replication/misc_heartbeats_on_master_changes_gh-3160.result
> > new file mode 100644
> > index 000000000..cdb463614
> > --- /dev/null
> > +++ b/test/replication/misc_heartbeats_on_master_changes_gh-3160.result
> > @@ -0,0 +1,76 @@
> > +uuid = require('uuid')
> > +---
> > +...
> 
> Same about UUID.
> 
> > +test_run = require('test_run').new()
> > +---
> > +...
> 
> The branch contains this change:
> 
> +...
> +replication_timeout = box.cfg.replication_timeout
> +---
> +...
> +replication_connect_timeout = box.cfg.replication_connect_timeout
> +---
> +...
> +box.cfg{replication_timeout=0.05, replication_connect_timeout=0.05,
> replication={}}
> +---
> +...
> 
> Which is extraneous in this test. The original testcase uses default timeout
> values,
> and I don't think we should change them.
> 
> > +-- gh-3160 - Send heartbeats if there are changes from a remote master only
> > +SERVERS = { 'autobootstrap1', 'autobootstrap2', 'autobootstrap3' }
> > +---
> > +...
> > +-- Deploy a cluster.
> > +test_run:create_cluster(SERVERS, "replication", {args="0.03"})
> > +---
> > +...
> > +test_run:wait_fullmesh(SERVERS)
> > +---
> > +...
> > +test_run:cmd("switch autobootstrap3")
> > +---
> > +- true
> > +...
> > +test_run = require('test_run').new()
> > +---
> > +...
> > +fiber = require('fiber')
> > +---
> > +...
> > +_ = box.schema.space.create('test_timeout'):create_index('pk')
> > +---
> > +...
> > +test_run:cmd("setopt delimiter ';'")
> > +---
> > +- true
> > +...
> > +function wait_not_follow(replicaA, replicaB)
> > +    return test_run:wait_cond(function()
> > +        return replicaA.status ~= 'follow' or replicaB.status ~= 'follow'
> > +    end, box.cfg.replication_timeout)
> > +end;
> > +---
> > +...
> > +function test_timeout()
> > +    local replicaA = box.info.replication[1].upstream or box.info.replication[2].upstream
> > +    local replicaB = box.info.replication[3].upstream or box.info.replication[2].upstream
> > +    local follows = test_run:wait_cond(function()
> > +        return replicaA.status == 'follow' or replicaB.status == 'follow'
> > +    end)
> > +    if not follows then error('replicas are not in the follow status') end
> > +    for i = 0, 99 do
> > +        box.space.test_timeout:replace({1})
> > +        if wait_not_follow(replicaA, replicaB) then
> > +            return error(box.info.replication)
> > +        end
> > +    end
> > +    return true
> > +end;
> > +---
> > +...
> > +test_run:cmd("setopt delimiter ''");
> > +---
> > +- true
> > +...
> > +test_timeout()
> > +---
> > +- true
> > +...
> > +test_run:cmd("switch default")
> > +---
> > +- true
> > +...
> > +test_run:drop_cluster(SERVERS)
> > +---
> > +...
> > +test_run:cleanup_cluster()
> 
> No need to call cleanup_cluster(). The 'default' instance wasn't part of the
> cluster
> you ran in this test.
> 
> 
> +...
> +box.cfg{replication = "", replication_timeout = replication_timeout, \
> +        replication_connect_timeout = replication_connect_timeout}
> +---
> +...
> 
> One more extraneous change, related to the one above.
> 
> 
> > +---
> > +...
> > diff --git a/test/replication/misc_heartbeats_on_master_changes_gh-3160.test.lua b/test/replication/misc_heartbeats_on_master_changes_gh-3160.test.lua
> > new file mode 100644
> > index 000000000..eba8a7725
> > --- /dev/null
> > +++ b/test/replication/misc_heartbeats_on_master_changes_gh-3160.test.lua
> > @@ -0,0 +1,40 @@
> > +uuid = require('uuid')
> > +test_run = require('test_run').new()
> > +
> > +-- gh-3160 - Send heartbeats if there are changes from a remote master only
> > +SERVERS = { 'autobootstrap1', 'autobootstrap2', 'autobootstrap3' }
> > +
> > +-- Deploy a cluster.
> > +test_run:create_cluster(SERVERS, "replication", {args="0.03"})
> > +test_run:wait_fullmesh(SERVERS)
> > +test_run:cmd("switch autobootstrap3")
> > +test_run = require('test_run').new()
> > +fiber = require('fiber')
> > +_ = box.schema.space.create('test_timeout'):create_index('pk')
> > +test_run:cmd("setopt delimiter ';'")
> > +function wait_not_follow(replicaA, replicaB)
> > +    return test_run:wait_cond(function()
> > +        return replicaA.status ~= 'follow' or replicaB.status ~= 'follow'
> > +    end, box.cfg.replication_timeout)
> > +end;
> > +function test_timeout()
> > +    local replicaA = box.info.replication[1].upstream or box.info.replication[2].upstream
> > +    local replicaB = box.info.replication[3].upstream or box.info.replication[2].upstream
> > +    local follows = test_run:wait_cond(function()
> > +        return replicaA.status == 'follow' or replicaB.status == 'follow'
> > +    end)
> > +    if not follows then error('replicas are not in the follow status') end
> > +    for i = 0, 99 do
> > +        box.space.test_timeout:replace({1})
> > +        if wait_not_follow(replicaA, replicaB) then
> > +            return error(box.info.replication)
> > +        end
> > +    end
> > +    return true
> > +end;
> > +test_run:cmd("setopt delimiter ''");
> > +test_timeout()
> > +
> > +test_run:cmd("switch default")
> > +test_run:drop_cluster(SERVERS)
> > +test_run:cleanup_cluster()
> > diff --git a/test/replication/misc_no_failure_on_error_reading_wal_gh-4399.result b/test/replication/misc_no_failure_on_error_reading_wal_gh-4399.result
> > new file mode 100644
> > index 000000000..64a4de26c
> > --- /dev/null
> > +++ b/test/replication/misc_no_failure_on_error_reading_wal_gh-4399.result
> > @@ -0,0 +1,97 @@
> > +test_run = require('test_run').new()
> > +---
> > +...
> > +test_run:cmd("restart server default")
> > +uuid = require('uuid')
> > +---
> > +...
> > +fiber = require('fiber')
> Same about restarting the server, requiring uuid and fiber. Not needed.
> > +---
> > +...
> > +--
> > +-- gh-4399 Check that an error reading WAL directory on subscribe
> > +-- doesn't lead to a permanent replication failure.
> > +--
> > +box.schema.user.grant("guest", "replication")
> > +---
> > +...
> > +test_run:cmd("create server replica with rpl_master=default, script='replication/replica.lua'")
> > +---
> > +- true
> > +...
> > +test_run:cmd("start server replica")
> > +---
> > +- true
> > +...
> > +-- Make the WAL directory inaccessible.
> > +fio = require('fio')
> > +---
> > +...
> > +path = fio.abspath(box.cfg.wal_dir)
> > +---
> > +...
> > +fio.chmod(path, 0)
> > +---
> > +- true
> > +...
> > +-- Break replication on timeout.
> > +replication_timeout = box.cfg.replication_timeout
> > +---
> > +...
> > +box.cfg{replication_timeout = 9000}
> > +---
> > +...
> > +test_run:cmd("switch replica")
> > +---
> > +- true
> > +...
> > +test_run:wait_cond(function() return box.info.replication[1].upstream.status ~= 'follow' end)
> > +---
> > +- true
> > +...
> > +require('fiber').sleep(box.cfg.replication_timeout)
> > +---
> > +...
> > +test_run:cmd("switch default")
> > +---
> > +- true
> > +...
> > +box.cfg{replication_timeout = replication_timeout}
> > +---
> > +...
> > +-- Restore access to the WAL directory.
> > +-- Wait for replication to be reestablished.
> > +fio.chmod(path, tonumber('777', 8))
> > +---
> > +- true
> > +...
> > +test_run:cmd("switch replica")
> > +---
> > +- true
> > +...
> > +test_run:wait_cond(function() return box.info.replication[1].upstream.status == 'follow' end)
> > +---
> > +- true
> > +...
> > +test_run:cmd("switch default")
> > +---
> > +- true
> > +...
> > +test_run:cmd("stop server replica")
> > +---
> > +- true
> > +...
> > +test_run:cmd("cleanup server replica")
> > +---
> > +- true
> > +...
> > +test_run:cmd("delete server replica")
> > +---
> > +- true
> > +...
> > +test_run:cleanup_cluster()
> > +---
> > +...
> > +box.schema.user.revoke('guest', 'replication')
> > +---
> > +...
> > diff --git a/test/replication/misc_no_failure_on_error_reading_wal_gh-4399.test.lua b/test/replication/misc_no_failure_on_error_reading_wal_gh-4399.test.lua
> > new file mode 100644
> > index 000000000..15e19a211
> > --- /dev/null
> > +++ b/test/replication/misc_no_failure_on_error_reading_wal_gh-4399.test.lua
> > @@ -0,0 +1,39 @@
> > +test_run = require('test_run').new()
> > +test_run:cmd("restart server default")
> > +uuid = require('uuid')
> > +fiber = require('fiber')
> > +
> > +--
> > +-- gh-4399 Check that an error reading WAL directory on subscribe
> > +-- doesn't lead to a permanent replication failure.
> > +--
> > +box.schema.user.grant("guest", "replication")
> > +test_run:cmd("create server replica with rpl_master=default, script='replication/replica.lua'")
> > +test_run:cmd("start server replica")
> > +
> > +-- Make the WAL directory inaccessible.
> > +fio = require('fio')
> > +path = fio.abspath(box.cfg.wal_dir)
> > +fio.chmod(path, 0)
> > +
> > +-- Break replication on timeout.
> > +replication_timeout = box.cfg.replication_timeout
> > +box.cfg{replication_timeout = 9000}
> > +test_run:cmd("switch replica")
> > +test_run:wait_cond(function() return box.info.replication[1].upstream.status ~= 'follow' end)
> > +require('fiber').sleep(box.cfg.replication_timeout)
> > +test_run:cmd("switch default")
> > +box.cfg{replication_timeout = replication_timeout}
> > +
> > +-- Restore access to the WAL directory.
> > +-- Wait for replication to be reestablished.
> > +fio.chmod(path, tonumber('777', 8))
> > +test_run:cmd("switch replica")
> > +test_run:wait_cond(function() return box.info.replication[1].upstream.status == 'follow' end)
> > +test_run:cmd("switch default")
> > +
> > +test_run:cmd("stop server replica")
> > +test_run:cmd("cleanup server replica")
> > +test_run:cmd("delete server replica")
> > +test_run:cleanup_cluster()
> > +box.schema.user.revoke('guest', 'replication')
> > diff --git a/test/replication/misc_no_panic_on_connected_gh-3637.result b/test/replication/misc_no_panic_on_connected_gh-3637.result
> > new file mode 100644
> > index 000000000..693b18c42
> > --- /dev/null
> > +++ b/test/replication/misc_no_panic_on_connected_gh-3637.result
> > @@ -0,0 +1,72 @@
> > +test_run = require('test_run').new()
> > +---
> > +...
> > +test_run:cmd("restart server default")
> > +uuid = require('uuid')
> > +---
> > +...
> Again, no need for uuid here. And no need for restart.
> > +--
> > +-- Test case for gh-3637, gh-4550. Before the fix replica would
> > +-- exit with an error if a user does not exist or a password is
> > +-- incorrect. Now check that we don't hang/panic and successfully
> > +-- connect.
> > +--
> > +fiber = require('fiber')
> > +---
> > +...
> > +test_run:cmd("create server replica_auth with rpl_master=default, script='replication/replica_auth.lua'")
> > +---
> > +- true
> > +...
> > +test_run:cmd("start server replica_auth with wait=False, wait_load=False, args='cluster:pass 0.05'")
> > +---
> > +- true
> > +...
> > +-- Wait a bit to make sure replica waits till user is created.
> > +fiber.sleep(0.1)
> > +---
> > +...
> > +box.schema.user.create('cluster')
> > +---
> > +...
> > +-- The user is created. Let the replica fail auth request due to
> > +-- a wrong password.
> > +fiber.sleep(0.1)
> > +---
> > +...
> > +box.schema.user.passwd('cluster', 'pass')
> > +---
> > +...
> > +box.schema.user.grant('cluster', 'replication')
> > +---
> > +...
> > +while box.info.replication[2] == nil do fiber.sleep(0.01) end
> > +---
> > +...
> > +vclock = test_run:get_vclock('default')
> > +---
> > +...
> > +vclock[0] = nil
> > +---
> > +...
> > +_ = test_run:wait_vclock('replica_auth', vclock)
> > +---
> > +...
> > +test_run:cmd("stop server replica_auth")
> > +---
> > +- true
> > +...
> > +test_run:cmd("cleanup server replica_auth")
> > +---
> > +- true
> > +...
> > +test_run:cmd("delete server replica_auth")
> > +---
> > +- true
> > +...
> > +test_run:cleanup_cluster()
> > +---
> > +...
> > +box.schema.user.drop('cluster')
> > +---
> > +...
> > diff --git a/test/replication/misc_no_panic_on_connected_gh-3637.test.lua b/test/replication/misc_no_panic_on_connected_gh-3637.test.lua
> > new file mode 100644
> > index 000000000..a1e51198b
> > --- /dev/null
> > +++ b/test/replication/misc_no_panic_on_connected_gh-3637.test.lua
> > @@ -0,0 +1,33 @@
> > +test_run = require('test_run').new()
> > +test_run:cmd("restart server default")
> > +uuid = require('uuid')
> > +
> > +--
> > +-- Test case for gh-3637, gh-4550. Before the fix replica would
> > +-- exit with an error if a user does not exist or a password is
> > +-- incorrect. Now check that we don't hang/panic and successfully
> > +-- connect.
> > +--
> > +fiber = require('fiber')
> > +test_run:cmd("create server replica_auth with rpl_master=default, script='replication/replica_auth.lua'")
> > +test_run:cmd("start server replica_auth with wait=False, wait_load=False, args='cluster:pass 0.05'")
> > +-- Wait a bit to make sure replica waits till user is created.
> > +fiber.sleep(0.1)
> > +box.schema.user.create('cluster')
> > +-- The user is created. Let the replica fail auth request due to
> > +-- a wrong password.
> > +fiber.sleep(0.1)
> > +box.schema.user.passwd('cluster', 'pass')
> > +box.schema.user.grant('cluster', 'replication')
> > +
> > +while box.info.replication[2] == nil do fiber.sleep(0.01) end
> > +vclock = test_run:get_vclock('default')
> > +vclock[0] = nil
> > +_ = test_run:wait_vclock('replica_auth', vclock)
> > +
> > +test_run:cmd("stop server replica_auth")
> > +test_run:cmd("cleanup server replica_auth")
> > +test_run:cmd("delete server replica_auth")
> > +test_run:cleanup_cluster()
> > +
> > +box.schema.user.drop('cluster')
> > diff --git a/test/replication/misc_no_restart_on_same_configuration_gh-3711.result b/test/replication/misc_no_restart_on_same_configuration_gh-3711.result
> > new file mode 100644
> > index 000000000..dd9ef848c
> > --- /dev/null
> > +++ b/test/replication/misc_no_restart_on_same_configuration_gh-3711.result
> > @@ -0,0 +1,107 @@
> > +test_run = require('test_run').new()
> > +---
> > +...
> > +test_run:cmd("restart server default")
> > +uuid = require('uuid')
> > +---
> > +...
> > +fiber = require('fiber')
> Same about restart, uuid and fiber.
> > +---
> > +...
> > +--
> > +-- gh-3711 Do not restart replication on box.cfg in case the
> > +-- configuration didn't change.
> > +--
> > +box.schema.user.grant('guest', 'replication')
> > +---
> > +...
> > +test_run:cmd("create server replica with rpl_master=default, script='replication/replica.lua'")
> > +---
> > +- true
> > +...
> > +test_run:cmd("start server replica")
> > +---
> > +- true
> > +...
> > +-- Access rights are checked only during reconnect. If the new
> > +-- and old configurations are equivalent, no reconnect will be
> > +-- issued and replication should continue working.
> > +box.schema.user.revoke('guest', 'replication')
> > +---
> > +...
> > +test_run:cmd("switch replica")
> > +---
> > +- true
> > +...
> > +replication = box.cfg.replication[1]
> > +---
> > +...
> > +box.cfg{replication = {replication}}
> > +---
> > +...
> > +box.info.status == 'running'
> > +---
> > +- true
> > +...
> > +box.cfg{replication = replication}
> > +---
> > +...
> > +box.info.status == 'running'
> > +---
> > +- true
> > +...
> > +-- Check that comparison of tables works as expected as well.
> > +test_run:cmd("switch default")
> > +---
> > +- true
> > +...
> > +box.schema.user.grant('guest', 'replication')
> > +---
> > +...
> > +test_run:cmd("switch replica")
> > +---
> > +- true
> > +...
> > +replication = box.cfg.replication
> > +---
> > +...
> > +table.insert(replication, box.cfg.listen)
> > +---
> > +...
> > +test_run:cmd("switch default")
> > +---
> > +- true
> > +...
> > +box.schema.user.revoke('guest', 'replication')
> > +---
> > +...
> > +test_run:cmd("switch replica")
> > +---
> > +- true
> > +...
> > +box.cfg{replication = replication}
> > +---
> > +...
> > +box.info.status == 'running'
> > +---
> > +- true
> > +...
> > +test_run:cmd("switch default")
> > +---
> > +- true
> > +...
> > +test_run:cmd("stop server replica")
> > +---
> > +- true
> > +...
> > +test_run:cmd("cleanup server replica")
> > +---
> > +- true
> > +...
> > +test_run:cmd("delete server replica")
> > +---
> > +- true
> > +...
> > +test_run:cleanup_cluster()
> > +---
> > +...
> > diff --git a/test/replication/misc_no_restart_on_same_configuration_gh-3711.test.lua b/test/replication/misc_no_restart_on_same_configuration_gh-3711.test.lua
> > new file mode 100644
> > index 000000000..14b522ead
> > --- /dev/null
> > +++ b/test/replication/misc_no_restart_on_same_configuration_gh-3711.test.lua
> > @@ -0,0 +1,41 @@
> > +test_run = require('test_run').new()
> > +test_run:cmd("restart server default")
> > +uuid = require('uuid')
> > +fiber = require('fiber')
> > +
> > +--
> > +-- gh-3711 Do not restart replication on box.cfg in case the
> > +-- configuration didn't change.
> > +--
> > +box.schema.user.grant('guest', 'replication')
> > +test_run:cmd("create server replica with rpl_master=default, script='replication/replica.lua'")
> > +test_run:cmd("start server replica")
> > +
> > +-- Access rights are checked only during reconnect. If the new
> > +-- and old configurations are equivalent, no reconnect will be
> > +-- issued and replication should continue working.
> > +box.schema.user.revoke('guest', 'replication')
> > +test_run:cmd("switch replica")
> > +replication = box.cfg.replication[1]
> > +box.cfg{replication = {replication}}
> > +box.info.status == 'running'
> > +box.cfg{replication = replication}
> > +box.info.status == 'running'
> > +
> > +-- Check that comparison of tables works as expected as well.
> > +test_run:cmd("switch default")
> > +box.schema.user.grant('guest', 'replication')
> > +test_run:cmd("switch replica")
> > +replication = box.cfg.replication
> > +table.insert(replication, box.cfg.listen)
> > +test_run:cmd("switch default")
> > +box.schema.user.revoke('guest', 'replication')
> > +test_run:cmd("switch replica")
> > +box.cfg{replication = replication}
> > +box.info.status == 'running'
> > +
> > +test_run:cmd("switch default")
> > +test_run:cmd("stop server replica")
> > +test_run:cmd("cleanup server replica")
> > +test_run:cmd("delete server replica")
> > +test_run:cleanup_cluster()
> > diff --git a/test/replication/misc_no_socket_leak_on_replica_disconnect_gh-3642.result b/test/replication/misc_no_socket_leak_on_replica_disconnect_gh-3642.result
> > new file mode 100644
> > index 000000000..3680bcebb
> > --- /dev/null
> > +++ b/test/replication/misc_no_socket_leak_on_replica_disconnect_gh-3642.result
> > @@ -0,0 +1,98 @@
> > +test_run = require('test_run').new()
> > +---
> > +...
> > +test_run:cmd("restart server default")
> 
> Unneeded restart. Correct me if I'm wrong.
> 
> > +uuid = require('uuid')
> 
> You don't need the uuid module in this testcase.
> 
> > +---
> > +...
> > +box.schema.user.grant('guest', 'replication')
> > +---
> > +...
> > +-- gh-3642 - Check that socket file descriptor doesn't leak
> > +-- when a replica is disconnected.
> > +rlimit = require('rlimit')
> > +---
> > +...
> > +lim = rlimit.limit()
> > +---
> > +...
> > +rlimit.getrlimit(rlimit.RLIMIT_NOFILE, lim)
> > +---
> > +...
> > +old_fno = lim.rlim_cur
> > +---
> > +...
> > +lim.rlim_cur = 64
> > +---
> > +...
> > +rlimit.setrlimit(rlimit.RLIMIT_NOFILE, lim)
> > +---
> > +...
> > +test_run:cmd('create server sock with rpl_master=default, script="replication/replica.lua"')
> > +---
> > +- true
> > +...
> > +test_run:cmd('start server sock')
> > +---
> > +- true
> > +...
> > +test_run:cmd('switch sock')
> > +---
> > +- true
> > +...
> > +test_run = require('test_run').new()
> > +---
> > +...
> > +fiber = require('fiber')
> > +---
> > +...
> > +test_run:cmd("setopt delimiter ';'")
> > +---
> > +- true
> > +...
> > +for i = 1, 64 do
> > +    local replication = box.cfg.replication
> > +    box.cfg{replication = {}}
> > +    box.cfg{replication = replication}
> > +    while box.info.replication[1].upstream.status ~= 'follow' do
> > +        fiber.sleep(0.001)
> > +    end
> > +end;
> > +---
> > +...
> > +test_run:cmd("setopt delimiter ''");
> > +---
> > +- true
> > +...
> > +box.info.replication[1].upstream.status
> > +---
> > +- follow
> > +...
> > +test_run:cmd('switch default')
> > +---
> > +- true
> > +...
> > +lim.rlim_cur = old_fno
> > +---
> > +...
> > +rlimit.setrlimit(rlimit.RLIMIT_NOFILE, lim)
> > +---
> > +...
> > +test_run:cmd("stop server sock")
> > +---
> > +- true
> > +...
> > +test_run:cmd("cleanup server sock")
> > +---
> > +- true
> > +...
> > +test_run:cmd("delete server sock")
> > +---
> > +- true
> > +...
> > +test_run:cleanup_cluster()
> > +---
> > +...
> > +box.schema.user.revoke('guest', 'replication')
> > +---
> > +...
> > diff --git a/test/replication/misc_no_socket_leak_on_replica_disconnect_gh-3642.test.lua b/test/replication/misc_no_socket_leak_on_replica_disconnect_gh-3642.test.lua
> > new file mode 100644
> > index 000000000..08ef9ec0d
> > --- /dev/null
> > +++ b/test/replication/misc_no_socket_leak_on_replica_disconnect_gh-3642.test.lua
> > @@ -0,0 +1,44 @@
> > +test_run = require('test_run').new()
> > +test_run:cmd("restart server default")
> > +uuid = require('uuid')
> > +
> > +box.schema.user.grant('guest', 'replication')
> > +
> > +-- gh-3642 - Check that socket file descriptor doesn't leak
> > +-- when a replica is disconnected.
> > +rlimit = require('rlimit')
> > +lim = rlimit.limit()
> > +rlimit.getrlimit(rlimit.RLIMIT_NOFILE, lim)
> > +old_fno = lim.rlim_cur
> > +lim.rlim_cur = 64
> > +rlimit.setrlimit(rlimit.RLIMIT_NOFILE, lim)
> > +
> > +test_run:cmd('create server sock with rpl_master=default, script="replication/replica.lua"')
> > +test_run:cmd('start server sock')
> > +test_run:cmd('switch sock')
> > +test_run = require('test_run').new()
> > +fiber = require('fiber')
> > +test_run:cmd("setopt delimiter ';'")
> > +for i = 1, 64 do
> > +    local replication = box.cfg.replication
> > +    box.cfg{replication = {}}
> > +    box.cfg{replication = replication}
> > +    while box.info.replication[1].upstream.status ~= 'follow' do
> > +        fiber.sleep(0.001)
> > +    end
> > +end;
> > +test_run:cmd("setopt delimiter ''");
> > +
> > +box.info.replication[1].upstream.status
> > +
> > +test_run:cmd('switch default')
> > +
> > +lim.rlim_cur = old_fno
> > +rlimit.setrlimit(rlimit.RLIMIT_NOFILE, lim)
> > +
> > +test_run:cmd("stop server sock")
> > +test_run:cmd("cleanup server sock")
> > +test_run:cmd("delete server sock")
> > +test_run:cleanup_cluster()
> > +
> > +box.schema.user.revoke('guest', 'replication')
> > diff --git a/test/replication/misc_orphan_on_reconfiguration_error_gh-4424.result b/test/replication/misc_orphan_on_reconfiguration_error_gh-4424.result
> > new file mode 100644
> > index 000000000..6a9d1148d
> > --- /dev/null
> > +++ b/test/replication/misc_orphan_on_reconfiguration_error_gh-4424.result
> > @@ -0,0 +1,88 @@
> > +uuid = require('uuid')
> > +---
> > +...
> > +test_run = require('test_run').new()
> > +---
> > +...
> > +fiber = require('fiber')
> 
> UUID and fiber aren't used in this testcase
> 
> > +---
> > +...
> > +--
> > +-- gh-4424 Always enter orphan mode on error in replication
> > +-- configuration change.
> > +--
> > +replication_connect_timeout = box.cfg.replication_connect_timeout
> > +---
> > +...
> > +replication_connect_quorum = box.cfg.replication_connect_quorum
> > +---
> > +...
> > +box.cfg{replication="12345", replication_connect_timeout=0.1, replication_connect_quorum=1}
> > +---
> > +...
> > +box.info.status
> > +---
> > +- orphan
> > +...
> > +box.info.ro
> > +---
> > +- true
> > +...
> > +-- reset replication => leave orphan mode
> > +box.cfg{replication=""}
> > +---
> > +...
> > +box.info.status
> > +---
> > +- running
> > +...
> > +box.info.ro
> > +---
> > +- false
> > +...
> > +-- no switch to orphan when quorum == 0
> > +box.cfg{replication="12345", replication_connect_quorum=0}
> > +---
> > +...
> > +box.info.status
> > +---
> > +- running
> > +...
> > +box.info.ro
> > +---
> > +- false
> > +...
> > +-- we could connect to one out of two replicas. Set orphan.
> > +box.cfg{replication_connect_quorum=2}
> > +---
> > +...
> > +box.cfg{replication={box.cfg.listen, "12345"}}
> > +---
> > +...
> > +box.info.status
> > +---
> > +- orphan
> > +...
> > +box.info.ro
> > +---
> > +- true
> > +...
> > +-- lower quorum => leave orphan mode
> > +box.cfg{replication_connect_quorum=1}
> > +---
> > +...
> > +box.info.status
> > +---
> > +- running
> > +...
> > +box.info.ro
> > +---
> > +- false
> > +...
> > +box.cfg{                                                        \
> > +    replication = {},                                           \
> > +    replication_connect_quorum = replication_connect_quorum,    \
> > +    replication_connect_timeout = replication_connect_timeout   \
> > +}
> > +---
> > +...
> > diff --git a/test/replication/misc_orphan_on_reconfiguration_error_gh-4424.test.lua b/test/replication/misc_orphan_on_reconfiguration_error_gh-4424.test.lua
> > new file mode 100644
> > index 000000000..7d1a70a36
> > --- /dev/null
> > +++ b/test/replication/misc_orphan_on_reconfiguration_error_gh-4424.test.lua
> > @@ -0,0 +1,37 @@
> > +uuid = require('uuid')
> > +test_run = require('test_run').new()
> > +fiber = require('fiber')
> > +
> > +--
> > +-- gh-4424 Always enter orphan mode on error in replication
> > +-- configuration change.
> > +--
> > +replication_connect_timeout = box.cfg.replication_connect_timeout
> > +replication_connect_quorum = box.cfg.replication_connect_quorum
> > +box.cfg{replication="12345", replication_connect_timeout=0.1, replication_connect_quorum=1}
> > +box.info.status
> > +box.info.ro
> > +-- reset replication => leave orphan mode
> > +box.cfg{replication=""}
> > +box.info.status
> > +box.info.ro
> > +-- no switch to orphan when quorum == 0
> > +box.cfg{replication="12345", replication_connect_quorum=0}
> > +box.info.status
> > +box.info.ro
> > +
> > +-- we could connect to one out of two replicas. Set orphan.
> > +box.cfg{replication_connect_quorum=2}
> > +box.cfg{replication={box.cfg.listen, "12345"}}
> > +box.info.status
> > +box.info.ro
> > +-- lower quorum => leave orphan mode
> > +box.cfg{replication_connect_quorum=1}
> > +box.info.status
> > +box.info.ro
> > +
> > +box.cfg{                                                        \
> > +    replication = {},                                           \
> > +    replication_connect_quorum = replication_connect_quorum,    \
> > +    replication_connect_timeout = replication_connect_timeout   \
> > +}
> > diff --git a/test/replication/misc_rebootstrap_from_ro_master_gh-3111.result b/test/replication/misc_rebootstrap_from_ro_master_gh-3111.result
> > new file mode 100644
> > index 000000000..7ffca1585
> > --- /dev/null
> > +++ b/test/replication/misc_rebootstrap_from_ro_master_gh-3111.result
> > @@ -0,0 +1,58 @@
> > +test_run = require('test_run').new()
> > +---
> > +...
> > +test_run:cmd("restart server default")
> 
> Why do you need it? Seems like the test passes just fine without the
> restart.
> Restarting the server will take some time if this is executed after some
> test that
> performs lots of DML operations.So You better remove the change
> 
> > +uuid = require('uuid')
> > +---
> > +...
> > +box.schema.user.grant('guest', 'replication')
> > +---
> > +...
> > +-- gh-3111 - Allow to rebootstrap a replica from a read-only master
> > +replica_uuid = uuid.new()
> > +---
> > +...
> > +test_run:cmd('create server test with rpl_master=default, script="replication/replica_uuid.lua"')
> > +---
> > +- true
> > +...
> > +test_run:cmd(string.format('start server test with args="%s"', replica_uuid))
> > +---
> > +- true
> > +...
> > +test_run:cmd('stop server test')
> > +---
> > +- true
> > +...
> > +test_run:cmd('cleanup server test')
> > +---
> > +- true
> > +...
> > +box.cfg{read_only = true}
> > +---
> > +...
> > +test_run:cmd(string.format('start server test with args="%s"', replica_uuid))
> > +---
> > +- true
> > +...
> > +test_run:cmd('stop server test')
> > +---
> > +- true
> > +...
> > +test_run:cmd('cleanup server test')
> > +---
> > +- true
> > +...
> > +box.cfg{read_only = false}
> > +---
> > +...
> > +test_run:cmd('delete server test')
> > +---
> > +- true
> > +...
> > +test_run:cleanup_cluster()
> > +---
> > +...
> > +box.schema.user.revoke('guest', 'replication')
> > +---
> > +...
> > diff --git a/test/replication/misc_rebootstrap_from_ro_master_gh-3111.test.lua b/test/replication/misc_rebootstrap_from_ro_master_gh-3111.test.lua
> > new file mode 100644
> > index 000000000..bb9b4a80f
> > --- /dev/null
> > +++ b/test/replication/misc_rebootstrap_from_ro_master_gh-3111.test.lua
> > @@ -0,0 +1,20 @@
> > +test_run = require('test_run').new()
> > +test_run:cmd("restart server default")
> > +uuid = require('uuid')
> > +
> > +box.schema.user.grant('guest', 'replication')
> > +
> > +-- gh-3111 - Allow to rebootstrap a replica from a read-only master
> > +replica_uuid = uuid.new()
> > +test_run:cmd('create server test with rpl_master=default, script="replication/replica_uuid.lua"')
> > +test_run:cmd(string.format('start server test with args="%s"', replica_uuid))
> > +test_run:cmd('stop server test')
> > +test_run:cmd('cleanup server test')
> > +box.cfg{read_only = true}
> > +test_run:cmd(string.format('start server test with args="%s"', replica_uuid))
> > +test_run:cmd('stop server test')
> > +test_run:cmd('cleanup server test')
> > +box.cfg{read_only = false}
> > +test_run:cmd('delete server test')
> > +test_run:cleanup_cluster()
> > +box.schema.user.revoke('guest', 'replication')
> > diff --git a/test/replication/misc_replica_checks_cluster_id_gh-3704.result b/test/replication/misc_replica_checks_cluster_id_gh-3704.result
> > new file mode 100644
> > index 000000000..e6bc8b4d8
> > --- /dev/null
> > +++ b/test/replication/misc_replica_checks_cluster_id_gh-3704.result
> > @@ -0,0 +1,71 @@
> > +test_run = require('test_run').new()
> > +---
> > +...
> > +test_run:cmd("restart server default")
> > +uuid = require('uuid')
> > +---
> > +...
> > +fiber = require('fiber')
> 
> No need to restart the instance, and require fiber module.
> 
> UUID is needed this time.
> 
> > +---
> > +...
> > +--
> > +-- gh-3704 move cluster id check to replica
> > +--
> > +test_run:cmd("create server replica with rpl_master=default, script='replication/replica.lua'")
> > +---
> > +- true
> > +...
> > +box.schema.user.grant("guest", "replication")
> > +---
> > +...
> > +test_run:cmd("start server replica")
> > +---
> > +- true
> > +...
> > +test_run:grep_log("replica", "REPLICASET_UUID_MISMATCH")
> > +---
> > +- null
> > +...
> > +box.info.replication[2].downstream.status
> > +---
> > +- follow
> > +...
> > +-- change master's cluster uuid and check that replica doesn't connect.
> > +test_run:cmd("stop server replica")
> > +---
> > +- true
> > +...
> > +_ = box.space._schema:replace{'cluster', tostring(uuid.new())}
> > +---
> > +...
> > +-- master believes replica is in cluster, but their cluster UUIDs differ.
> > +test_run:cmd("start server replica")
> > +---
> > +- true
> > +...
> > +test_run:wait_log("replica", "REPLICASET_UUID_MISMATCH", nil, 1.0)
> > +---
> > +- REPLICASET_UUID_MISMATCH
> > +...
> > +test_run:wait_downstream(2, {status = 'stopped'})
> > +---
> > +- true
> > +...
> > +test_run:cmd("stop server replica")
> > +---
> > +- true
> > +...
> > +test_run:cmd("cleanup server replica")
> > +---
> > +- true
> > +...
> > +test_run:cmd("delete server replica")
> > +---
> > +- true
> > +...
> > +test_run:cleanup_cluster()
> > +---
> > +...
> > +box.schema.user.revoke('guest', 'replication')
> > +---
> > +...
> > diff --git a/test/replication/misc_replica_checks_cluster_id_gh-3704.test.lua b/test/replication/misc_replica_checks_cluster_id_gh-3704.test.lua
> > new file mode 100644
> > index 000000000..8a23cc1fe
> > --- /dev/null
> > +++ b/test/replication/misc_replica_checks_cluster_id_gh-3704.test.lua
> > @@ -0,0 +1,26 @@
> > +test_run = require('test_run').new()
> > +test_run:cmd("restart server default")
> > +uuid = require('uuid')
> > +fiber = require('fiber')
> > +
> > +--
> > +-- gh-3704 move cluster id check to replica
> > +--
> > +test_run:cmd("create server replica with rpl_master=default, script='replication/replica.lua'")
> > +box.schema.user.grant("guest", "replication")
> > +test_run:cmd("start server replica")
> > +test_run:grep_log("replica", "REPLICASET_UUID_MISMATCH")
> > +box.info.replication[2].downstream.status
> > +-- change master's cluster uuid and check that replica doesn't connect.
> > +test_run:cmd("stop server replica")
> > +_ = box.space._schema:replace{'cluster', tostring(uuid.new())}
> > +-- master believes replica is in cluster, but their cluster UUIDs differ.
> > +test_run:cmd("start server replica")
> > +test_run:wait_log("replica", "REPLICASET_UUID_MISMATCH", nil, 1.0)
> > +test_run:wait_downstream(2, {status = 'stopped'})
> > +
> > +test_run:cmd("stop server replica")
> > +test_run:cmd("cleanup server replica")
> > +test_run:cmd("delete server replica")
> > +test_run:cleanup_cluster()
> > +box.schema.user.revoke('guest', 'replication')
> > diff --git a/test/replication/misc_return_on_quorum_0_gh-3760.result b/test/replication/misc_return_on_quorum_0_gh-3760.result
> > new file mode 100644
> > index 000000000..2eb622896
> > --- /dev/null
> > +++ b/test/replication/misc_return_on_quorum_0_gh-3760.result
> > @@ -0,0 +1,48 @@
> > +uuid = require('uuid')
> > +---
> > +...
> > +test_run = require('test_run').new()
> > +---
> > +...
> > +fiber = require('fiber')
> 
> Unneeded uuid and fiber
> 
> > +---
> > +...
> > +replication_connect_timeout = box.cfg.replication_connect_timeout
> > +---
> > +...
> > +replication_connect_quorum = box.cfg.replication_connect_quorum
> > +---
> > +...
> > +box.cfg{replication="12345", replication_connect_timeout=0.1, replication_connect_quorum=1}
> 
> This box.cfg call is extraneous. The test'll be just fine without it.
> 
> > +---
> > +...
> > +--
> > +-- gh-3760: replication quorum 0 on reconfiguration should return
> > +-- from box.cfg immediately.
> > +--
> > +replication = box.cfg.replication
> > +---
> > +...
> > +box.cfg{                                                        \
> > +    replication = {},                                           \
> > +    replication_connect_quorum = 0,                             \
> > +    replication_connect_timeout = 1000000                       \
> > +}
> > +---
> > +...
> > +-- The call below would hang, if quorum 0 is ignored, or checked
> > +-- too late.
> > +box.cfg{replication = {'localhost:12345'}}
> > +---
> > +...
> > +box.info.status
> > +---
> > +- running
> > +...
> > +box.cfg{                                                        \
> > +    replication = {},                                           \
> > +    replication_connect_quorum = replication_connect_quorum,    \
> > +    replication_connect_timeout = replication_connect_timeout   \
> > +}
> > +---
> > +...
> > diff --git a/test/replication/misc_return_on_quorum_0_gh-3760.test.lua b/test/replication/misc_return_on_quorum_0_gh-3760.test.lua
> > new file mode 100644
> > index 000000000..9e0651032
> > --- /dev/null
> > +++ b/test/replication/misc_return_on_quorum_0_gh-3760.test.lua
> > @@ -0,0 +1,27 @@
> > +uuid = require('uuid')
> > +test_run = require('test_run').new()
> > +fiber = require('fiber')
> > +
> > +replication_connect_timeout = box.cfg.replication_connect_timeout
> > +replication_connect_quorum = box.cfg.replication_connect_quorum
> > +box.cfg{replication="12345", replication_connect_timeout=0.1, replication_connect_quorum=1}
> > +
> > +--
> > +-- gh-3760: replication quorum 0 on reconfiguration should return
> > +-- from box.cfg immediately.
> > +--
> > +replication = box.cfg.replication
> > +box.cfg{                                                        \
> > +    replication = {},                                           \
> > +    replication_connect_quorum = 0,                             \
> > +    replication_connect_timeout = 1000000                       \
> > +}
> > +-- The call below would hang, if quorum 0 is ignored, or checked
> > +-- too late.
> > +box.cfg{replication = {'localhost:12345'}}
> > +box.info.status
> > +box.cfg{                                                        \
> > +    replication = {},                                           \
> > +    replication_connect_quorum = replication_connect_quorum,    \
> > +    replication_connect_timeout = replication_connect_timeout   \
> > +}
> > diff --git a/test/replication/misc_value_not_replicated_on_iproto_request_gh-3247.result b/test/replication/misc_value_not_replicated_on_iproto_request_gh-3247.result
> > new file mode 100644
> > index 000000000..d416bd9a6
> > --- /dev/null
> > +++ b/test/replication/misc_value_not_replicated_on_iproto_request_gh-3247.result
> > @@ -0,0 +1,90 @@
> > +test_run = require('test_run').new()
> > +---
> > +...
> > +test_run:cmd("restart server default")
> > +uuid = require('uuid')
> > +---
> > +...
> 
> You don't need  the uuid module in this testcase.
> 
> 
> > +-- Deploy a cluster.
> > +SERVERS = { 'autobootstrap1', 'autobootstrap2', 'autobootstrap3' }
> > +---
> > +...
> > +test_run:create_cluster(SERVERS, "replication", {args="0.03"})
> > +---
> > +...
> > +test_run:wait_fullmesh(SERVERS)
> > +---
> > +...
> > +-- gh-3247 - Sequence-generated value is not replicated in case
> > +-- the request was sent via iproto.
> > +test_run:cmd("switch autobootstrap1")
> > +---
> > +- true
> > +...
> > +net_box = require('net.box')
> > +---
> > +...
> > +_ = box.schema.space.create('space1')
> > +---
> > +...
> > +_ = box.schema.sequence.create('seq')
> > +---
> > +...
> > +_ = box.space.space1:create_index('primary', {sequence = true} )
> > +---
> > +...
> > +_ = box.space.space1:create_index('secondary', {parts = {2, 'unsigned'}})
> > +---
> > +...
> > +box.schema.user.grant('guest', 'read,write', 'space', 'space1')
> > +---
> > +...
> > +c = net_box.connect(box.cfg.listen)
> > +---
> > +...
> > +c.space.space1:insert{box.NULL, "data"} -- fails, but bumps sequence value
> > +---
> > +- error: 'Tuple field 2 type does not match one required by operation: expected unsigned'
> > +...
> > +c.space.space1:insert{box.NULL, 1, "data"}
> > +---
> > +- [2, 1, 'data']
> > +...
> > +box.space.space1:select{}
> > +---
> > +- - [2, 1, 'data']
> > +...
> > +vclock = test_run:get_vclock("autobootstrap1")
> > +---
> > +...
> > +vclock[0] = nil
> > +---
> > +...
> > +_ = test_run:wait_vclock("autobootstrap2", vclock)
> > +---
> > +...
> > +test_run:cmd("switch autobootstrap2")
> > +---
> > +- true
> > +...
> > +box.space.space1:select{}
> > +---
> > +- - [2, 1, 'data']
> > +...
> > +test_run:cmd("switch autobootstrap1")
> > +---
> > +- true
> > +...
> > +box.space.space1:drop()
> > +---
> > +...
> > +test_run:cmd("switch default")
> > +---
> > +- true
> > +...
> > +test_run:drop_cluster(SERVERS)
> > +---
> > +...
> > +test_run:cleanup_cluster()
> > +---
> > +...
> > diff --git a/test/replication/misc_value_not_replicated_on_iproto_request_gh-3247.test.lua b/test/replication/misc_value_not_replicated_on_iproto_request_gh-3247.test.lua
> > new file mode 100644
> > index 000000000..586e8b997
> > --- /dev/null
> > +++ b/test/replication/misc_value_not_replicated_on_iproto_request_gh-3247.test.lua
> > @@ -0,0 +1,33 @@
> > +test_run = require('test_run').new()
> > +test_run:cmd("restart server default")
> > +uuid = require('uuid')
> > +
> > +-- Deploy a cluster.
> > +SERVERS = { 'autobootstrap1', 'autobootstrap2', 'autobootstrap3' }
> > +test_run:create_cluster(SERVERS, "replication", {args="0.03"})
> > +test_run:wait_fullmesh(SERVERS)
> > +
> > +-- gh-3247 - Sequence-generated value is not replicated in case
> > +-- the request was sent via iproto.
> > +test_run:cmd("switch autobootstrap1")
> > +net_box = require('net.box')
> > +_ = box.schema.space.create('space1')
> > +_ = box.schema.sequence.create('seq')
> > +_ = box.space.space1:create_index('primary', {sequence = true} )
> > +_ = box.space.space1:create_index('secondary', {parts = {2, 'unsigned'}})
> > +box.schema.user.grant('guest', 'read,write', 'space', 'space1')
> > +c = net_box.connect(box.cfg.listen)
> > +c.space.space1:insert{box.NULL, "data"} -- fails, but bumps sequence value
> > +c.space.space1:insert{box.NULL, 1, "data"}
> > +box.space.space1:select{}
> > +vclock = test_run:get_vclock("autobootstrap1")
> > +vclock[0] = nil
> > +_ = test_run:wait_vclock("autobootstrap2", vclock)
> > +test_run:cmd("switch autobootstrap2")
> > +box.space.space1:select{}
> > +test_run:cmd("switch autobootstrap1")
> > +box.space.space1:drop()
> > +
> > +test_run:cmd("switch default")
> > +test_run:drop_cluster(SERVERS)
> > +test_run:cleanup_cluster()
> > diff --git a/test/replication/suite.cfg b/test/replication/suite.cfg
> > index f357b07da..e21daa5ad 100644
> > --- a/test/replication/suite.cfg
> > +++ b/test/replication/suite.cfg
> > @@ -1,6 +1,19 @@
> >   {
> >       "anon.test.lua": {},
> > -    "misc.test.lua": {},
> > +    "misc_assert_connecting_master_twice_gh-3610.test.lua": {},
> > +    "misc_assert_on_server_die_gh-2991.test.lua": {},
> > +    "misc_assert_replica_on_applier_disconnect_gh-3510.test.lua": {},
> > +    "misc_crash_on_box_concurrent_update_gh-3606.test.lua": {},
> > +    "misc_heartbeats_on_master_changes_gh-3160.test.lua": {},
> > +    "misc_no_failure_on_error_reading_wal_gh-4399.test.lua": {},
> > +    "misc_no_panic_on_connected_gh-3637.test.lua": {},
> > +    "misc_no_restart_on_same_configuration_gh-3711.test.lua": {},
> > +    "misc_no_socket_leak_on_replica_disconnect_gh-3642.test.lua": {},
> > +    "misc_orphan_on_reconfiguration_error_gh-4424.test.lua": {},
> > +    "misc_rebootstrap_from_ro_master_gh-3111.test.lua": {},
> > +    "misc_replica_checks_cluster_id_gh-3704.test.lua": {},
> > +    "misc_return_on_quorum_0_gh-3760.test.lua": {},
> > +    "misc_value_not_replicated_on_iproto_request_gh-3247.test.lua": {},
> >       "once.test.lua": {},
> >       "on_replace.test.lua": {},
> >       "status.test.lua": {},
> 
> -- 
> Serge Petrenko
> 

  reply	other threads:[~2020-09-02 10:08 UTC|newest]

Thread overview: 6+ messages / expand[flat|nested]  mbox.gz  Atom feed  top
2020-07-17  9:25 Alexander V. Tikhonov
2020-09-02  8:09 ` Serge Petrenko
2020-09-02 10:08   ` Alexander V. Tikhonov [this message]
2020-09-03  7:45     ` Serge Petrenko
2020-09-03 13:42       ` Alexander V. Tikhonov
2020-09-03 14:18         ` Serge Petrenko

Reply instructions:

You may reply publicly to this message via plain-text email
using any one of the following methods:

* Save the following mbox file, import it into your mail client,
  and reply-to-all from there: mbox

  Avoid top-posting and favor interleaved quoting:
  https://en.wikipedia.org/wiki/Posting_style#Interleaved_style

* Reply using the --to, --cc, and --in-reply-to
  switches of git-send-email(1):

  git send-email \
    --in-reply-to=20200902100837.GA22588@hpalx \
    --to=avtikhon@tarantool.org \
    --cc=imeevma@gmail.com \
    --cc=sergepetrenko@tarantool.org \
    --cc=tarantool-patches@dev.tarantool.org \
    --subject='Re: [Tarantool-patches] [PATCH v1] Divide replication/mist.test.lua' \
    /path/to/YOUR_REPLY

  https://kernel.org/pub/software/scm/git/docs/git-send-email.html

* If your mail client supports setting the In-Reply-To header
  via mailto: links, try the mailto: link

This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox