[Tarantool-patches] [PATCH v1] Divide replication/mist.test.lua
Serge Petrenko
sergepetrenko at tarantool.org
Thu Sep 3 10:45:17 MSK 2020
02.09.2020 13:08, Alexander V. Tikhonov пишет:
> Hi Sergey, thanks for the review. The major comment on tests naming
> that it was discussed during fixes in issue [1], and decided to use
> this naming format, check as an example of it with:
>
> ls -al test/box/net.box*.test.lua
>
> [1] - https://github.com/tarantool/tarantool/issues/4880
Hi, Alex!
Strange. I was guided by our earlier decision and replication tests:
ls test/replication/gh-*.test.lua
test/replication/gh-4114-local-space-replication.test.lua
test/replication/gh-4402-info-errno.test.lua
test/replication/gh-4605-empty-password.test.lua
test/replication/gh-4606-admin-creds.test.lua
test/replication/gh-4729-netbox-group-id.test.lua
test/replication/gh-4730-applier-rollback.test.lua
test/replication/gh-4739-vclock-assert.test.lua
test/replication/gh-4928-tx-boundaries.test.lua
test/replication/gh-5140-qsync-casc-rollback.test.lua
test/replication/gh-5144-qsync-dup-confirm.test.lua
test/replication/gh-5163-qsync-restart-crash.test.lua
test/replication/gh-5167-qsync-rollback-snap.test.lua
test/replication/gh-5195-qsync-replica-write.test.lua
Anyway, if someone can confirm that we're using another naming
scheme now, I'm ok with it.
>
> On Wed, Sep 02, 2020 at 11:09:30AM +0300, Serge Petrenko wrote:
>> Hi! Thanks for the patch!
>>
>> Please see my comments below.
>>
>> 17.07.2020 12:25, Alexander V. Tikhonov пишет:
>>> To fix flaky issues of replication/misc.test.lua the test had to be
>>> divided into smaller tests to be able to localize the flaky results:
>>>
>>> misc_assert_connecting_master_twice_gh-3610.test.lua
>>> misc_assert_on_server_die_gh-2991.test.lua
>>> misc_assert_replica_on_applier_disconnect_gh-3510.test.lua
>>> misc_crash_on_box_concurrent_update_gh-3606.test.lua
>>> misc_heartbeats_on_master_changes_gh-3160.test.lua
>>> misc_no_failure_on_error_reading_wal_gh-4399.test.lua
>>> misc_no_panic_on_connected_gh-3637.test.lua
>>> misc_no_restart_on_same_configuration_gh-3711.test.lua
>>> misc_no_socket_leak_on_replica_disconnect_gh-3642.test.lua
>>> misc_orphan_on_reconfiguration_error_gh-4424.test.lua
>>> misc_rebootstrap_from_ro_master_gh-3111.test.lua
>>> misc_replica_checks_cluster_id_gh-3704.test.lua
>>> misc_return_on_quorum_0_gh-3760.test.lua
>>> misc_value_not_replicated_on_iproto_request_gh-3247.test.lua
>> We use a different naming scheme for regression tests.
>> It should be `gh-xxxx-bug-description`, also notice the dashes
>> between words instead of the underscores.
>>
>>> Needed for #4940
>>> ---
>>>
>>> Github: https://github.com/tarantool/tarantool/tree/avtikhon/gh-4940-replication-misc
>>> Issue: https://github.com/tarantool/tarantool/issues/4940
>>>
>>> test/replication/misc.result | 866 ------------------
>>> test/replication/misc.test.lua | 356 -------
>>> ...ert_connecting_master_twice_gh-3610.result | 86 ++
>>> ...t_connecting_master_twice_gh-3610.test.lua | 34 +
>>> .../misc_assert_on_server_die_gh-2991.result | 31 +
>>> ...misc_assert_on_server_die_gh-2991.test.lua | 12 +
>>> ...plica_on_applier_disconnect_gh-3510.result | 49 +
>>> ...ica_on_applier_disconnect_gh-3510.test.lua | 17 +
>>> ...sh_on_box_concurrent_update_gh-3606.result | 50 +
>>> ..._on_box_concurrent_update_gh-3606.test.lua | 19 +
>>> ...eartbeats_on_master_changes_gh-3160.result | 76 ++
>>> ...rtbeats_on_master_changes_gh-3160.test.lua | 40 +
>>> ...ailure_on_error_reading_wal_gh-4399.result | 97 ++
>>> ...lure_on_error_reading_wal_gh-4399.test.lua | 39 +
>>> .../misc_no_panic_on_connected_gh-3637.result | 72 ++
>>> ...isc_no_panic_on_connected_gh-3637.test.lua | 33 +
>>> ...start_on_same_configuration_gh-3711.result | 107 +++
>>> ...art_on_same_configuration_gh-3711.test.lua | 41 +
>>> ..._leak_on_replica_disconnect_gh-3642.result | 98 ++
>>> ...eak_on_replica_disconnect_gh-3642.test.lua | 44 +
>>> ...an_on_reconfiguration_error_gh-4424.result | 88 ++
>>> ..._on_reconfiguration_error_gh-4424.test.lua | 37 +
>>> ..._rebootstrap_from_ro_master_gh-3111.result | 58 ++
>>> ...ebootstrap_from_ro_master_gh-3111.test.lua | 20 +
>>> ...c_replica_checks_cluster_id_gh-3704.result | 71 ++
>>> ...replica_checks_cluster_id_gh-3704.test.lua | 26 +
>>> .../misc_return_on_quorum_0_gh-3760.result | 48 +
>>> .../misc_return_on_quorum_0_gh-3760.test.lua | 27 +
>>> ...eplicated_on_iproto_request_gh-3247.result | 90 ++
>>> ...licated_on_iproto_request_gh-3247.test.lua | 33 +
>>> test/replication/suite.cfg | 15 +-
>>> 31 files changed, 1457 insertions(+), 1223 deletions(-)
>>> delete mode 100644 test/replication/misc.result
>>> delete mode 100644 test/replication/misc.test.lua
>>> create mode 100644 test/replication/misc_assert_connecting_master_twice_gh-3610.result
>>> create mode 100644 test/replication/misc_assert_connecting_master_twice_gh-3610.test.lua
>>> create mode 100644 test/replication/misc_assert_on_server_die_gh-2991.result
>>> create mode 100644 test/replication/misc_assert_on_server_die_gh-2991.test.lua
>>> create mode 100644 test/replication/misc_assert_replica_on_applier_disconnect_gh-3510.result
>>> create mode 100644 test/replication/misc_assert_replica_on_applier_disconnect_gh-3510.test.lua
>>> create mode 100644 test/replication/misc_crash_on_box_concurrent_update_gh-3606.result
>>> create mode 100644 test/replication/misc_crash_on_box_concurrent_update_gh-3606.test.lua
>>> create mode 100644 test/replication/misc_heartbeats_on_master_changes_gh-3160.result
>>> create mode 100644 test/replication/misc_heartbeats_on_master_changes_gh-3160.test.lua
>>> create mode 100644 test/replication/misc_no_failure_on_error_reading_wal_gh-4399.result
>>> create mode 100644 test/replication/misc_no_failure_on_error_reading_wal_gh-4399.test.lua
>>> create mode 100644 test/replication/misc_no_panic_on_connected_gh-3637.result
>>> create mode 100644 test/replication/misc_no_panic_on_connected_gh-3637.test.lua
>>> create mode 100644 test/replication/misc_no_restart_on_same_configuration_gh-3711.result
>>> create mode 100644 test/replication/misc_no_restart_on_same_configuration_gh-3711.test.lua
>>> create mode 100644 test/replication/misc_no_socket_leak_on_replica_disconnect_gh-3642.result
>>> create mode 100644 test/replication/misc_no_socket_leak_on_replica_disconnect_gh-3642.test.lua
>>> create mode 100644 test/replication/misc_orphan_on_reconfiguration_error_gh-4424.result
>>> create mode 100644 test/replication/misc_orphan_on_reconfiguration_error_gh-4424.test.lua
>>> create mode 100644 test/replication/misc_rebootstrap_from_ro_master_gh-3111.result
>>> create mode 100644 test/replication/misc_rebootstrap_from_ro_master_gh-3111.test.lua
>>> create mode 100644 test/replication/misc_replica_checks_cluster_id_gh-3704.result
>>> create mode 100644 test/replication/misc_replica_checks_cluster_id_gh-3704.test.lua
>>> create mode 100644 test/replication/misc_return_on_quorum_0_gh-3760.result
>>> create mode 100644 test/replication/misc_return_on_quorum_0_gh-3760.test.lua
>>> create mode 100644 test/replication/misc_value_not_replicated_on_iproto_request_gh-3247.result
>>> create mode 100644 test/replication/misc_value_not_replicated_on_iproto_request_gh-3247.test.lua
>>>
>>> diff --git a/test/replication/misc.result b/test/replication/misc.result
>>> deleted file mode 100644
>>> index e5d1f560e..000000000
>>> --- a/test/replication/misc.result
>>> +++ /dev/null
>>> @@ -1,866 +0,0 @@
>>> -uuid = require('uuid')
>>> ----
>>> -...
>>> -test_run = require('test_run').new()
>>> ----
>>> -...
>>> -box.schema.user.grant('guest', 'replication')
>>> ----
>>> -...
>>> --- gh-2991 - Tarantool asserts on box.cfg.replication update if one of
>>> --- servers is dead
>>> -replication_timeout = box.cfg.replication_timeout
>>> ----
>>> -...
>>> -replication_connect_timeout = box.cfg.replication_connect_timeout
>>> ----
>>> -...
>>> -box.cfg{replication_timeout=0.05, replication_connect_timeout=0.05, replication={}}
>>> ----
>>> -...
>>> -box.cfg{replication_connect_quorum=2}
>>> ----
>>> -...
>>> -box.cfg{replication = {'127.0.0.1:12345', box.cfg.listen}}
>>> ----
>>> -...
>>> -box.info.status
>>> ----
>>> -- orphan
>>> -...
>>> -box.info.ro
>>> ----
>>> -- true
>>> -...
>>> --- gh-3606 - Tarantool crashes if box.cfg.replication is updated concurrently
>>> -fiber = require('fiber')
>>> ----
>>> -...
>>> -c = fiber.channel(2)
>>> ----
>>> -...
>>> -f = function() fiber.create(function() pcall(box.cfg, {replication = {12345}}) c:put(true) end) end
>>> ----
>>> -...
>>> -f()
>>> ----
>>> -...
>>> -f()
>>> ----
>>> -...
>>> -c:get()
>>> ----
>>> -- true
>>> -...
>>> -c:get()
>>> ----
>>> -- true
>>> -...
>>> -box.cfg{replication = "", replication_timeout = replication_timeout, replication_connect_timeout = replication_connect_timeout}
>>> ----
>>> -...
>>> -box.info.status
>>> ----
>>> -- running
>>> -...
>>> -box.info.ro
>>> ----
>>> -- false
>>> -...
>>> --- gh-3111 - Allow to rebootstrap a replica from a read-only master
>>> -replica_uuid = uuid.new()
>>> ----
>>> -...
>>> -test_run:cmd('create server test with rpl_master=default, script="replication/replica_uuid.lua"')
>>> ----
>>> -- true
>>> -...
>>> -test_run:cmd(string.format('start server test with args="%s"', replica_uuid))
>>> ----
>>> -- true
>>> -...
>>> -test_run:cmd('stop server test')
>>> ----
>>> -- true
>>> -...
>>> -test_run:cmd('cleanup server test')
>>> ----
>>> -- true
>>> -...
>>> -box.cfg{read_only = true}
>>> ----
>>> -...
>>> -test_run:cmd(string.format('start server test with args="%s"', replica_uuid))
>>> ----
>>> -- true
>>> -...
>>> -test_run:cmd('stop server test')
>>> ----
>>> -- true
>>> -...
>>> -test_run:cmd('cleanup server test')
>>> ----
>>> -- true
>>> -...
>>> -box.cfg{read_only = false}
>>> ----
>>> -...
>>> -test_run:cmd('delete server test')
>>> ----
>>> -- true
>>> -...
>>> -test_run:cleanup_cluster()
>>> ----
>>> -...
>>> --- gh-3160 - Send heartbeats if there are changes from a remote master only
>>> -SERVERS = { 'autobootstrap1', 'autobootstrap2', 'autobootstrap3' }
>>> ----
>>> -...
>>> --- Deploy a cluster.
>>> -test_run:create_cluster(SERVERS, "replication", {args="0.03"})
>>> ----
>>> -...
>>> -test_run:wait_fullmesh(SERVERS)
>>> ----
>>> -...
>>> -test_run:cmd("switch autobootstrap3")
>>> ----
>>> -- true
>>> -...
>>> -test_run = require('test_run').new()
>>> ----
>>> -...
>>> -fiber = require('fiber')
>>> ----
>>> -...
>>> -_ = box.schema.space.create('test_timeout'):create_index('pk')
>>> ----
>>> -...
>>> -test_run:cmd("setopt delimiter ';'")
>>> ----
>>> -- true
>>> -...
>>> -function wait_not_follow(replicaA, replicaB)
>>> - return test_run:wait_cond(function()
>>> - return replicaA.status ~= 'follow' or replicaB.status ~= 'follow'
>>> - end, box.cfg.replication_timeout)
>>> -end;
>>> ----
>>> -...
>>> -function test_timeout()
>>> - local replicaA = box.info.replication[1].upstream or box.info.replication[2].upstream
>>> - local replicaB = box.info.replication[3].upstream or box.info.replication[2].upstream
>>> - local follows = test_run:wait_cond(function()
>>> - return replicaA.status == 'follow' or replicaB.status == 'follow'
>>> - end)
>>> - if not follows then error('replicas are not in the follow status') end
>>> - for i = 0, 99 do
>>> - box.space.test_timeout:replace({1})
>>> - if wait_not_follow(replicaA, replicaB) then
>>> - return error(box.info.replication)
>>> - end
>>> - end
>>> - return true
>>> -end;
>>> ----
>>> -...
>>> -test_run:cmd("setopt delimiter ''");
>>> ----
>>> -- true
>>> -...
>>> -test_timeout()
>>> ----
>>> -- true
>>> -...
>>> --- gh-3247 - Sequence-generated value is not replicated in case
>>> --- the request was sent via iproto.
>>> -test_run:cmd("switch autobootstrap1")
>>> ----
>>> -- true
>>> -...
>>> -net_box = require('net.box')
>>> ----
>>> -...
>>> -_ = box.schema.space.create('space1')
>>> ----
>>> -...
>>> -_ = box.schema.sequence.create('seq')
>>> ----
>>> -...
>>> -_ = box.space.space1:create_index('primary', {sequence = true} )
>>> ----
>>> -...
>>> -_ = box.space.space1:create_index('secondary', {parts = {2, 'unsigned'}})
>>> ----
>>> -...
>>> -box.schema.user.grant('guest', 'read,write', 'space', 'space1')
>>> ----
>>> -...
>>> -c = net_box.connect(box.cfg.listen)
>>> ----
>>> -...
>>> -c.space.space1:insert{box.NULL, "data"} -- fails, but bumps sequence value
>>> ----
>>> -- error: 'Tuple field 2 type does not match one required by operation: expected unsigned'
>>> -...
>>> -c.space.space1:insert{box.NULL, 1, "data"}
>>> ----
>>> -- [2, 1, 'data']
>>> -...
>>> -box.space.space1:select{}
>>> ----
>>> -- - [2, 1, 'data']
>>> -...
>>> -vclock = test_run:get_vclock("autobootstrap1")
>>> ----
>>> -...
>>> -vclock[0] = nil
>>> ----
>>> -...
>>> -_ = test_run:wait_vclock("autobootstrap2", vclock)
>>> ----
>>> -...
>>> -test_run:cmd("switch autobootstrap2")
>>> ----
>>> -- true
>>> -...
>>> -box.space.space1:select{}
>>> ----
>>> -- - [2, 1, 'data']
>>> -...
>>> -test_run:cmd("switch autobootstrap1")
>>> ----
>>> -- true
>>> -...
>>> -box.space.space1:drop()
>>> ----
>>> -...
>>> -test_run:cmd("switch default")
>>> ----
>>> -- true
>>> -...
>>> -test_run:drop_cluster(SERVERS)
>>> ----
>>> -...
>>> -test_run:cleanup_cluster()
>>> ----
>>> -...
>>> --- gh-3642 - Check that socket file descriptor doesn't leak
>>> --- when a replica is disconnected.
>>> -rlimit = require('rlimit')
>>> ----
>>> -...
>>> -lim = rlimit.limit()
>>> ----
>>> -...
>>> -rlimit.getrlimit(rlimit.RLIMIT_NOFILE, lim)
>>> ----
>>> -...
>>> -old_fno = lim.rlim_cur
>>> ----
>>> -...
>>> -lim.rlim_cur = 64
>>> ----
>>> -...
>>> -rlimit.setrlimit(rlimit.RLIMIT_NOFILE, lim)
>>> ----
>>> -...
>>> -test_run:cmd('create server sock with rpl_master=default, script="replication/replica.lua"')
>>> ----
>>> -- true
>>> -...
>>> -test_run:cmd('start server sock')
>>> ----
>>> -- true
>>> -...
>>> -test_run:cmd('switch sock')
>>> ----
>>> -- true
>>> -...
>>> -test_run = require('test_run').new()
>>> ----
>>> -...
>>> -fiber = require('fiber')
>>> ----
>>> -...
>>> -test_run:cmd("setopt delimiter ';'")
>>> ----
>>> -- true
>>> -...
>>> -for i = 1, 64 do
>>> - local replication = box.cfg.replication
>>> - box.cfg{replication = {}}
>>> - box.cfg{replication = replication}
>>> - while box.info.replication[1].upstream.status ~= 'follow' do
>>> - fiber.sleep(0.001)
>>> - end
>>> -end;
>>> ----
>>> -...
>>> -test_run:cmd("setopt delimiter ''");
>>> ----
>>> -- true
>>> -...
>>> -box.info.replication[1].upstream.status
>>> ----
>>> -- follow
>>> -...
>>> -test_run:cmd('switch default')
>>> ----
>>> -- true
>>> -...
>>> -lim.rlim_cur = old_fno
>>> ----
>>> -...
>>> -rlimit.setrlimit(rlimit.RLIMIT_NOFILE, lim)
>>> ----
>>> -...
>>> -test_run:cmd("stop server sock")
>>> ----
>>> -- true
>>> -...
>>> -test_run:cmd("cleanup server sock")
>>> ----
>>> -- true
>>> -...
>>> -test_run:cmd("delete server sock")
>>> ----
>>> -- true
>>> -...
>>> -test_run:cleanup_cluster()
>>> ----
>>> -...
>>> -box.schema.user.revoke('guest', 'replication')
>>> ----
>>> -...
>>> --- gh-3510 assertion failure in replica_on_applier_disconnect()
>>> -test_run:cmd('create server er_load1 with script="replication/er_load1.lua"')
>>> ----
>>> -- true
>>> -...
>>> -test_run:cmd('create server er_load2 with script="replication/er_load2.lua"')
>>> ----
>>> -- true
>>> -...
>>> -test_run:cmd('start server er_load1 with wait=False, wait_load=False')
>>> ----
>>> -- true
>>> -...
>>> --- Instance er_load2 will fail with error ER_REPLICASET_UUID_MISMATCH.
>>> --- This is OK since we only test here that er_load1 doesn't assert.
>>> -test_run:cmd('start server er_load2 with wait=True, wait_load=True, crash_expected = True')
>>> ----
>>> -- false
>>> -...
>>> -test_run:cmd('stop server er_load1')
>>> ----
>>> -- true
>>> -...
>>> --- er_load2 exits automatically.
>>> -test_run:cmd('cleanup server er_load1')
>>> ----
>>> -- true
>>> -...
>>> -test_run:cmd('cleanup server er_load2')
>>> ----
>>> -- true
>>> -...
>>> -test_run:cmd('delete server er_load1')
>>> ----
>>> -- true
>>> -...
>>> -test_run:cmd('delete server er_load2')
>>> ----
>>> -- true
>>> -...
>>> -test_run:cleanup_cluster()
>>> ----
>>> -...
>>> ---
>>> --- Test case for gh-3637, gh-4550. Before the fix replica would
>>> --- exit with an error if a user does not exist or a password is
>>> --- incorrect. Now check that we don't hang/panic and successfully
>>> --- connect.
>>> ---
>>> -fiber = require('fiber')
>>> ----
>>> -...
>>> -test_run:cmd("create server replica_auth with rpl_master=default, script='replication/replica_auth.lua'")
>>> ----
>>> -- true
>>> -...
>>> -test_run:cmd("start server replica_auth with wait=False, wait_load=False, args='cluster:pass 0.05'")
>>> ----
>>> -- true
>>> -...
>>> --- Wait a bit to make sure replica waits till user is created.
>>> -fiber.sleep(0.1)
>>> ----
>>> -...
>>> -box.schema.user.create('cluster')
>>> ----
>>> -...
>>> --- The user is created. Let the replica fail auth request due to
>>> --- a wrong password.
>>> -fiber.sleep(0.1)
>>> ----
>>> -...
>>> -box.schema.user.passwd('cluster', 'pass')
>>> ----
>>> -...
>>> -box.schema.user.grant('cluster', 'replication')
>>> ----
>>> -...
>>> -while box.info.replication[2] == nil do fiber.sleep(0.01) end
>>> ----
>>> -...
>>> -vclock = test_run:get_vclock('default')
>>> ----
>>> -...
>>> -vclock[0] = nil
>>> ----
>>> -...
>>> -_ = test_run:wait_vclock('replica_auth', vclock)
>>> ----
>>> -...
>>> -test_run:cmd("stop server replica_auth")
>>> ----
>>> -- true
>>> -...
>>> -test_run:cmd("cleanup server replica_auth")
>>> ----
>>> -- true
>>> -...
>>> -test_run:cmd("delete server replica_auth")
>>> ----
>>> -- true
>>> -...
>>> -test_run:cleanup_cluster()
>>> ----
>>> -...
>>> -box.schema.user.drop('cluster')
>>> ----
>>> -...
>>> ---
>>> --- Test case for gh-3610. Before the fix replica would fail with the assertion
>>> --- when trying to connect to the same master twice.
>>> ---
>>> -box.schema.user.grant('guest', 'replication')
>>> ----
>>> -...
>>> -test_run:cmd("create server replica with rpl_master=default, script='replication/replica.lua'")
>>> ----
>>> -- true
>>> -...
>>> -test_run:cmd("start server replica")
>>> ----
>>> -- true
>>> -...
>>> -test_run:cmd("switch replica")
>>> ----
>>> -- true
>>> -...
>>> -replication = box.cfg.replication[1]
>>> ----
>>> -...
>>> -box.cfg{replication = {replication, replication}}
>>> ----
>>> -- error: 'Incorrect value for option ''replication'': duplicate connection to the
>>> - same replica'
>>> -...
>>> --- Check the case when duplicate connection is detected in the background.
>>> -test_run:cmd("switch default")
>>> ----
>>> -- true
>>> -...
>>> -listen = box.cfg.listen
>>> ----
>>> -...
>>> -box.cfg{listen = ''}
>>> ----
>>> -...
>>> -test_run:cmd("switch replica")
>>> ----
>>> -- true
>>> -...
>>> -box.cfg{replication_connect_quorum = 0, replication_connect_timeout = 0.01}
>>> ----
>>> -...
>>> -box.cfg{replication = {replication, replication}}
>>> ----
>>> -...
>>> -test_run:cmd("switch default")
>>> ----
>>> -- true
>>> -...
>>> -box.cfg{listen = listen}
>>> ----
>>> -...
>>> -while test_run:grep_log('replica', 'duplicate connection') == nil do fiber.sleep(0.01) end
>>> ----
>>> -...
>>> -test_run:cmd("stop server replica")
>>> ----
>>> -- true
>>> -...
>>> -test_run:cmd("cleanup server replica")
>>> ----
>>> -- true
>>> -...
>>> -test_run:cmd("delete server replica")
>>> ----
>>> -- true
>>> -...
>>> -test_run:cleanup_cluster()
>>> ----
>>> -...
>>> -box.schema.user.revoke('guest', 'replication')
>>> ----
>>> -...
>>> ---
>>> --- gh-3711 Do not restart replication on box.cfg in case the
>>> --- configuration didn't change.
>>> ---
>>> -box.schema.user.grant('guest', 'replication')
>>> ----
>>> -...
>>> -test_run:cmd("create server replica with rpl_master=default, script='replication/replica.lua'")
>>> ----
>>> -- true
>>> -...
>>> -test_run:cmd("start server replica")
>>> ----
>>> -- true
>>> -...
>>> --- Access rights are checked only during reconnect. If the new
>>> --- and old configurations are equivalent, no reconnect will be
>>> --- issued and replication should continue working.
>>> -box.schema.user.revoke('guest', 'replication')
>>> ----
>>> -...
>>> -test_run:cmd("switch replica")
>>> ----
>>> -- true
>>> -...
>>> -replication = box.cfg.replication[1]
>>> ----
>>> -...
>>> -box.cfg{replication = {replication}}
>>> ----
>>> -...
>>> -box.info.status == 'running'
>>> ----
>>> -- true
>>> -...
>>> -box.cfg{replication = replication}
>>> ----
>>> -...
>>> -box.info.status == 'running'
>>> ----
>>> -- true
>>> -...
>>> --- Check that comparison of tables works as expected as well.
>>> -test_run:cmd("switch default")
>>> ----
>>> -- true
>>> -...
>>> -box.schema.user.grant('guest', 'replication')
>>> ----
>>> -...
>>> -test_run:cmd("switch replica")
>>> ----
>>> -- true
>>> -...
>>> -replication = box.cfg.replication
>>> ----
>>> -...
>>> -table.insert(replication, box.cfg.listen)
>>> ----
>>> -...
>>> -test_run:cmd("switch default")
>>> ----
>>> -- true
>>> -...
>>> -box.schema.user.revoke('guest', 'replication')
>>> ----
>>> -...
>>> -test_run:cmd("switch replica")
>>> ----
>>> -- true
>>> -...
>>> -box.cfg{replication = replication}
>>> ----
>>> -...
>>> -box.info.status == 'running'
>>> ----
>>> -- true
>>> -...
>>> -test_run:cmd("switch default")
>>> ----
>>> -- true
>>> -...
>>> -test_run:cmd("stop server replica")
>>> ----
>>> -- true
>>> -...
>>> -test_run:cmd("cleanup server replica")
>>> ----
>>> -- true
>>> -...
>>> -test_run:cmd("delete server replica")
>>> ----
>>> -- true
>>> -...
>>> -test_run:cleanup_cluster()
>>> ----
>>> -...
>>> ---
>>> --- gh-3704 move cluster id check to replica
>>> ---
>>> -test_run:cmd("create server replica with rpl_master=default, script='replication/replica.lua'")
>>> ----
>>> -- true
>>> -...
>>> -box.schema.user.grant("guest", "replication")
>>> ----
>>> -...
>>> -test_run:cmd("start server replica")
>>> ----
>>> -- true
>>> -...
>>> -test_run:grep_log("replica", "REPLICASET_UUID_MISMATCH")
>>> ----
>>> -- null
>>> -...
>>> -box.info.replication[2].downstream.status
>>> ----
>>> -- follow
>>> -...
>>> --- change master's cluster uuid and check that replica doesn't connect.
>>> -test_run:cmd("stop server replica")
>>> ----
>>> -- true
>>> -...
>>> -_ = box.space._schema:replace{'cluster', tostring(uuid.new())}
>>> ----
>>> -...
>>> --- master believes replica is in cluster, but their cluster UUIDs differ.
>>> -test_run:cmd("start server replica")
>>> ----
>>> -- true
>>> -...
>>> -test_run:wait_log("replica", "REPLICASET_UUID_MISMATCH", nil, 1.0)
>>> ----
>>> -- REPLICASET_UUID_MISMATCH
>>> -...
>>> -test_run:wait_downstream(2, {status = 'stopped'})
>>> ----
>>> -- true
>>> -...
>>> -test_run:cmd("stop server replica")
>>> ----
>>> -- true
>>> -...
>>> -test_run:cmd("cleanup server replica")
>>> ----
>>> -- true
>>> -...
>>> -test_run:cmd("delete server replica")
>>> ----
>>> -- true
>>> -...
>>> -test_run:cleanup_cluster()
>>> ----
>>> -...
>>> -box.schema.user.revoke('guest', 'replication')
>>> ----
>>> -...
>>> ---
>>> --- gh-4399 Check that an error reading WAL directory on subscribe
>>> --- doesn't lead to a permanent replication failure.
>>> ---
>>> -box.schema.user.grant("guest", "replication")
>>> ----
>>> -...
>>> -test_run:cmd("create server replica with rpl_master=default, script='replication/replica.lua'")
>>> ----
>>> -- true
>>> -...
>>> -test_run:cmd("start server replica")
>>> ----
>>> -- true
>>> -...
>>> --- Make the WAL directory inaccessible.
>>> -fio = require('fio')
>>> ----
>>> -...
>>> -path = fio.abspath(box.cfg.wal_dir)
>>> ----
>>> -...
>>> -fio.chmod(path, 0)
>>> ----
>>> -- true
>>> -...
>>> --- Break replication on timeout.
>>> -replication_timeout = box.cfg.replication_timeout
>>> ----
>>> -...
>>> -box.cfg{replication_timeout = 9000}
>>> ----
>>> -...
>>> -test_run:cmd("switch replica")
>>> ----
>>> -- true
>>> -...
>>> -test_run:wait_cond(function() return box.info.replication[1].upstream.status ~= 'follow' end)
>>> ----
>>> -- true
>>> -...
>>> -require('fiber').sleep(box.cfg.replication_timeout)
>>> ----
>>> -...
>>> -test_run:cmd("switch default")
>>> ----
>>> -- true
>>> -...
>>> -box.cfg{replication_timeout = replication_timeout}
>>> ----
>>> -...
>>> --- Restore access to the WAL directory.
>>> --- Wait for replication to be reestablished.
>>> -fio.chmod(path, tonumber('777', 8))
>>> ----
>>> -- true
>>> -...
>>> -test_run:cmd("switch replica")
>>> ----
>>> -- true
>>> -...
>>> -test_run:wait_cond(function() return box.info.replication[1].upstream.status == 'follow' end)
>>> ----
>>> -- true
>>> -...
>>> -test_run:cmd("switch default")
>>> ----
>>> -- true
>>> -...
>>> -test_run:cmd("stop server replica")
>>> ----
>>> -- true
>>> -...
>>> -test_run:cmd("cleanup server replica")
>>> ----
>>> -- true
>>> -...
>>> -test_run:cmd("delete server replica")
>>> ----
>>> -- true
>>> -...
>>> -test_run:cleanup_cluster()
>>> ----
>>> -...
>>> -box.schema.user.revoke('guest', 'replication')
>>> ----
>>> -...
>>> ---
>>> --- gh-4424 Always enter orphan mode on error in replication
>>> --- configuration change.
>>> ---
>>> -replication_connect_timeout = box.cfg.replication_connect_timeout
>>> ----
>>> -...
>>> -replication_connect_quorum = box.cfg.replication_connect_quorum
>>> ----
>>> -...
>>> -box.cfg{replication="12345", replication_connect_timeout=0.1, replication_connect_quorum=1}
>>> ----
>>> -...
>>> -box.info.status
>>> ----
>>> -- orphan
>>> -...
>>> -box.info.ro
>>> ----
>>> -- true
>>> -...
>>> --- reset replication => leave orphan mode
>>> -box.cfg{replication=""}
>>> ----
>>> -...
>>> -box.info.status
>>> ----
>>> -- running
>>> -...
>>> -box.info.ro
>>> ----
>>> -- false
>>> -...
>>> --- no switch to orphan when quorum == 0
>>> -box.cfg{replication="12345", replication_connect_quorum=0}
>>> ----
>>> -...
>>> -box.info.status
>>> ----
>>> -- running
>>> -...
>>> -box.info.ro
>>> ----
>>> -- false
>>> -...
>>> --- we could connect to one out of two replicas. Set orphan.
>>> -box.cfg{replication_connect_quorum=2}
>>> ----
>>> -...
>>> -box.cfg{replication={box.cfg.listen, "12345"}}
>>> ----
>>> -...
>>> -box.info.status
>>> ----
>>> -- orphan
>>> -...
>>> -box.info.ro
>>> ----
>>> -- true
>>> -...
>>> --- lower quorum => leave orphan mode
>>> -box.cfg{replication_connect_quorum=1}
>>> ----
>>> -...
>>> -box.info.status
>>> ----
>>> -- running
>>> -...
>>> -box.info.ro
>>> ----
>>> -- false
>>> -...
>>> ---
>>> --- gh-3760: replication quorum 0 on reconfiguration should return
>>> --- from box.cfg immediately.
>>> ---
>>> -replication = box.cfg.replication
>>> ----
>>> -...
>>> -box.cfg{ \
>>> - replication = {}, \
>>> - replication_connect_quorum = 0, \
>>> - replication_connect_timeout = 1000000 \
>>> -}
>>> ----
>>> -...
>>> --- The call below would hang, if quorum 0 is ignored, or checked
>>> --- too late.
>>> -box.cfg{replication = {'localhost:12345'}}
>>> ----
>>> -...
>>> -box.info.status
>>> ----
>>> -- running
>>> -...
>>> -box.cfg{ \
>>> - replication = {}, \
>>> - replication_connect_quorum = replication_connect_quorum, \
>>> - replication_connect_timeout = replication_connect_timeout \
>>> -}
>>> ----
>>> -...
>>> diff --git a/test/replication/misc.test.lua b/test/replication/misc.test.lua
>>> deleted file mode 100644
>>> index d285b014a..000000000
>>> --- a/test/replication/misc.test.lua
>>> +++ /dev/null
>>> @@ -1,356 +0,0 @@
>>> -uuid = require('uuid')
>>> -test_run = require('test_run').new()
>>> -
>>> -box.schema.user.grant('guest', 'replication')
>>> -
>>> --- gh-2991 - Tarantool asserts on box.cfg.replication update if one of
>>> --- servers is dead
>>> -replication_timeout = box.cfg.replication_timeout
>>> -replication_connect_timeout = box.cfg.replication_connect_timeout
>>> -box.cfg{replication_timeout=0.05, replication_connect_timeout=0.05, replication={}}
>>> -box.cfg{replication_connect_quorum=2}
>>> -box.cfg{replication = {'127.0.0.1:12345', box.cfg.listen}}
>>> -box.info.status
>>> -box.info.ro
>>> -
>>> --- gh-3606 - Tarantool crashes if box.cfg.replication is updated concurrently
>>> -fiber = require('fiber')
>>> -c = fiber.channel(2)
>>> -f = function() fiber.create(function() pcall(box.cfg, {replication = {12345}}) c:put(true) end) end
>>> -f()
>>> -f()
>>> -c:get()
>>> -c:get()
>>> -
>>> -box.cfg{replication = "", replication_timeout = replication_timeout, replication_connect_timeout = replication_connect_timeout}
>>> -box.info.status
>>> -box.info.ro
>>> -
>>> --- gh-3111 - Allow to rebootstrap a replica from a read-only master
>>> -replica_uuid = uuid.new()
>>> -test_run:cmd('create server test with rpl_master=default, script="replication/replica_uuid.lua"')
>>> -test_run:cmd(string.format('start server test with args="%s"', replica_uuid))
>>> -test_run:cmd('stop server test')
>>> -test_run:cmd('cleanup server test')
>>> -box.cfg{read_only = true}
>>> -test_run:cmd(string.format('start server test with args="%s"', replica_uuid))
>>> -test_run:cmd('stop server test')
>>> -test_run:cmd('cleanup server test')
>>> -box.cfg{read_only = false}
>>> -test_run:cmd('delete server test')
>>> -test_run:cleanup_cluster()
>>> -
>>> --- gh-3160 - Send heartbeats if there are changes from a remote master only
>>> -SERVERS = { 'autobootstrap1', 'autobootstrap2', 'autobootstrap3' }
>>> -
>>> --- Deploy a cluster.
>>> -test_run:create_cluster(SERVERS, "replication", {args="0.03"})
>>> -test_run:wait_fullmesh(SERVERS)
>>> -test_run:cmd("switch autobootstrap3")
>>> -test_run = require('test_run').new()
>>> -fiber = require('fiber')
>>> -_ = box.schema.space.create('test_timeout'):create_index('pk')
>>> -test_run:cmd("setopt delimiter ';'")
>>> -function wait_not_follow(replicaA, replicaB)
>>> - return test_run:wait_cond(function()
>>> - return replicaA.status ~= 'follow' or replicaB.status ~= 'follow'
>>> - end, box.cfg.replication_timeout)
>>> -end;
>>> -function test_timeout()
>>> - local replicaA = box.info.replication[1].upstream or box.info.replication[2].upstream
>>> - local replicaB = box.info.replication[3].upstream or box.info.replication[2].upstream
>>> - local follows = test_run:wait_cond(function()
>>> - return replicaA.status == 'follow' or replicaB.status == 'follow'
>>> - end)
>>> - if not follows then error('replicas are not in the follow status') end
>>> - for i = 0, 99 do
>>> - box.space.test_timeout:replace({1})
>>> - if wait_not_follow(replicaA, replicaB) then
>>> - return error(box.info.replication)
>>> - end
>>> - end
>>> - return true
>>> -end;
>>> -test_run:cmd("setopt delimiter ''");
>>> -test_timeout()
>>> -
>>> --- gh-3247 - Sequence-generated value is not replicated in case
>>> --- the request was sent via iproto.
>>> -test_run:cmd("switch autobootstrap1")
>>> -net_box = require('net.box')
>>> -_ = box.schema.space.create('space1')
>>> -_ = box.schema.sequence.create('seq')
>>> -_ = box.space.space1:create_index('primary', {sequence = true} )
>>> -_ = box.space.space1:create_index('secondary', {parts = {2, 'unsigned'}})
>>> -box.schema.user.grant('guest', 'read,write', 'space', 'space1')
>>> -c = net_box.connect(box.cfg.listen)
>>> -c.space.space1:insert{box.NULL, "data"} -- fails, but bumps sequence value
>>> -c.space.space1:insert{box.NULL, 1, "data"}
>>> -box.space.space1:select{}
>>> -vclock = test_run:get_vclock("autobootstrap1")
>>> -vclock[0] = nil
>>> -_ = test_run:wait_vclock("autobootstrap2", vclock)
>>> -test_run:cmd("switch autobootstrap2")
>>> -box.space.space1:select{}
>>> -test_run:cmd("switch autobootstrap1")
>>> -box.space.space1:drop()
>>> -
>>> -test_run:cmd("switch default")
>>> -test_run:drop_cluster(SERVERS)
>>> -test_run:cleanup_cluster()
>>> -
>>> --- gh-3642 - Check that socket file descriptor doesn't leak
>>> --- when a replica is disconnected.
>>> -rlimit = require('rlimit')
>>> -lim = rlimit.limit()
>>> -rlimit.getrlimit(rlimit.RLIMIT_NOFILE, lim)
>>> -old_fno = lim.rlim_cur
>>> -lim.rlim_cur = 64
>>> -rlimit.setrlimit(rlimit.RLIMIT_NOFILE, lim)
>>> -
>>> -test_run:cmd('create server sock with rpl_master=default, script="replication/replica.lua"')
>>> -test_run:cmd('start server sock')
>>> -test_run:cmd('switch sock')
>>> -test_run = require('test_run').new()
>>> -fiber = require('fiber')
>>> -test_run:cmd("setopt delimiter ';'")
>>> -for i = 1, 64 do
>>> - local replication = box.cfg.replication
>>> - box.cfg{replication = {}}
>>> - box.cfg{replication = replication}
>>> - while box.info.replication[1].upstream.status ~= 'follow' do
>>> - fiber.sleep(0.001)
>>> - end
>>> -end;
>>> -test_run:cmd("setopt delimiter ''");
>>> -
>>> -box.info.replication[1].upstream.status
>>> -
>>> -test_run:cmd('switch default')
>>> -
>>> -lim.rlim_cur = old_fno
>>> -rlimit.setrlimit(rlimit.RLIMIT_NOFILE, lim)
>>> -
>>> -test_run:cmd("stop server sock")
>>> -test_run:cmd("cleanup server sock")
>>> -test_run:cmd("delete server sock")
>>> -test_run:cleanup_cluster()
>>> -
>>> -box.schema.user.revoke('guest', 'replication')
>>> -
>>> --- gh-3510 assertion failure in replica_on_applier_disconnect()
>>> -test_run:cmd('create server er_load1 with script="replication/er_load1.lua"')
>>> -test_run:cmd('create server er_load2 with script="replication/er_load2.lua"')
>>> -test_run:cmd('start server er_load1 with wait=False, wait_load=False')
>>> --- Instance er_load2 will fail with error ER_REPLICASET_UUID_MISMATCH.
>>> --- This is OK since we only test here that er_load1 doesn't assert.
>>> -test_run:cmd('start server er_load2 with wait=True, wait_load=True, crash_expected = True')
>>> -test_run:cmd('stop server er_load1')
>>> --- er_load2 exits automatically.
>>> -test_run:cmd('cleanup server er_load1')
>>> -test_run:cmd('cleanup server er_load2')
>>> -test_run:cmd('delete server er_load1')
>>> -test_run:cmd('delete server er_load2')
>>> -test_run:cleanup_cluster()
>>> -
>>> ---
>>> --- Test case for gh-3637, gh-4550. Before the fix replica would
>>> --- exit with an error if a user does not exist or a password is
>>> --- incorrect. Now check that we don't hang/panic and successfully
>>> --- connect.
>>> ---
>>> -fiber = require('fiber')
>>> -test_run:cmd("create server replica_auth with rpl_master=default, script='replication/replica_auth.lua'")
>>> -test_run:cmd("start server replica_auth with wait=False, wait_load=False, args='cluster:pass 0.05'")
>>> --- Wait a bit to make sure replica waits till user is created.
>>> -fiber.sleep(0.1)
>>> -box.schema.user.create('cluster')
>>> --- The user is created. Let the replica fail auth request due to
>>> --- a wrong password.
>>> -fiber.sleep(0.1)
>>> -box.schema.user.passwd('cluster', 'pass')
>>> -box.schema.user.grant('cluster', 'replication')
>>> -
>>> -while box.info.replication[2] == nil do fiber.sleep(0.01) end
>>> -vclock = test_run:get_vclock('default')
>>> -vclock[0] = nil
>>> -_ = test_run:wait_vclock('replica_auth', vclock)
>>> -
>>> -test_run:cmd("stop server replica_auth")
>>> -test_run:cmd("cleanup server replica_auth")
>>> -test_run:cmd("delete server replica_auth")
>>> -test_run:cleanup_cluster()
>>> -
>>> -box.schema.user.drop('cluster')
>>> -
>>> ---
>>> --- Test case for gh-3610. Before the fix replica would fail with the assertion
>>> --- when trying to connect to the same master twice.
>>> ---
>>> -box.schema.user.grant('guest', 'replication')
>>> -test_run:cmd("create server replica with rpl_master=default, script='replication/replica.lua'")
>>> -test_run:cmd("start server replica")
>>> -test_run:cmd("switch replica")
>>> -replication = box.cfg.replication[1]
>>> -box.cfg{replication = {replication, replication}}
>>> -
>>> --- Check the case when duplicate connection is detected in the background.
>>> -test_run:cmd("switch default")
>>> -listen = box.cfg.listen
>>> -box.cfg{listen = ''}
>>> -
>>> -test_run:cmd("switch replica")
>>> -box.cfg{replication_connect_quorum = 0, replication_connect_timeout = 0.01}
>>> -box.cfg{replication = {replication, replication}}
>>> -
>>> -test_run:cmd("switch default")
>>> -box.cfg{listen = listen}
>>> -while test_run:grep_log('replica', 'duplicate connection') == nil do fiber.sleep(0.01) end
>>> -
>>> -test_run:cmd("stop server replica")
>>> -test_run:cmd("cleanup server replica")
>>> -test_run:cmd("delete server replica")
>>> -test_run:cleanup_cluster()
>>> -box.schema.user.revoke('guest', 'replication')
>>> -
>>> ---
>>> --- gh-3711 Do not restart replication on box.cfg in case the
>>> --- configuration didn't change.
>>> ---
>>> -box.schema.user.grant('guest', 'replication')
>>> -test_run:cmd("create server replica with rpl_master=default, script='replication/replica.lua'")
>>> -test_run:cmd("start server replica")
>>> -
>>> --- Access rights are checked only during reconnect. If the new
>>> --- and old configurations are equivalent, no reconnect will be
>>> --- issued and replication should continue working.
>>> -box.schema.user.revoke('guest', 'replication')
>>> -test_run:cmd("switch replica")
>>> -replication = box.cfg.replication[1]
>>> -box.cfg{replication = {replication}}
>>> -box.info.status == 'running'
>>> -box.cfg{replication = replication}
>>> -box.info.status == 'running'
>>> -
>>> --- Check that comparison of tables works as expected as well.
>>> -test_run:cmd("switch default")
>>> -box.schema.user.grant('guest', 'replication')
>>> -test_run:cmd("switch replica")
>>> -replication = box.cfg.replication
>>> -table.insert(replication, box.cfg.listen)
>>> -test_run:cmd("switch default")
>>> -box.schema.user.revoke('guest', 'replication')
>>> -test_run:cmd("switch replica")
>>> -box.cfg{replication = replication}
>>> -box.info.status == 'running'
>>> -
>>> -test_run:cmd("switch default")
>>> -test_run:cmd("stop server replica")
>>> -test_run:cmd("cleanup server replica")
>>> -test_run:cmd("delete server replica")
>>> -test_run:cleanup_cluster()
>>> -
>>> ---
>>> --- gh-3704 move cluster id check to replica
>>> ---
>>> -test_run:cmd("create server replica with rpl_master=default, script='replication/replica.lua'")
>>> -box.schema.user.grant("guest", "replication")
>>> -test_run:cmd("start server replica")
>>> -test_run:grep_log("replica", "REPLICASET_UUID_MISMATCH")
>>> -box.info.replication[2].downstream.status
>>> --- change master's cluster uuid and check that replica doesn't connect.
>>> -test_run:cmd("stop server replica")
>>> -_ = box.space._schema:replace{'cluster', tostring(uuid.new())}
>>> --- master believes replica is in cluster, but their cluster UUIDs differ.
>>> -test_run:cmd("start server replica")
>>> -test_run:wait_log("replica", "REPLICASET_UUID_MISMATCH", nil, 1.0)
>>> -test_run:wait_downstream(2, {status = 'stopped'})
>>> -
>>> -test_run:cmd("stop server replica")
>>> -test_run:cmd("cleanup server replica")
>>> -test_run:cmd("delete server replica")
>>> -test_run:cleanup_cluster()
>>> -box.schema.user.revoke('guest', 'replication')
>>> -
>>> ---
>>> --- gh-4399 Check that an error reading WAL directory on subscribe
>>> --- doesn't lead to a permanent replication failure.
>>> ---
>>> -box.schema.user.grant("guest", "replication")
>>> -test_run:cmd("create server replica with rpl_master=default, script='replication/replica.lua'")
>>> -test_run:cmd("start server replica")
>>> -
>>> --- Make the WAL directory inaccessible.
>>> -fio = require('fio')
>>> -path = fio.abspath(box.cfg.wal_dir)
>>> -fio.chmod(path, 0)
>>> -
>>> --- Break replication on timeout.
>>> -replication_timeout = box.cfg.replication_timeout
>>> -box.cfg{replication_timeout = 9000}
>>> -test_run:cmd("switch replica")
>>> -test_run:wait_cond(function() return box.info.replication[1].upstream.status ~= 'follow' end)
>>> -require('fiber').sleep(box.cfg.replication_timeout)
>>> -test_run:cmd("switch default")
>>> -box.cfg{replication_timeout = replication_timeout}
>>> -
>>> --- Restore access to the WAL directory.
>>> --- Wait for replication to be reestablished.
>>> -fio.chmod(path, tonumber('777', 8))
>>> -test_run:cmd("switch replica")
>>> -test_run:wait_cond(function() return box.info.replication[1].upstream.status == 'follow' end)
>>> -test_run:cmd("switch default")
>>> -
>>> -test_run:cmd("stop server replica")
>>> -test_run:cmd("cleanup server replica")
>>> -test_run:cmd("delete server replica")
>>> -test_run:cleanup_cluster()
>>> -box.schema.user.revoke('guest', 'replication')
>>> -
>>> ---
>>> --- gh-4424 Always enter orphan mode on error in replication
>>> --- configuration change.
>>> ---
>>> -replication_connect_timeout = box.cfg.replication_connect_timeout
>>> -replication_connect_quorum = box.cfg.replication_connect_quorum
>>> -box.cfg{replication="12345", replication_connect_timeout=0.1, replication_connect_quorum=1}
>>> -box.info.status
>>> -box.info.ro
>>> --- reset replication => leave orphan mode
>>> -box.cfg{replication=""}
>>> -box.info.status
>>> -box.info.ro
>>> --- no switch to orphan when quorum == 0
>>> -box.cfg{replication="12345", replication_connect_quorum=0}
>>> -box.info.status
>>> -box.info.ro
>>> -
>>> --- we could connect to one out of two replicas. Set orphan.
>>> -box.cfg{replication_connect_quorum=2}
>>> -box.cfg{replication={box.cfg.listen, "12345"}}
>>> -box.info.status
>>> -box.info.ro
>>> --- lower quorum => leave orphan mode
>>> -box.cfg{replication_connect_quorum=1}
>>> -box.info.status
>>> -box.info.ro
>>> -
>>> ---
>>> --- gh-3760: replication quorum 0 on reconfiguration should return
>>> --- from box.cfg immediately.
>>> ---
>>> -replication = box.cfg.replication
>>> -box.cfg{ \
>>> - replication = {}, \
>>> - replication_connect_quorum = 0, \
>>> - replication_connect_timeout = 1000000 \
>>> -}
>>> --- The call below would hang, if quorum 0 is ignored, or checked
>>> --- too late.
>>> -box.cfg{replication = {'localhost:12345'}}
>>> -box.info.status
>>> -box.cfg{ \
>>> - replication = {}, \
>>> - replication_connect_quorum = replication_connect_quorum, \
>>> - replication_connect_timeout = replication_connect_timeout \
>>> -}
>>> diff --git a/test/replication/misc_assert_connecting_master_twice_gh-3610.result b/test/replication/misc_assert_connecting_master_twice_gh-3610.result
>>> new file mode 100644
>>> index 000000000..d7b7cc25b
>>> --- /dev/null
>>> +++ b/test/replication/misc_assert_connecting_master_twice_gh-3610.result
>>> @@ -0,0 +1,86 @@
>>> +test_run = require('test_run').new()
>>> +---
>>> +...
>>> +test_run:cmd("restart server default")
>> I don't think you should restart the server here.
>> The test seems fine without it. Same about almost all the
>> testcases below.
>>> +uuid = require('uuid')
>>> +---
>>> +...
>>> +fiber = require('fiber')
>>> +---
>>> +...
>> UUID and fiber modules aren't used in this test, so please remove
>> these two extra lines. Same about almost all the testcases below.
>>
>>
>> The branch contains this extraneous change:
>>
>> +replication_connect_quorum = box.cfg.replication_connect_quorum
>>
>> +---
>> +...
>> +replication_connect_timeout = box.cfg.replication_connect_timeout
>> +---
>>
>> +...
>>
>> You don't need to save these options, since you only change them on a
>> replica.
>>
>>> +--
>>> +-- Test case for gh-3610. Before the fix replica would fail with the assertion
>>> +-- when trying to connect to the same master twice.
>>> +--
>>> +box.schema.user.grant('guest', 'replication')
>>> +---
>>> +...
>>> +test_run:cmd("create server replica with rpl_master=default, script='replication/replica.lua'")
>>> +---
>>> +- true
>>> +...
>>> +test_run:cmd("start server replica")
>>> +---
>>> +- true
>>> +...
>>> +test_run:cmd("switch replica")
>>> +---
>>> +- true
>>> +...
>>> +replication = box.cfg.replication[1]
>>> +---
>>> +...
>>> +box.cfg{replication = {replication, replication}}
>>> +---
>>> +- error: 'Incorrect value for option ''replication'': duplicate connection to the
>>> + same replica'
>>> +...
>>> +-- Check the case when duplicate connection is detected in the background.
>>> +test_run:cmd("switch default")
>>> +---
>>> +- true
>>> +...
>>> +listen = box.cfg.listen
>>> +---
>>> +...
>>> +box.cfg{listen = ''}
>>> +---
>>> +...
>>> +test_run:cmd("switch replica")
>>> +---
>>> +- true
>>> +...
>>> +box.cfg{replication_connect_quorum = 0, replication_connect_timeout = 0.01}
>>> +---
>>> +...
>>> +box.cfg{replication = {replication, replication}}
>>> +---
>>> +...
>>> +test_run:cmd("switch default")
>>> +---
>>> +- true
>>> +...
>>> +box.cfg{listen = listen}
>>> +---
>>> +...
>>> +while test_run:grep_log('replica', 'duplicate connection') == nil do fiber.sleep(0.01) end
>>> +---
>>> +...
>>> +test_run:cmd("stop server replica")
>>> +---
>>> +- true
>>> +...
>>> +test_run:cmd("cleanup server replica")
>>> +---
>>> +- true
>>> +...
>>> +test_run:cmd("delete server replica")
>>> +---
>>> +- true
>>> +...
>>> +test_run:cleanup_cluster()
>>> +---
>>> +...
>>> +box.schema.user.revoke('guest', 'replication')
>>> +---
>>> +...
>> +box.cfg{replication = "", \
>> + replication_connect_quorum = replication_connect_quorum, \
>> + replication_connect_timeout = replication_connect_timeout}
>> +---
>> +...
>>
>> The change I spoke of above. This piece is also unneeded.
>>
>>> diff --git a/test/replication/misc_assert_connecting_master_twice_gh-3610.test.lua b/test/replication/misc_assert_connecting_master_twice_gh-3610.test.lua
>>> new file mode 100644
>>> index 000000000..9b12e623b
>>> --- /dev/null
>>> +++ b/test/replication/misc_assert_connecting_master_twice_gh-3610.test.lua
>>> @@ -0,0 +1,34 @@
>>> +test_run = require('test_run').new()
>>> +test_run:cmd("restart server default")
>>> +uuid = require('uuid')
>>> +fiber = require('fiber')
>>> +
>>> +--
>>> +-- Test case for gh-3610. Before the fix replica would fail with the assertion
>>> +-- when trying to connect to the same master twice.
>>> +--
>>> +box.schema.user.grant('guest', 'replication')
>>> +test_run:cmd("create server replica with rpl_master=default, script='replication/replica.lua'")
>>> +test_run:cmd("start server replica")
>>> +test_run:cmd("switch replica")
>>> +replication = box.cfg.replication[1]
>>> +box.cfg{replication = {replication, replication}}
>>> +
>>> +-- Check the case when duplicate connection is detected in the background.
>>> +test_run:cmd("switch default")
>>> +listen = box.cfg.listen
>>> +box.cfg{listen = ''}
>>> +
>>> +test_run:cmd("switch replica")
>>> +box.cfg{replication_connect_quorum = 0, replication_connect_timeout = 0.01}
>>> +box.cfg{replication = {replication, replication}}
>>> +
>>> +test_run:cmd("switch default")
>>> +box.cfg{listen = listen}
>>> +while test_run:grep_log('replica', 'duplicate connection') == nil do fiber.sleep(0.01) end
>>> +
>>> +test_run:cmd("stop server replica")
>>> +test_run:cmd("cleanup server replica")
>>> +test_run:cmd("delete server replica")
>>> +test_run:cleanup_cluster()
>>> +box.schema.user.revoke('guest', 'replication')
>>> diff --git a/test/replication/misc_assert_on_server_die_gh-2991.result b/test/replication/misc_assert_on_server_die_gh-2991.result
>>> new file mode 100644
>>> index 000000000..ea9e80f6b
>>> --- /dev/null
>>> +++ b/test/replication/misc_assert_on_server_die_gh-2991.result
>>> @@ -0,0 +1,31 @@
>>> +uuid = require('uuid')
>>> +---
>>> +...
>> UUID module isn't used in this testcase, please remove the extra line.
>>> +test_run = require('test_run').new()
>>> +---
>>> +...
>>> +-- gh-2991 - Tarantool asserts on box.cfg.replication update if one of
>>> +-- servers is dead
>>> +replication_timeout = box.cfg.replication_timeout
>>> +---
>>> +...
>>> +replication_connect_timeout = box.cfg.replication_connect_timeout
>>> +---
>>> +...
>>> +box.cfg{replication_timeout=0.05, replication_connect_timeout=0.05, replication={}}
>>> +---
>>> +...
>>> +box.cfg{replication_connect_quorum=2}
>>> +---
>>> +...
>>> +box.cfg{replication = {'127.0.0.1:12345', box.cfg.listen}}
>>> +---
>>> +...
>>> +box.info.status
>>> +---
>>> +- orphan
>>> +...
>>> +box.info.ro
>>> +---
>>> +- true
>>> +...
>>> diff --git a/test/replication/misc_assert_on_server_die_gh-2991.test.lua b/test/replication/misc_assert_on_server_die_gh-2991.test.lua
>>> new file mode 100644
>>> index 000000000..adda839f7
>>> --- /dev/null
>>> +++ b/test/replication/misc_assert_on_server_die_gh-2991.test.lua
>>> @@ -0,0 +1,12 @@
>>> +uuid = require('uuid')
>>> +test_run = require('test_run').new()
>>> +
>>> +-- gh-2991 - Tarantool asserts on box.cfg.replication update if one of
>>> +-- servers is dead
>>> +replication_timeout = box.cfg.replication_timeout
>>> +replication_connect_timeout = box.cfg.replication_connect_timeout
>>> +box.cfg{replication_timeout=0.05, replication_connect_timeout=0.05, replication={}}
>>> +box.cfg{replication_connect_quorum=2}
>>> +box.cfg{replication = {'127.0.0.1:12345', box.cfg.listen}}
>>> +box.info.status
>>> +box.info.ro
>>> diff --git a/test/replication/misc_assert_replica_on_applier_disconnect_gh-3510.result b/test/replication/misc_assert_replica_on_applier_disconnect_gh-3510.result
>>> new file mode 100644
>>> index 000000000..82f5d9a23
>>> --- /dev/null
>>> +++ b/test/replication/misc_assert_replica_on_applier_disconnect_gh-3510.result
>>> @@ -0,0 +1,49 @@
>>> +uuid = require('uuid')
>>> +---
>>> +...
>> Same about the uuid.
>>> +test_run = require('test_run').new()
>>> +---
>>> +...
>>> +-- gh-3510 assertion failure in replica_on_applier_disconnect()
>>> +test_run:cmd('create server er_load1 with script="replication/er_load1.lua"')
>>> +---
>>> +- true
>>> +...
>>> +test_run:cmd('create server er_load2 with script="replication/er_load2.lua"')
>>> +---
>>> +- true
>>> +...
>>> +test_run:cmd('start server er_load1 with wait=False, wait_load=False')
>>> +---
>>> +- true
>>> +...
>>> +-- Instance er_load2 will fail with error ER_REPLICASET_UUID_MISMATCH.
>>> +-- This is OK since we only test here that er_load1 doesn't assert.
>>> +test_run:cmd('start server er_load2 with wait=True, wait_load=True, crash_expected = True')
>>> +---
>>> +- false
>>> +...
>>> +test_run:cmd('stop server er_load1')
>>> +---
>>> +- true
>>> +...
>>> +-- er_load2 exits automatically.
>>> +test_run:cmd('cleanup server er_load1')
>>> +---
>>> +- true
>>> +...
>>> +test_run:cmd('cleanup server er_load2')
>>> +---
>>> +- true
>>> +...
>>> +test_run:cmd('delete server er_load1')
>>> +---
>>> +- true
>>> +...
>>> +test_run:cmd('delete server er_load2')
>>> +---
>>> +- true
>>> +...
>>> +test_run:cleanup_cluster()
>>> +---
>>> +...
>>> diff --git a/test/replication/misc_assert_replica_on_applier_disconnect_gh-3510.test.lua b/test/replication/misc_assert_replica_on_applier_disconnect_gh-3510.test.lua
>>> new file mode 100644
>>> index 000000000..4e1d2a41e
>>> --- /dev/null
>>> +++ b/test/replication/misc_assert_replica_on_applier_disconnect_gh-3510.test.lua
>>> @@ -0,0 +1,17 @@
>>> +uuid = require('uuid')
>>> +test_run = require('test_run').new()
>>> +
>>> +-- gh-3510 assertion failure in replica_on_applier_disconnect()
>>> +test_run:cmd('create server er_load1 with script="replication/er_load1.lua"')
>>> +test_run:cmd('create server er_load2 with script="replication/er_load2.lua"')
>>> +test_run:cmd('start server er_load1 with wait=False, wait_load=False')
>>> +-- Instance er_load2 will fail with error ER_REPLICASET_UUID_MISMATCH.
>>> +-- This is OK since we only test here that er_load1 doesn't assert.
>>> +test_run:cmd('start server er_load2 with wait=True, wait_load=True, crash_expected = True')
>>> +test_run:cmd('stop server er_load1')
>>> +-- er_load2 exits automatically.
>>> +test_run:cmd('cleanup server er_load1')
>>> +test_run:cmd('cleanup server er_load2')
>>> +test_run:cmd('delete server er_load1')
>>> +test_run:cmd('delete server er_load2')
>>> +test_run:cleanup_cluster()
>>> diff --git a/test/replication/misc_crash_on_box_concurrent_update_gh-3606.result b/test/replication/misc_crash_on_box_concurrent_update_gh-3606.result
>>> new file mode 100644
>>> index 000000000..b43b00849
>>> --- /dev/null
>>> +++ b/test/replication/misc_crash_on_box_concurrent_update_gh-3606.result
>>> @@ -0,0 +1,50 @@
>>> +uuid = require('uuid')
>>> +---
>>> +...
>> Same about UUID.
>>> +test_run = require('test_run').new()
>>> +---
>>> +...
>>> +replication_timeout = box.cfg.replication_timeout
>>> +---
>>> +...
>>> +replication_connect_timeout = box.cfg.replication_connect_timeout
>>> +---
>>> +...
>>> +box.cfg{replication_timeout=0.05, replication_connect_timeout=0.05, replication={}}
>>> +---
>>> +...
>>> +-- gh-3606 - Tarantool crashes if box.cfg.replication is updated concurrently
>>> +fiber = require('fiber')
>>> +---
>>> +...
>>> +c = fiber.channel(2)
>>> +---
>>> +...
>>> +f = function() fiber.create(function() pcall(box.cfg, {replication = {12345}}) c:put(true) end) end
>>> +---
>>> +...
>>> +f()
>>> +---
>>> +...
>>> +f()
>>> +---
>>> +...
>>> +c:get()
>>> +---
>>> +- true
>>> +...
>>> +c:get()
>>> +---
>>> +- true
>>> +...
>>> +box.cfg{replication = "", replication_timeout = replication_timeout, replication_connect_timeout = replication_connect_timeout}
>>> +---
>>> +...
>>> +box.info.status
>>> +---
>>> +- running
>>> +...
>>> +box.info.ro
>>> +---
>>> +- false
>>> +...
>>> diff --git a/test/replication/misc_crash_on_box_concurrent_update_gh-3606.test.lua b/test/replication/misc_crash_on_box_concurrent_update_gh-3606.test.lua
>>> new file mode 100644
>>> index 000000000..17f9c6bc6
>>> --- /dev/null
>>> +++ b/test/replication/misc_crash_on_box_concurrent_update_gh-3606.test.lua
>>> @@ -0,0 +1,19 @@
>>> +uuid = require('uuid')
>>> +test_run = require('test_run').new()
>>> +
>>> +replication_timeout = box.cfg.replication_timeout
>>> +replication_connect_timeout = box.cfg.replication_connect_timeout
>>> +box.cfg{replication_timeout=0.05, replication_connect_timeout=0.05, replication={}}
>>> +
>>> +-- gh-3606 - Tarantool crashes if box.cfg.replication is updated concurrently
>>> +fiber = require('fiber')
>>> +c = fiber.channel(2)
>>> +f = function() fiber.create(function() pcall(box.cfg, {replication = {12345}}) c:put(true) end) end
>>> +f()
>>> +f()
>>> +c:get()
>>> +c:get()
>>> +
>>> +box.cfg{replication = "", replication_timeout = replication_timeout, replication_connect_timeout = replication_connect_timeout}
>>> +box.info.status
>>> +box.info.ro
>>> diff --git a/test/replication/misc_heartbeats_on_master_changes_gh-3160.result b/test/replication/misc_heartbeats_on_master_changes_gh-3160.result
>>> new file mode 100644
>>> index 000000000..cdb463614
>>> --- /dev/null
>>> +++ b/test/replication/misc_heartbeats_on_master_changes_gh-3160.result
>>> @@ -0,0 +1,76 @@
>>> +uuid = require('uuid')
>>> +---
>>> +...
>> Same about UUID.
>>
>>> +test_run = require('test_run').new()
>>> +---
>>> +...
>> The branch contains this change:
>>
>> +...
>> +replication_timeout = box.cfg.replication_timeout
>> +---
>> +...
>> +replication_connect_timeout = box.cfg.replication_connect_timeout
>> +---
>> +...
>> +box.cfg{replication_timeout=0.05, replication_connect_timeout=0.05,
>> replication={}}
>> +---
>> +...
>>
>> Which is extraneous in this test. The original testcase uses default timeout
>> values,
>> and I don't think we should change them.
>>
>>> +-- gh-3160 - Send heartbeats if there are changes from a remote master only
>>> +SERVERS = { 'autobootstrap1', 'autobootstrap2', 'autobootstrap3' }
>>> +---
>>> +...
>>> +-- Deploy a cluster.
>>> +test_run:create_cluster(SERVERS, "replication", {args="0.03"})
>>> +---
>>> +...
>>> +test_run:wait_fullmesh(SERVERS)
>>> +---
>>> +...
>>> +test_run:cmd("switch autobootstrap3")
>>> +---
>>> +- true
>>> +...
>>> +test_run = require('test_run').new()
>>> +---
>>> +...
>>> +fiber = require('fiber')
>>> +---
>>> +...
>>> +_ = box.schema.space.create('test_timeout'):create_index('pk')
>>> +---
>>> +...
>>> +test_run:cmd("setopt delimiter ';'")
>>> +---
>>> +- true
>>> +...
>>> +function wait_not_follow(replicaA, replicaB)
>>> + return test_run:wait_cond(function()
>>> + return replicaA.status ~= 'follow' or replicaB.status ~= 'follow'
>>> + end, box.cfg.replication_timeout)
>>> +end;
>>> +---
>>> +...
>>> +function test_timeout()
>>> + local replicaA = box.info.replication[1].upstream or box.info.replication[2].upstream
>>> + local replicaB = box.info.replication[3].upstream or box.info.replication[2].upstream
>>> + local follows = test_run:wait_cond(function()
>>> + return replicaA.status == 'follow' or replicaB.status == 'follow'
>>> + end)
>>> + if not follows then error('replicas are not in the follow status') end
>>> + for i = 0, 99 do
>>> + box.space.test_timeout:replace({1})
>>> + if wait_not_follow(replicaA, replicaB) then
>>> + return error(box.info.replication)
>>> + end
>>> + end
>>> + return true
>>> +end;
>>> +---
>>> +...
>>> +test_run:cmd("setopt delimiter ''");
>>> +---
>>> +- true
>>> +...
>>> +test_timeout()
>>> +---
>>> +- true
>>> +...
>>> +test_run:cmd("switch default")
>>> +---
>>> +- true
>>> +...
>>> +test_run:drop_cluster(SERVERS)
>>> +---
>>> +...
>>> +test_run:cleanup_cluster()
>> No need to call cleanup_cluster(). The 'default' instance wasn't part of the
>> cluster
>> you ran in this test.
>>
>>
>> +...
>> +box.cfg{replication = "", replication_timeout = replication_timeout, \
>> + replication_connect_timeout = replication_connect_timeout}
>> +---
>> +...
>>
>> One more extraneous change, related to the one above.
>>
>>
>>> +---
>>> +...
>>> diff --git a/test/replication/misc_heartbeats_on_master_changes_gh-3160.test.lua b/test/replication/misc_heartbeats_on_master_changes_gh-3160.test.lua
>>> new file mode 100644
>>> index 000000000..eba8a7725
>>> --- /dev/null
>>> +++ b/test/replication/misc_heartbeats_on_master_changes_gh-3160.test.lua
>>> @@ -0,0 +1,40 @@
>>> +uuid = require('uuid')
>>> +test_run = require('test_run').new()
>>> +
>>> +-- gh-3160 - Send heartbeats if there are changes from a remote master only
>>> +SERVERS = { 'autobootstrap1', 'autobootstrap2', 'autobootstrap3' }
>>> +
>>> +-- Deploy a cluster.
>>> +test_run:create_cluster(SERVERS, "replication", {args="0.03"})
>>> +test_run:wait_fullmesh(SERVERS)
>>> +test_run:cmd("switch autobootstrap3")
>>> +test_run = require('test_run').new()
>>> +fiber = require('fiber')
>>> +_ = box.schema.space.create('test_timeout'):create_index('pk')
>>> +test_run:cmd("setopt delimiter ';'")
>>> +function wait_not_follow(replicaA, replicaB)
>>> + return test_run:wait_cond(function()
>>> + return replicaA.status ~= 'follow' or replicaB.status ~= 'follow'
>>> + end, box.cfg.replication_timeout)
>>> +end;
>>> +function test_timeout()
>>> + local replicaA = box.info.replication[1].upstream or box.info.replication[2].upstream
>>> + local replicaB = box.info.replication[3].upstream or box.info.replication[2].upstream
>>> + local follows = test_run:wait_cond(function()
>>> + return replicaA.status == 'follow' or replicaB.status == 'follow'
>>> + end)
>>> + if not follows then error('replicas are not in the follow status') end
>>> + for i = 0, 99 do
>>> + box.space.test_timeout:replace({1})
>>> + if wait_not_follow(replicaA, replicaB) then
>>> + return error(box.info.replication)
>>> + end
>>> + end
>>> + return true
>>> +end;
>>> +test_run:cmd("setopt delimiter ''");
>>> +test_timeout()
>>> +
>>> +test_run:cmd("switch default")
>>> +test_run:drop_cluster(SERVERS)
>>> +test_run:cleanup_cluster()
>>> diff --git a/test/replication/misc_no_failure_on_error_reading_wal_gh-4399.result b/test/replication/misc_no_failure_on_error_reading_wal_gh-4399.result
>>> new file mode 100644
>>> index 000000000..64a4de26c
>>> --- /dev/null
>>> +++ b/test/replication/misc_no_failure_on_error_reading_wal_gh-4399.result
>>> @@ -0,0 +1,97 @@
>>> +test_run = require('test_run').new()
>>> +---
>>> +...
>>> +test_run:cmd("restart server default")
>>> +uuid = require('uuid')
>>> +---
>>> +...
>>> +fiber = require('fiber')
>> Same about restarting the server, requiring uuid and fiber. Not needed.
>>> +---
>>> +...
>>> +--
>>> +-- gh-4399 Check that an error reading WAL directory on subscribe
>>> +-- doesn't lead to a permanent replication failure.
>>> +--
>>> +box.schema.user.grant("guest", "replication")
>>> +---
>>> +...
>>> +test_run:cmd("create server replica with rpl_master=default, script='replication/replica.lua'")
>>> +---
>>> +- true
>>> +...
>>> +test_run:cmd("start server replica")
>>> +---
>>> +- true
>>> +...
>>> +-- Make the WAL directory inaccessible.
>>> +fio = require('fio')
>>> +---
>>> +...
>>> +path = fio.abspath(box.cfg.wal_dir)
>>> +---
>>> +...
>>> +fio.chmod(path, 0)
>>> +---
>>> +- true
>>> +...
>>> +-- Break replication on timeout.
>>> +replication_timeout = box.cfg.replication_timeout
>>> +---
>>> +...
>>> +box.cfg{replication_timeout = 9000}
>>> +---
>>> +...
>>> +test_run:cmd("switch replica")
>>> +---
>>> +- true
>>> +...
>>> +test_run:wait_cond(function() return box.info.replication[1].upstream.status ~= 'follow' end)
>>> +---
>>> +- true
>>> +...
>>> +require('fiber').sleep(box.cfg.replication_timeout)
>>> +---
>>> +...
>>> +test_run:cmd("switch default")
>>> +---
>>> +- true
>>> +...
>>> +box.cfg{replication_timeout = replication_timeout}
>>> +---
>>> +...
>>> +-- Restore access to the WAL directory.
>>> +-- Wait for replication to be reestablished.
>>> +fio.chmod(path, tonumber('777', 8))
>>> +---
>>> +- true
>>> +...
>>> +test_run:cmd("switch replica")
>>> +---
>>> +- true
>>> +...
>>> +test_run:wait_cond(function() return box.info.replication[1].upstream.status == 'follow' end)
>>> +---
>>> +- true
>>> +...
>>> +test_run:cmd("switch default")
>>> +---
>>> +- true
>>> +...
>>> +test_run:cmd("stop server replica")
>>> +---
>>> +- true
>>> +...
>>> +test_run:cmd("cleanup server replica")
>>> +---
>>> +- true
>>> +...
>>> +test_run:cmd("delete server replica")
>>> +---
>>> +- true
>>> +...
>>> +test_run:cleanup_cluster()
>>> +---
>>> +...
>>> +box.schema.user.revoke('guest', 'replication')
>>> +---
>>> +...
>>> diff --git a/test/replication/misc_no_failure_on_error_reading_wal_gh-4399.test.lua b/test/replication/misc_no_failure_on_error_reading_wal_gh-4399.test.lua
>>> new file mode 100644
>>> index 000000000..15e19a211
>>> --- /dev/null
>>> +++ b/test/replication/misc_no_failure_on_error_reading_wal_gh-4399.test.lua
>>> @@ -0,0 +1,39 @@
>>> +test_run = require('test_run').new()
>>> +test_run:cmd("restart server default")
>>> +uuid = require('uuid')
>>> +fiber = require('fiber')
>>> +
>>> +--
>>> +-- gh-4399 Check that an error reading WAL directory on subscribe
>>> +-- doesn't lead to a permanent replication failure.
>>> +--
>>> +box.schema.user.grant("guest", "replication")
>>> +test_run:cmd("create server replica with rpl_master=default, script='replication/replica.lua'")
>>> +test_run:cmd("start server replica")
>>> +
>>> +-- Make the WAL directory inaccessible.
>>> +fio = require('fio')
>>> +path = fio.abspath(box.cfg.wal_dir)
>>> +fio.chmod(path, 0)
>>> +
>>> +-- Break replication on timeout.
>>> +replication_timeout = box.cfg.replication_timeout
>>> +box.cfg{replication_timeout = 9000}
>>> +test_run:cmd("switch replica")
>>> +test_run:wait_cond(function() return box.info.replication[1].upstream.status ~= 'follow' end)
>>> +require('fiber').sleep(box.cfg.replication_timeout)
>>> +test_run:cmd("switch default")
>>> +box.cfg{replication_timeout = replication_timeout}
>>> +
>>> +-- Restore access to the WAL directory.
>>> +-- Wait for replication to be reestablished.
>>> +fio.chmod(path, tonumber('777', 8))
>>> +test_run:cmd("switch replica")
>>> +test_run:wait_cond(function() return box.info.replication[1].upstream.status == 'follow' end)
>>> +test_run:cmd("switch default")
>>> +
>>> +test_run:cmd("stop server replica")
>>> +test_run:cmd("cleanup server replica")
>>> +test_run:cmd("delete server replica")
>>> +test_run:cleanup_cluster()
>>> +box.schema.user.revoke('guest', 'replication')
>>> diff --git a/test/replication/misc_no_panic_on_connected_gh-3637.result b/test/replication/misc_no_panic_on_connected_gh-3637.result
>>> new file mode 100644
>>> index 000000000..693b18c42
>>> --- /dev/null
>>> +++ b/test/replication/misc_no_panic_on_connected_gh-3637.result
>>> @@ -0,0 +1,72 @@
>>> +test_run = require('test_run').new()
>>> +---
>>> +...
>>> +test_run:cmd("restart server default")
>>> +uuid = require('uuid')
>>> +---
>>> +...
>> Again, no need for uuid here. And no need for restart.
>>> +--
>>> +-- Test case for gh-3637, gh-4550. Before the fix replica would
>>> +-- exit with an error if a user does not exist or a password is
>>> +-- incorrect. Now check that we don't hang/panic and successfully
>>> +-- connect.
>>> +--
>>> +fiber = require('fiber')
>>> +---
>>> +...
>>> +test_run:cmd("create server replica_auth with rpl_master=default, script='replication/replica_auth.lua'")
>>> +---
>>> +- true
>>> +...
>>> +test_run:cmd("start server replica_auth with wait=False, wait_load=False, args='cluster:pass 0.05'")
>>> +---
>>> +- true
>>> +...
>>> +-- Wait a bit to make sure replica waits till user is created.
>>> +fiber.sleep(0.1)
>>> +---
>>> +...
>>> +box.schema.user.create('cluster')
>>> +---
>>> +...
>>> +-- The user is created. Let the replica fail auth request due to
>>> +-- a wrong password.
>>> +fiber.sleep(0.1)
>>> +---
>>> +...
>>> +box.schema.user.passwd('cluster', 'pass')
>>> +---
>>> +...
>>> +box.schema.user.grant('cluster', 'replication')
>>> +---
>>> +...
>>> +while box.info.replication[2] == nil do fiber.sleep(0.01) end
>>> +---
>>> +...
>>> +vclock = test_run:get_vclock('default')
>>> +---
>>> +...
>>> +vclock[0] = nil
>>> +---
>>> +...
>>> +_ = test_run:wait_vclock('replica_auth', vclock)
>>> +---
>>> +...
>>> +test_run:cmd("stop server replica_auth")
>>> +---
>>> +- true
>>> +...
>>> +test_run:cmd("cleanup server replica_auth")
>>> +---
>>> +- true
>>> +...
>>> +test_run:cmd("delete server replica_auth")
>>> +---
>>> +- true
>>> +...
>>> +test_run:cleanup_cluster()
>>> +---
>>> +...
>>> +box.schema.user.drop('cluster')
>>> +---
>>> +...
>>> diff --git a/test/replication/misc_no_panic_on_connected_gh-3637.test.lua b/test/replication/misc_no_panic_on_connected_gh-3637.test.lua
>>> new file mode 100644
>>> index 000000000..a1e51198b
>>> --- /dev/null
>>> +++ b/test/replication/misc_no_panic_on_connected_gh-3637.test.lua
>>> @@ -0,0 +1,33 @@
>>> +test_run = require('test_run').new()
>>> +test_run:cmd("restart server default")
>>> +uuid = require('uuid')
>>> +
>>> +--
>>> +-- Test case for gh-3637, gh-4550. Before the fix replica would
>>> +-- exit with an error if a user does not exist or a password is
>>> +-- incorrect. Now check that we don't hang/panic and successfully
>>> +-- connect.
>>> +--
>>> +fiber = require('fiber')
>>> +test_run:cmd("create server replica_auth with rpl_master=default, script='replication/replica_auth.lua'")
>>> +test_run:cmd("start server replica_auth with wait=False, wait_load=False, args='cluster:pass 0.05'")
>>> +-- Wait a bit to make sure replica waits till user is created.
>>> +fiber.sleep(0.1)
>>> +box.schema.user.create('cluster')
>>> +-- The user is created. Let the replica fail auth request due to
>>> +-- a wrong password.
>>> +fiber.sleep(0.1)
>>> +box.schema.user.passwd('cluster', 'pass')
>>> +box.schema.user.grant('cluster', 'replication')
>>> +
>>> +while box.info.replication[2] == nil do fiber.sleep(0.01) end
>>> +vclock = test_run:get_vclock('default')
>>> +vclock[0] = nil
>>> +_ = test_run:wait_vclock('replica_auth', vclock)
>>> +
>>> +test_run:cmd("stop server replica_auth")
>>> +test_run:cmd("cleanup server replica_auth")
>>> +test_run:cmd("delete server replica_auth")
>>> +test_run:cleanup_cluster()
>>> +
>>> +box.schema.user.drop('cluster')
>>> diff --git a/test/replication/misc_no_restart_on_same_configuration_gh-3711.result b/test/replication/misc_no_restart_on_same_configuration_gh-3711.result
>>> new file mode 100644
>>> index 000000000..dd9ef848c
>>> --- /dev/null
>>> +++ b/test/replication/misc_no_restart_on_same_configuration_gh-3711.result
>>> @@ -0,0 +1,107 @@
>>> +test_run = require('test_run').new()
>>> +---
>>> +...
>>> +test_run:cmd("restart server default")
>>> +uuid = require('uuid')
>>> +---
>>> +...
>>> +fiber = require('fiber')
>> Same about restart, uuid and fiber.
>>> +---
>>> +...
>>> +--
>>> +-- gh-3711 Do not restart replication on box.cfg in case the
>>> +-- configuration didn't change.
>>> +--
>>> +box.schema.user.grant('guest', 'replication')
>>> +---
>>> +...
>>> +test_run:cmd("create server replica with rpl_master=default, script='replication/replica.lua'")
>>> +---
>>> +- true
>>> +...
>>> +test_run:cmd("start server replica")
>>> +---
>>> +- true
>>> +...
>>> +-- Access rights are checked only during reconnect. If the new
>>> +-- and old configurations are equivalent, no reconnect will be
>>> +-- issued and replication should continue working.
>>> +box.schema.user.revoke('guest', 'replication')
>>> +---
>>> +...
>>> +test_run:cmd("switch replica")
>>> +---
>>> +- true
>>> +...
>>> +replication = box.cfg.replication[1]
>>> +---
>>> +...
>>> +box.cfg{replication = {replication}}
>>> +---
>>> +...
>>> +box.info.status == 'running'
>>> +---
>>> +- true
>>> +...
>>> +box.cfg{replication = replication}
>>> +---
>>> +...
>>> +box.info.status == 'running'
>>> +---
>>> +- true
>>> +...
>>> +-- Check that comparison of tables works as expected as well.
>>> +test_run:cmd("switch default")
>>> +---
>>> +- true
>>> +...
>>> +box.schema.user.grant('guest', 'replication')
>>> +---
>>> +...
>>> +test_run:cmd("switch replica")
>>> +---
>>> +- true
>>> +...
>>> +replication = box.cfg.replication
>>> +---
>>> +...
>>> +table.insert(replication, box.cfg.listen)
>>> +---
>>> +...
>>> +test_run:cmd("switch default")
>>> +---
>>> +- true
>>> +...
>>> +box.schema.user.revoke('guest', 'replication')
>>> +---
>>> +...
>>> +test_run:cmd("switch replica")
>>> +---
>>> +- true
>>> +...
>>> +box.cfg{replication = replication}
>>> +---
>>> +...
>>> +box.info.status == 'running'
>>> +---
>>> +- true
>>> +...
>>> +test_run:cmd("switch default")
>>> +---
>>> +- true
>>> +...
>>> +test_run:cmd("stop server replica")
>>> +---
>>> +- true
>>> +...
>>> +test_run:cmd("cleanup server replica")
>>> +---
>>> +- true
>>> +...
>>> +test_run:cmd("delete server replica")
>>> +---
>>> +- true
>>> +...
>>> +test_run:cleanup_cluster()
>>> +---
>>> +...
>>> diff --git a/test/replication/misc_no_restart_on_same_configuration_gh-3711.test.lua b/test/replication/misc_no_restart_on_same_configuration_gh-3711.test.lua
>>> new file mode 100644
>>> index 000000000..14b522ead
>>> --- /dev/null
>>> +++ b/test/replication/misc_no_restart_on_same_configuration_gh-3711.test.lua
>>> @@ -0,0 +1,41 @@
>>> +test_run = require('test_run').new()
>>> +test_run:cmd("restart server default")
>>> +uuid = require('uuid')
>>> +fiber = require('fiber')
>>> +
>>> +--
>>> +-- gh-3711 Do not restart replication on box.cfg in case the
>>> +-- configuration didn't change.
>>> +--
>>> +box.schema.user.grant('guest', 'replication')
>>> +test_run:cmd("create server replica with rpl_master=default, script='replication/replica.lua'")
>>> +test_run:cmd("start server replica")
>>> +
>>> +-- Access rights are checked only during reconnect. If the new
>>> +-- and old configurations are equivalent, no reconnect will be
>>> +-- issued and replication should continue working.
>>> +box.schema.user.revoke('guest', 'replication')
>>> +test_run:cmd("switch replica")
>>> +replication = box.cfg.replication[1]
>>> +box.cfg{replication = {replication}}
>>> +box.info.status == 'running'
>>> +box.cfg{replication = replication}
>>> +box.info.status == 'running'
>>> +
>>> +-- Check that comparison of tables works as expected as well.
>>> +test_run:cmd("switch default")
>>> +box.schema.user.grant('guest', 'replication')
>>> +test_run:cmd("switch replica")
>>> +replication = box.cfg.replication
>>> +table.insert(replication, box.cfg.listen)
>>> +test_run:cmd("switch default")
>>> +box.schema.user.revoke('guest', 'replication')
>>> +test_run:cmd("switch replica")
>>> +box.cfg{replication = replication}
>>> +box.info.status == 'running'
>>> +
>>> +test_run:cmd("switch default")
>>> +test_run:cmd("stop server replica")
>>> +test_run:cmd("cleanup server replica")
>>> +test_run:cmd("delete server replica")
>>> +test_run:cleanup_cluster()
>>> diff --git a/test/replication/misc_no_socket_leak_on_replica_disconnect_gh-3642.result b/test/replication/misc_no_socket_leak_on_replica_disconnect_gh-3642.result
>>> new file mode 100644
>>> index 000000000..3680bcebb
>>> --- /dev/null
>>> +++ b/test/replication/misc_no_socket_leak_on_replica_disconnect_gh-3642.result
>>> @@ -0,0 +1,98 @@
>>> +test_run = require('test_run').new()
>>> +---
>>> +...
>>> +test_run:cmd("restart server default")
>> Unneeded restart. Correct me if I'm wrong.
>>
>>> +uuid = require('uuid')
>> You don't need the uuid module in this testcase.
>>
>>> +---
>>> +...
>>> +box.schema.user.grant('guest', 'replication')
>>> +---
>>> +...
>>> +-- gh-3642 - Check that socket file descriptor doesn't leak
>>> +-- when a replica is disconnected.
>>> +rlimit = require('rlimit')
>>> +---
>>> +...
>>> +lim = rlimit.limit()
>>> +---
>>> +...
>>> +rlimit.getrlimit(rlimit.RLIMIT_NOFILE, lim)
>>> +---
>>> +...
>>> +old_fno = lim.rlim_cur
>>> +---
>>> +...
>>> +lim.rlim_cur = 64
>>> +---
>>> +...
>>> +rlimit.setrlimit(rlimit.RLIMIT_NOFILE, lim)
>>> +---
>>> +...
>>> +test_run:cmd('create server sock with rpl_master=default, script="replication/replica.lua"')
>>> +---
>>> +- true
>>> +...
>>> +test_run:cmd('start server sock')
>>> +---
>>> +- true
>>> +...
>>> +test_run:cmd('switch sock')
>>> +---
>>> +- true
>>> +...
>>> +test_run = require('test_run').new()
>>> +---
>>> +...
>>> +fiber = require('fiber')
>>> +---
>>> +...
>>> +test_run:cmd("setopt delimiter ';'")
>>> +---
>>> +- true
>>> +...
>>> +for i = 1, 64 do
>>> + local replication = box.cfg.replication
>>> + box.cfg{replication = {}}
>>> + box.cfg{replication = replication}
>>> + while box.info.replication[1].upstream.status ~= 'follow' do
>>> + fiber.sleep(0.001)
>>> + end
>>> +end;
>>> +---
>>> +...
>>> +test_run:cmd("setopt delimiter ''");
>>> +---
>>> +- true
>>> +...
>>> +box.info.replication[1].upstream.status
>>> +---
>>> +- follow
>>> +...
>>> +test_run:cmd('switch default')
>>> +---
>>> +- true
>>> +...
>>> +lim.rlim_cur = old_fno
>>> +---
>>> +...
>>> +rlimit.setrlimit(rlimit.RLIMIT_NOFILE, lim)
>>> +---
>>> +...
>>> +test_run:cmd("stop server sock")
>>> +---
>>> +- true
>>> +...
>>> +test_run:cmd("cleanup server sock")
>>> +---
>>> +- true
>>> +...
>>> +test_run:cmd("delete server sock")
>>> +---
>>> +- true
>>> +...
>>> +test_run:cleanup_cluster()
>>> +---
>>> +...
>>> +box.schema.user.revoke('guest', 'replication')
>>> +---
>>> +...
>>> diff --git a/test/replication/misc_no_socket_leak_on_replica_disconnect_gh-3642.test.lua b/test/replication/misc_no_socket_leak_on_replica_disconnect_gh-3642.test.lua
>>> new file mode 100644
>>> index 000000000..08ef9ec0d
>>> --- /dev/null
>>> +++ b/test/replication/misc_no_socket_leak_on_replica_disconnect_gh-3642.test.lua
>>> @@ -0,0 +1,44 @@
>>> +test_run = require('test_run').new()
>>> +test_run:cmd("restart server default")
>>> +uuid = require('uuid')
>>> +
>>> +box.schema.user.grant('guest', 'replication')
>>> +
>>> +-- gh-3642 - Check that socket file descriptor doesn't leak
>>> +-- when a replica is disconnected.
>>> +rlimit = require('rlimit')
>>> +lim = rlimit.limit()
>>> +rlimit.getrlimit(rlimit.RLIMIT_NOFILE, lim)
>>> +old_fno = lim.rlim_cur
>>> +lim.rlim_cur = 64
>>> +rlimit.setrlimit(rlimit.RLIMIT_NOFILE, lim)
>>> +
>>> +test_run:cmd('create server sock with rpl_master=default, script="replication/replica.lua"')
>>> +test_run:cmd('start server sock')
>>> +test_run:cmd('switch sock')
>>> +test_run = require('test_run').new()
>>> +fiber = require('fiber')
>>> +test_run:cmd("setopt delimiter ';'")
>>> +for i = 1, 64 do
>>> + local replication = box.cfg.replication
>>> + box.cfg{replication = {}}
>>> + box.cfg{replication = replication}
>>> + while box.info.replication[1].upstream.status ~= 'follow' do
>>> + fiber.sleep(0.001)
>>> + end
>>> +end;
>>> +test_run:cmd("setopt delimiter ''");
>>> +
>>> +box.info.replication[1].upstream.status
>>> +
>>> +test_run:cmd('switch default')
>>> +
>>> +lim.rlim_cur = old_fno
>>> +rlimit.setrlimit(rlimit.RLIMIT_NOFILE, lim)
>>> +
>>> +test_run:cmd("stop server sock")
>>> +test_run:cmd("cleanup server sock")
>>> +test_run:cmd("delete server sock")
>>> +test_run:cleanup_cluster()
>>> +
>>> +box.schema.user.revoke('guest', 'replication')
>>> diff --git a/test/replication/misc_orphan_on_reconfiguration_error_gh-4424.result b/test/replication/misc_orphan_on_reconfiguration_error_gh-4424.result
>>> new file mode 100644
>>> index 000000000..6a9d1148d
>>> --- /dev/null
>>> +++ b/test/replication/misc_orphan_on_reconfiguration_error_gh-4424.result
>>> @@ -0,0 +1,88 @@
>>> +uuid = require('uuid')
>>> +---
>>> +...
>>> +test_run = require('test_run').new()
>>> +---
>>> +...
>>> +fiber = require('fiber')
>> UUID and fiber aren't used in this testcase
>>
>>> +---
>>> +...
>>> +--
>>> +-- gh-4424 Always enter orphan mode on error in replication
>>> +-- configuration change.
>>> +--
>>> +replication_connect_timeout = box.cfg.replication_connect_timeout
>>> +---
>>> +...
>>> +replication_connect_quorum = box.cfg.replication_connect_quorum
>>> +---
>>> +...
>>> +box.cfg{replication="12345", replication_connect_timeout=0.1, replication_connect_quorum=1}
>>> +---
>>> +...
>>> +box.info.status
>>> +---
>>> +- orphan
>>> +...
>>> +box.info.ro
>>> +---
>>> +- true
>>> +...
>>> +-- reset replication => leave orphan mode
>>> +box.cfg{replication=""}
>>> +---
>>> +...
>>> +box.info.status
>>> +---
>>> +- running
>>> +...
>>> +box.info.ro
>>> +---
>>> +- false
>>> +...
>>> +-- no switch to orphan when quorum == 0
>>> +box.cfg{replication="12345", replication_connect_quorum=0}
>>> +---
>>> +...
>>> +box.info.status
>>> +---
>>> +- running
>>> +...
>>> +box.info.ro
>>> +---
>>> +- false
>>> +...
>>> +-- we could connect to one out of two replicas. Set orphan.
>>> +box.cfg{replication_connect_quorum=2}
>>> +---
>>> +...
>>> +box.cfg{replication={box.cfg.listen, "12345"}}
>>> +---
>>> +...
>>> +box.info.status
>>> +---
>>> +- orphan
>>> +...
>>> +box.info.ro
>>> +---
>>> +- true
>>> +...
>>> +-- lower quorum => leave orphan mode
>>> +box.cfg{replication_connect_quorum=1}
>>> +---
>>> +...
>>> +box.info.status
>>> +---
>>> +- running
>>> +...
>>> +box.info.ro
>>> +---
>>> +- false
>>> +...
>>> +box.cfg{ \
>>> + replication = {}, \
>>> + replication_connect_quorum = replication_connect_quorum, \
>>> + replication_connect_timeout = replication_connect_timeout \
>>> +}
>>> +---
>>> +...
>>> diff --git a/test/replication/misc_orphan_on_reconfiguration_error_gh-4424.test.lua b/test/replication/misc_orphan_on_reconfiguration_error_gh-4424.test.lua
>>> new file mode 100644
>>> index 000000000..7d1a70a36
>>> --- /dev/null
>>> +++ b/test/replication/misc_orphan_on_reconfiguration_error_gh-4424.test.lua
>>> @@ -0,0 +1,37 @@
>>> +uuid = require('uuid')
>>> +test_run = require('test_run').new()
>>> +fiber = require('fiber')
>>> +
>>> +--
>>> +-- gh-4424 Always enter orphan mode on error in replication
>>> +-- configuration change.
>>> +--
>>> +replication_connect_timeout = box.cfg.replication_connect_timeout
>>> +replication_connect_quorum = box.cfg.replication_connect_quorum
>>> +box.cfg{replication="12345", replication_connect_timeout=0.1, replication_connect_quorum=1}
>>> +box.info.status
>>> +box.info.ro
>>> +-- reset replication => leave orphan mode
>>> +box.cfg{replication=""}
>>> +box.info.status
>>> +box.info.ro
>>> +-- no switch to orphan when quorum == 0
>>> +box.cfg{replication="12345", replication_connect_quorum=0}
>>> +box.info.status
>>> +box.info.ro
>>> +
>>> +-- we could connect to one out of two replicas. Set orphan.
>>> +box.cfg{replication_connect_quorum=2}
>>> +box.cfg{replication={box.cfg.listen, "12345"}}
>>> +box.info.status
>>> +box.info.ro
>>> +-- lower quorum => leave orphan mode
>>> +box.cfg{replication_connect_quorum=1}
>>> +box.info.status
>>> +box.info.ro
>>> +
>>> +box.cfg{ \
>>> + replication = {}, \
>>> + replication_connect_quorum = replication_connect_quorum, \
>>> + replication_connect_timeout = replication_connect_timeout \
>>> +}
>>> diff --git a/test/replication/misc_rebootstrap_from_ro_master_gh-3111.result b/test/replication/misc_rebootstrap_from_ro_master_gh-3111.result
>>> new file mode 100644
>>> index 000000000..7ffca1585
>>> --- /dev/null
>>> +++ b/test/replication/misc_rebootstrap_from_ro_master_gh-3111.result
>>> @@ -0,0 +1,58 @@
>>> +test_run = require('test_run').new()
>>> +---
>>> +...
>>> +test_run:cmd("restart server default")
>> Why do you need it? Seems like the test passes just fine without the
>> restart.
>> Restarting the server will take some time if this is executed after some
>> test that
>> performs lots of DML operations.So You better remove the change
>>
>>> +uuid = require('uuid')
>>> +---
>>> +...
>>> +box.schema.user.grant('guest', 'replication')
>>> +---
>>> +...
>>> +-- gh-3111 - Allow to rebootstrap a replica from a read-only master
>>> +replica_uuid = uuid.new()
>>> +---
>>> +...
>>> +test_run:cmd('create server test with rpl_master=default, script="replication/replica_uuid.lua"')
>>> +---
>>> +- true
>>> +...
>>> +test_run:cmd(string.format('start server test with args="%s"', replica_uuid))
>>> +---
>>> +- true
>>> +...
>>> +test_run:cmd('stop server test')
>>> +---
>>> +- true
>>> +...
>>> +test_run:cmd('cleanup server test')
>>> +---
>>> +- true
>>> +...
>>> +box.cfg{read_only = true}
>>> +---
>>> +...
>>> +test_run:cmd(string.format('start server test with args="%s"', replica_uuid))
>>> +---
>>> +- true
>>> +...
>>> +test_run:cmd('stop server test')
>>> +---
>>> +- true
>>> +...
>>> +test_run:cmd('cleanup server test')
>>> +---
>>> +- true
>>> +...
>>> +box.cfg{read_only = false}
>>> +---
>>> +...
>>> +test_run:cmd('delete server test')
>>> +---
>>> +- true
>>> +...
>>> +test_run:cleanup_cluster()
>>> +---
>>> +...
>>> +box.schema.user.revoke('guest', 'replication')
>>> +---
>>> +...
>>> diff --git a/test/replication/misc_rebootstrap_from_ro_master_gh-3111.test.lua b/test/replication/misc_rebootstrap_from_ro_master_gh-3111.test.lua
>>> new file mode 100644
>>> index 000000000..bb9b4a80f
>>> --- /dev/null
>>> +++ b/test/replication/misc_rebootstrap_from_ro_master_gh-3111.test.lua
>>> @@ -0,0 +1,20 @@
>>> +test_run = require('test_run').new()
>>> +test_run:cmd("restart server default")
>>> +uuid = require('uuid')
>>> +
>>> +box.schema.user.grant('guest', 'replication')
>>> +
>>> +-- gh-3111 - Allow to rebootstrap a replica from a read-only master
>>> +replica_uuid = uuid.new()
>>> +test_run:cmd('create server test with rpl_master=default, script="replication/replica_uuid.lua"')
>>> +test_run:cmd(string.format('start server test with args="%s"', replica_uuid))
>>> +test_run:cmd('stop server test')
>>> +test_run:cmd('cleanup server test')
>>> +box.cfg{read_only = true}
>>> +test_run:cmd(string.format('start server test with args="%s"', replica_uuid))
>>> +test_run:cmd('stop server test')
>>> +test_run:cmd('cleanup server test')
>>> +box.cfg{read_only = false}
>>> +test_run:cmd('delete server test')
>>> +test_run:cleanup_cluster()
>>> +box.schema.user.revoke('guest', 'replication')
>>> diff --git a/test/replication/misc_replica_checks_cluster_id_gh-3704.result b/test/replication/misc_replica_checks_cluster_id_gh-3704.result
>>> new file mode 100644
>>> index 000000000..e6bc8b4d8
>>> --- /dev/null
>>> +++ b/test/replication/misc_replica_checks_cluster_id_gh-3704.result
>>> @@ -0,0 +1,71 @@
>>> +test_run = require('test_run').new()
>>> +---
>>> +...
>>> +test_run:cmd("restart server default")
>>> +uuid = require('uuid')
>>> +---
>>> +...
>>> +fiber = require('fiber')
>> No need to restart the instance, and require fiber module.
>>
>> UUID is needed this time.
>>
>>> +---
>>> +...
>>> +--
>>> +-- gh-3704 move cluster id check to replica
>>> +--
>>> +test_run:cmd("create server replica with rpl_master=default, script='replication/replica.lua'")
>>> +---
>>> +- true
>>> +...
>>> +box.schema.user.grant("guest", "replication")
>>> +---
>>> +...
>>> +test_run:cmd("start server replica")
>>> +---
>>> +- true
>>> +...
>>> +test_run:grep_log("replica", "REPLICASET_UUID_MISMATCH")
>>> +---
>>> +- null
>>> +...
>>> +box.info.replication[2].downstream.status
>>> +---
>>> +- follow
>>> +...
>>> +-- change master's cluster uuid and check that replica doesn't connect.
>>> +test_run:cmd("stop server replica")
>>> +---
>>> +- true
>>> +...
>>> +_ = box.space._schema:replace{'cluster', tostring(uuid.new())}
>>> +---
>>> +...
>>> +-- master believes replica is in cluster, but their cluster UUIDs differ.
>>> +test_run:cmd("start server replica")
>>> +---
>>> +- true
>>> +...
>>> +test_run:wait_log("replica", "REPLICASET_UUID_MISMATCH", nil, 1.0)
>>> +---
>>> +- REPLICASET_UUID_MISMATCH
>>> +...
>>> +test_run:wait_downstream(2, {status = 'stopped'})
>>> +---
>>> +- true
>>> +...
>>> +test_run:cmd("stop server replica")
>>> +---
>>> +- true
>>> +...
>>> +test_run:cmd("cleanup server replica")
>>> +---
>>> +- true
>>> +...
>>> +test_run:cmd("delete server replica")
>>> +---
>>> +- true
>>> +...
>>> +test_run:cleanup_cluster()
>>> +---
>>> +...
>>> +box.schema.user.revoke('guest', 'replication')
>>> +---
>>> +...
>>> diff --git a/test/replication/misc_replica_checks_cluster_id_gh-3704.test.lua b/test/replication/misc_replica_checks_cluster_id_gh-3704.test.lua
>>> new file mode 100644
>>> index 000000000..8a23cc1fe
>>> --- /dev/null
>>> +++ b/test/replication/misc_replica_checks_cluster_id_gh-3704.test.lua
>>> @@ -0,0 +1,26 @@
>>> +test_run = require('test_run').new()
>>> +test_run:cmd("restart server default")
>>> +uuid = require('uuid')
>>> +fiber = require('fiber')
>>> +
>>> +--
>>> +-- gh-3704 move cluster id check to replica
>>> +--
>>> +test_run:cmd("create server replica with rpl_master=default, script='replication/replica.lua'")
>>> +box.schema.user.grant("guest", "replication")
>>> +test_run:cmd("start server replica")
>>> +test_run:grep_log("replica", "REPLICASET_UUID_MISMATCH")
>>> +box.info.replication[2].downstream.status
>>> +-- change master's cluster uuid and check that replica doesn't connect.
>>> +test_run:cmd("stop server replica")
>>> +_ = box.space._schema:replace{'cluster', tostring(uuid.new())}
>>> +-- master believes replica is in cluster, but their cluster UUIDs differ.
>>> +test_run:cmd("start server replica")
>>> +test_run:wait_log("replica", "REPLICASET_UUID_MISMATCH", nil, 1.0)
>>> +test_run:wait_downstream(2, {status = 'stopped'})
>>> +
>>> +test_run:cmd("stop server replica")
>>> +test_run:cmd("cleanup server replica")
>>> +test_run:cmd("delete server replica")
>>> +test_run:cleanup_cluster()
>>> +box.schema.user.revoke('guest', 'replication')
>>> diff --git a/test/replication/misc_return_on_quorum_0_gh-3760.result b/test/replication/misc_return_on_quorum_0_gh-3760.result
>>> new file mode 100644
>>> index 000000000..2eb622896
>>> --- /dev/null
>>> +++ b/test/replication/misc_return_on_quorum_0_gh-3760.result
>>> @@ -0,0 +1,48 @@
>>> +uuid = require('uuid')
>>> +---
>>> +...
>>> +test_run = require('test_run').new()
>>> +---
>>> +...
>>> +fiber = require('fiber')
>> Unneeded uuid and fiber
>>
>>> +---
>>> +...
>>> +replication_connect_timeout = box.cfg.replication_connect_timeout
>>> +---
>>> +...
>>> +replication_connect_quorum = box.cfg.replication_connect_quorum
>>> +---
>>> +...
>>> +box.cfg{replication="12345", replication_connect_timeout=0.1, replication_connect_quorum=1}
>> This box.cfg call is extraneous. The test'll be just fine without it.
>>
>>> +---
>>> +...
>>> +--
>>> +-- gh-3760: replication quorum 0 on reconfiguration should return
>>> +-- from box.cfg immediately.
>>> +--
>>> +replication = box.cfg.replication
>>> +---
>>> +...
>>> +box.cfg{ \
>>> + replication = {}, \
>>> + replication_connect_quorum = 0, \
>>> + replication_connect_timeout = 1000000 \
>>> +}
>>> +---
>>> +...
>>> +-- The call below would hang, if quorum 0 is ignored, or checked
>>> +-- too late.
>>> +box.cfg{replication = {'localhost:12345'}}
>>> +---
>>> +...
>>> +box.info.status
>>> +---
>>> +- running
>>> +...
>>> +box.cfg{ \
>>> + replication = {}, \
>>> + replication_connect_quorum = replication_connect_quorum, \
>>> + replication_connect_timeout = replication_connect_timeout \
>>> +}
>>> +---
>>> +...
>>> diff --git a/test/replication/misc_return_on_quorum_0_gh-3760.test.lua b/test/replication/misc_return_on_quorum_0_gh-3760.test.lua
>>> new file mode 100644
>>> index 000000000..9e0651032
>>> --- /dev/null
>>> +++ b/test/replication/misc_return_on_quorum_0_gh-3760.test.lua
>>> @@ -0,0 +1,27 @@
>>> +uuid = require('uuid')
>>> +test_run = require('test_run').new()
>>> +fiber = require('fiber')
>>> +
>>> +replication_connect_timeout = box.cfg.replication_connect_timeout
>>> +replication_connect_quorum = box.cfg.replication_connect_quorum
>>> +box.cfg{replication="12345", replication_connect_timeout=0.1, replication_connect_quorum=1}
>>> +
>>> +--
>>> +-- gh-3760: replication quorum 0 on reconfiguration should return
>>> +-- from box.cfg immediately.
>>> +--
>>> +replication = box.cfg.replication
>>> +box.cfg{ \
>>> + replication = {}, \
>>> + replication_connect_quorum = 0, \
>>> + replication_connect_timeout = 1000000 \
>>> +}
>>> +-- The call below would hang, if quorum 0 is ignored, or checked
>>> +-- too late.
>>> +box.cfg{replication = {'localhost:12345'}}
>>> +box.info.status
>>> +box.cfg{ \
>>> + replication = {}, \
>>> + replication_connect_quorum = replication_connect_quorum, \
>>> + replication_connect_timeout = replication_connect_timeout \
>>> +}
>>> diff --git a/test/replication/misc_value_not_replicated_on_iproto_request_gh-3247.result b/test/replication/misc_value_not_replicated_on_iproto_request_gh-3247.result
>>> new file mode 100644
>>> index 000000000..d416bd9a6
>>> --- /dev/null
>>> +++ b/test/replication/misc_value_not_replicated_on_iproto_request_gh-3247.result
>>> @@ -0,0 +1,90 @@
>>> +test_run = require('test_run').new()
>>> +---
>>> +...
>>> +test_run:cmd("restart server default")
>>> +uuid = require('uuid')
>>> +---
>>> +...
>> You don't need the uuid module in this testcase.
>>
>>
>>> +-- Deploy a cluster.
>>> +SERVERS = { 'autobootstrap1', 'autobootstrap2', 'autobootstrap3' }
>>> +---
>>> +...
>>> +test_run:create_cluster(SERVERS, "replication", {args="0.03"})
>>> +---
>>> +...
>>> +test_run:wait_fullmesh(SERVERS)
>>> +---
>>> +...
>>> +-- gh-3247 - Sequence-generated value is not replicated in case
>>> +-- the request was sent via iproto.
>>> +test_run:cmd("switch autobootstrap1")
>>> +---
>>> +- true
>>> +...
>>> +net_box = require('net.box')
>>> +---
>>> +...
>>> +_ = box.schema.space.create('space1')
>>> +---
>>> +...
>>> +_ = box.schema.sequence.create('seq')
>>> +---
>>> +...
>>> +_ = box.space.space1:create_index('primary', {sequence = true} )
>>> +---
>>> +...
>>> +_ = box.space.space1:create_index('secondary', {parts = {2, 'unsigned'}})
>>> +---
>>> +...
>>> +box.schema.user.grant('guest', 'read,write', 'space', 'space1')
>>> +---
>>> +...
>>> +c = net_box.connect(box.cfg.listen)
>>> +---
>>> +...
>>> +c.space.space1:insert{box.NULL, "data"} -- fails, but bumps sequence value
>>> +---
>>> +- error: 'Tuple field 2 type does not match one required by operation: expected unsigned'
>>> +...
>>> +c.space.space1:insert{box.NULL, 1, "data"}
>>> +---
>>> +- [2, 1, 'data']
>>> +...
>>> +box.space.space1:select{}
>>> +---
>>> +- - [2, 1, 'data']
>>> +...
>>> +vclock = test_run:get_vclock("autobootstrap1")
>>> +---
>>> +...
>>> +vclock[0] = nil
>>> +---
>>> +...
>>> +_ = test_run:wait_vclock("autobootstrap2", vclock)
>>> +---
>>> +...
>>> +test_run:cmd("switch autobootstrap2")
>>> +---
>>> +- true
>>> +...
>>> +box.space.space1:select{}
>>> +---
>>> +- - [2, 1, 'data']
>>> +...
>>> +test_run:cmd("switch autobootstrap1")
>>> +---
>>> +- true
>>> +...
>>> +box.space.space1:drop()
>>> +---
>>> +...
>>> +test_run:cmd("switch default")
>>> +---
>>> +- true
>>> +...
>>> +test_run:drop_cluster(SERVERS)
>>> +---
>>> +...
>>> +test_run:cleanup_cluster()
>>> +---
>>> +...
>>> diff --git a/test/replication/misc_value_not_replicated_on_iproto_request_gh-3247.test.lua b/test/replication/misc_value_not_replicated_on_iproto_request_gh-3247.test.lua
>>> new file mode 100644
>>> index 000000000..586e8b997
>>> --- /dev/null
>>> +++ b/test/replication/misc_value_not_replicated_on_iproto_request_gh-3247.test.lua
>>> @@ -0,0 +1,33 @@
>>> +test_run = require('test_run').new()
>>> +test_run:cmd("restart server default")
>>> +uuid = require('uuid')
>>> +
>>> +-- Deploy a cluster.
>>> +SERVERS = { 'autobootstrap1', 'autobootstrap2', 'autobootstrap3' }
>>> +test_run:create_cluster(SERVERS, "replication", {args="0.03"})
>>> +test_run:wait_fullmesh(SERVERS)
>>> +
>>> +-- gh-3247 - Sequence-generated value is not replicated in case
>>> +-- the request was sent via iproto.
>>> +test_run:cmd("switch autobootstrap1")
>>> +net_box = require('net.box')
>>> +_ = box.schema.space.create('space1')
>>> +_ = box.schema.sequence.create('seq')
>>> +_ = box.space.space1:create_index('primary', {sequence = true} )
>>> +_ = box.space.space1:create_index('secondary', {parts = {2, 'unsigned'}})
>>> +box.schema.user.grant('guest', 'read,write', 'space', 'space1')
>>> +c = net_box.connect(box.cfg.listen)
>>> +c.space.space1:insert{box.NULL, "data"} -- fails, but bumps sequence value
>>> +c.space.space1:insert{box.NULL, 1, "data"}
>>> +box.space.space1:select{}
>>> +vclock = test_run:get_vclock("autobootstrap1")
>>> +vclock[0] = nil
>>> +_ = test_run:wait_vclock("autobootstrap2", vclock)
>>> +test_run:cmd("switch autobootstrap2")
>>> +box.space.space1:select{}
>>> +test_run:cmd("switch autobootstrap1")
>>> +box.space.space1:drop()
>>> +
>>> +test_run:cmd("switch default")
>>> +test_run:drop_cluster(SERVERS)
>>> +test_run:cleanup_cluster()
>>> diff --git a/test/replication/suite.cfg b/test/replication/suite.cfg
>>> index f357b07da..e21daa5ad 100644
>>> --- a/test/replication/suite.cfg
>>> +++ b/test/replication/suite.cfg
>>> @@ -1,6 +1,19 @@
>>> {
>>> "anon.test.lua": {},
>>> - "misc.test.lua": {},
>>> + "misc_assert_connecting_master_twice_gh-3610.test.lua": {},
>>> + "misc_assert_on_server_die_gh-2991.test.lua": {},
>>> + "misc_assert_replica_on_applier_disconnect_gh-3510.test.lua": {},
>>> + "misc_crash_on_box_concurrent_update_gh-3606.test.lua": {},
>>> + "misc_heartbeats_on_master_changes_gh-3160.test.lua": {},
>>> + "misc_no_failure_on_error_reading_wal_gh-4399.test.lua": {},
>>> + "misc_no_panic_on_connected_gh-3637.test.lua": {},
>>> + "misc_no_restart_on_same_configuration_gh-3711.test.lua": {},
>>> + "misc_no_socket_leak_on_replica_disconnect_gh-3642.test.lua": {},
>>> + "misc_orphan_on_reconfiguration_error_gh-4424.test.lua": {},
>>> + "misc_rebootstrap_from_ro_master_gh-3111.test.lua": {},
>>> + "misc_replica_checks_cluster_id_gh-3704.test.lua": {},
>>> + "misc_return_on_quorum_0_gh-3760.test.lua": {},
>>> + "misc_value_not_replicated_on_iproto_request_gh-3247.test.lua": {},
>>> "once.test.lua": {},
>>> "on_replace.test.lua": {},
>>> "status.test.lua": {},
>> --
>> Serge Petrenko
>>
--
Serge Petrenko
More information about the Tarantool-patches
mailing list