From: Serge Petrenko <sergepetrenko@tarantool.org> To: Sergey Bronnikov <sergeyb@tarantool.org>, "Alexander V. Tikhonov" <avtikhon@tarantool.org> Cc: tarantool-patches@dev.tarantool.org Subject: Re: [Tarantool-patches] [sergepetrenko@tarantool.org: Re: [PATCH v1] Divide replication/mist.test.lua] Date: Thu, 3 Sep 2020 14:30:18 +0300 [thread overview] Message-ID: <3ca7abf0-08fd-009f-be4a-af43ce260930@tarantool.org> (raw) In-Reply-To: <13c35d18-e188-a68a-417e-8ae605599979@tarantool.org> 03.09.2020 12:32, Sergey Bronnikov пишет: > Hi, > > I remember that we already discussed naming format for regression > tests when > > Alexander sent patches for box/net.box [1] and decided to add gh issue > number as a postfix to the name. > > BTW, ddl.test.lua was splitted to the tests with gh issue number as a > postfix - [2]. > > > 1. > https://lists.tarantool.org/pipermail/tarantool-patches/2020-April/015355.html > > 2. > https://github.com/tarantool/tarantool/commit/4a8d1ebda3e0eb7b7d6d14f66b77a9c6d8d0815e Ok, I see. I'm ok with it then. > > On 03.09.2020 10:54, Alexander V. Tikhonov wrote: >> Hi Sergey, could you please confirm the the new tests style used >> in dividing box/net.box*.test.lua test. >> >> >> ----- Forwarded message from Serge Petrenko >> <sergepetrenko@tarantool.org> ----- >> >> Date: Thu, 3 Sep 2020 10:45:17 +0300 >> From: Serge Petrenko <sergepetrenko@tarantool.org> >> To: "Alexander V. Tikhonov" <avtikhon@tarantool.org> >> Cc: Mergen Imeev <imeevma@gmail.com>, >> tarantool-patches@dev.tarantool.org, Alexander Turenko >> <alexander.turenko@tarantool.org>, Yukhin Kirill <k.yukhin@corp.mail.ru> >> Subject: Re: [Tarantool-patches] [PATCH v1] Divide >> replication/mist.test.lua >> User-Agent: Mozilla/5.0 (Macintosh; Intel Mac OS X 10.15; rv:68.0) >> Gecko/20100101 Thunderbird/68.12.0 >> >> >> 02.09.2020 13:08, Alexander V. Tikhonov пишет: >>> Hi Sergey, thanks for the review. The major comment on tests naming >>> that it was discussed during fixes in issue [1], and decided to use >>> this naming format, check as an example of it with: >>> >>> ls -al test/box/net.box*.test.lua >>> >>> [1] - https://github.com/tarantool/tarantool/issues/4880 >> Hi, Alex! >> >> Strange. I was guided by our earlier decision and replication tests: >> >> ls test/replication/gh-*.test.lua >> test/replication/gh-4114-local-space-replication.test.lua >> test/replication/gh-4402-info-errno.test.lua >> test/replication/gh-4605-empty-password.test.lua >> test/replication/gh-4606-admin-creds.test.lua >> test/replication/gh-4729-netbox-group-id.test.lua >> test/replication/gh-4730-applier-rollback.test.lua >> test/replication/gh-4739-vclock-assert.test.lua >> test/replication/gh-4928-tx-boundaries.test.lua >> test/replication/gh-5140-qsync-casc-rollback.test.lua >> test/replication/gh-5144-qsync-dup-confirm.test.lua >> test/replication/gh-5163-qsync-restart-crash.test.lua >> test/replication/gh-5167-qsync-rollback-snap.test.lua >> test/replication/gh-5195-qsync-replica-write.test.lua >> >> Anyway, if someone can confirm that we're using another naming >> scheme now, I'm ok with it. >> >>> On Wed, Sep 02, 2020 at 11:09:30AM +0300, Serge Petrenko wrote: >>>> Hi! Thanks for the patch! >>>> >>>> Please see my comments below. >>>> >>>> 17.07.2020 12:25, Alexander V. Tikhonov пишет: >>>>> To fix flaky issues of replication/misc.test.lua the test had to be >>>>> divided into smaller tests to be able to localize the flaky results: >>>>> >>>>> misc_assert_connecting_master_twice_gh-3610.test.lua >>>>> misc_assert_on_server_die_gh-2991.test.lua >>>>> misc_assert_replica_on_applier_disconnect_gh-3510.test.lua >>>>> misc_crash_on_box_concurrent_update_gh-3606.test.lua >>>>> misc_heartbeats_on_master_changes_gh-3160.test.lua >>>>> misc_no_failure_on_error_reading_wal_gh-4399.test.lua >>>>> misc_no_panic_on_connected_gh-3637.test.lua >>>>> misc_no_restart_on_same_configuration_gh-3711.test.lua >>>>> misc_no_socket_leak_on_replica_disconnect_gh-3642.test.lua >>>>> misc_orphan_on_reconfiguration_error_gh-4424.test.lua >>>>> misc_rebootstrap_from_ro_master_gh-3111.test.lua >>>>> misc_replica_checks_cluster_id_gh-3704.test.lua >>>>> misc_return_on_quorum_0_gh-3760.test.lua >>>>> misc_value_not_replicated_on_iproto_request_gh-3247.test.lua >>>> We use a different naming scheme for regression tests. >>>> It should be `gh-xxxx-bug-description`, also notice the dashes >>>> between words instead of the underscores. >>>> >>>>> Needed for #4940 >>>>> --- >>>>> >>>>> Github: >>>>> https://github.com/tarantool/tarantool/tree/avtikhon/gh-4940-replication-misc >>>>> Issue: https://github.com/tarantool/tarantool/issues/4940 >>>>> >>>>> test/replication/misc.result | 866 >>>>> ------------------ >>>>> test/replication/misc.test.lua | 356 ------- >>>>> ...ert_connecting_master_twice_gh-3610.result | 86 ++ >>>>> ...t_connecting_master_twice_gh-3610.test.lua | 34 + >>>>> .../misc_assert_on_server_die_gh-2991.result | 31 + >>>>> ...misc_assert_on_server_die_gh-2991.test.lua | 12 + >>>>> ...plica_on_applier_disconnect_gh-3510.result | 49 + >>>>> ...ica_on_applier_disconnect_gh-3510.test.lua | 17 + >>>>> ...sh_on_box_concurrent_update_gh-3606.result | 50 + >>>>> ..._on_box_concurrent_update_gh-3606.test.lua | 19 + >>>>> ...eartbeats_on_master_changes_gh-3160.result | 76 ++ >>>>> ...rtbeats_on_master_changes_gh-3160.test.lua | 40 + >>>>> ...ailure_on_error_reading_wal_gh-4399.result | 97 ++ >>>>> ...lure_on_error_reading_wal_gh-4399.test.lua | 39 + >>>>> .../misc_no_panic_on_connected_gh-3637.result | 72 ++ >>>>> ...isc_no_panic_on_connected_gh-3637.test.lua | 33 + >>>>> ...start_on_same_configuration_gh-3711.result | 107 +++ >>>>> ...art_on_same_configuration_gh-3711.test.lua | 41 + >>>>> ..._leak_on_replica_disconnect_gh-3642.result | 98 ++ >>>>> ...eak_on_replica_disconnect_gh-3642.test.lua | 44 + >>>>> ...an_on_reconfiguration_error_gh-4424.result | 88 ++ >>>>> ..._on_reconfiguration_error_gh-4424.test.lua | 37 + >>>>> ..._rebootstrap_from_ro_master_gh-3111.result | 58 ++ >>>>> ...ebootstrap_from_ro_master_gh-3111.test.lua | 20 + >>>>> ...c_replica_checks_cluster_id_gh-3704.result | 71 ++ >>>>> ...replica_checks_cluster_id_gh-3704.test.lua | 26 + >>>>> .../misc_return_on_quorum_0_gh-3760.result | 48 + >>>>> .../misc_return_on_quorum_0_gh-3760.test.lua | 27 + >>>>> ...eplicated_on_iproto_request_gh-3247.result | 90 ++ >>>>> ...licated_on_iproto_request_gh-3247.test.lua | 33 + >>>>> test/replication/suite.cfg | 15 +- >>>>> 31 files changed, 1457 insertions(+), 1223 deletions(-) >>>>> delete mode 100644 test/replication/misc.result >>>>> delete mode 100644 test/replication/misc.test.lua >>>>> create mode 100644 >>>>> test/replication/misc_assert_connecting_master_twice_gh-3610.result >>>>> create mode 100644 >>>>> test/replication/misc_assert_connecting_master_twice_gh-3610.test.lua >>>>> create mode 100644 >>>>> test/replication/misc_assert_on_server_die_gh-2991.result >>>>> create mode 100644 >>>>> test/replication/misc_assert_on_server_die_gh-2991.test.lua >>>>> create mode 100644 >>>>> test/replication/misc_assert_replica_on_applier_disconnect_gh-3510.result >>>>> create mode 100644 >>>>> test/replication/misc_assert_replica_on_applier_disconnect_gh-3510.test.lua >>>>> create mode 100644 >>>>> test/replication/misc_crash_on_box_concurrent_update_gh-3606.result >>>>> create mode 100644 >>>>> test/replication/misc_crash_on_box_concurrent_update_gh-3606.test.lua >>>>> create mode 100644 >>>>> test/replication/misc_heartbeats_on_master_changes_gh-3160.result >>>>> create mode 100644 >>>>> test/replication/misc_heartbeats_on_master_changes_gh-3160.test.lua >>>>> create mode 100644 >>>>> test/replication/misc_no_failure_on_error_reading_wal_gh-4399.result >>>>> create mode 100644 >>>>> test/replication/misc_no_failure_on_error_reading_wal_gh-4399.test.lua >>>>> >>>>> create mode 100644 >>>>> test/replication/misc_no_panic_on_connected_gh-3637.result >>>>> create mode 100644 >>>>> test/replication/misc_no_panic_on_connected_gh-3637.test.lua >>>>> create mode 100644 >>>>> test/replication/misc_no_restart_on_same_configuration_gh-3711.result >>>>> create mode 100644 >>>>> test/replication/misc_no_restart_on_same_configuration_gh-3711.test.lua >>>>> >>>>> create mode 100644 >>>>> test/replication/misc_no_socket_leak_on_replica_disconnect_gh-3642.result >>>>> create mode 100644 >>>>> test/replication/misc_no_socket_leak_on_replica_disconnect_gh-3642.test.lua >>>>> create mode 100644 >>>>> test/replication/misc_orphan_on_reconfiguration_error_gh-4424.result >>>>> create mode 100644 >>>>> test/replication/misc_orphan_on_reconfiguration_error_gh-4424.test.lua >>>>> >>>>> create mode 100644 >>>>> test/replication/misc_rebootstrap_from_ro_master_gh-3111.result >>>>> create mode 100644 >>>>> test/replication/misc_rebootstrap_from_ro_master_gh-3111.test.lua >>>>> create mode 100644 >>>>> test/replication/misc_replica_checks_cluster_id_gh-3704.result >>>>> create mode 100644 >>>>> test/replication/misc_replica_checks_cluster_id_gh-3704.test.lua >>>>> create mode 100644 >>>>> test/replication/misc_return_on_quorum_0_gh-3760.result >>>>> create mode 100644 >>>>> test/replication/misc_return_on_quorum_0_gh-3760.test.lua >>>>> create mode 100644 >>>>> test/replication/misc_value_not_replicated_on_iproto_request_gh-3247.result >>>>> create mode 100644 >>>>> test/replication/misc_value_not_replicated_on_iproto_request_gh-3247.test.lua >>>>> >>>>> diff --git a/test/replication/misc.result >>>>> b/test/replication/misc.result >>>>> deleted file mode 100644 >>>>> index e5d1f560e..000000000 >>>>> --- a/test/replication/misc.result >>>>> +++ /dev/null >>>>> @@ -1,866 +0,0 @@ >>>>> -uuid = require('uuid') >>>>> ---- >>>>> -... >>>>> -test_run = require('test_run').new() >>>>> ---- >>>>> -... >>>>> -box.schema.user.grant('guest', 'replication') >>>>> ---- >>>>> -... >>>>> --- gh-2991 - Tarantool asserts on box.cfg.replication update if >>>>> one of >>>>> --- servers is dead >>>>> -replication_timeout = box.cfg.replication_timeout >>>>> ---- >>>>> -... >>>>> -replication_connect_timeout = box.cfg.replication_connect_timeout >>>>> ---- >>>>> -... >>>>> -box.cfg{replication_timeout=0.05, >>>>> replication_connect_timeout=0.05, replication={}} >>>>> ---- >>>>> -... >>>>> -box.cfg{replication_connect_quorum=2} >>>>> ---- >>>>> -... >>>>> -box.cfg{replication = {'127.0.0.1:12345', box.cfg.listen}} >>>>> ---- >>>>> -... >>>>> -box.info.status >>>>> ---- >>>>> -- orphan >>>>> -... >>>>> -box.info.ro >>>>> ---- >>>>> -- true >>>>> -... >>>>> --- gh-3606 - Tarantool crashes if box.cfg.replication is updated >>>>> concurrently >>>>> -fiber = require('fiber') >>>>> ---- >>>>> -... >>>>> -c = fiber.channel(2) >>>>> ---- >>>>> -... >>>>> -f = function() fiber.create(function() pcall(box.cfg, >>>>> {replication = {12345}}) c:put(true) end) end >>>>> ---- >>>>> -... >>>>> -f() >>>>> ---- >>>>> -... >>>>> -f() >>>>> ---- >>>>> -... >>>>> -c:get() >>>>> ---- >>>>> -- true >>>>> -... >>>>> -c:get() >>>>> ---- >>>>> -- true >>>>> -... >>>>> -box.cfg{replication = "", replication_timeout = >>>>> replication_timeout, replication_connect_timeout = >>>>> replication_connect_timeout} >>>>> ---- >>>>> -... >>>>> -box.info.status >>>>> ---- >>>>> -- running >>>>> -... >>>>> -box.info.ro >>>>> ---- >>>>> -- false >>>>> -... >>>>> --- gh-3111 - Allow to rebootstrap a replica from a read-only master >>>>> -replica_uuid = uuid.new() >>>>> ---- >>>>> -... >>>>> -test_run:cmd('create server test with rpl_master=default, >>>>> script="replication/replica_uuid.lua"') >>>>> ---- >>>>> -- true >>>>> -... >>>>> -test_run:cmd(string.format('start server test with args="%s"', >>>>> replica_uuid)) >>>>> ---- >>>>> -- true >>>>> -... >>>>> -test_run:cmd('stop server test') >>>>> ---- >>>>> -- true >>>>> -... >>>>> -test_run:cmd('cleanup server test') >>>>> ---- >>>>> -- true >>>>> -... >>>>> -box.cfg{read_only = true} >>>>> ---- >>>>> -... >>>>> -test_run:cmd(string.format('start server test with args="%s"', >>>>> replica_uuid)) >>>>> ---- >>>>> -- true >>>>> -... >>>>> -test_run:cmd('stop server test') >>>>> ---- >>>>> -- true >>>>> -... >>>>> -test_run:cmd('cleanup server test') >>>>> ---- >>>>> -- true >>>>> -... >>>>> -box.cfg{read_only = false} >>>>> ---- >>>>> -... >>>>> -test_run:cmd('delete server test') >>>>> ---- >>>>> -- true >>>>> -... >>>>> -test_run:cleanup_cluster() >>>>> ---- >>>>> -... >>>>> --- gh-3160 - Send heartbeats if there are changes from a remote >>>>> master only >>>>> -SERVERS = { 'autobootstrap1', 'autobootstrap2', 'autobootstrap3' } >>>>> ---- >>>>> -... >>>>> --- Deploy a cluster. >>>>> -test_run:create_cluster(SERVERS, "replication", {args="0.03"}) >>>>> ---- >>>>> -... >>>>> -test_run:wait_fullmesh(SERVERS) >>>>> ---- >>>>> -... >>>>> -test_run:cmd("switch autobootstrap3") >>>>> ---- >>>>> -- true >>>>> -... >>>>> -test_run = require('test_run').new() >>>>> ---- >>>>> -... >>>>> -fiber = require('fiber') >>>>> ---- >>>>> -... >>>>> -_ = box.schema.space.create('test_timeout'):create_index('pk') >>>>> ---- >>>>> -... >>>>> -test_run:cmd("setopt delimiter ';'") >>>>> ---- >>>>> -- true >>>>> -... >>>>> -function wait_not_follow(replicaA, replicaB) >>>>> - return test_run:wait_cond(function() >>>>> - return replicaA.status ~= 'follow' or replicaB.status ~= >>>>> 'follow' >>>>> - end, box.cfg.replication_timeout) >>>>> -end; >>>>> ---- >>>>> -... >>>>> -function test_timeout() >>>>> - local replicaA = box.info.replication[1].upstream or >>>>> box.info.replication[2].upstream >>>>> - local replicaB = box.info.replication[3].upstream or >>>>> box.info.replication[2].upstream >>>>> - local follows = test_run:wait_cond(function() >>>>> - return replicaA.status == 'follow' or replicaB.status == >>>>> 'follow' >>>>> - end) >>>>> - if not follows then error('replicas are not in the follow >>>>> status') end >>>>> - for i = 0, 99 do >>>>> - box.space.test_timeout:replace({1}) >>>>> - if wait_not_follow(replicaA, replicaB) then >>>>> - return error(box.info.replication) >>>>> - end >>>>> - end >>>>> - return true >>>>> -end; >>>>> ---- >>>>> -... >>>>> -test_run:cmd("setopt delimiter ''"); >>>>> ---- >>>>> -- true >>>>> -... >>>>> -test_timeout() >>>>> ---- >>>>> -- true >>>>> -... >>>>> --- gh-3247 - Sequence-generated value is not replicated in case >>>>> --- the request was sent via iproto. >>>>> -test_run:cmd("switch autobootstrap1") >>>>> ---- >>>>> -- true >>>>> -... >>>>> -net_box = require('net.box') >>>>> ---- >>>>> -... >>>>> -_ = box.schema.space.create('space1') >>>>> ---- >>>>> -... >>>>> -_ = box.schema.sequence.create('seq') >>>>> ---- >>>>> -... >>>>> -_ = box.space.space1:create_index('primary', {sequence = true} ) >>>>> ---- >>>>> -... >>>>> -_ = box.space.space1:create_index('secondary', {parts = {2, >>>>> 'unsigned'}}) >>>>> ---- >>>>> -... >>>>> -box.schema.user.grant('guest', 'read,write', 'space', 'space1') >>>>> ---- >>>>> -... >>>>> -c = net_box.connect(box.cfg.listen) >>>>> ---- >>>>> -... >>>>> -c.space.space1:insert{box.NULL, "data"} -- fails, but bumps >>>>> sequence value >>>>> ---- >>>>> -- error: 'Tuple field 2 type does not match one required by >>>>> operation: expected unsigned' >>>>> -... >>>>> -c.space.space1:insert{box.NULL, 1, "data"} >>>>> ---- >>>>> -- [2, 1, 'data'] >>>>> -... >>>>> -box.space.space1:select{} >>>>> ---- >>>>> -- - [2, 1, 'data'] >>>>> -... >>>>> -vclock = test_run:get_vclock("autobootstrap1") >>>>> ---- >>>>> -... >>>>> -vclock[0] = nil >>>>> ---- >>>>> -... >>>>> -_ = test_run:wait_vclock("autobootstrap2", vclock) >>>>> ---- >>>>> -... >>>>> -test_run:cmd("switch autobootstrap2") >>>>> ---- >>>>> -- true >>>>> -... >>>>> -box.space.space1:select{} >>>>> ---- >>>>> -- - [2, 1, 'data'] >>>>> -... >>>>> -test_run:cmd("switch autobootstrap1") >>>>> ---- >>>>> -- true >>>>> -... >>>>> -box.space.space1:drop() >>>>> ---- >>>>> -... >>>>> -test_run:cmd("switch default") >>>>> ---- >>>>> -- true >>>>> -... >>>>> -test_run:drop_cluster(SERVERS) >>>>> ---- >>>>> -... >>>>> -test_run:cleanup_cluster() >>>>> ---- >>>>> -... >>>>> --- gh-3642 - Check that socket file descriptor doesn't leak >>>>> --- when a replica is disconnected. >>>>> -rlimit = require('rlimit') >>>>> ---- >>>>> -... >>>>> -lim = rlimit.limit() >>>>> ---- >>>>> -... >>>>> -rlimit.getrlimit(rlimit.RLIMIT_NOFILE, lim) >>>>> ---- >>>>> -... >>>>> -old_fno = lim.rlim_cur >>>>> ---- >>>>> -... >>>>> -lim.rlim_cur = 64 >>>>> ---- >>>>> -... >>>>> -rlimit.setrlimit(rlimit.RLIMIT_NOFILE, lim) >>>>> ---- >>>>> -... >>>>> -test_run:cmd('create server sock with rpl_master=default, >>>>> script="replication/replica.lua"') >>>>> ---- >>>>> -- true >>>>> -... >>>>> -test_run:cmd('start server sock') >>>>> ---- >>>>> -- true >>>>> -... >>>>> -test_run:cmd('switch sock') >>>>> ---- >>>>> -- true >>>>> -... >>>>> -test_run = require('test_run').new() >>>>> ---- >>>>> -... >>>>> -fiber = require('fiber') >>>>> ---- >>>>> -... >>>>> -test_run:cmd("setopt delimiter ';'") >>>>> ---- >>>>> -- true >>>>> -... >>>>> -for i = 1, 64 do >>>>> - local replication = box.cfg.replication >>>>> - box.cfg{replication = {}} >>>>> - box.cfg{replication = replication} >>>>> - while box.info.replication[1].upstream.status ~= 'follow' do >>>>> - fiber.sleep(0.001) >>>>> - end >>>>> -end; >>>>> ---- >>>>> -... >>>>> -test_run:cmd("setopt delimiter ''"); >>>>> ---- >>>>> -- true >>>>> -... >>>>> -box.info.replication[1].upstream.status >>>>> ---- >>>>> -- follow >>>>> -... >>>>> -test_run:cmd('switch default') >>>>> ---- >>>>> -- true >>>>> -... >>>>> -lim.rlim_cur = old_fno >>>>> ---- >>>>> -... >>>>> -rlimit.setrlimit(rlimit.RLIMIT_NOFILE, lim) >>>>> ---- >>>>> -... >>>>> -test_run:cmd("stop server sock") >>>>> ---- >>>>> -- true >>>>> -... >>>>> -test_run:cmd("cleanup server sock") >>>>> ---- >>>>> -- true >>>>> -... >>>>> -test_run:cmd("delete server sock") >>>>> ---- >>>>> -- true >>>>> -... >>>>> -test_run:cleanup_cluster() >>>>> ---- >>>>> -... >>>>> -box.schema.user.revoke('guest', 'replication') >>>>> ---- >>>>> -... >>>>> --- gh-3510 assertion failure in replica_on_applier_disconnect() >>>>> -test_run:cmd('create server er_load1 with >>>>> script="replication/er_load1.lua"') >>>>> ---- >>>>> -- true >>>>> -... >>>>> -test_run:cmd('create server er_load2 with >>>>> script="replication/er_load2.lua"') >>>>> ---- >>>>> -- true >>>>> -... >>>>> -test_run:cmd('start server er_load1 with wait=False, >>>>> wait_load=False') >>>>> ---- >>>>> -- true >>>>> -... >>>>> --- Instance er_load2 will fail with error >>>>> ER_REPLICASET_UUID_MISMATCH. >>>>> --- This is OK since we only test here that er_load1 doesn't assert. >>>>> -test_run:cmd('start server er_load2 with wait=True, >>>>> wait_load=True, crash_expected = True') >>>>> ---- >>>>> -- false >>>>> -... >>>>> -test_run:cmd('stop server er_load1') >>>>> ---- >>>>> -- true >>>>> -... >>>>> --- er_load2 exits automatically. >>>>> -test_run:cmd('cleanup server er_load1') >>>>> ---- >>>>> -- true >>>>> -... >>>>> -test_run:cmd('cleanup server er_load2') >>>>> ---- >>>>> -- true >>>>> -... >>>>> -test_run:cmd('delete server er_load1') >>>>> ---- >>>>> -- true >>>>> -... >>>>> -test_run:cmd('delete server er_load2') >>>>> ---- >>>>> -- true >>>>> -... >>>>> -test_run:cleanup_cluster() >>>>> ---- >>>>> -... >>>>> --- >>>>> --- Test case for gh-3637, gh-4550. Before the fix replica would >>>>> --- exit with an error if a user does not exist or a password is >>>>> --- incorrect. Now check that we don't hang/panic and successfully >>>>> --- connect. >>>>> --- >>>>> -fiber = require('fiber') >>>>> ---- >>>>> -... >>>>> -test_run:cmd("create server replica_auth with rpl_master=default, >>>>> script='replication/replica_auth.lua'") >>>>> ---- >>>>> -- true >>>>> -... >>>>> -test_run:cmd("start server replica_auth with wait=False, >>>>> wait_load=False, args='cluster:pass 0.05'") >>>>> ---- >>>>> -- true >>>>> -... >>>>> --- Wait a bit to make sure replica waits till user is created. >>>>> -fiber.sleep(0.1) >>>>> ---- >>>>> -... >>>>> -box.schema.user.create('cluster') >>>>> ---- >>>>> -... >>>>> --- The user is created. Let the replica fail auth request due to >>>>> --- a wrong password. >>>>> -fiber.sleep(0.1) >>>>> ---- >>>>> -... >>>>> -box.schema.user.passwd('cluster', 'pass') >>>>> ---- >>>>> -... >>>>> -box.schema.user.grant('cluster', 'replication') >>>>> ---- >>>>> -... >>>>> -while box.info.replication[2] == nil do fiber.sleep(0.01) end >>>>> ---- >>>>> -... >>>>> -vclock = test_run:get_vclock('default') >>>>> ---- >>>>> -... >>>>> -vclock[0] = nil >>>>> ---- >>>>> -... >>>>> -_ = test_run:wait_vclock('replica_auth', vclock) >>>>> ---- >>>>> -... >>>>> -test_run:cmd("stop server replica_auth") >>>>> ---- >>>>> -- true >>>>> -... >>>>> -test_run:cmd("cleanup server replica_auth") >>>>> ---- >>>>> -- true >>>>> -... >>>>> -test_run:cmd("delete server replica_auth") >>>>> ---- >>>>> -- true >>>>> -... >>>>> -test_run:cleanup_cluster() >>>>> ---- >>>>> -... >>>>> -box.schema.user.drop('cluster') >>>>> ---- >>>>> -... >>>>> --- >>>>> --- Test case for gh-3610. Before the fix replica would fail with >>>>> the assertion >>>>> --- when trying to connect to the same master twice. >>>>> --- >>>>> -box.schema.user.grant('guest', 'replication') >>>>> ---- >>>>> -... >>>>> -test_run:cmd("create server replica with rpl_master=default, >>>>> script='replication/replica.lua'") >>>>> ---- >>>>> -- true >>>>> -... >>>>> -test_run:cmd("start server replica") >>>>> ---- >>>>> -- true >>>>> -... >>>>> -test_run:cmd("switch replica") >>>>> ---- >>>>> -- true >>>>> -... >>>>> -replication = box.cfg.replication[1] >>>>> ---- >>>>> -... >>>>> -box.cfg{replication = {replication, replication}} >>>>> ---- >>>>> -- error: 'Incorrect value for option ''replication'': duplicate >>>>> connection to the >>>>> - same replica' >>>>> -... >>>>> --- Check the case when duplicate connection is detected in the >>>>> background. >>>>> -test_run:cmd("switch default") >>>>> ---- >>>>> -- true >>>>> -... >>>>> -listen = box.cfg.listen >>>>> ---- >>>>> -... >>>>> -box.cfg{listen = ''} >>>>> ---- >>>>> -... >>>>> -test_run:cmd("switch replica") >>>>> ---- >>>>> -- true >>>>> -... >>>>> -box.cfg{replication_connect_quorum = 0, >>>>> replication_connect_timeout = 0.01} >>>>> ---- >>>>> -... >>>>> -box.cfg{replication = {replication, replication}} >>>>> ---- >>>>> -... >>>>> -test_run:cmd("switch default") >>>>> ---- >>>>> -- true >>>>> -... >>>>> -box.cfg{listen = listen} >>>>> ---- >>>>> -... >>>>> -while test_run:grep_log('replica', 'duplicate connection') == nil >>>>> do fiber.sleep(0.01) end >>>>> ---- >>>>> -... >>>>> -test_run:cmd("stop server replica") >>>>> ---- >>>>> -- true >>>>> -... >>>>> -test_run:cmd("cleanup server replica") >>>>> ---- >>>>> -- true >>>>> -... >>>>> -test_run:cmd("delete server replica") >>>>> ---- >>>>> -- true >>>>> -... >>>>> -test_run:cleanup_cluster() >>>>> ---- >>>>> -... >>>>> -box.schema.user.revoke('guest', 'replication') >>>>> ---- >>>>> -... >>>>> --- >>>>> --- gh-3711 Do not restart replication on box.cfg in case the >>>>> --- configuration didn't change. >>>>> --- >>>>> -box.schema.user.grant('guest', 'replication') >>>>> ---- >>>>> -... >>>>> -test_run:cmd("create server replica with rpl_master=default, >>>>> script='replication/replica.lua'") >>>>> ---- >>>>> -- true >>>>> -... >>>>> -test_run:cmd("start server replica") >>>>> ---- >>>>> -- true >>>>> -... >>>>> --- Access rights are checked only during reconnect. If the new >>>>> --- and old configurations are equivalent, no reconnect will be >>>>> --- issued and replication should continue working. >>>>> -box.schema.user.revoke('guest', 'replication') >>>>> ---- >>>>> -... >>>>> -test_run:cmd("switch replica") >>>>> ---- >>>>> -- true >>>>> -... >>>>> -replication = box.cfg.replication[1] >>>>> ---- >>>>> -... >>>>> -box.cfg{replication = {replication}} >>>>> ---- >>>>> -... >>>>> -box.info.status == 'running' >>>>> ---- >>>>> -- true >>>>> -... >>>>> -box.cfg{replication = replication} >>>>> ---- >>>>> -... >>>>> -box.info.status == 'running' >>>>> ---- >>>>> -- true >>>>> -... >>>>> --- Check that comparison of tables works as expected as well. >>>>> -test_run:cmd("switch default") >>>>> ---- >>>>> -- true >>>>> -... >>>>> -box.schema.user.grant('guest', 'replication') >>>>> ---- >>>>> -... >>>>> -test_run:cmd("switch replica") >>>>> ---- >>>>> -- true >>>>> -... >>>>> -replication = box.cfg.replication >>>>> ---- >>>>> -... >>>>> -table.insert(replication, box.cfg.listen) >>>>> ---- >>>>> -... >>>>> -test_run:cmd("switch default") >>>>> ---- >>>>> -- true >>>>> -... >>>>> -box.schema.user.revoke('guest', 'replication') >>>>> ---- >>>>> -... >>>>> -test_run:cmd("switch replica") >>>>> ---- >>>>> -- true >>>>> -... >>>>> -box.cfg{replication = replication} >>>>> ---- >>>>> -... >>>>> -box.info.status == 'running' >>>>> ---- >>>>> -- true >>>>> -... >>>>> -test_run:cmd("switch default") >>>>> ---- >>>>> -- true >>>>> -... >>>>> -test_run:cmd("stop server replica") >>>>> ---- >>>>> -- true >>>>> -... >>>>> -test_run:cmd("cleanup server replica") >>>>> ---- >>>>> -- true >>>>> -... >>>>> -test_run:cmd("delete server replica") >>>>> ---- >>>>> -- true >>>>> -... >>>>> -test_run:cleanup_cluster() >>>>> ---- >>>>> -... >>>>> --- >>>>> --- gh-3704 move cluster id check to replica >>>>> --- >>>>> -test_run:cmd("create server replica with rpl_master=default, >>>>> script='replication/replica.lua'") >>>>> ---- >>>>> -- true >>>>> -... >>>>> -box.schema.user.grant("guest", "replication") >>>>> ---- >>>>> -... >>>>> -test_run:cmd("start server replica") >>>>> ---- >>>>> -- true >>>>> -... >>>>> -test_run:grep_log("replica", "REPLICASET_UUID_MISMATCH") >>>>> ---- >>>>> -- null >>>>> -... >>>>> -box.info.replication[2].downstream.status >>>>> ---- >>>>> -- follow >>>>> -... >>>>> --- change master's cluster uuid and check that replica doesn't >>>>> connect. >>>>> -test_run:cmd("stop server replica") >>>>> ---- >>>>> -- true >>>>> -... >>>>> -_ = box.space._schema:replace{'cluster', tostring(uuid.new())} >>>>> ---- >>>>> -... >>>>> --- master believes replica is in cluster, but their cluster UUIDs >>>>> differ. >>>>> -test_run:cmd("start server replica") >>>>> ---- >>>>> -- true >>>>> -... >>>>> -test_run:wait_log("replica", "REPLICASET_UUID_MISMATCH", nil, 1.0) >>>>> ---- >>>>> -- REPLICASET_UUID_MISMATCH >>>>> -... >>>>> -test_run:wait_downstream(2, {status = 'stopped'}) >>>>> ---- >>>>> -- true >>>>> -... >>>>> -test_run:cmd("stop server replica") >>>>> ---- >>>>> -- true >>>>> -... >>>>> -test_run:cmd("cleanup server replica") >>>>> ---- >>>>> -- true >>>>> -... >>>>> -test_run:cmd("delete server replica") >>>>> ---- >>>>> -- true >>>>> -... >>>>> -test_run:cleanup_cluster() >>>>> ---- >>>>> -... >>>>> -box.schema.user.revoke('guest', 'replication') >>>>> ---- >>>>> -... >>>>> --- >>>>> --- gh-4399 Check that an error reading WAL directory on subscribe >>>>> --- doesn't lead to a permanent replication failure. >>>>> --- >>>>> -box.schema.user.grant("guest", "replication") >>>>> ---- >>>>> -... >>>>> -test_run:cmd("create server replica with rpl_master=default, >>>>> script='replication/replica.lua'") >>>>> ---- >>>>> -- true >>>>> -... >>>>> -test_run:cmd("start server replica") >>>>> ---- >>>>> -- true >>>>> -... >>>>> --- Make the WAL directory inaccessible. >>>>> -fio = require('fio') >>>>> ---- >>>>> -... >>>>> -path = fio.abspath(box.cfg.wal_dir) >>>>> ---- >>>>> -... >>>>> -fio.chmod(path, 0) >>>>> ---- >>>>> -- true >>>>> -... >>>>> --- Break replication on timeout. >>>>> -replication_timeout = box.cfg.replication_timeout >>>>> ---- >>>>> -... >>>>> -box.cfg{replication_timeout = 9000} >>>>> ---- >>>>> -... >>>>> -test_run:cmd("switch replica") >>>>> ---- >>>>> -- true >>>>> -... >>>>> -test_run:wait_cond(function() return >>>>> box.info.replication[1].upstream.status ~= 'follow' end) >>>>> ---- >>>>> -- true >>>>> -... >>>>> -require('fiber').sleep(box.cfg.replication_timeout) >>>>> ---- >>>>> -... >>>>> -test_run:cmd("switch default") >>>>> ---- >>>>> -- true >>>>> -... >>>>> -box.cfg{replication_timeout = replication_timeout} >>>>> ---- >>>>> -... >>>>> --- Restore access to the WAL directory. >>>>> --- Wait for replication to be reestablished. >>>>> -fio.chmod(path, tonumber('777', 8)) >>>>> ---- >>>>> -- true >>>>> -... >>>>> -test_run:cmd("switch replica") >>>>> ---- >>>>> -- true >>>>> -... >>>>> -test_run:wait_cond(function() return >>>>> box.info.replication[1].upstream.status == 'follow' end) >>>>> ---- >>>>> -- true >>>>> -... >>>>> -test_run:cmd("switch default") >>>>> ---- >>>>> -- true >>>>> -... >>>>> -test_run:cmd("stop server replica") >>>>> ---- >>>>> -- true >>>>> -... >>>>> -test_run:cmd("cleanup server replica") >>>>> ---- >>>>> -- true >>>>> -... >>>>> -test_run:cmd("delete server replica") >>>>> ---- >>>>> -- true >>>>> -... >>>>> -test_run:cleanup_cluster() >>>>> ---- >>>>> -... >>>>> -box.schema.user.revoke('guest', 'replication') >>>>> ---- >>>>> -... >>>>> --- >>>>> --- gh-4424 Always enter orphan mode on error in replication >>>>> --- configuration change. >>>>> --- >>>>> -replication_connect_timeout = box.cfg.replication_connect_timeout >>>>> ---- >>>>> -... >>>>> -replication_connect_quorum = box.cfg.replication_connect_quorum >>>>> ---- >>>>> -... >>>>> -box.cfg{replication="12345", replication_connect_timeout=0.1, >>>>> replication_connect_quorum=1} >>>>> ---- >>>>> -... >>>>> -box.info.status >>>>> ---- >>>>> -- orphan >>>>> -... >>>>> -box.info.ro >>>>> ---- >>>>> -- true >>>>> -... >>>>> --- reset replication => leave orphan mode >>>>> -box.cfg{replication=""} >>>>> ---- >>>>> -... >>>>> -box.info.status >>>>> ---- >>>>> -- running >>>>> -... >>>>> -box.info.ro >>>>> ---- >>>>> -- false >>>>> -... >>>>> --- no switch to orphan when quorum == 0 >>>>> -box.cfg{replication="12345", replication_connect_quorum=0} >>>>> ---- >>>>> -... >>>>> -box.info.status >>>>> ---- >>>>> -- running >>>>> -... >>>>> -box.info.ro >>>>> ---- >>>>> -- false >>>>> -... >>>>> --- we could connect to one out of two replicas. Set orphan. >>>>> -box.cfg{replication_connect_quorum=2} >>>>> ---- >>>>> -... >>>>> -box.cfg{replication={box.cfg.listen, "12345"}} >>>>> ---- >>>>> -... >>>>> -box.info.status >>>>> ---- >>>>> -- orphan >>>>> -... >>>>> -box.info.ro >>>>> ---- >>>>> -- true >>>>> -... >>>>> --- lower quorum => leave orphan mode >>>>> -box.cfg{replication_connect_quorum=1} >>>>> ---- >>>>> -... >>>>> -box.info.status >>>>> ---- >>>>> -- running >>>>> -... >>>>> -box.info.ro >>>>> ---- >>>>> -- false >>>>> -... >>>>> --- >>>>> --- gh-3760: replication quorum 0 on reconfiguration should return >>>>> --- from box.cfg immediately. >>>>> --- >>>>> -replication = box.cfg.replication >>>>> ---- >>>>> -... >>>>> -box.cfg{ \ >>>>> - replication = {}, \ >>>>> - replication_connect_quorum = 0, \ >>>>> - replication_connect_timeout = 1000000 \ >>>>> -} >>>>> ---- >>>>> -... >>>>> --- The call below would hang, if quorum 0 is ignored, or checked >>>>> --- too late. >>>>> -box.cfg{replication = {'localhost:12345'}} >>>>> ---- >>>>> -... >>>>> -box.info.status >>>>> ---- >>>>> -- running >>>>> -... >>>>> -box.cfg{ \ >>>>> - replication = {}, \ >>>>> - replication_connect_quorum = replication_connect_quorum, \ >>>>> - replication_connect_timeout = replication_connect_timeout \ >>>>> -} >>>>> ---- >>>>> -... >>>>> diff --git a/test/replication/misc.test.lua >>>>> b/test/replication/misc.test.lua >>>>> deleted file mode 100644 >>>>> index d285b014a..000000000 >>>>> --- a/test/replication/misc.test.lua >>>>> +++ /dev/null >>>>> @@ -1,356 +0,0 @@ >>>>> -uuid = require('uuid') >>>>> -test_run = require('test_run').new() >>>>> - >>>>> -box.schema.user.grant('guest', 'replication') >>>>> - >>>>> --- gh-2991 - Tarantool asserts on box.cfg.replication update if >>>>> one of >>>>> --- servers is dead >>>>> -replication_timeout = box.cfg.replication_timeout >>>>> -replication_connect_timeout = box.cfg.replication_connect_timeout >>>>> -box.cfg{replication_timeout=0.05, >>>>> replication_connect_timeout=0.05, replication={}} >>>>> -box.cfg{replication_connect_quorum=2} >>>>> -box.cfg{replication = {'127.0.0.1:12345', box.cfg.listen}} >>>>> -box.info.status >>>>> -box.info.ro >>>>> - >>>>> --- gh-3606 - Tarantool crashes if box.cfg.replication is updated >>>>> concurrently >>>>> -fiber = require('fiber') >>>>> -c = fiber.channel(2) >>>>> -f = function() fiber.create(function() pcall(box.cfg, >>>>> {replication = {12345}}) c:put(true) end) end >>>>> -f() >>>>> -f() >>>>> -c:get() >>>>> -c:get() >>>>> - >>>>> -box.cfg{replication = "", replication_timeout = >>>>> replication_timeout, replication_connect_timeout = >>>>> replication_connect_timeout} >>>>> -box.info.status >>>>> -box.info.ro >>>>> - >>>>> --- gh-3111 - Allow to rebootstrap a replica from a read-only master >>>>> -replica_uuid = uuid.new() >>>>> -test_run:cmd('create server test with rpl_master=default, >>>>> script="replication/replica_uuid.lua"') >>>>> -test_run:cmd(string.format('start server test with args="%s"', >>>>> replica_uuid)) >>>>> -test_run:cmd('stop server test') >>>>> -test_run:cmd('cleanup server test') >>>>> -box.cfg{read_only = true} >>>>> -test_run:cmd(string.format('start server test with args="%s"', >>>>> replica_uuid)) >>>>> -test_run:cmd('stop server test') >>>>> -test_run:cmd('cleanup server test') >>>>> -box.cfg{read_only = false} >>>>> -test_run:cmd('delete server test') >>>>> -test_run:cleanup_cluster() >>>>> - >>>>> --- gh-3160 - Send heartbeats if there are changes from a remote >>>>> master only >>>>> -SERVERS = { 'autobootstrap1', 'autobootstrap2', 'autobootstrap3' } >>>>> - >>>>> --- Deploy a cluster. >>>>> -test_run:create_cluster(SERVERS, "replication", {args="0.03"}) >>>>> -test_run:wait_fullmesh(SERVERS) >>>>> -test_run:cmd("switch autobootstrap3") >>>>> -test_run = require('test_run').new() >>>>> -fiber = require('fiber') >>>>> -_ = box.schema.space.create('test_timeout'):create_index('pk') >>>>> -test_run:cmd("setopt delimiter ';'") >>>>> -function wait_not_follow(replicaA, replicaB) >>>>> - return test_run:wait_cond(function() >>>>> - return replicaA.status ~= 'follow' or replicaB.status ~= >>>>> 'follow' >>>>> - end, box.cfg.replication_timeout) >>>>> -end; >>>>> -function test_timeout() >>>>> - local replicaA = box.info.replication[1].upstream or >>>>> box.info.replication[2].upstream >>>>> - local replicaB = box.info.replication[3].upstream or >>>>> box.info.replication[2].upstream >>>>> - local follows = test_run:wait_cond(function() >>>>> - return replicaA.status == 'follow' or replicaB.status == >>>>> 'follow' >>>>> - end) >>>>> - if not follows then error('replicas are not in the follow >>>>> status') end >>>>> - for i = 0, 99 do >>>>> - box.space.test_timeout:replace({1}) >>>>> - if wait_not_follow(replicaA, replicaB) then >>>>> - return error(box.info.replication) >>>>> - end >>>>> - end >>>>> - return true >>>>> -end; >>>>> -test_run:cmd("setopt delimiter ''"); >>>>> -test_timeout() >>>>> - >>>>> --- gh-3247 - Sequence-generated value is not replicated in case >>>>> --- the request was sent via iproto. >>>>> -test_run:cmd("switch autobootstrap1") >>>>> -net_box = require('net.box') >>>>> -_ = box.schema.space.create('space1') >>>>> -_ = box.schema.sequence.create('seq') >>>>> -_ = box.space.space1:create_index('primary', {sequence = true} ) >>>>> -_ = box.space.space1:create_index('secondary', {parts = {2, >>>>> 'unsigned'}}) >>>>> -box.schema.user.grant('guest', 'read,write', 'space', 'space1') >>>>> -c = net_box.connect(box.cfg.listen) >>>>> -c.space.space1:insert{box.NULL, "data"} -- fails, but bumps >>>>> sequence value >>>>> -c.space.space1:insert{box.NULL, 1, "data"} >>>>> -box.space.space1:select{} >>>>> -vclock = test_run:get_vclock("autobootstrap1") >>>>> -vclock[0] = nil >>>>> -_ = test_run:wait_vclock("autobootstrap2", vclock) >>>>> -test_run:cmd("switch autobootstrap2") >>>>> -box.space.space1:select{} >>>>> -test_run:cmd("switch autobootstrap1") >>>>> -box.space.space1:drop() >>>>> - >>>>> -test_run:cmd("switch default") >>>>> -test_run:drop_cluster(SERVERS) >>>>> -test_run:cleanup_cluster() >>>>> - >>>>> --- gh-3642 - Check that socket file descriptor doesn't leak >>>>> --- when a replica is disconnected. >>>>> -rlimit = require('rlimit') >>>>> -lim = rlimit.limit() >>>>> -rlimit.getrlimit(rlimit.RLIMIT_NOFILE, lim) >>>>> -old_fno = lim.rlim_cur >>>>> -lim.rlim_cur = 64 >>>>> -rlimit.setrlimit(rlimit.RLIMIT_NOFILE, lim) >>>>> - >>>>> -test_run:cmd('create server sock with rpl_master=default, >>>>> script="replication/replica.lua"') >>>>> -test_run:cmd('start server sock') >>>>> -test_run:cmd('switch sock') >>>>> -test_run = require('test_run').new() >>>>> -fiber = require('fiber') >>>>> -test_run:cmd("setopt delimiter ';'") >>>>> -for i = 1, 64 do >>>>> - local replication = box.cfg.replication >>>>> - box.cfg{replication = {}} >>>>> - box.cfg{replication = replication} >>>>> - while box.info.replication[1].upstream.status ~= 'follow' do >>>>> - fiber.sleep(0.001) >>>>> - end >>>>> -end; >>>>> -test_run:cmd("setopt delimiter ''"); >>>>> - >>>>> -box.info.replication[1].upstream.status >>>>> - >>>>> -test_run:cmd('switch default') >>>>> - >>>>> -lim.rlim_cur = old_fno >>>>> -rlimit.setrlimit(rlimit.RLIMIT_NOFILE, lim) >>>>> - >>>>> -test_run:cmd("stop server sock") >>>>> -test_run:cmd("cleanup server sock") >>>>> -test_run:cmd("delete server sock") >>>>> -test_run:cleanup_cluster() >>>>> - >>>>> -box.schema.user.revoke('guest', 'replication') >>>>> - >>>>> --- gh-3510 assertion failure in replica_on_applier_disconnect() >>>>> -test_run:cmd('create server er_load1 with >>>>> script="replication/er_load1.lua"') >>>>> -test_run:cmd('create server er_load2 with >>>>> script="replication/er_load2.lua"') >>>>> -test_run:cmd('start server er_load1 with wait=False, >>>>> wait_load=False') >>>>> --- Instance er_load2 will fail with error >>>>> ER_REPLICASET_UUID_MISMATCH. >>>>> --- This is OK since we only test here that er_load1 doesn't assert. >>>>> -test_run:cmd('start server er_load2 with wait=True, >>>>> wait_load=True, crash_expected = True') >>>>> -test_run:cmd('stop server er_load1') >>>>> --- er_load2 exits automatically. >>>>> -test_run:cmd('cleanup server er_load1') >>>>> -test_run:cmd('cleanup server er_load2') >>>>> -test_run:cmd('delete server er_load1') >>>>> -test_run:cmd('delete server er_load2') >>>>> -test_run:cleanup_cluster() >>>>> - >>>>> --- >>>>> --- Test case for gh-3637, gh-4550. Before the fix replica would >>>>> --- exit with an error if a user does not exist or a password is >>>>> --- incorrect. Now check that we don't hang/panic and successfully >>>>> --- connect. >>>>> --- >>>>> -fiber = require('fiber') >>>>> -test_run:cmd("create server replica_auth with rpl_master=default, >>>>> script='replication/replica_auth.lua'") >>>>> -test_run:cmd("start server replica_auth with wait=False, >>>>> wait_load=False, args='cluster:pass 0.05'") >>>>> --- Wait a bit to make sure replica waits till user is created. >>>>> -fiber.sleep(0.1) >>>>> -box.schema.user.create('cluster') >>>>> --- The user is created. Let the replica fail auth request due to >>>>> --- a wrong password. >>>>> -fiber.sleep(0.1) >>>>> -box.schema.user.passwd('cluster', 'pass') >>>>> -box.schema.user.grant('cluster', 'replication') >>>>> - >>>>> -while box.info.replication[2] == nil do fiber.sleep(0.01) end >>>>> -vclock = test_run:get_vclock('default') >>>>> -vclock[0] = nil >>>>> -_ = test_run:wait_vclock('replica_auth', vclock) >>>>> - >>>>> -test_run:cmd("stop server replica_auth") >>>>> -test_run:cmd("cleanup server replica_auth") >>>>> -test_run:cmd("delete server replica_auth") >>>>> -test_run:cleanup_cluster() >>>>> - >>>>> -box.schema.user.drop('cluster') >>>>> - >>>>> --- >>>>> --- Test case for gh-3610. Before the fix replica would fail with >>>>> the assertion >>>>> --- when trying to connect to the same master twice. >>>>> --- >>>>> -box.schema.user.grant('guest', 'replication') >>>>> -test_run:cmd("create server replica with rpl_master=default, >>>>> script='replication/replica.lua'") >>>>> -test_run:cmd("start server replica") >>>>> -test_run:cmd("switch replica") >>>>> -replication = box.cfg.replication[1] >>>>> -box.cfg{replication = {replication, replication}} >>>>> - >>>>> --- Check the case when duplicate connection is detected in the >>>>> background. >>>>> -test_run:cmd("switch default") >>>>> -listen = box.cfg.listen >>>>> -box.cfg{listen = ''} >>>>> - >>>>> -test_run:cmd("switch replica") >>>>> -box.cfg{replication_connect_quorum = 0, >>>>> replication_connect_timeout = 0.01} >>>>> -box.cfg{replication = {replication, replication}} >>>>> - >>>>> -test_run:cmd("switch default") >>>>> -box.cfg{listen = listen} >>>>> -while test_run:grep_log('replica', 'duplicate connection') == nil >>>>> do fiber.sleep(0.01) end >>>>> - >>>>> -test_run:cmd("stop server replica") >>>>> -test_run:cmd("cleanup server replica") >>>>> -test_run:cmd("delete server replica") >>>>> -test_run:cleanup_cluster() >>>>> -box.schema.user.revoke('guest', 'replication') >>>>> - >>>>> --- >>>>> --- gh-3711 Do not restart replication on box.cfg in case the >>>>> --- configuration didn't change. >>>>> --- >>>>> -box.schema.user.grant('guest', 'replication') >>>>> -test_run:cmd("create server replica with rpl_master=default, >>>>> script='replication/replica.lua'") >>>>> -test_run:cmd("start server replica") >>>>> - >>>>> --- Access rights are checked only during reconnect. If the new >>>>> --- and old configurations are equivalent, no reconnect will be >>>>> --- issued and replication should continue working. >>>>> -box.schema.user.revoke('guest', 'replication') >>>>> -test_run:cmd("switch replica") >>>>> -replication = box.cfg.replication[1] >>>>> -box.cfg{replication = {replication}} >>>>> -box.info.status == 'running' >>>>> -box.cfg{replication = replication} >>>>> -box.info.status == 'running' >>>>> - >>>>> --- Check that comparison of tables works as expected as well. >>>>> -test_run:cmd("switch default") >>>>> -box.schema.user.grant('guest', 'replication') >>>>> -test_run:cmd("switch replica") >>>>> -replication = box.cfg.replication >>>>> -table.insert(replication, box.cfg.listen) >>>>> -test_run:cmd("switch default") >>>>> -box.schema.user.revoke('guest', 'replication') >>>>> -test_run:cmd("switch replica") >>>>> -box.cfg{replication = replication} >>>>> -box.info.status == 'running' >>>>> - >>>>> -test_run:cmd("switch default") >>>>> -test_run:cmd("stop server replica") >>>>> -test_run:cmd("cleanup server replica") >>>>> -test_run:cmd("delete server replica") >>>>> -test_run:cleanup_cluster() >>>>> - >>>>> --- >>>>> --- gh-3704 move cluster id check to replica >>>>> --- >>>>> -test_run:cmd("create server replica with rpl_master=default, >>>>> script='replication/replica.lua'") >>>>> -box.schema.user.grant("guest", "replication") >>>>> -test_run:cmd("start server replica") >>>>> -test_run:grep_log("replica", "REPLICASET_UUID_MISMATCH") >>>>> -box.info.replication[2].downstream.status >>>>> --- change master's cluster uuid and check that replica doesn't >>>>> connect. >>>>> -test_run:cmd("stop server replica") >>>>> -_ = box.space._schema:replace{'cluster', tostring(uuid.new())} >>>>> --- master believes replica is in cluster, but their cluster UUIDs >>>>> differ. >>>>> -test_run:cmd("start server replica") >>>>> -test_run:wait_log("replica", "REPLICASET_UUID_MISMATCH", nil, 1.0) >>>>> -test_run:wait_downstream(2, {status = 'stopped'}) >>>>> - >>>>> -test_run:cmd("stop server replica") >>>>> -test_run:cmd("cleanup server replica") >>>>> -test_run:cmd("delete server replica") >>>>> -test_run:cleanup_cluster() >>>>> -box.schema.user.revoke('guest', 'replication') >>>>> - >>>>> --- >>>>> --- gh-4399 Check that an error reading WAL directory on subscribe >>>>> --- doesn't lead to a permanent replication failure. >>>>> --- >>>>> -box.schema.user.grant("guest", "replication") >>>>> -test_run:cmd("create server replica with rpl_master=default, >>>>> script='replication/replica.lua'") >>>>> -test_run:cmd("start server replica") >>>>> - >>>>> --- Make the WAL directory inaccessible. >>>>> -fio = require('fio') >>>>> -path = fio.abspath(box.cfg.wal_dir) >>>>> -fio.chmod(path, 0) >>>>> - >>>>> --- Break replication on timeout. >>>>> -replication_timeout = box.cfg.replication_timeout >>>>> -box.cfg{replication_timeout = 9000} >>>>> -test_run:cmd("switch replica") >>>>> -test_run:wait_cond(function() return >>>>> box.info.replication[1].upstream.status ~= 'follow' end) >>>>> -require('fiber').sleep(box.cfg.replication_timeout) >>>>> -test_run:cmd("switch default") >>>>> -box.cfg{replication_timeout = replication_timeout} >>>>> - >>>>> --- Restore access to the WAL directory. >>>>> --- Wait for replication to be reestablished. >>>>> -fio.chmod(path, tonumber('777', 8)) >>>>> -test_run:cmd("switch replica") >>>>> -test_run:wait_cond(function() return >>>>> box.info.replication[1].upstream.status == 'follow' end) >>>>> -test_run:cmd("switch default") >>>>> - >>>>> -test_run:cmd("stop server replica") >>>>> -test_run:cmd("cleanup server replica") >>>>> -test_run:cmd("delete server replica") >>>>> -test_run:cleanup_cluster() >>>>> -box.schema.user.revoke('guest', 'replication') >>>>> - >>>>> --- >>>>> --- gh-4424 Always enter orphan mode on error in replication >>>>> --- configuration change. >>>>> --- >>>>> -replication_connect_timeout = box.cfg.replication_connect_timeout >>>>> -replication_connect_quorum = box.cfg.replication_connect_quorum >>>>> -box.cfg{replication="12345", replication_connect_timeout=0.1, >>>>> replication_connect_quorum=1} >>>>> -box.info.status >>>>> -box.info.ro >>>>> --- reset replication => leave orphan mode >>>>> -box.cfg{replication=""} >>>>> -box.info.status >>>>> -box.info.ro >>>>> --- no switch to orphan when quorum == 0 >>>>> -box.cfg{replication="12345", replication_connect_quorum=0} >>>>> -box.info.status >>>>> -box.info.ro >>>>> - >>>>> --- we could connect to one out of two replicas. Set orphan. >>>>> -box.cfg{replication_connect_quorum=2} >>>>> -box.cfg{replication={box.cfg.listen, "12345"}} >>>>> -box.info.status >>>>> -box.info.ro >>>>> --- lower quorum => leave orphan mode >>>>> -box.cfg{replication_connect_quorum=1} >>>>> -box.info.status >>>>> -box.info.ro >>>>> - >>>>> --- >>>>> --- gh-3760: replication quorum 0 on reconfiguration should return >>>>> --- from box.cfg immediately. >>>>> --- >>>>> -replication = box.cfg.replication >>>>> -box.cfg{ \ >>>>> - replication = {}, \ >>>>> - replication_connect_quorum = 0, \ >>>>> - replication_connect_timeout = 1000000 \ >>>>> -} >>>>> --- The call below would hang, if quorum 0 is ignored, or checked >>>>> --- too late. >>>>> -box.cfg{replication = {'localhost:12345'}} >>>>> -box.info.status >>>>> -box.cfg{ \ >>>>> - replication = {}, \ >>>>> - replication_connect_quorum = replication_connect_quorum, \ >>>>> - replication_connect_timeout = replication_connect_timeout \ >>>>> -} >>>>> diff --git >>>>> a/test/replication/misc_assert_connecting_master_twice_gh-3610.result >>>>> b/test/replication/misc_assert_connecting_master_twice_gh-3610.result >>>>> new file mode 100644 >>>>> index 000000000..d7b7cc25b >>>>> --- /dev/null >>>>> +++ >>>>> b/test/replication/misc_assert_connecting_master_twice_gh-3610.result >>>>> @@ -0,0 +1,86 @@ >>>>> +test_run = require('test_run').new() >>>>> +--- >>>>> +... >>>>> +test_run:cmd("restart server default") >>>> I don't think you should restart the server here. >>>> The test seems fine without it. Same about almost all the >>>> testcases below. >>>>> +uuid = require('uuid') >>>>> +--- >>>>> +... >>>>> +fiber = require('fiber') >>>>> +--- >>>>> +... >>>> UUID and fiber modules aren't used in this test, so please remove >>>> these two extra lines. Same about almost all the testcases below. >>>> >>>> >>>> The branch contains this extraneous change: >>>> >>>> +replication_connect_quorum = box.cfg.replication_connect_quorum >>>> >>>> +--- >>>> +... >>>> +replication_connect_timeout = box.cfg.replication_connect_timeout >>>> +--- >>>> >>>> +... >>>> >>>> You don't need to save these options, since you only change them on a >>>> replica. >>>> >>>>> +-- >>>>> +-- Test case for gh-3610. Before the fix replica would fail with >>>>> the assertion >>>>> +-- when trying to connect to the same master twice. >>>>> +-- >>>>> +box.schema.user.grant('guest', 'replication') >>>>> +--- >>>>> +... >>>>> +test_run:cmd("create server replica with rpl_master=default, >>>>> script='replication/replica.lua'") >>>>> +--- >>>>> +- true >>>>> +... >>>>> +test_run:cmd("start server replica") >>>>> +--- >>>>> +- true >>>>> +... >>>>> +test_run:cmd("switch replica") >>>>> +--- >>>>> +- true >>>>> +... >>>>> +replication = box.cfg.replication[1] >>>>> +--- >>>>> +... >>>>> +box.cfg{replication = {replication, replication}} >>>>> +--- >>>>> +- error: 'Incorrect value for option ''replication'': duplicate >>>>> connection to the >>>>> + same replica' >>>>> +... >>>>> +-- Check the case when duplicate connection is detected in the >>>>> background. >>>>> +test_run:cmd("switch default") >>>>> +--- >>>>> +- true >>>>> +... >>>>> +listen = box.cfg.listen >>>>> +--- >>>>> +... >>>>> +box.cfg{listen = ''} >>>>> +--- >>>>> +... >>>>> +test_run:cmd("switch replica") >>>>> +--- >>>>> +- true >>>>> +... >>>>> +box.cfg{replication_connect_quorum = 0, >>>>> replication_connect_timeout = 0.01} >>>>> +--- >>>>> +... >>>>> +box.cfg{replication = {replication, replication}} >>>>> +--- >>>>> +... >>>>> +test_run:cmd("switch default") >>>>> +--- >>>>> +- true >>>>> +... >>>>> +box.cfg{listen = listen} >>>>> +--- >>>>> +... >>>>> +while test_run:grep_log('replica', 'duplicate connection') == nil >>>>> do fiber.sleep(0.01) end >>>>> +--- >>>>> +... >>>>> +test_run:cmd("stop server replica") >>>>> +--- >>>>> +- true >>>>> +... >>>>> +test_run:cmd("cleanup server replica") >>>>> +--- >>>>> +- true >>>>> +... >>>>> +test_run:cmd("delete server replica") >>>>> +--- >>>>> +- true >>>>> +... >>>>> +test_run:cleanup_cluster() >>>>> +--- >>>>> +... >>>>> +box.schema.user.revoke('guest', 'replication') >>>>> +--- >>>>> +... >>>> +box.cfg{replication = "", \ >>>> + replication_connect_quorum = replication_connect_quorum, \ >>>> + replication_connect_timeout = replication_connect_timeout} >>>> +--- >>>> +... >>>> >>>> The change I spoke of above. This piece is also unneeded. >>>> >>>>> diff --git >>>>> a/test/replication/misc_assert_connecting_master_twice_gh-3610.test.lua >>>>> b/test/replication/misc_assert_connecting_master_twice_gh-3610.test.lua >>>>> >>>>> new file mode 100644 >>>>> index 000000000..9b12e623b >>>>> --- /dev/null >>>>> +++ >>>>> b/test/replication/misc_assert_connecting_master_twice_gh-3610.test.lua >>>>> >>>>> @@ -0,0 +1,34 @@ >>>>> +test_run = require('test_run').new() >>>>> +test_run:cmd("restart server default") >>>>> +uuid = require('uuid') >>>>> +fiber = require('fiber') >>>>> + >>>>> +-- >>>>> +-- Test case for gh-3610. Before the fix replica would fail with >>>>> the assertion >>>>> +-- when trying to connect to the same master twice. >>>>> +-- >>>>> +box.schema.user.grant('guest', 'replication') >>>>> +test_run:cmd("create server replica with rpl_master=default, >>>>> script='replication/replica.lua'") >>>>> +test_run:cmd("start server replica") >>>>> +test_run:cmd("switch replica") >>>>> +replication = box.cfg.replication[1] >>>>> +box.cfg{replication = {replication, replication}} >>>>> + >>>>> +-- Check the case when duplicate connection is detected in the >>>>> background. >>>>> +test_run:cmd("switch default") >>>>> +listen = box.cfg.listen >>>>> +box.cfg{listen = ''} >>>>> + >>>>> +test_run:cmd("switch replica") >>>>> +box.cfg{replication_connect_quorum = 0, >>>>> replication_connect_timeout = 0.01} >>>>> +box.cfg{replication = {replication, replication}} >>>>> + >>>>> +test_run:cmd("switch default") >>>>> +box.cfg{listen = listen} >>>>> +while test_run:grep_log('replica', 'duplicate connection') == nil >>>>> do fiber.sleep(0.01) end >>>>> + >>>>> +test_run:cmd("stop server replica") >>>>> +test_run:cmd("cleanup server replica") >>>>> +test_run:cmd("delete server replica") >>>>> +test_run:cleanup_cluster() >>>>> +box.schema.user.revoke('guest', 'replication') >>>>> diff --git >>>>> a/test/replication/misc_assert_on_server_die_gh-2991.result >>>>> b/test/replication/misc_assert_on_server_die_gh-2991.result >>>>> new file mode 100644 >>>>> index 000000000..ea9e80f6b >>>>> --- /dev/null >>>>> +++ b/test/replication/misc_assert_on_server_die_gh-2991.result >>>>> @@ -0,0 +1,31 @@ >>>>> +uuid = require('uuid') >>>>> +--- >>>>> +... >>>> UUID module isn't used in this testcase, please remove the extra line. >>>>> +test_run = require('test_run').new() >>>>> +--- >>>>> +... >>>>> +-- gh-2991 - Tarantool asserts on box.cfg.replication update if >>>>> one of >>>>> +-- servers is dead >>>>> +replication_timeout = box.cfg.replication_timeout >>>>> +--- >>>>> +... >>>>> +replication_connect_timeout = box.cfg.replication_connect_timeout >>>>> +--- >>>>> +... >>>>> +box.cfg{replication_timeout=0.05, >>>>> replication_connect_timeout=0.05, replication={}} >>>>> +--- >>>>> +... >>>>> +box.cfg{replication_connect_quorum=2} >>>>> +--- >>>>> +... >>>>> +box.cfg{replication = {'127.0.0.1:12345', box.cfg.listen}} >>>>> +--- >>>>> +... >>>>> +box.info.status >>>>> +--- >>>>> +- orphan >>>>> +... >>>>> +box.info.ro >>>>> +--- >>>>> +- true >>>>> +... >>>>> diff --git >>>>> a/test/replication/misc_assert_on_server_die_gh-2991.test.lua >>>>> b/test/replication/misc_assert_on_server_die_gh-2991.test.lua >>>>> new file mode 100644 >>>>> index 000000000..adda839f7 >>>>> --- /dev/null >>>>> +++ b/test/replication/misc_assert_on_server_die_gh-2991.test.lua >>>>> @@ -0,0 +1,12 @@ >>>>> +uuid = require('uuid') >>>>> +test_run = require('test_run').new() >>>>> + >>>>> +-- gh-2991 - Tarantool asserts on box.cfg.replication update if >>>>> one of >>>>> +-- servers is dead >>>>> +replication_timeout = box.cfg.replication_timeout >>>>> +replication_connect_timeout = box.cfg.replication_connect_timeout >>>>> +box.cfg{replication_timeout=0.05, >>>>> replication_connect_timeout=0.05, replication={}} >>>>> +box.cfg{replication_connect_quorum=2} >>>>> +box.cfg{replication = {'127.0.0.1:12345', box.cfg.listen}} >>>>> +box.info.status >>>>> +box.info.ro >>>>> diff --git >>>>> a/test/replication/misc_assert_replica_on_applier_disconnect_gh-3510.result >>>>> b/test/replication/misc_assert_replica_on_applier_disconnect_gh-3510.result >>>>> >>>>> new file mode 100644 >>>>> index 000000000..82f5d9a23 >>>>> --- /dev/null >>>>> +++ >>>>> b/test/replication/misc_assert_replica_on_applier_disconnect_gh-3510.result >>>>> @@ -0,0 +1,49 @@ >>>>> +uuid = require('uuid') >>>>> +--- >>>>> +... >>>> Same about the uuid. >>>>> +test_run = require('test_run').new() >>>>> +--- >>>>> +... >>>>> +-- gh-3510 assertion failure in replica_on_applier_disconnect() >>>>> +test_run:cmd('create server er_load1 with >>>>> script="replication/er_load1.lua"') >>>>> +--- >>>>> +- true >>>>> +... >>>>> +test_run:cmd('create server er_load2 with >>>>> script="replication/er_load2.lua"') >>>>> +--- >>>>> +- true >>>>> +... >>>>> +test_run:cmd('start server er_load1 with wait=False, >>>>> wait_load=False') >>>>> +--- >>>>> +- true >>>>> +... >>>>> +-- Instance er_load2 will fail with error >>>>> ER_REPLICASET_UUID_MISMATCH. >>>>> +-- This is OK since we only test here that er_load1 doesn't assert. >>>>> +test_run:cmd('start server er_load2 with wait=True, >>>>> wait_load=True, crash_expected = True') >>>>> +--- >>>>> +- false >>>>> +... >>>>> +test_run:cmd('stop server er_load1') >>>>> +--- >>>>> +- true >>>>> +... >>>>> +-- er_load2 exits automatically. >>>>> +test_run:cmd('cleanup server er_load1') >>>>> +--- >>>>> +- true >>>>> +... >>>>> +test_run:cmd('cleanup server er_load2') >>>>> +--- >>>>> +- true >>>>> +... >>>>> +test_run:cmd('delete server er_load1') >>>>> +--- >>>>> +- true >>>>> +... >>>>> +test_run:cmd('delete server er_load2') >>>>> +--- >>>>> +- true >>>>> +... >>>>> +test_run:cleanup_cluster() >>>>> +--- >>>>> +... >>>>> diff --git >>>>> a/test/replication/misc_assert_replica_on_applier_disconnect_gh-3510.test.lua >>>>> b/test/replication/misc_assert_replica_on_applier_disconnect_gh-3510.test.lua >>>>> >>>>> new file mode 100644 >>>>> index 000000000..4e1d2a41e >>>>> --- /dev/null >>>>> +++ >>>>> b/test/replication/misc_assert_replica_on_applier_disconnect_gh-3510.test.lua >>>>> @@ -0,0 +1,17 @@ >>>>> +uuid = require('uuid') >>>>> +test_run = require('test_run').new() >>>>> + >>>>> +-- gh-3510 assertion failure in replica_on_applier_disconnect() >>>>> +test_run:cmd('create server er_load1 with >>>>> script="replication/er_load1.lua"') >>>>> +test_run:cmd('create server er_load2 with >>>>> script="replication/er_load2.lua"') >>>>> +test_run:cmd('start server er_load1 with wait=False, >>>>> wait_load=False') >>>>> +-- Instance er_load2 will fail with error >>>>> ER_REPLICASET_UUID_MISMATCH. >>>>> +-- This is OK since we only test here that er_load1 doesn't assert. >>>>> +test_run:cmd('start server er_load2 with wait=True, >>>>> wait_load=True, crash_expected = True') >>>>> +test_run:cmd('stop server er_load1') >>>>> +-- er_load2 exits automatically. >>>>> +test_run:cmd('cleanup server er_load1') >>>>> +test_run:cmd('cleanup server er_load2') >>>>> +test_run:cmd('delete server er_load1') >>>>> +test_run:cmd('delete server er_load2') >>>>> +test_run:cleanup_cluster() >>>>> diff --git >>>>> a/test/replication/misc_crash_on_box_concurrent_update_gh-3606.result >>>>> b/test/replication/misc_crash_on_box_concurrent_update_gh-3606.result >>>>> new file mode 100644 >>>>> index 000000000..b43b00849 >>>>> --- /dev/null >>>>> +++ >>>>> b/test/replication/misc_crash_on_box_concurrent_update_gh-3606.result >>>>> @@ -0,0 +1,50 @@ >>>>> +uuid = require('uuid') >>>>> +--- >>>>> +... >>>> Same about UUID. >>>>> +test_run = require('test_run').new() >>>>> +--- >>>>> +... >>>>> +replication_timeout = box.cfg.replication_timeout >>>>> +--- >>>>> +... >>>>> +replication_connect_timeout = box.cfg.replication_connect_timeout >>>>> +--- >>>>> +... >>>>> +box.cfg{replication_timeout=0.05, >>>>> replication_connect_timeout=0.05, replication={}} >>>>> +--- >>>>> +... >>>>> +-- gh-3606 - Tarantool crashes if box.cfg.replication is updated >>>>> concurrently >>>>> +fiber = require('fiber') >>>>> +--- >>>>> +... >>>>> +c = fiber.channel(2) >>>>> +--- >>>>> +... >>>>> +f = function() fiber.create(function() pcall(box.cfg, >>>>> {replication = {12345}}) c:put(true) end) end >>>>> +--- >>>>> +... >>>>> +f() >>>>> +--- >>>>> +... >>>>> +f() >>>>> +--- >>>>> +... >>>>> +c:get() >>>>> +--- >>>>> +- true >>>>> +... >>>>> +c:get() >>>>> +--- >>>>> +- true >>>>> +... >>>>> +box.cfg{replication = "", replication_timeout = >>>>> replication_timeout, replication_connect_timeout = >>>>> replication_connect_timeout} >>>>> +--- >>>>> +... >>>>> +box.info.status >>>>> +--- >>>>> +- running >>>>> +... >>>>> +box.info.ro >>>>> +--- >>>>> +- false >>>>> +... >>>>> diff --git >>>>> a/test/replication/misc_crash_on_box_concurrent_update_gh-3606.test.lua >>>>> b/test/replication/misc_crash_on_box_concurrent_update_gh-3606.test.lua >>>>> >>>>> new file mode 100644 >>>>> index 000000000..17f9c6bc6 >>>>> --- /dev/null >>>>> +++ >>>>> b/test/replication/misc_crash_on_box_concurrent_update_gh-3606.test.lua >>>>> >>>>> @@ -0,0 +1,19 @@ >>>>> +uuid = require('uuid') >>>>> +test_run = require('test_run').new() >>>>> + >>>>> +replication_timeout = box.cfg.replication_timeout >>>>> +replication_connect_timeout = box.cfg.replication_connect_timeout >>>>> +box.cfg{replication_timeout=0.05, >>>>> replication_connect_timeout=0.05, replication={}} >>>>> + >>>>> +-- gh-3606 - Tarantool crashes if box.cfg.replication is updated >>>>> concurrently >>>>> +fiber = require('fiber') >>>>> +c = fiber.channel(2) >>>>> +f = function() fiber.create(function() pcall(box.cfg, >>>>> {replication = {12345}}) c:put(true) end) end >>>>> +f() >>>>> +f() >>>>> +c:get() >>>>> +c:get() >>>>> + >>>>> +box.cfg{replication = "", replication_timeout = >>>>> replication_timeout, replication_connect_timeout = >>>>> replication_connect_timeout} >>>>> +box.info.status >>>>> +box.info.ro >>>>> diff --git >>>>> a/test/replication/misc_heartbeats_on_master_changes_gh-3160.result >>>>> b/test/replication/misc_heartbeats_on_master_changes_gh-3160.result >>>>> new file mode 100644 >>>>> index 000000000..cdb463614 >>>>> --- /dev/null >>>>> +++ >>>>> b/test/replication/misc_heartbeats_on_master_changes_gh-3160.result >>>>> @@ -0,0 +1,76 @@ >>>>> +uuid = require('uuid') >>>>> +--- >>>>> +... >>>> Same about UUID. >>>> >>>>> +test_run = require('test_run').new() >>>>> +--- >>>>> +... >>>> The branch contains this change: >>>> >>>> +... >>>> +replication_timeout = box.cfg.replication_timeout >>>> +--- >>>> +... >>>> +replication_connect_timeout = box.cfg.replication_connect_timeout >>>> +--- >>>> +... >>>> +box.cfg{replication_timeout=0.05, replication_connect_timeout=0.05, >>>> replication={}} >>>> +--- >>>> +... >>>> >>>> Which is extraneous in this test. The original testcase uses >>>> default timeout >>>> values, >>>> and I don't think we should change them. >>>> >>>>> +-- gh-3160 - Send heartbeats if there are changes from a remote >>>>> master only >>>>> +SERVERS = { 'autobootstrap1', 'autobootstrap2', 'autobootstrap3' } >>>>> +--- >>>>> +... >>>>> +-- Deploy a cluster. >>>>> +test_run:create_cluster(SERVERS, "replication", {args="0.03"}) >>>>> +--- >>>>> +... >>>>> +test_run:wait_fullmesh(SERVERS) >>>>> +--- >>>>> +... >>>>> +test_run:cmd("switch autobootstrap3") >>>>> +--- >>>>> +- true >>>>> +... >>>>> +test_run = require('test_run').new() >>>>> +--- >>>>> +... >>>>> +fiber = require('fiber') >>>>> +--- >>>>> +... >>>>> +_ = box.schema.space.create('test_timeout'):create_index('pk') >>>>> +--- >>>>> +... >>>>> +test_run:cmd("setopt delimiter ';'") >>>>> +--- >>>>> +- true >>>>> +... >>>>> +function wait_not_follow(replicaA, replicaB) >>>>> + return test_run:wait_cond(function() >>>>> + return replicaA.status ~= 'follow' or replicaB.status ~= >>>>> 'follow' >>>>> + end, box.cfg.replication_timeout) >>>>> +end; >>>>> +--- >>>>> +... >>>>> +function test_timeout() >>>>> + local replicaA = box.info.replication[1].upstream or >>>>> box.info.replication[2].upstream >>>>> + local replicaB = box.info.replication[3].upstream or >>>>> box.info.replication[2].upstream >>>>> + local follows = test_run:wait_cond(function() >>>>> + return replicaA.status == 'follow' or replicaB.status == >>>>> 'follow' >>>>> + end) >>>>> + if not follows then error('replicas are not in the follow >>>>> status') end >>>>> + for i = 0, 99 do >>>>> + box.space.test_timeout:replace({1}) >>>>> + if wait_not_follow(replicaA, replicaB) then >>>>> + return error(box.info.replication) >>>>> + end >>>>> + end >>>>> + return true >>>>> +end; >>>>> +--- >>>>> +... >>>>> +test_run:cmd("setopt delimiter ''"); >>>>> +--- >>>>> +- true >>>>> +... >>>>> +test_timeout() >>>>> +--- >>>>> +- true >>>>> +... >>>>> +test_run:cmd("switch default") >>>>> +--- >>>>> +- true >>>>> +... >>>>> +test_run:drop_cluster(SERVERS) >>>>> +--- >>>>> +... >>>>> +test_run:cleanup_cluster() >>>> No need to call cleanup_cluster(). The 'default' instance wasn't >>>> part of the >>>> cluster >>>> you ran in this test. >>>> >>>> >>>> +... >>>> +box.cfg{replication = "", replication_timeout = >>>> replication_timeout, \ >>>> + replication_connect_timeout = replication_connect_timeout} >>>> +--- >>>> +... >>>> >>>> One more extraneous change, related to the one above. >>>> >>>> >>>>> +--- >>>>> +... >>>>> diff --git >>>>> a/test/replication/misc_heartbeats_on_master_changes_gh-3160.test.lua >>>>> b/test/replication/misc_heartbeats_on_master_changes_gh-3160.test.lua >>>>> new file mode 100644 >>>>> index 000000000..eba8a7725 >>>>> --- /dev/null >>>>> +++ >>>>> b/test/replication/misc_heartbeats_on_master_changes_gh-3160.test.lua >>>>> @@ -0,0 +1,40 @@ >>>>> +uuid = require('uuid') >>>>> +test_run = require('test_run').new() >>>>> + >>>>> +-- gh-3160 - Send heartbeats if there are changes from a remote >>>>> master only >>>>> +SERVERS = { 'autobootstrap1', 'autobootstrap2', 'autobootstrap3' } >>>>> + >>>>> +-- Deploy a cluster. >>>>> +test_run:create_cluster(SERVERS, "replication", {args="0.03"}) >>>>> +test_run:wait_fullmesh(SERVERS) >>>>> +test_run:cmd("switch autobootstrap3") >>>>> +test_run = require('test_run').new() >>>>> +fiber = require('fiber') >>>>> +_ = box.schema.space.create('test_timeout'):create_index('pk') >>>>> +test_run:cmd("setopt delimiter ';'") >>>>> +function wait_not_follow(replicaA, replicaB) >>>>> + return test_run:wait_cond(function() >>>>> + return replicaA.status ~= 'follow' or replicaB.status ~= >>>>> 'follow' >>>>> + end, box.cfg.replication_timeout) >>>>> +end; >>>>> +function test_timeout() >>>>> + local replicaA = box.info.replication[1].upstream or >>>>> box.info.replication[2].upstream >>>>> + local replicaB = box.info.replication[3].upstream or >>>>> box.info.replication[2].upstream >>>>> + local follows = test_run:wait_cond(function() >>>>> + return replicaA.status == 'follow' or replicaB.status == >>>>> 'follow' >>>>> + end) >>>>> + if not follows then error('replicas are not in the follow >>>>> status') end >>>>> + for i = 0, 99 do >>>>> + box.space.test_timeout:replace({1}) >>>>> + if wait_not_follow(replicaA, replicaB) then >>>>> + return error(box.info.replication) >>>>> + end >>>>> + end >>>>> + return true >>>>> +end; >>>>> +test_run:cmd("setopt delimiter ''"); >>>>> +test_timeout() >>>>> + >>>>> +test_run:cmd("switch default") >>>>> +test_run:drop_cluster(SERVERS) >>>>> +test_run:cleanup_cluster() >>>>> diff --git >>>>> a/test/replication/misc_no_failure_on_error_reading_wal_gh-4399.result >>>>> b/test/replication/misc_no_failure_on_error_reading_wal_gh-4399.result >>>>> >>>>> new file mode 100644 >>>>> index 000000000..64a4de26c >>>>> --- /dev/null >>>>> +++ >>>>> b/test/replication/misc_no_failure_on_error_reading_wal_gh-4399.result >>>>> >>>>> @@ -0,0 +1,97 @@ >>>>> +test_run = require('test_run').new() >>>>> +--- >>>>> +... >>>>> +test_run:cmd("restart server default") >>>>> +uuid = require('uuid') >>>>> +--- >>>>> +... >>>>> +fiber = require('fiber') >>>> Same about restarting the server, requiring uuid and fiber. Not >>>> needed. >>>>> +--- >>>>> +... >>>>> +-- >>>>> +-- gh-4399 Check that an error reading WAL directory on subscribe >>>>> +-- doesn't lead to a permanent replication failure. >>>>> +-- >>>>> +box.schema.user.grant("guest", "replication") >>>>> +--- >>>>> +... >>>>> +test_run:cmd("create server replica with rpl_master=default, >>>>> script='replication/replica.lua'") >>>>> +--- >>>>> +- true >>>>> +... >>>>> +test_run:cmd("start server replica") >>>>> +--- >>>>> +- true >>>>> +... >>>>> +-- Make the WAL directory inaccessible. >>>>> +fio = require('fio') >>>>> +--- >>>>> +... >>>>> +path = fio.abspath(box.cfg.wal_dir) >>>>> +--- >>>>> +... >>>>> +fio.chmod(path, 0) >>>>> +--- >>>>> +- true >>>>> +... >>>>> +-- Break replication on timeout. >>>>> +replication_timeout = box.cfg.replication_timeout >>>>> +--- >>>>> +... >>>>> +box.cfg{replication_timeout = 9000} >>>>> +--- >>>>> +... >>>>> +test_run:cmd("switch replica") >>>>> +--- >>>>> +- true >>>>> +... >>>>> +test_run:wait_cond(function() return >>>>> box.info.replication[1].upstream.status ~= 'follow' end) >>>>> +--- >>>>> +- true >>>>> +... >>>>> +require('fiber').sleep(box.cfg.replication_timeout) >>>>> +--- >>>>> +... >>>>> +test_run:cmd("switch default") >>>>> +--- >>>>> +- true >>>>> +... >>>>> +box.cfg{replication_timeout = replication_timeout} >>>>> +--- >>>>> +... >>>>> +-- Restore access to the WAL directory. >>>>> +-- Wait for replication to be reestablished. >>>>> +fio.chmod(path, tonumber('777', 8)) >>>>> +--- >>>>> +- true >>>>> +... >>>>> +test_run:cmd("switch replica") >>>>> +--- >>>>> +- true >>>>> +... >>>>> +test_run:wait_cond(function() return >>>>> box.info.replication[1].upstream.status == 'follow' end) >>>>> +--- >>>>> +- true >>>>> +... >>>>> +test_run:cmd("switch default") >>>>> +--- >>>>> +- true >>>>> +... >>>>> +test_run:cmd("stop server replica") >>>>> +--- >>>>> +- true >>>>> +... >>>>> +test_run:cmd("cleanup server replica") >>>>> +--- >>>>> +- true >>>>> +... >>>>> +test_run:cmd("delete server replica") >>>>> +--- >>>>> +- true >>>>> +... >>>>> +test_run:cleanup_cluster() >>>>> +--- >>>>> +... >>>>> +box.schema.user.revoke('guest', 'replication') >>>>> +--- >>>>> +... >>>>> diff --git >>>>> a/test/replication/misc_no_failure_on_error_reading_wal_gh-4399.test.lua >>>>> b/test/replication/misc_no_failure_on_error_reading_wal_gh-4399.test.lua >>>>> >>>>> new file mode 100644 >>>>> index 000000000..15e19a211 >>>>> --- /dev/null >>>>> +++ >>>>> b/test/replication/misc_no_failure_on_error_reading_wal_gh-4399.test.lua >>>>> >>>>> @@ -0,0 +1,39 @@ >>>>> +test_run = require('test_run').new() >>>>> +test_run:cmd("restart server default") >>>>> +uuid = require('uuid') >>>>> +fiber = require('fiber') >>>>> + >>>>> +-- >>>>> +-- gh-4399 Check that an error reading WAL directory on subscribe >>>>> +-- doesn't lead to a permanent replication failure. >>>>> +-- >>>>> +box.schema.user.grant("guest", "replication") >>>>> +test_run:cmd("create server replica with rpl_master=default, >>>>> script='replication/replica.lua'") >>>>> +test_run:cmd("start server replica") >>>>> + >>>>> +-- Make the WAL directory inaccessible. >>>>> +fio = require('fio') >>>>> +path = fio.abspath(box.cfg.wal_dir) >>>>> +fio.chmod(path, 0) >>>>> + >>>>> +-- Break replication on timeout. >>>>> +replication_timeout = box.cfg.replication_timeout >>>>> +box.cfg{replication_timeout = 9000} >>>>> +test_run:cmd("switch replica") >>>>> +test_run:wait_cond(function() return >>>>> box.info.replication[1].upstream.status ~= 'follow' end) >>>>> +require('fiber').sleep(box.cfg.replication_timeout) >>>>> +test_run:cmd("switch default") >>>>> +box.cfg{replication_timeout = replication_timeout} >>>>> + >>>>> +-- Restore access to the WAL directory. >>>>> +-- Wait for replication to be reestablished. >>>>> +fio.chmod(path, tonumber('777', 8)) >>>>> +test_run:cmd("switch replica") >>>>> +test_run:wait_cond(function() return >>>>> box.info.replication[1].upstream.status == 'follow' end) >>>>> +test_run:cmd("switch default") >>>>> + >>>>> +test_run:cmd("stop server replica") >>>>> +test_run:cmd("cleanup server replica") >>>>> +test_run:cmd("delete server replica") >>>>> +test_run:cleanup_cluster() >>>>> +box.schema.user.revoke('guest', 'replication') >>>>> diff --git >>>>> a/test/replication/misc_no_panic_on_connected_gh-3637.result >>>>> b/test/replication/misc_no_panic_on_connected_gh-3637.result >>>>> new file mode 100644 >>>>> index 000000000..693b18c42 >>>>> --- /dev/null >>>>> +++ b/test/replication/misc_no_panic_on_connected_gh-3637.result >>>>> @@ -0,0 +1,72 @@ >>>>> +test_run = require('test_run').new() >>>>> +--- >>>>> +... >>>>> +test_run:cmd("restart server default") >>>>> +uuid = require('uuid') >>>>> +--- >>>>> +... >>>> Again, no need for uuid here. And no need for restart. >>>>> +-- >>>>> +-- Test case for gh-3637, gh-4550. Before the fix replica would >>>>> +-- exit with an error if a user does not exist or a password is >>>>> +-- incorrect. Now check that we don't hang/panic and successfully >>>>> +-- connect. >>>>> +-- >>>>> +fiber = require('fiber') >>>>> +--- >>>>> +... >>>>> +test_run:cmd("create server replica_auth with rpl_master=default, >>>>> script='replication/replica_auth.lua'") >>>>> +--- >>>>> +- true >>>>> +... >>>>> +test_run:cmd("start server replica_auth with wait=False, >>>>> wait_load=False, args='cluster:pass 0.05'") >>>>> +--- >>>>> +- true >>>>> +... >>>>> +-- Wait a bit to make sure replica waits till user is created. >>>>> +fiber.sleep(0.1) >>>>> +--- >>>>> +... >>>>> +box.schema.user.create('cluster') >>>>> +--- >>>>> +... >>>>> +-- The user is created. Let the replica fail auth request due to >>>>> +-- a wrong password. >>>>> +fiber.sleep(0.1) >>>>> +--- >>>>> +... >>>>> +box.schema.user.passwd('cluster', 'pass') >>>>> +--- >>>>> +... >>>>> +box.schema.user.grant('cluster', 'replication') >>>>> +--- >>>>> +... >>>>> +while box.info.replication[2] == nil do fiber.sleep(0.01) end >>>>> +--- >>>>> +... >>>>> +vclock = test_run:get_vclock('default') >>>>> +--- >>>>> +... >>>>> +vclock[0] = nil >>>>> +--- >>>>> +... >>>>> +_ = test_run:wait_vclock('replica_auth', vclock) >>>>> +--- >>>>> +... >>>>> +test_run:cmd("stop server replica_auth") >>>>> +--- >>>>> +- true >>>>> +... >>>>> +test_run:cmd("cleanup server replica_auth") >>>>> +--- >>>>> +- true >>>>> +... >>>>> +test_run:cmd("delete server replica_auth") >>>>> +--- >>>>> +- true >>>>> +... >>>>> +test_run:cleanup_cluster() >>>>> +--- >>>>> +... >>>>> +box.schema.user.drop('cluster') >>>>> +--- >>>>> +... >>>>> diff --git >>>>> a/test/replication/misc_no_panic_on_connected_gh-3637.test.lua >>>>> b/test/replication/misc_no_panic_on_connected_gh-3637.test.lua >>>>> new file mode 100644 >>>>> index 000000000..a1e51198b >>>>> --- /dev/null >>>>> +++ b/test/replication/misc_no_panic_on_connected_gh-3637.test.lua >>>>> @@ -0,0 +1,33 @@ >>>>> +test_run = require('test_run').new() >>>>> +test_run:cmd("restart server default") >>>>> +uuid = require('uuid') >>>>> + >>>>> +-- >>>>> +-- Test case for gh-3637, gh-4550. Before the fix replica would >>>>> +-- exit with an error if a user does not exist or a password is >>>>> +-- incorrect. Now check that we don't hang/panic and successfully >>>>> +-- connect. >>>>> +-- >>>>> +fiber = require('fiber') >>>>> +test_run:cmd("create server replica_auth with rpl_master=default, >>>>> script='replication/replica_auth.lua'") >>>>> +test_run:cmd("start server replica_auth with wait=False, >>>>> wait_load=False, args='cluster:pass 0.05'") >>>>> +-- Wait a bit to make sure replica waits till user is created. >>>>> +fiber.sleep(0.1) >>>>> +box.schema.user.create('cluster') >>>>> +-- The user is created. Let the replica fail auth request due to >>>>> +-- a wrong password. >>>>> +fiber.sleep(0.1) >>>>> +box.schema.user.passwd('cluster', 'pass') >>>>> +box.schema.user.grant('cluster', 'replication') >>>>> + >>>>> +while box.info.replication[2] == nil do fiber.sleep(0.01) end >>>>> +vclock = test_run:get_vclock('default') >>>>> +vclock[0] = nil >>>>> +_ = test_run:wait_vclock('replica_auth', vclock) >>>>> + >>>>> +test_run:cmd("stop server replica_auth") >>>>> +test_run:cmd("cleanup server replica_auth") >>>>> +test_run:cmd("delete server replica_auth") >>>>> +test_run:cleanup_cluster() >>>>> + >>>>> +box.schema.user.drop('cluster') >>>>> diff --git >>>>> a/test/replication/misc_no_restart_on_same_configuration_gh-3711.result >>>>> b/test/replication/misc_no_restart_on_same_configuration_gh-3711.result >>>>> >>>>> new file mode 100644 >>>>> index 000000000..dd9ef848c >>>>> --- /dev/null >>>>> +++ >>>>> b/test/replication/misc_no_restart_on_same_configuration_gh-3711.result >>>>> >>>>> @@ -0,0 +1,107 @@ >>>>> +test_run = require('test_run').new() >>>>> +--- >>>>> +... >>>>> +test_run:cmd("restart server default") >>>>> +uuid = require('uuid') >>>>> +--- >>>>> +... >>>>> +fiber = require('fiber') >>>> Same about restart, uuid and fiber. >>>>> +--- >>>>> +... >>>>> +-- >>>>> +-- gh-3711 Do not restart replication on box.cfg in case the >>>>> +-- configuration didn't change. >>>>> +-- >>>>> +box.schema.user.grant('guest', 'replication') >>>>> +--- >>>>> +... >>>>> +test_run:cmd("create server replica with rpl_master=default, >>>>> script='replication/replica.lua'") >>>>> +--- >>>>> +- true >>>>> +... >>>>> +test_run:cmd("start server replica") >>>>> +--- >>>>> +- true >>>>> +... >>>>> +-- Access rights are checked only during reconnect. If the new >>>>> +-- and old configurations are equivalent, no reconnect will be >>>>> +-- issued and replication should continue working. >>>>> +box.schema.user.revoke('guest', 'replication') >>>>> +--- >>>>> +... >>>>> +test_run:cmd("switch replica") >>>>> +--- >>>>> +- true >>>>> +... >>>>> +replication = box.cfg.replication[1] >>>>> +--- >>>>> +... >>>>> +box.cfg{replication = {replication}} >>>>> +--- >>>>> +... >>>>> +box.info.status == 'running' >>>>> +--- >>>>> +- true >>>>> +... >>>>> +box.cfg{replication = replication} >>>>> +--- >>>>> +... >>>>> +box.info.status == 'running' >>>>> +--- >>>>> +- true >>>>> +... >>>>> +-- Check that comparison of tables works as expected as well. >>>>> +test_run:cmd("switch default") >>>>> +--- >>>>> +- true >>>>> +... >>>>> +box.schema.user.grant('guest', 'replication') >>>>> +--- >>>>> +... >>>>> +test_run:cmd("switch replica") >>>>> +--- >>>>> +- true >>>>> +... >>>>> +replication = box.cfg.replication >>>>> +--- >>>>> +... >>>>> +table.insert(replication, box.cfg.listen) >>>>> +--- >>>>> +... >>>>> +test_run:cmd("switch default") >>>>> +--- >>>>> +- true >>>>> +... >>>>> +box.schema.user.revoke('guest', 'replication') >>>>> +--- >>>>> +... >>>>> +test_run:cmd("switch replica") >>>>> +--- >>>>> +- true >>>>> +... >>>>> +box.cfg{replication = replication} >>>>> +--- >>>>> +... >>>>> +box.info.status == 'running' >>>>> +--- >>>>> +- true >>>>> +... >>>>> +test_run:cmd("switch default") >>>>> +--- >>>>> +- true >>>>> +... >>>>> +test_run:cmd("stop server replica") >>>>> +--- >>>>> +- true >>>>> +... >>>>> +test_run:cmd("cleanup server replica") >>>>> +--- >>>>> +- true >>>>> +... >>>>> +test_run:cmd("delete server replica") >>>>> +--- >>>>> +- true >>>>> +... >>>>> +test_run:cleanup_cluster() >>>>> +--- >>>>> +... >>>>> diff --git >>>>> a/test/replication/misc_no_restart_on_same_configuration_gh-3711.test.lua >>>>> b/test/replication/misc_no_restart_on_same_configuration_gh-3711.test.lua >>>>> >>>>> new file mode 100644 >>>>> index 000000000..14b522ead >>>>> --- /dev/null >>>>> +++ >>>>> b/test/replication/misc_no_restart_on_same_configuration_gh-3711.test.lua >>>>> @@ -0,0 +1,41 @@ >>>>> +test_run = require('test_run').new() >>>>> +test_run:cmd("restart server default") >>>>> +uuid = require('uuid') >>>>> +fiber = require('fiber') >>>>> + >>>>> +-- >>>>> +-- gh-3711 Do not restart replication on box.cfg in case the >>>>> +-- configuration didn't change. >>>>> +-- >>>>> +box.schema.user.grant('guest', 'replication') >>>>> +test_run:cmd("create server replica with rpl_master=default, >>>>> script='replication/replica.lua'") >>>>> +test_run:cmd("start server replica") >>>>> + >>>>> +-- Access rights are checked only during reconnect. If the new >>>>> +-- and old configurations are equivalent, no reconnect will be >>>>> +-- issued and replication should continue working. >>>>> +box.schema.user.revoke('guest', 'replication') >>>>> +test_run:cmd("switch replica") >>>>> +replication = box.cfg.replication[1] >>>>> +box.cfg{replication = {replication}} >>>>> +box.info.status == 'running' >>>>> +box.cfg{replication = replication} >>>>> +box.info.status == 'running' >>>>> + >>>>> +-- Check that comparison of tables works as expected as well. >>>>> +test_run:cmd("switch default") >>>>> +box.schema.user.grant('guest', 'replication') >>>>> +test_run:cmd("switch replica") >>>>> +replication = box.cfg.replication >>>>> +table.insert(replication, box.cfg.listen) >>>>> +test_run:cmd("switch default") >>>>> +box.schema.user.revoke('guest', 'replication') >>>>> +test_run:cmd("switch replica") >>>>> +box.cfg{replication = replication} >>>>> +box.info.status == 'running' >>>>> + >>>>> +test_run:cmd("switch default") >>>>> +test_run:cmd("stop server replica") >>>>> +test_run:cmd("cleanup server replica") >>>>> +test_run:cmd("delete server replica") >>>>> +test_run:cleanup_cluster() >>>>> diff --git >>>>> a/test/replication/misc_no_socket_leak_on_replica_disconnect_gh-3642.result >>>>> b/test/replication/misc_no_socket_leak_on_replica_disconnect_gh-3642.result >>>>> >>>>> new file mode 100644 >>>>> index 000000000..3680bcebb >>>>> --- /dev/null >>>>> +++ >>>>> b/test/replication/misc_no_socket_leak_on_replica_disconnect_gh-3642.result >>>>> @@ -0,0 +1,98 @@ >>>>> +test_run = require('test_run').new() >>>>> +--- >>>>> +... >>>>> +test_run:cmd("restart server default") >>>> Unneeded restart. Correct me if I'm wrong. >>>> >>>>> +uuid = require('uuid') >>>> You don't need the uuid module in this testcase. >>>> >>>>> +--- >>>>> +... >>>>> +box.schema.user.grant('guest', 'replication') >>>>> +--- >>>>> +... >>>>> +-- gh-3642 - Check that socket file descriptor doesn't leak >>>>> +-- when a replica is disconnected. >>>>> +rlimit = require('rlimit') >>>>> +--- >>>>> +... >>>>> +lim = rlimit.limit() >>>>> +--- >>>>> +... >>>>> +rlimit.getrlimit(rlimit.RLIMIT_NOFILE, lim) >>>>> +--- >>>>> +... >>>>> +old_fno = lim.rlim_cur >>>>> +--- >>>>> +... >>>>> +lim.rlim_cur = 64 >>>>> +--- >>>>> +... >>>>> +rlimit.setrlimit(rlimit.RLIMIT_NOFILE, lim) >>>>> +--- >>>>> +... >>>>> +test_run:cmd('create server sock with rpl_master=default, >>>>> script="replication/replica.lua"') >>>>> +--- >>>>> +- true >>>>> +... >>>>> +test_run:cmd('start server sock') >>>>> +--- >>>>> +- true >>>>> +... >>>>> +test_run:cmd('switch sock') >>>>> +--- >>>>> +- true >>>>> +... >>>>> +test_run = require('test_run').new() >>>>> +--- >>>>> +... >>>>> +fiber = require('fiber') >>>>> +--- >>>>> +... >>>>> +test_run:cmd("setopt delimiter ';'") >>>>> +--- >>>>> +- true >>>>> +... >>>>> +for i = 1, 64 do >>>>> + local replication = box.cfg.replication >>>>> + box.cfg{replication = {}} >>>>> + box.cfg{replication = replication} >>>>> + while box.info.replication[1].upstream.status ~= 'follow' do >>>>> + fiber.sleep(0.001) >>>>> + end >>>>> +end; >>>>> +--- >>>>> +... >>>>> +test_run:cmd("setopt delimiter ''"); >>>>> +--- >>>>> +- true >>>>> +... >>>>> +box.info.replication[1].upstream.status >>>>> +--- >>>>> +- follow >>>>> +... >>>>> +test_run:cmd('switch default') >>>>> +--- >>>>> +- true >>>>> +... >>>>> +lim.rlim_cur = old_fno >>>>> +--- >>>>> +... >>>>> +rlimit.setrlimit(rlimit.RLIMIT_NOFILE, lim) >>>>> +--- >>>>> +... >>>>> +test_run:cmd("stop server sock") >>>>> +--- >>>>> +- true >>>>> +... >>>>> +test_run:cmd("cleanup server sock") >>>>> +--- >>>>> +- true >>>>> +... >>>>> +test_run:cmd("delete server sock") >>>>> +--- >>>>> +- true >>>>> +... >>>>> +test_run:cleanup_cluster() >>>>> +--- >>>>> +... >>>>> +box.schema.user.revoke('guest', 'replication') >>>>> +--- >>>>> +... >>>>> diff --git >>>>> a/test/replication/misc_no_socket_leak_on_replica_disconnect_gh-3642.test.lua >>>>> b/test/replication/misc_no_socket_leak_on_replica_disconnect_gh-3642.test.lua >>>>> >>>>> new file mode 100644 >>>>> index 000000000..08ef9ec0d >>>>> --- /dev/null >>>>> +++ >>>>> b/test/replication/misc_no_socket_leak_on_replica_disconnect_gh-3642.test.lua >>>>> @@ -0,0 +1,44 @@ >>>>> +test_run = require('test_run').new() >>>>> +test_run:cmd("restart server default") >>>>> +uuid = require('uuid') >>>>> + >>>>> +box.schema.user.grant('guest', 'replication') >>>>> + >>>>> +-- gh-3642 - Check that socket file descriptor doesn't leak >>>>> +-- when a replica is disconnected. >>>>> +rlimit = require('rlimit') >>>>> +lim = rlimit.limit() >>>>> +rlimit.getrlimit(rlimit.RLIMIT_NOFILE, lim) >>>>> +old_fno = lim.rlim_cur >>>>> +lim.rlim_cur = 64 >>>>> +rlimit.setrlimit(rlimit.RLIMIT_NOFILE, lim) >>>>> + >>>>> +test_run:cmd('create server sock with rpl_master=default, >>>>> script="replication/replica.lua"') >>>>> +test_run:cmd('start server sock') >>>>> +test_run:cmd('switch sock') >>>>> +test_run = require('test_run').new() >>>>> +fiber = require('fiber') >>>>> +test_run:cmd("setopt delimiter ';'") >>>>> +for i = 1, 64 do >>>>> + local replication = box.cfg.replication >>>>> + box.cfg{replication = {}} >>>>> + box.cfg{replication = replication} >>>>> + while box.info.replication[1].upstream.status ~= 'follow' do >>>>> + fiber.sleep(0.001) >>>>> + end >>>>> +end; >>>>> +test_run:cmd("setopt delimiter ''"); >>>>> + >>>>> +box.info.replication[1].upstream.status >>>>> + >>>>> +test_run:cmd('switch default') >>>>> + >>>>> +lim.rlim_cur = old_fno >>>>> +rlimit.setrlimit(rlimit.RLIMIT_NOFILE, lim) >>>>> + >>>>> +test_run:cmd("stop server sock") >>>>> +test_run:cmd("cleanup server sock") >>>>> +test_run:cmd("delete server sock") >>>>> +test_run:cleanup_cluster() >>>>> + >>>>> +box.schema.user.revoke('guest', 'replication') >>>>> diff --git >>>>> a/test/replication/misc_orphan_on_reconfiguration_error_gh-4424.result >>>>> b/test/replication/misc_orphan_on_reconfiguration_error_gh-4424.result >>>>> >>>>> new file mode 100644 >>>>> index 000000000..6a9d1148d >>>>> --- /dev/null >>>>> +++ >>>>> b/test/replication/misc_orphan_on_reconfiguration_error_gh-4424.result >>>>> >>>>> @@ -0,0 +1,88 @@ >>>>> +uuid = require('uuid') >>>>> +--- >>>>> +... >>>>> +test_run = require('test_run').new() >>>>> +--- >>>>> +... >>>>> +fiber = require('fiber') >>>> UUID and fiber aren't used in this testcase >>>> >>>>> +--- >>>>> +... >>>>> +-- >>>>> +-- gh-4424 Always enter orphan mode on error in replication >>>>> +-- configuration change. >>>>> +-- >>>>> +replication_connect_timeout = box.cfg.replication_connect_timeout >>>>> +--- >>>>> +... >>>>> +replication_connect_quorum = box.cfg.replication_connect_quorum >>>>> +--- >>>>> +... >>>>> +box.cfg{replication="12345", replication_connect_timeout=0.1, >>>>> replication_connect_quorum=1} >>>>> +--- >>>>> +... >>>>> +box.info.status >>>>> +--- >>>>> +- orphan >>>>> +... >>>>> +box.info.ro >>>>> +--- >>>>> +- true >>>>> +... >>>>> +-- reset replication => leave orphan mode >>>>> +box.cfg{replication=""} >>>>> +--- >>>>> +... >>>>> +box.info.status >>>>> +--- >>>>> +- running >>>>> +... >>>>> +box.info.ro >>>>> +--- >>>>> +- false >>>>> +... >>>>> +-- no switch to orphan when quorum == 0 >>>>> +box.cfg{replication="12345", replication_connect_quorum=0} >>>>> +--- >>>>> +... >>>>> +box.info.status >>>>> +--- >>>>> +- running >>>>> +... >>>>> +box.info.ro >>>>> +--- >>>>> +- false >>>>> +... >>>>> +-- we could connect to one out of two replicas. Set orphan. >>>>> +box.cfg{replication_connect_quorum=2} >>>>> +--- >>>>> +... >>>>> +box.cfg{replication={box.cfg.listen, "12345"}} >>>>> +--- >>>>> +... >>>>> +box.info.status >>>>> +--- >>>>> +- orphan >>>>> +... >>>>> +box.info.ro >>>>> +--- >>>>> +- true >>>>> +... >>>>> +-- lower quorum => leave orphan mode >>>>> +box.cfg{replication_connect_quorum=1} >>>>> +--- >>>>> +... >>>>> +box.info.status >>>>> +--- >>>>> +- running >>>>> +... >>>>> +box.info.ro >>>>> +--- >>>>> +- false >>>>> +... >>>>> +box.cfg{ \ >>>>> + replication = {}, \ >>>>> + replication_connect_quorum = replication_connect_quorum, \ >>>>> + replication_connect_timeout = replication_connect_timeout \ >>>>> +} >>>>> +--- >>>>> +... >>>>> diff --git >>>>> a/test/replication/misc_orphan_on_reconfiguration_error_gh-4424.test.lua >>>>> b/test/replication/misc_orphan_on_reconfiguration_error_gh-4424.test.lua >>>>> >>>>> new file mode 100644 >>>>> index 000000000..7d1a70a36 >>>>> --- /dev/null >>>>> +++ >>>>> b/test/replication/misc_orphan_on_reconfiguration_error_gh-4424.test.lua >>>>> >>>>> @@ -0,0 +1,37 @@ >>>>> +uuid = require('uuid') >>>>> +test_run = require('test_run').new() >>>>> +fiber = require('fiber') >>>>> + >>>>> +-- >>>>> +-- gh-4424 Always enter orphan mode on error in replication >>>>> +-- configuration change. >>>>> +-- >>>>> +replication_connect_timeout = box.cfg.replication_connect_timeout >>>>> +replication_connect_quorum = box.cfg.replication_connect_quorum >>>>> +box.cfg{replication="12345", replication_connect_timeout=0.1, >>>>> replication_connect_quorum=1} >>>>> +box.info.status >>>>> +box.info.ro >>>>> +-- reset replication => leave orphan mode >>>>> +box.cfg{replication=""} >>>>> +box.info.status >>>>> +box.info.ro >>>>> +-- no switch to orphan when quorum == 0 >>>>> +box.cfg{replication="12345", replication_connect_quorum=0} >>>>> +box.info.status >>>>> +box.info.ro >>>>> + >>>>> +-- we could connect to one out of two replicas. Set orphan. >>>>> +box.cfg{replication_connect_quorum=2} >>>>> +box.cfg{replication={box.cfg.listen, "12345"}} >>>>> +box.info.status >>>>> +box.info.ro >>>>> +-- lower quorum => leave orphan mode >>>>> +box.cfg{replication_connect_quorum=1} >>>>> +box.info.status >>>>> +box.info.ro >>>>> + >>>>> +box.cfg{ \ >>>>> + replication = {}, \ >>>>> + replication_connect_quorum = replication_connect_quorum, \ >>>>> + replication_connect_timeout = replication_connect_timeout \ >>>>> +} >>>>> diff --git >>>>> a/test/replication/misc_rebootstrap_from_ro_master_gh-3111.result >>>>> b/test/replication/misc_rebootstrap_from_ro_master_gh-3111.result >>>>> new file mode 100644 >>>>> index 000000000..7ffca1585 >>>>> --- /dev/null >>>>> +++ b/test/replication/misc_rebootstrap_from_ro_master_gh-3111.result >>>>> @@ -0,0 +1,58 @@ >>>>> +test_run = require('test_run').new() >>>>> +--- >>>>> +... >>>>> +test_run:cmd("restart server default") >>>> Why do you need it? Seems like the test passes just fine without the >>>> restart. >>>> Restarting the server will take some time if this is executed after >>>> some >>>> test that >>>> performs lots of DML operations.So You better remove the change >>>> >>>>> +uuid = require('uuid') >>>>> +--- >>>>> +... >>>>> +box.schema.user.grant('guest', 'replication') >>>>> +--- >>>>> +... >>>>> +-- gh-3111 - Allow to rebootstrap a replica from a read-only master >>>>> +replica_uuid = uuid.new() >>>>> +--- >>>>> +... >>>>> +test_run:cmd('create server test with rpl_master=default, >>>>> script="replication/replica_uuid.lua"') >>>>> +--- >>>>> +- true >>>>> +... >>>>> +test_run:cmd(string.format('start server test with args="%s"', >>>>> replica_uuid)) >>>>> +--- >>>>> +- true >>>>> +... >>>>> +test_run:cmd('stop server test') >>>>> +--- >>>>> +- true >>>>> +... >>>>> +test_run:cmd('cleanup server test') >>>>> +--- >>>>> +- true >>>>> +... >>>>> +box.cfg{read_only = true} >>>>> +--- >>>>> +... >>>>> +test_run:cmd(string.format('start server test with args="%s"', >>>>> replica_uuid)) >>>>> +--- >>>>> +- true >>>>> +... >>>>> +test_run:cmd('stop server test') >>>>> +--- >>>>> +- true >>>>> +... >>>>> +test_run:cmd('cleanup server test') >>>>> +--- >>>>> +- true >>>>> +... >>>>> +box.cfg{read_only = false} >>>>> +--- >>>>> +... >>>>> +test_run:cmd('delete server test') >>>>> +--- >>>>> +- true >>>>> +... >>>>> +test_run:cleanup_cluster() >>>>> +--- >>>>> +... >>>>> +box.schema.user.revoke('guest', 'replication') >>>>> +--- >>>>> +... >>>>> diff --git >>>>> a/test/replication/misc_rebootstrap_from_ro_master_gh-3111.test.lua >>>>> b/test/replication/misc_rebootstrap_from_ro_master_gh-3111.test.lua >>>>> new file mode 100644 >>>>> index 000000000..bb9b4a80f >>>>> --- /dev/null >>>>> +++ >>>>> b/test/replication/misc_rebootstrap_from_ro_master_gh-3111.test.lua >>>>> @@ -0,0 +1,20 @@ >>>>> +test_run = require('test_run').new() >>>>> +test_run:cmd("restart server default") >>>>> +uuid = require('uuid') >>>>> + >>>>> +box.schema.user.grant('guest', 'replication') >>>>> + >>>>> +-- gh-3111 - Allow to rebootstrap a replica from a read-only master >>>>> +replica_uuid = uuid.new() >>>>> +test_run:cmd('create server test with rpl_master=default, >>>>> script="replication/replica_uuid.lua"') >>>>> +test_run:cmd(string.format('start server test with args="%s"', >>>>> replica_uuid)) >>>>> +test_run:cmd('stop server test') >>>>> +test_run:cmd('cleanup server test') >>>>> +box.cfg{read_only = true} >>>>> +test_run:cmd(string.format('start server test with args="%s"', >>>>> replica_uuid)) >>>>> +test_run:cmd('stop server test') >>>>> +test_run:cmd('cleanup server test') >>>>> +box.cfg{read_only = false} >>>>> +test_run:cmd('delete server test') >>>>> +test_run:cleanup_cluster() >>>>> +box.schema.user.revoke('guest', 'replication') >>>>> diff --git >>>>> a/test/replication/misc_replica_checks_cluster_id_gh-3704.result >>>>> b/test/replication/misc_replica_checks_cluster_id_gh-3704.result >>>>> new file mode 100644 >>>>> index 000000000..e6bc8b4d8 >>>>> --- /dev/null >>>>> +++ b/test/replication/misc_replica_checks_cluster_id_gh-3704.result >>>>> @@ -0,0 +1,71 @@ >>>>> +test_run = require('test_run').new() >>>>> +--- >>>>> +... >>>>> +test_run:cmd("restart server default") >>>>> +uuid = require('uuid') >>>>> +--- >>>>> +... >>>>> +fiber = require('fiber') >>>> No need to restart the instance, and require fiber module. >>>> >>>> UUID is needed this time. >>>> >>>>> +--- >>>>> +... >>>>> +-- >>>>> +-- gh-3704 move cluster id check to replica >>>>> +-- >>>>> +test_run:cmd("create server replica with rpl_master=default, >>>>> script='replication/replica.lua'") >>>>> +--- >>>>> +- true >>>>> +... >>>>> +box.schema.user.grant("guest", "replication") >>>>> +--- >>>>> +... >>>>> +test_run:cmd("start server replica") >>>>> +--- >>>>> +- true >>>>> +... >>>>> +test_run:grep_log("replica", "REPLICASET_UUID_MISMATCH") >>>>> +--- >>>>> +- null >>>>> +... >>>>> +box.info.replication[2].downstream.status >>>>> +--- >>>>> +- follow >>>>> +... >>>>> +-- change master's cluster uuid and check that replica doesn't >>>>> connect. >>>>> +test_run:cmd("stop server replica") >>>>> +--- >>>>> +- true >>>>> +... >>>>> +_ = box.space._schema:replace{'cluster', tostring(uuid.new())} >>>>> +--- >>>>> +... >>>>> +-- master believes replica is in cluster, but their cluster UUIDs >>>>> differ. >>>>> +test_run:cmd("start server replica") >>>>> +--- >>>>> +- true >>>>> +... >>>>> +test_run:wait_log("replica", "REPLICASET_UUID_MISMATCH", nil, 1.0) >>>>> +--- >>>>> +- REPLICASET_UUID_MISMATCH >>>>> +... >>>>> +test_run:wait_downstream(2, {status = 'stopped'}) >>>>> +--- >>>>> +- true >>>>> +... >>>>> +test_run:cmd("stop server replica") >>>>> +--- >>>>> +- true >>>>> +... >>>>> +test_run:cmd("cleanup server replica") >>>>> +--- >>>>> +- true >>>>> +... >>>>> +test_run:cmd("delete server replica") >>>>> +--- >>>>> +- true >>>>> +... >>>>> +test_run:cleanup_cluster() >>>>> +--- >>>>> +... >>>>> +box.schema.user.revoke('guest', 'replication') >>>>> +--- >>>>> +... >>>>> diff --git >>>>> a/test/replication/misc_replica_checks_cluster_id_gh-3704.test.lua >>>>> b/test/replication/misc_replica_checks_cluster_id_gh-3704.test.lua >>>>> new file mode 100644 >>>>> index 000000000..8a23cc1fe >>>>> --- /dev/null >>>>> +++ >>>>> b/test/replication/misc_replica_checks_cluster_id_gh-3704.test.lua >>>>> @@ -0,0 +1,26 @@ >>>>> +test_run = require('test_run').new() >>>>> +test_run:cmd("restart server default") >>>>> +uuid = require('uuid') >>>>> +fiber = require('fiber') >>>>> + >>>>> +-- >>>>> +-- gh-3704 move cluster id check to replica >>>>> +-- >>>>> +test_run:cmd("create server replica with rpl_master=default, >>>>> script='replication/replica.lua'") >>>>> +box.schema.user.grant("guest", "replication") >>>>> +test_run:cmd("start server replica") >>>>> +test_run:grep_log("replica", "REPLICASET_UUID_MISMATCH") >>>>> +box.info.replication[2].downstream.status >>>>> +-- change master's cluster uuid and check that replica doesn't >>>>> connect. >>>>> +test_run:cmd("stop server replica") >>>>> +_ = box.space._schema:replace{'cluster', tostring(uuid.new())} >>>>> +-- master believes replica is in cluster, but their cluster UUIDs >>>>> differ. >>>>> +test_run:cmd("start server replica") >>>>> +test_run:wait_log("replica", "REPLICASET_UUID_MISMATCH", nil, 1.0) >>>>> +test_run:wait_downstream(2, {status = 'stopped'}) >>>>> + >>>>> +test_run:cmd("stop server replica") >>>>> +test_run:cmd("cleanup server replica") >>>>> +test_run:cmd("delete server replica") >>>>> +test_run:cleanup_cluster() >>>>> +box.schema.user.revoke('guest', 'replication') >>>>> diff --git >>>>> a/test/replication/misc_return_on_quorum_0_gh-3760.result >>>>> b/test/replication/misc_return_on_quorum_0_gh-3760.result >>>>> new file mode 100644 >>>>> index 000000000..2eb622896 >>>>> --- /dev/null >>>>> +++ b/test/replication/misc_return_on_quorum_0_gh-3760.result >>>>> @@ -0,0 +1,48 @@ >>>>> +uuid = require('uuid') >>>>> +--- >>>>> +... >>>>> +test_run = require('test_run').new() >>>>> +--- >>>>> +... >>>>> +fiber = require('fiber') >>>> Unneeded uuid and fiber >>>> >>>>> +--- >>>>> +... >>>>> +replication_connect_timeout = box.cfg.replication_connect_timeout >>>>> +--- >>>>> +... >>>>> +replication_connect_quorum = box.cfg.replication_connect_quorum >>>>> +--- >>>>> +... >>>>> +box.cfg{replication="12345", replication_connect_timeout=0.1, >>>>> replication_connect_quorum=1} >>>> This box.cfg call is extraneous. The test'll be just fine without it. >>>> >>>>> +--- >>>>> +... >>>>> +-- >>>>> +-- gh-3760: replication quorum 0 on reconfiguration should return >>>>> +-- from box.cfg immediately. >>>>> +-- >>>>> +replication = box.cfg.replication >>>>> +--- >>>>> +... >>>>> +box.cfg{ \ >>>>> + replication = {}, \ >>>>> + replication_connect_quorum = 0, \ >>>>> + replication_connect_timeout = 1000000 \ >>>>> +} >>>>> +--- >>>>> +... >>>>> +-- The call below would hang, if quorum 0 is ignored, or checked >>>>> +-- too late. >>>>> +box.cfg{replication = {'localhost:12345'}} >>>>> +--- >>>>> +... >>>>> +box.info.status >>>>> +--- >>>>> +- running >>>>> +... >>>>> +box.cfg{ \ >>>>> + replication = {}, \ >>>>> + replication_connect_quorum = replication_connect_quorum, \ >>>>> + replication_connect_timeout = replication_connect_timeout \ >>>>> +} >>>>> +--- >>>>> +... >>>>> diff --git >>>>> a/test/replication/misc_return_on_quorum_0_gh-3760.test.lua >>>>> b/test/replication/misc_return_on_quorum_0_gh-3760.test.lua >>>>> new file mode 100644 >>>>> index 000000000..9e0651032 >>>>> --- /dev/null >>>>> +++ b/test/replication/misc_return_on_quorum_0_gh-3760.test.lua >>>>> @@ -0,0 +1,27 @@ >>>>> +uuid = require('uuid') >>>>> +test_run = require('test_run').new() >>>>> +fiber = require('fiber') >>>>> + >>>>> +replication_connect_timeout = box.cfg.replication_connect_timeout >>>>> +replication_connect_quorum = box.cfg.replication_connect_quorum >>>>> +box.cfg{replication="12345", replication_connect_timeout=0.1, >>>>> replication_connect_quorum=1} >>>>> + >>>>> +-- >>>>> +-- gh-3760: replication quorum 0 on reconfiguration should return >>>>> +-- from box.cfg immediately. >>>>> +-- >>>>> +replication = box.cfg.replication >>>>> +box.cfg{ \ >>>>> + replication = {}, \ >>>>> + replication_connect_quorum = 0, \ >>>>> + replication_connect_timeout = 1000000 \ >>>>> +} >>>>> +-- The call below would hang, if quorum 0 is ignored, or checked >>>>> +-- too late. >>>>> +box.cfg{replication = {'localhost:12345'}} >>>>> +box.info.status >>>>> +box.cfg{ \ >>>>> + replication = {}, \ >>>>> + replication_connect_quorum = replication_connect_quorum, \ >>>>> + replication_connect_timeout = replication_connect_timeout \ >>>>> +} >>>>> diff --git >>>>> a/test/replication/misc_value_not_replicated_on_iproto_request_gh-3247.result >>>>> b/test/replication/misc_value_not_replicated_on_iproto_request_gh-3247.result >>>>> >>>>> new file mode 100644 >>>>> index 000000000..d416bd9a6 >>>>> --- /dev/null >>>>> +++ >>>>> b/test/replication/misc_value_not_replicated_on_iproto_request_gh-3247.result >>>>> @@ -0,0 +1,90 @@ >>>>> +test_run = require('test_run').new() >>>>> +--- >>>>> +... >>>>> +test_run:cmd("restart server default") >>>>> +uuid = require('uuid') >>>>> +--- >>>>> +... >>>> You don't need the uuid module in this testcase. >>>> >>>> >>>>> +-- Deploy a cluster. >>>>> +SERVERS = { 'autobootstrap1', 'autobootstrap2', 'autobootstrap3' } >>>>> +--- >>>>> +... >>>>> +test_run:create_cluster(SERVERS, "replication", {args="0.03"}) >>>>> +--- >>>>> +... >>>>> +test_run:wait_fullmesh(SERVERS) >>>>> +--- >>>>> +... >>>>> +-- gh-3247 - Sequence-generated value is not replicated in case >>>>> +-- the request was sent via iproto. >>>>> +test_run:cmd("switch autobootstrap1") >>>>> +--- >>>>> +- true >>>>> +... >>>>> +net_box = require('net.box') >>>>> +--- >>>>> +... >>>>> +_ = box.schema.space.create('space1') >>>>> +--- >>>>> +... >>>>> +_ = box.schema.sequence.create('seq') >>>>> +--- >>>>> +... >>>>> +_ = box.space.space1:create_index('primary', {sequence = true} ) >>>>> +--- >>>>> +... >>>>> +_ = box.space.space1:create_index('secondary', {parts = {2, >>>>> 'unsigned'}}) >>>>> +--- >>>>> +... >>>>> +box.schema.user.grant('guest', 'read,write', 'space', 'space1') >>>>> +--- >>>>> +... >>>>> +c = net_box.connect(box.cfg.listen) >>>>> +--- >>>>> +... >>>>> +c.space.space1:insert{box.NULL, "data"} -- fails, but bumps >>>>> sequence value >>>>> +--- >>>>> +- error: 'Tuple field 2 type does not match one required by >>>>> operation: expected unsigned' >>>>> +... >>>>> +c.space.space1:insert{box.NULL, 1, "data"} >>>>> +--- >>>>> +- [2, 1, 'data'] >>>>> +... >>>>> +box.space.space1:select{} >>>>> +--- >>>>> +- - [2, 1, 'data'] >>>>> +... >>>>> +vclock = test_run:get_vclock("autobootstrap1") >>>>> +--- >>>>> +... >>>>> +vclock[0] = nil >>>>> +--- >>>>> +... >>>>> +_ = test_run:wait_vclock("autobootstrap2", vclock) >>>>> +--- >>>>> +... >>>>> +test_run:cmd("switch autobootstrap2") >>>>> +--- >>>>> +- true >>>>> +... >>>>> +box.space.space1:select{} >>>>> +--- >>>>> +- - [2, 1, 'data'] >>>>> +... >>>>> +test_run:cmd("switch autobootstrap1") >>>>> +--- >>>>> +- true >>>>> +... >>>>> +box.space.space1:drop() >>>>> +--- >>>>> +... >>>>> +test_run:cmd("switch default") >>>>> +--- >>>>> +- true >>>>> +... >>>>> +test_run:drop_cluster(SERVERS) >>>>> +--- >>>>> +... >>>>> +test_run:cleanup_cluster() >>>>> +--- >>>>> +... >>>>> diff --git >>>>> a/test/replication/misc_value_not_replicated_on_iproto_request_gh-3247.test.lua >>>>> b/test/replication/misc_value_not_replicated_on_iproto_request_gh-3247.test.lua >>>>> >>>>> new file mode 100644 >>>>> index 000000000..586e8b997 >>>>> --- /dev/null >>>>> +++ >>>>> b/test/replication/misc_value_not_replicated_on_iproto_request_gh-3247.test.lua >>>>> @@ -0,0 +1,33 @@ >>>>> +test_run = require('test_run').new() >>>>> +test_run:cmd("restart server default") >>>>> +uuid = require('uuid') >>>>> + >>>>> +-- Deploy a cluster. >>>>> +SERVERS = { 'autobootstrap1', 'autobootstrap2', 'autobootstrap3' } >>>>> +test_run:create_cluster(SERVERS, "replication", {args="0.03"}) >>>>> +test_run:wait_fullmesh(SERVERS) >>>>> + >>>>> +-- gh-3247 - Sequence-generated value is not replicated in case >>>>> +-- the request was sent via iproto. >>>>> +test_run:cmd("switch autobootstrap1") >>>>> +net_box = require('net.box') >>>>> +_ = box.schema.space.create('space1') >>>>> +_ = box.schema.sequence.create('seq') >>>>> +_ = box.space.space1:create_index('primary', {sequence = true} ) >>>>> +_ = box.space.space1:create_index('secondary', {parts = {2, >>>>> 'unsigned'}}) >>>>> +box.schema.user.grant('guest', 'read,write', 'space', 'space1') >>>>> +c = net_box.connect(box.cfg.listen) >>>>> +c.space.space1:insert{box.NULL, "data"} -- fails, but bumps >>>>> sequence value >>>>> +c.space.space1:insert{box.NULL, 1, "data"} >>>>> +box.space.space1:select{} >>>>> +vclock = test_run:get_vclock("autobootstrap1") >>>>> +vclock[0] = nil >>>>> +_ = test_run:wait_vclock("autobootstrap2", vclock) >>>>> +test_run:cmd("switch autobootstrap2") >>>>> +box.space.space1:select{} >>>>> +test_run:cmd("switch autobootstrap1") >>>>> +box.space.space1:drop() >>>>> + >>>>> +test_run:cmd("switch default") >>>>> +test_run:drop_cluster(SERVERS) >>>>> +test_run:cleanup_cluster() >>>>> diff --git a/test/replication/suite.cfg b/test/replication/suite.cfg >>>>> index f357b07da..e21daa5ad 100644 >>>>> --- a/test/replication/suite.cfg >>>>> +++ b/test/replication/suite.cfg >>>>> @@ -1,6 +1,19 @@ >>>>> { >>>>> "anon.test.lua": {}, >>>>> - "misc.test.lua": {}, >>>>> + "misc_assert_connecting_master_twice_gh-3610.test.lua": {}, >>>>> + "misc_assert_on_server_die_gh-2991.test.lua": {}, >>>>> + "misc_assert_replica_on_applier_disconnect_gh-3510.test.lua": {}, >>>>> + "misc_crash_on_box_concurrent_update_gh-3606.test.lua": {}, >>>>> + "misc_heartbeats_on_master_changes_gh-3160.test.lua": {}, >>>>> + "misc_no_failure_on_error_reading_wal_gh-4399.test.lua": {}, >>>>> + "misc_no_panic_on_connected_gh-3637.test.lua": {}, >>>>> + "misc_no_restart_on_same_configuration_gh-3711.test.lua": {}, >>>>> + "misc_no_socket_leak_on_replica_disconnect_gh-3642.test.lua": {}, >>>>> + "misc_orphan_on_reconfiguration_error_gh-4424.test.lua": {}, >>>>> + "misc_rebootstrap_from_ro_master_gh-3111.test.lua": {}, >>>>> + "misc_replica_checks_cluster_id_gh-3704.test.lua": {}, >>>>> + "misc_return_on_quorum_0_gh-3760.test.lua": {}, >>>>> + "misc_value_not_replicated_on_iproto_request_gh-3247.test.lua": {}, >>>>> "once.test.lua": {}, >>>>> "on_replace.test.lua": {}, >>>>> "status.test.lua": {}, >>>> -- >>>> Serge Petrenko >>>> -- Serge Petrenko
prev parent reply other threads:[~2020-09-03 11:30 UTC|newest] Thread overview: 2+ messages / expand[flat|nested] mbox.gz Atom feed top [not found] <20200903075417.GA24428@hpalx> 2020-09-03 9:32 ` Sergey Bronnikov 2020-09-03 11:30 ` Serge Petrenko [this message]
Reply instructions: You may reply publicly to this message via plain-text email using any one of the following methods: * Save the following mbox file, import it into your mail client, and reply-to-all from there: mbox Avoid top-posting and favor interleaved quoting: https://en.wikipedia.org/wiki/Posting_style#Interleaved_style * Reply using the --to, --cc, and --in-reply-to switches of git-send-email(1): git send-email \ --in-reply-to=3ca7abf0-08fd-009f-be4a-af43ce260930@tarantool.org \ --to=sergepetrenko@tarantool.org \ --cc=avtikhon@tarantool.org \ --cc=sergeyb@tarantool.org \ --cc=tarantool-patches@dev.tarantool.org \ --subject='Re: [Tarantool-patches] [sergepetrenko@tarantool.org: Re: [PATCH v1] Divide replication/mist.test.lua]' \ /path/to/YOUR_REPLY https://kernel.org/pub/software/scm/git/docs/git-send-email.html * If your mail client supports setting the In-Reply-To header via mailto: links, try the mailto: link
This is a public inbox, see mirroring instructions for how to clone and mirror all data and code used for this inbox