* [Tarantool-patches] [PATCH v3] replication: fix join vclock obtainment in relay_initial_join
@ 2019-10-29 14:23 Ilya Kosarev
2019-10-29 17:02 ` Konstantin Osipov
0 siblings, 1 reply; 8+ messages in thread
From: Ilya Kosarev @ 2019-10-29 14:23 UTC (permalink / raw)
To: tarantool-patches; +Cc: tarantool-patches, v.shpilevoy
join_vclock test could fail on huge load due to vclock advance
comparing to an actual WAL. In order to fix this we updated wal_sync so
that now we can obtain up to date vclock on the flushed state using it.
It is also better to get max index and index count in single request in
join_vclock test. With fixes mentioned above it is not fragile anymore.
Closes #4160
---
https://github.com/tarantool/tarantool/tree/i.kosarev/gh-4160-fix-join-vclock
https://github.com/tarantool/tarantool/issues/4160
src/box/relay.cc | 7 +++----
src/box/vinyl.c | 4 ++--
src/box/wal.c | 23 +++++++++++++++++------
src/box/wal.h | 2 +-
test/replication/join_vclock.result | 5 +----
test/replication/join_vclock.test.lua | 3 +--
test/replication/suite.ini | 1 -
7 files changed, 25 insertions(+), 20 deletions(-)
diff --git a/src/box/relay.cc b/src/box/relay.cc
index e849fcf4f..b9ed27503 100644
--- a/src/box/relay.cc
+++ b/src/box/relay.cc
@@ -307,13 +307,12 @@ relay_initial_join(int fd, uint64_t sync, struct vclock *vclock)
/*
* Sync WAL to make sure that all changes visible from
- * the frozen read view are successfully committed.
+ * the frozen read view are successfully committed and
+ * obtain corresponding vclock.
*/
- if (wal_sync() != 0)
+ if (wal_sync(vclock) != 0)
diag_raise();
- vclock_copy(vclock, &replicaset.vclock);
-
/* Respond to the JOIN request with the current vclock. */
struct xrow_header row;
xrow_encode_vclock_xc(&row, vclock);
diff --git a/src/box/vinyl.c b/src/box/vinyl.c
index a2bbaa529..e160f3764 100644
--- a/src/box/vinyl.c
+++ b/src/box/vinyl.c
@@ -1087,7 +1087,7 @@ vinyl_space_check_format(struct space *space, struct tuple_format *format)
*/
int rc;
if (need_wal_sync) {
- rc = wal_sync();
+ rc = wal_sync(NULL);
if (rc != 0)
goto out;
}
@@ -4168,7 +4168,7 @@ vinyl_space_build_index(struct space *src_space, struct index *new_index,
*/
int rc;
if (need_wal_sync) {
- rc = wal_sync();
+ rc = wal_sync(NULL);
if (rc != 0)
goto out;
}
diff --git a/src/box/wal.c b/src/box/wal.c
index 5e2c13e0e..eabc73fff 100644
--- a/src/box/wal.c
+++ b/src/box/wal.c
@@ -519,21 +519,28 @@ wal_free(void)
wal_writer_destroy(writer);
}
+struct wal_vclock_msg {
+ struct cbus_call_msg base;
+ struct vclock *vclock;
+};
+
static int
-wal_sync_f(struct cbus_call_msg *msg)
+wal_sync_f(struct cbus_call_msg *data)
{
- (void)msg;
+ struct wal_vclock_msg *msg = (struct wal_vclock_msg *) data;
struct wal_writer *writer = &wal_writer_singleton;
if (writer->in_rollback.route != NULL) {
/* We're rolling back a failed write. */
diag_set(ClientError, ER_WAL_IO);
return -1;
}
+ if (msg->vclock != NULL)
+ vclock_copy(msg->vclock, &writer->vclock);
return 0;
}
int
-wal_sync(void)
+wal_sync(struct vclock *vclock)
{
ERROR_INJECT(ERRINJ_WAL_SYNC, {
diag_set(ClientError, ER_INJECTION, "wal sync");
@@ -541,17 +548,21 @@ wal_sync(void)
});
struct wal_writer *writer = &wal_writer_singleton;
- if (writer->wal_mode == WAL_NONE)
+ if (writer->wal_mode == WAL_NONE) {
+ if (vclock != NULL)
+ vclock_copy(vclock, &writer->vclock);
return 0;
+ }
if (!stailq_empty(&writer->rollback)) {
/* We're rolling back a failed write. */
diag_set(ClientError, ER_WAL_IO);
return -1;
}
bool cancellable = fiber_set_cancellable(false);
- struct cbus_call_msg msg;
+ struct wal_vclock_msg msg;
+ msg.vclock = vclock;
int rc = cbus_call(&writer->wal_pipe, &writer->tx_prio_pipe,
- &msg, wal_sync_f, NULL, TIMEOUT_INFINITY);
+ &msg.base, wal_sync_f, NULL, TIMEOUT_INFINITY);
fiber_set_cancellable(cancellable);
return rc;
}
diff --git a/src/box/wal.h b/src/box/wal.h
index b76b0a41f..d82934ffc 100644
--- a/src/box/wal.h
+++ b/src/box/wal.h
@@ -184,7 +184,7 @@ wal_mode();
* to disk. Returns 0 on success, -1 if write failed.
*/
int
-wal_sync(void);
+wal_sync(struct vclock *vclock);
struct wal_checkpoint {
struct cbus_call_msg base;
diff --git a/test/replication/join_vclock.result b/test/replication/join_vclock.result
index a9781073d..d6d9af783 100644
--- a/test/replication/join_vclock.result
+++ b/test/replication/join_vclock.result
@@ -67,10 +67,7 @@ test_run:cmd("switch replica1")
---
- true
...
-cnt = box.space.test.index[0]:count()
----
-...
-box.space.test.index.primary:max()[1] == cnt - 1
+box.space.test.index.primary:max()[1] == box.space.test.index.primary:count() - 1
---
- true
...
diff --git a/test/replication/join_vclock.test.lua b/test/replication/join_vclock.test.lua
index 0b60dffc2..a813ba31f 100644
--- a/test/replication/join_vclock.test.lua
+++ b/test/replication/join_vclock.test.lua
@@ -26,8 +26,7 @@ ch:get()
errinj.set("ERRINJ_RELAY_FINAL_SLEEP", false)
test_run:cmd("switch replica1")
-cnt = box.space.test.index[0]:count()
-box.space.test.index.primary:max()[1] == cnt - 1
+box.space.test.index.primary:max()[1] == box.space.test.index.primary:count() - 1
test_run:cmd("switch default")
replica_set.drop_all(test_run)
diff --git a/test/replication/suite.ini b/test/replication/suite.ini
index 384dac677..ed1de3140 100644
--- a/test/replication/suite.ini
+++ b/test/replication/suite.ini
@@ -12,7 +12,6 @@ long_run = prune.test.lua
is_parallel = True
pretest_clean = True
fragile = errinj.test.lua ; gh-3870
- join_vclock.test.lua ; gh-4160
long_row_timeout.test.lua ; gh-4351
skip_conflict_row.test.lua ; gh-4457
sync.test.lua ; gh-3835 gh-3877
--
2.17.1
^ permalink raw reply [flat|nested] 8+ messages in thread
* Re: [Tarantool-patches] [PATCH v3] replication: fix join vclock obtainment in relay_initial_join
2019-10-29 14:23 [Tarantool-patches] [PATCH v3] replication: fix join vclock obtainment in relay_initial_join Ilya Kosarev
@ 2019-10-29 17:02 ` Konstantin Osipov
2019-10-29 21:20 ` [Tarantool-patches] [PATCH v3] relay: " Ilya Kosarev
` (2 more replies)
0 siblings, 3 replies; 8+ messages in thread
From: Konstantin Osipov @ 2019-10-29 17:02 UTC (permalink / raw)
To: Ilya Kosarev; +Cc: tarantool-patches, v.shpilevoy, tarantool-patches
* Ilya Kosarev <i.kosarev@tarantool.org> [19/10/29 19:59]:
> join_vclock test could fail on huge load due to vclock advance
> comparing to an actual WAL. In order to fix this we updated wal_sync so
> that now we can obtain up to date vclock on the flushed state using it.
> It is also better to get max index and index count in single request in
> join_vclock test. With fixes mentioned above it is not fragile anymore.
I suggest we copy vclock. it's not a good idea to make the two
threads depend on each other on memory. the relay thread may be
cancelled while the messasge is in the wal thread.
--
Konstantin Osipov, Moscow, Russia
^ permalink raw reply [flat|nested] 8+ messages in thread
* [Tarantool-patches] [PATCH v3] relay: fix join vclock obtainment in relay_initial_join
2019-10-29 17:02 ` Konstantin Osipov
@ 2019-10-29 21:20 ` Ilya Kosarev
2019-10-30 22:22 ` Vladislav Shpilevoy
2019-10-29 21:20 ` [Tarantool-patches] [PATCH v3] replication: " Ilya Kosarev
2019-10-30 14:33 ` [Tarantool-patches] [tarantool-patches] " Ilya Kosarev
2 siblings, 1 reply; 8+ messages in thread
From: Ilya Kosarev @ 2019-10-29 21:20 UTC (permalink / raw)
To: tarantool-patches; +Cc: tarantool-patches, v.shpilevoy
join_vclock test could fail on huge load due to vclock advance
comparing to an actual WAL. In order to fix this we updated wal_sync so
that now we can obtain up to date vclock on the flushed state using it.
It is also better to get max index and index count in single request in
join_vclock test. With fixes mentioned above it is not fragile anymore.
Closes #4160
---
https://github.com/tarantool/tarantool/tree/i.kosarev/gh-4160-fix-join-vclock
https://github.com/tarantool/tarantool/issues/4160
src/box/relay.cc | 7 +++----
src/box/vinyl.c | 4 ++--
src/box/wal.c | 23 +++++++++++++++++------
src/box/wal.h | 3 ++-
test/replication/join_vclock.result | 5 +----
test/replication/join_vclock.test.lua | 3 +--
test/replication/suite.ini | 1 -
7 files changed, 26 insertions(+), 20 deletions(-)
diff --git a/src/box/relay.cc b/src/box/relay.cc
index e849fcf4f..b9ed27503 100644
--- a/src/box/relay.cc
+++ b/src/box/relay.cc
@@ -307,13 +307,12 @@ relay_initial_join(int fd, uint64_t sync, struct vclock *vclock)
/*
* Sync WAL to make sure that all changes visible from
- * the frozen read view are successfully committed.
+ * the frozen read view are successfully committed and
+ * obtain corresponding vclock.
*/
- if (wal_sync() != 0)
+ if (wal_sync(vclock) != 0)
diag_raise();
- vclock_copy(vclock, &replicaset.vclock);
-
/* Respond to the JOIN request with the current vclock. */
struct xrow_header row;
xrow_encode_vclock_xc(&row, vclock);
diff --git a/src/box/vinyl.c b/src/box/vinyl.c
index a2bbaa529..e160f3764 100644
--- a/src/box/vinyl.c
+++ b/src/box/vinyl.c
@@ -1087,7 +1087,7 @@ vinyl_space_check_format(struct space *space, struct tuple_format *format)
*/
int rc;
if (need_wal_sync) {
- rc = wal_sync();
+ rc = wal_sync(NULL);
if (rc != 0)
goto out;
}
@@ -4168,7 +4168,7 @@ vinyl_space_build_index(struct space *src_space, struct index *new_index,
*/
int rc;
if (need_wal_sync) {
- rc = wal_sync();
+ rc = wal_sync(NULL);
if (rc != 0)
goto out;
}
diff --git a/src/box/wal.c b/src/box/wal.c
index 5e2c13e0e..6348ef456 100644
--- a/src/box/wal.c
+++ b/src/box/wal.c
@@ -519,21 +519,27 @@ wal_free(void)
wal_writer_destroy(writer);
}
+struct wal_vclock_msg {
+ struct cbus_call_msg base;
+ struct vclock vclock;
+};
+
static int
-wal_sync_f(struct cbus_call_msg *msg)
+wal_sync_f(struct cbus_call_msg *data)
{
- (void)msg;
+ struct wal_vclock_msg *msg = (struct wal_vclock_msg *) data;
struct wal_writer *writer = &wal_writer_singleton;
if (writer->in_rollback.route != NULL) {
/* We're rolling back a failed write. */
diag_set(ClientError, ER_WAL_IO);
return -1;
}
+ vclock_copy(&msg->vclock, &writer->vclock);
return 0;
}
int
-wal_sync(void)
+wal_sync(struct vclock *vclock)
{
ERROR_INJECT(ERRINJ_WAL_SYNC, {
diag_set(ClientError, ER_INJECTION, "wal sync");
@@ -541,18 +547,23 @@ wal_sync(void)
});
struct wal_writer *writer = &wal_writer_singleton;
- if (writer->wal_mode == WAL_NONE)
+ if (writer->wal_mode == WAL_NONE) {
+ if (vclock != NULL)
+ vclock_copy(vclock, &writer->vclock);
return 0;
+ }
if (!stailq_empty(&writer->rollback)) {
/* We're rolling back a failed write. */
diag_set(ClientError, ER_WAL_IO);
return -1;
}
bool cancellable = fiber_set_cancellable(false);
- struct cbus_call_msg msg;
+ struct wal_vclock_msg msg;
int rc = cbus_call(&writer->wal_pipe, &writer->tx_prio_pipe,
- &msg, wal_sync_f, NULL, TIMEOUT_INFINITY);
+ &msg.base, wal_sync_f, NULL, TIMEOUT_INFINITY);
fiber_set_cancellable(cancellable);
+ if (vclock != NULL)
+ vclock_copy(vclock, &msg.vclock);
return rc;
}
diff --git a/src/box/wal.h b/src/box/wal.h
index b76b0a41f..827b0fb85 100644
--- a/src/box/wal.h
+++ b/src/box/wal.h
@@ -182,9 +182,10 @@ wal_mode();
/**
* Wait until all submitted writes are successfully flushed
* to disk. Returns 0 on success, -1 if write failed.
+ * Corresponding vclock is returned in @vclock unless it is NULL.
*/
int
-wal_sync(void);
+wal_sync(struct vclock *vclock);
struct wal_checkpoint {
struct cbus_call_msg base;
diff --git a/test/replication/join_vclock.result b/test/replication/join_vclock.result
index a9781073d..d6d9af783 100644
--- a/test/replication/join_vclock.result
+++ b/test/replication/join_vclock.result
@@ -67,10 +67,7 @@ test_run:cmd("switch replica1")
---
- true
...
-cnt = box.space.test.index[0]:count()
----
-...
-box.space.test.index.primary:max()[1] == cnt - 1
+box.space.test.index.primary:max()[1] == box.space.test.index.primary:count() - 1
---
- true
...
diff --git a/test/replication/join_vclock.test.lua b/test/replication/join_vclock.test.lua
index 0b60dffc2..a813ba31f 100644
--- a/test/replication/join_vclock.test.lua
+++ b/test/replication/join_vclock.test.lua
@@ -26,8 +26,7 @@ ch:get()
errinj.set("ERRINJ_RELAY_FINAL_SLEEP", false)
test_run:cmd("switch replica1")
-cnt = box.space.test.index[0]:count()
-box.space.test.index.primary:max()[1] == cnt - 1
+box.space.test.index.primary:max()[1] == box.space.test.index.primary:count() - 1
test_run:cmd("switch default")
replica_set.drop_all(test_run)
diff --git a/test/replication/suite.ini b/test/replication/suite.ini
index 384dac677..ed1de3140 100644
--- a/test/replication/suite.ini
+++ b/test/replication/suite.ini
@@ -12,7 +12,6 @@ long_run = prune.test.lua
is_parallel = True
pretest_clean = True
fragile = errinj.test.lua ; gh-3870
- join_vclock.test.lua ; gh-4160
long_row_timeout.test.lua ; gh-4351
skip_conflict_row.test.lua ; gh-4457
sync.test.lua ; gh-3835 gh-3877
--
2.17.1
^ permalink raw reply [flat|nested] 8+ messages in thread
* Re: [Tarantool-patches] [PATCH v3] replication: fix join vclock obtainment in relay_initial_join
2019-10-29 17:02 ` Konstantin Osipov
2019-10-29 21:20 ` [Tarantool-patches] [PATCH v3] relay: " Ilya Kosarev
@ 2019-10-29 21:20 ` Ilya Kosarev
2019-10-30 14:33 ` [Tarantool-patches] [tarantool-patches] " Ilya Kosarev
2 siblings, 0 replies; 8+ messages in thread
From: Ilya Kosarev @ 2019-10-29 21:20 UTC (permalink / raw)
To: Konstantin Osipov; +Cc: tarantool-patches, v.shpilevoy, tarantool-patches
[-- Attachment #1: Type: text/plain, Size: 822 bytes --]
Right. Sent fixed patch in reply.
>Вторник, 29 октября 2019, 20:02 +03:00 от Konstantin Osipov <kostja.osipov@gmail.com>:
>
>* Ilya Kosarev < i.kosarev@tarantool.org > [19/10/29 19:59]:
>> join_vclock test could fail on huge load due to vclock advance
>> comparing to an actual WAL. In order to fix this we updated wal_sync so
>> that now we can obtain up to date vclock on the flushed state using it.
>> It is also better to get max index and index count in single request in
>> join_vclock test. With fixes mentioned above it is not fragile anymore.
>
> I suggest we copy vclock. it's not a good idea to make the two
> threads depend on each other on memory. the relay thread may be
> cancelled while the messasge is in the wal thread.
>
>
>--
>Konstantin Osipov, Moscow, Russia
--
Ilya Kosarev
[-- Attachment #2: Type: text/html, Size: 1321 bytes --]
^ permalink raw reply [flat|nested] 8+ messages in thread
* Re: [Tarantool-patches] [tarantool-patches] Re: [PATCH v3] replication: fix join vclock obtainment in relay_initial_join
2019-10-29 17:02 ` Konstantin Osipov
2019-10-29 21:20 ` [Tarantool-patches] [PATCH v3] relay: " Ilya Kosarev
2019-10-29 21:20 ` [Tarantool-patches] [PATCH v3] replication: " Ilya Kosarev
@ 2019-10-30 14:33 ` Ilya Kosarev
2 siblings, 0 replies; 8+ messages in thread
From: Ilya Kosarev @ 2019-10-30 14:33 UTC (permalink / raw)
To: tarantool-patches; +Cc: tarantool-patches, v.shpilevoy
[-- Attachment #1: Type: text/plain, Size: 958 bytes --]
For some reason i got this message second time. Here is fixed patch: https://lists.tarantool.org/pipermail/tarantool-patches/2019-October/012162.html
Guess it could be lost due to updated subject. среда, 30 октября 2019г., 17:03 +03:00 от Konstantin Osipov kostja.osipov@gmail.com :
>* Ilya Kosarev < i.kosarev@tarantool.org> [19/10/29 19:59]:
> join_vclock test could fail on huge load due to vclock advance
> comparing to an actual WAL. In order to fix this we updated wal_sync so
> that now we can obtain up to date vclock on the flushed state using it.
> It is also better to get max index and index count in single request in
> join_vclock test. With fixes mentioned above it is not fragile anymore.
>
> I suggest we copy vclock. it's not a good idea to make the two
> threads depend on each other on memory. the relay thread may be
> cancelled while the messasge is in the wal thread.
>
>
>--
>Konstantin Osipov, Moscow, Russia
>
[-- Attachment #2: Type: text/html, Size: 1640 bytes --]
^ permalink raw reply [flat|nested] 8+ messages in thread
* Re: [Tarantool-patches] [PATCH v3] relay: fix join vclock obtainment in relay_initial_join
2019-10-29 21:20 ` [Tarantool-patches] [PATCH v3] relay: " Ilya Kosarev
@ 2019-10-30 22:22 ` Vladislav Shpilevoy
2019-11-01 0:35 ` Ilya Kosarev
2019-11-01 0:38 ` Ilya Kosarev
0 siblings, 2 replies; 8+ messages in thread
From: Vladislav Shpilevoy @ 2019-10-30 22:22 UTC (permalink / raw)
To: Ilya Kosarev, tarantool-patches; +Cc: tarantool-patches
Hi! Thanks for the patch!
LGTM, except the comment below.
> diff --git a/src/box/wal.h b/src/box/wal.h
> index b76b0a41f..827b0fb85 100644
> --- a/src/box/wal.h
> +++ b/src/box/wal.h
> @@ -182,9 +182,10 @@ wal_mode();
> /**
> * Wait until all submitted writes are successfully flushed
> * to disk. Returns 0 on success, -1 if write failed.
> + * Corresponding vclock is returned in @vclock unless it is NULL.
> */
For parameter reference, please, use '@a <param_name>' syntax,
not '@<param_name>'. See doxygen documentation on @a.
> int
> -wal_sync(void);
> +wal_sync(struct vclock *vclock);
>
^ permalink raw reply [flat|nested] 8+ messages in thread
* [Tarantool-patches] [PATCH v3] relay: fix join vclock obtainment in relay_initial_join
2019-10-30 22:22 ` Vladislav Shpilevoy
@ 2019-11-01 0:35 ` Ilya Kosarev
2019-11-01 0:38 ` Ilya Kosarev
1 sibling, 0 replies; 8+ messages in thread
From: Ilya Kosarev @ 2019-11-01 0:35 UTC (permalink / raw)
To: tarantool-patches; +Cc: tarantool-patches, v.shpilevoy
join_vclock test could fail on huge load due to vclock advance
comparing to an actual WAL. In order to fix this we updated wal_sync so
that now we can obtain up to date vclock on the flushed state using it.
It is also better to get max index and index count in single request in
join_vclock test. With fixes mentioned above it is not fragile anymore.
Closes #4160
---
https://github.com/tarantool/tarantool/tree/i.kosarev/gh-4160-fix-join-vclock
https://github.com/tarantool/tarantool/issues/4160
src/box/relay.cc | 7 +++----
src/box/vinyl.c | 4 ++--
src/box/wal.c | 23 +++++++++++++++++------
src/box/wal.h | 4 +++-
test/replication/join_vclock.result | 5 +----
test/replication/join_vclock.test.lua | 3 +--
test/replication/suite.ini | 1 -
7 files changed, 27 insertions(+), 20 deletions(-)
diff --git a/src/box/relay.cc b/src/box/relay.cc
index e849fcf4f..b9ed27503 100644
--- a/src/box/relay.cc
+++ b/src/box/relay.cc
@@ -307,13 +307,12 @@ relay_initial_join(int fd, uint64_t sync, struct vclock *vclock)
/*
* Sync WAL to make sure that all changes visible from
- * the frozen read view are successfully committed.
+ * the frozen read view are successfully committed and
+ * obtain corresponding vclock.
*/
- if (wal_sync() != 0)
+ if (wal_sync(vclock) != 0)
diag_raise();
- vclock_copy(vclock, &replicaset.vclock);
-
/* Respond to the JOIN request with the current vclock. */
struct xrow_header row;
xrow_encode_vclock_xc(&row, vclock);
diff --git a/src/box/vinyl.c b/src/box/vinyl.c
index a2bbaa529..e160f3764 100644
--- a/src/box/vinyl.c
+++ b/src/box/vinyl.c
@@ -1087,7 +1087,7 @@ vinyl_space_check_format(struct space *space, struct tuple_format *format)
*/
int rc;
if (need_wal_sync) {
- rc = wal_sync();
+ rc = wal_sync(NULL);
if (rc != 0)
goto out;
}
@@ -4168,7 +4168,7 @@ vinyl_space_build_index(struct space *src_space, struct index *new_index,
*/
int rc;
if (need_wal_sync) {
- rc = wal_sync();
+ rc = wal_sync(NULL);
if (rc != 0)
goto out;
}
diff --git a/src/box/wal.c b/src/box/wal.c
index 5e2c13e0e..6348ef456 100644
--- a/src/box/wal.c
+++ b/src/box/wal.c
@@ -519,21 +519,27 @@ wal_free(void)
wal_writer_destroy(writer);
}
+struct wal_vclock_msg {
+ struct cbus_call_msg base;
+ struct vclock vclock;
+};
+
static int
-wal_sync_f(struct cbus_call_msg *msg)
+wal_sync_f(struct cbus_call_msg *data)
{
- (void)msg;
+ struct wal_vclock_msg *msg = (struct wal_vclock_msg *) data;
struct wal_writer *writer = &wal_writer_singleton;
if (writer->in_rollback.route != NULL) {
/* We're rolling back a failed write. */
diag_set(ClientError, ER_WAL_IO);
return -1;
}
+ vclock_copy(&msg->vclock, &writer->vclock);
return 0;
}
int
-wal_sync(void)
+wal_sync(struct vclock *vclock)
{
ERROR_INJECT(ERRINJ_WAL_SYNC, {
diag_set(ClientError, ER_INJECTION, "wal sync");
@@ -541,18 +547,23 @@ wal_sync(void)
});
struct wal_writer *writer = &wal_writer_singleton;
- if (writer->wal_mode == WAL_NONE)
+ if (writer->wal_mode == WAL_NONE) {
+ if (vclock != NULL)
+ vclock_copy(vclock, &writer->vclock);
return 0;
+ }
if (!stailq_empty(&writer->rollback)) {
/* We're rolling back a failed write. */
diag_set(ClientError, ER_WAL_IO);
return -1;
}
bool cancellable = fiber_set_cancellable(false);
- struct cbus_call_msg msg;
+ struct wal_vclock_msg msg;
int rc = cbus_call(&writer->wal_pipe, &writer->tx_prio_pipe,
- &msg, wal_sync_f, NULL, TIMEOUT_INFINITY);
+ &msg.base, wal_sync_f, NULL, TIMEOUT_INFINITY);
fiber_set_cancellable(cancellable);
+ if (vclock != NULL)
+ vclock_copy(vclock, &msg.vclock);
return rc;
}
diff --git a/src/box/wal.h b/src/box/wal.h
index b76b0a41f..76b44941a 100644
--- a/src/box/wal.h
+++ b/src/box/wal.h
@@ -182,9 +182,11 @@ wal_mode();
/**
* Wait until all submitted writes are successfully flushed
* to disk. Returns 0 on success, -1 if write failed.
+ * Corresponding vclock is returned in @a vclock unless it is
+ * NULL.
*/
int
-wal_sync(void);
+wal_sync(struct vclock *vclock);
struct wal_checkpoint {
struct cbus_call_msg base;
diff --git a/test/replication/join_vclock.result b/test/replication/join_vclock.result
index a9781073d..d6d9af783 100644
--- a/test/replication/join_vclock.result
+++ b/test/replication/join_vclock.result
@@ -67,10 +67,7 @@ test_run:cmd("switch replica1")
---
- true
...
-cnt = box.space.test.index[0]:count()
----
-...
-box.space.test.index.primary:max()[1] == cnt - 1
+box.space.test.index.primary:max()[1] == box.space.test.index.primary:count() - 1
---
- true
...
diff --git a/test/replication/join_vclock.test.lua b/test/replication/join_vclock.test.lua
index 0b60dffc2..a813ba31f 100644
--- a/test/replication/join_vclock.test.lua
+++ b/test/replication/join_vclock.test.lua
@@ -26,8 +26,7 @@ ch:get()
errinj.set("ERRINJ_RELAY_FINAL_SLEEP", false)
test_run:cmd("switch replica1")
-cnt = box.space.test.index[0]:count()
-box.space.test.index.primary:max()[1] == cnt - 1
+box.space.test.index.primary:max()[1] == box.space.test.index.primary:count() - 1
test_run:cmd("switch default")
replica_set.drop_all(test_run)
diff --git a/test/replication/suite.ini b/test/replication/suite.ini
index 384dac677..ed1de3140 100644
--- a/test/replication/suite.ini
+++ b/test/replication/suite.ini
@@ -12,7 +12,6 @@ long_run = prune.test.lua
is_parallel = True
pretest_clean = True
fragile = errinj.test.lua ; gh-3870
- join_vclock.test.lua ; gh-4160
long_row_timeout.test.lua ; gh-4351
skip_conflict_row.test.lua ; gh-4457
sync.test.lua ; gh-3835 gh-3877
--
2.17.1
^ permalink raw reply [flat|nested] 8+ messages in thread
* Re: [Tarantool-patches] [PATCH v3] relay: fix join vclock obtainment in relay_initial_join
2019-10-30 22:22 ` Vladislav Shpilevoy
2019-11-01 0:35 ` Ilya Kosarev
@ 2019-11-01 0:38 ` Ilya Kosarev
1 sibling, 0 replies; 8+ messages in thread
From: Ilya Kosarev @ 2019-11-01 0:38 UTC (permalink / raw)
To: Vladislav Shpilevoy; +Cc: tarantool-patches, tarantool-patches
[-- Attachment #1: Type: text/plain, Size: 914 bytes --]
Hi!
Thanks for your review.
Sent fixed patch in reply to this message:
https://lists.tarantool.org/pipermail/tarantool-patches/2019-November/012195.html
>Четверг, 31 октября 2019, 1:16 +03:00 от Vladislav Shpilevoy <v.shpilevoy@tarantool.org>:
>
>Hi! Thanks for the patch!
>
>LGTM, except the comment below.
>
>> diff --git a/src/box/wal.h b/src/box/wal.h
>> index b76b0a41f..827b0fb85 100644
>> --- a/src/box/wal.h
>> +++ b/src/box/wal.h
>> @@ -182,9 +182,10 @@ wal_mode();
>> /**
>> * Wait until all submitted writes are successfully flushed
>> * to disk. Returns 0 on success, -1 if write failed.
>> + * Corresponding vclock is returned in @vclock unless it is NULL.
>> */
>For parameter reference, please, use '@a <param_name>' syntax,
>not '@<param_name>'. See doxygen documentation on @a.
>
>> int
>> -wal_sync(void);
>> +wal_sync(struct vclock *vclock);
>>
--
Ilya Kosarev
[-- Attachment #2: Type: text/html, Size: 2049 bytes --]
^ permalink raw reply [flat|nested] 8+ messages in thread
end of thread, other threads:[~2019-11-01 0:38 UTC | newest]
Thread overview: 8+ messages (download: mbox.gz / follow: Atom feed)
-- links below jump to the message on this page --
2019-10-29 14:23 [Tarantool-patches] [PATCH v3] replication: fix join vclock obtainment in relay_initial_join Ilya Kosarev
2019-10-29 17:02 ` Konstantin Osipov
2019-10-29 21:20 ` [Tarantool-patches] [PATCH v3] relay: " Ilya Kosarev
2019-10-30 22:22 ` Vladislav Shpilevoy
2019-11-01 0:35 ` Ilya Kosarev
2019-11-01 0:38 ` Ilya Kosarev
2019-10-29 21:20 ` [Tarantool-patches] [PATCH v3] replication: " Ilya Kosarev
2019-10-30 14:33 ` [Tarantool-patches] [tarantool-patches] " Ilya Kosarev
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox