[tarantool-patches] Re: [PATCH v2 1/4] relay: adjust gc state on relay status update

Konstantin Osipov kostja.osipov at gmail.com
Fri Sep 20 10:26:50 MSK 2019

* Georgy Kirichenko <georgy at tarantool.org> [19/09/19 16:15]:
> On Thursday, September 19, 2019 12:45:47 AM MSK Vladislav Shpilevoy wrote:
> > Thanks for the patch!
> Thanks for the review. I'll try to explain the patch here.
> A relay collects ACK's from replica. Before a parallel applier was implemented 
> there was one ACK packet per transaction. And it was to expensive to update gc 
> state for each transaction. To overwhelm this issue a relay used a trigger 
> which fires when recovery finished with a file. So, when a relay received a 
> close-log event, it waits for the first ACK greather than 'closed' vclock and 
> then advances a gc.  In case of in-memory replication we definitely couldn't 
> rely on file boundaries and on_close trigger. Because we already have parallel 
> applier we shouldn't have to much ACK packets I decided to not to use 
> on_close_log more and pass ACK direct to garbage collector.
> In other words, a relay still continues gc advancing in both modes (file or 
> memory) but does it after each ACK. Also this required to change vclock 
> comparison because gc vclocks are not aligned by local xlog vclock timeline.
> Yes, this changed the gc behavior - now gc keeps only local changes (because 
> an INSTANCE_ID is used). Though, this could/would be changed back when we move 
> a relay to the wal thread (what is needed for synchronous replication 
> purposes)

By tracking only local changes you make it impossible to recover a
replica which has fallen behind in a 3-node setup. Essentially
each replica is now responsible for keeping track of its own
changes only - which means it will happily delete xlogs with
changes of a lost peer even if these changes are not synced to
other peers yet.

Konstantin Osipov, Moscow, Russia

More information about the Tarantool-patches mailing list