From mboxrd@z Thu Jan 1 00:00:00 1970 Return-Path: From: Vladimir Davydov Subject: [PATCH v2 7/8] vinyl: factor out helpers for accounting dump/compaction Date: Sun, 16 Sep 2018 20:06:50 +0300 Message-Id: <20d44f8f20bfb116f462054fb917a47ca59f271f.1537115208.git.vdavydov.dev@gmail.com> In-Reply-To: References: In-Reply-To: References: To: kostja@tarantool.org Cc: tarantool-patches@freelists.org List-ID: So that we can easily extend them to account the stats not only per LSM tree, but also globally, in vy_lsm_env. --- src/box/vy_lsm.c | 20 ++++++++++++++++++++ src/box/vy_lsm.h | 16 ++++++++++++++++ src/box/vy_scheduler.c | 29 +++++++++++++++-------------- 3 files changed, 51 insertions(+), 14 deletions(-) diff --git a/src/box/vy_lsm.c b/src/box/vy_lsm.c index 6b9d0e6d..dbea2898 100644 --- a/src/box/vy_lsm.c +++ b/src/box/vy_lsm.c @@ -747,6 +747,26 @@ vy_lsm_unacct_range(struct vy_lsm *lsm, struct vy_range *range) &range->compact_queue); } +void +vy_lsm_acct_dump(struct vy_lsm *lsm, + const struct vy_stmt_counter *in, + const struct vy_disk_stmt_counter *out) +{ + lsm->stat.disk.dump.count++; + vy_stmt_counter_add(&lsm->stat.disk.dump.in, in); + vy_disk_stmt_counter_add(&lsm->stat.disk.dump.out, out); +} + +void +vy_lsm_acct_compaction(struct vy_lsm *lsm, + const struct vy_disk_stmt_counter *in, + const struct vy_disk_stmt_counter *out) +{ + lsm->stat.disk.compact.count++; + vy_disk_stmt_counter_add(&lsm->stat.disk.compact.in, in); + vy_disk_stmt_counter_add(&lsm->stat.disk.compact.out, out); +} + int vy_lsm_rotate_mem(struct vy_lsm *lsm) { diff --git a/src/box/vy_lsm.h b/src/box/vy_lsm.h index ba2feeef..19f82e34 100644 --- a/src/box/vy_lsm.h +++ b/src/box/vy_lsm.h @@ -457,6 +457,22 @@ void vy_lsm_unacct_range(struct vy_lsm *lsm, struct vy_range *range); /** + * Account dump in LSM tree statistics. + */ +void +vy_lsm_acct_dump(struct vy_lsm *lsm, + const struct vy_stmt_counter *in, + const struct vy_disk_stmt_counter *out); + +/** + * Account compaction in LSM tree statistics. + */ +void +vy_lsm_acct_compaction(struct vy_lsm *lsm, + const struct vy_disk_stmt_counter *in, + const struct vy_disk_stmt_counter *out); + +/** * Allocate a new active in-memory index for an LSM tree while * moving the old one to the sealed list. Used by the dump task * in order not to bother about synchronization with concurrent diff --git a/src/box/vy_scheduler.c b/src/box/vy_scheduler.c index e4afeafd..2f85424a 100644 --- a/src/box/vy_scheduler.c +++ b/src/box/vy_scheduler.c @@ -1090,6 +1090,8 @@ vy_task_dump_complete(struct vy_task *task) struct vy_lsm *lsm = task->lsm; struct vy_run *new_run = task->new_run; int64_t dump_lsn = new_run->dump_lsn; + struct vy_disk_stmt_counter dump_out = new_run->count; + struct vy_stmt_counter dump_in; struct tuple_format *key_format = lsm->env->key_format; struct vy_mem *mem, *next_mem; struct vy_slice **new_slices, *slice; @@ -1178,12 +1180,8 @@ vy_task_dump_complete(struct vy_task *task) if (vy_log_tx_commit() < 0) goto fail_free_slices; - /* - * Account the new run. - */ + /* Account the new run. */ vy_lsm_add_run(lsm, new_run); - vy_disk_stmt_counter_add(&lsm->stat.disk.dump.out, &new_run->count); - /* Drop the reference held by the task. */ vy_run_unref(new_run); @@ -1212,16 +1210,18 @@ vy_task_dump_complete(struct vy_task *task) delete_mems: /* - * Delete dumped in-memory trees. + * Delete dumped in-memory trees and account dump in + * LSM tree statistics. */ + vy_stmt_counter_reset(&dump_in); rlist_foreach_entry_safe(mem, &lsm->sealed, in_sealed, next_mem) { if (mem->generation > scheduler->dump_generation) continue; - vy_stmt_counter_add(&lsm->stat.disk.dump.in, &mem->count); + vy_stmt_counter_add(&dump_in, &mem->count); vy_lsm_delete_mem(lsm, mem); } lsm->dump_lsn = MAX(lsm->dump_lsn, dump_lsn); - lsm->stat.disk.dump.count++; + vy_lsm_acct_dump(lsm, &dump_in, &dump_out); /* The iterator has been cleaned up in a worker thread. */ task->wi->iface->close(task->wi); @@ -1443,6 +1443,8 @@ vy_task_compact_complete(struct vy_task *task) struct vy_lsm *lsm = task->lsm; struct vy_range *range = task->range; struct vy_run *new_run = task->new_run; + struct vy_disk_stmt_counter compact_out = new_run->count; + struct vy_disk_stmt_counter compact_in; struct vy_slice *first_slice = task->first_slice; struct vy_slice *last_slice = task->last_slice; struct vy_slice *slice, *next_slice, *new_slice = NULL; @@ -1526,15 +1528,14 @@ vy_task_compact_complete(struct vy_task *task) */ if (new_slice != NULL) { vy_lsm_add_run(lsm, new_run); - vy_disk_stmt_counter_add(&lsm->stat.disk.compact.out, - &new_run->count); /* Drop the reference held by the task. */ vy_run_unref(new_run); } else vy_run_discard(new_run); /* - * Replace compacted slices with the resulting slice. + * Replace compacted slices with the resulting slice and + * account compaction in LSM tree statistics. * * Note, since a slice might have been added to the range * by a concurrent dump while compaction was in progress, @@ -1545,12 +1546,12 @@ vy_task_compact_complete(struct vy_task *task) vy_lsm_unacct_range(lsm, range); if (new_slice != NULL) vy_range_add_slice_before(range, new_slice, first_slice); + vy_disk_stmt_counter_reset(&compact_in); for (slice = first_slice; ; slice = next_slice) { next_slice = rlist_next_entry(slice, in_range); vy_range_remove_slice(range, slice); rlist_add_entry(&compacted_slices, slice, in_range); - vy_disk_stmt_counter_add(&lsm->stat.disk.compact.in, - &slice->count); + vy_disk_stmt_counter_add(&compact_in, &slice->count); if (slice == last_slice) break; } @@ -1558,7 +1559,7 @@ vy_task_compact_complete(struct vy_task *task) range->version++; vy_range_update_compact_priority(range, &lsm->opts); vy_lsm_acct_range(lsm, range); - lsm->stat.disk.compact.count++; + vy_lsm_acct_compaction(lsm, &compact_in, &compact_out); /* * Unaccount unused runs and delete compacted slices. -- 2.11.0