From: Kirill Yukhin <kyukhin@tarantool.org>
To: tarantool-patches@dev.tarantool.org
Subject: [Tarantool-patches] [PATCH 3/3] Apply clang-format
Date: Wed, 7 Oct 2020 17:11:33 +0300 [thread overview]
Message-ID: <96ac986f4ecc27e2bdff835e9fe2fef5a5a89a7a.1602076356.git.kyukhin@tarantool.org> (raw)
In-Reply-To: <cover.1602076356.git.kyukhin@tarantool.org>
---
src/box/alter.cc | 1565 ++++++++++++++++++++---------------------
src/box/applier.cc | 140 ++--
src/box/applier.h | 36 +-
src/box/authentication.cc | 8 +-
src/box/authentication.h | 1 -
src/box/bind.c | 13 +-
src/box/bind.h | 3 +-
src/box/blackhole.c | 15 +-
src/box/box.cc | 192 +++--
src/box/box.h | 104 ++-
src/box/call.c | 9 +-
src/box/checkpoint_schedule.c | 4 +-
src/box/checkpoint_schedule.h | 4 +-
src/box/ck_constraint.c | 21 +-
src/box/ck_constraint.h | 4 +-
src/box/coll_id.c | 2 +-
src/box/coll_id_cache.c | 9 +-
src/box/coll_id_def.c | 17 +-
src/box/column_mask.h | 8 +-
src/box/constraint_id.c | 8 +-
src/box/engine.c | 56 +-
src/box/engine.h | 84 ++-
src/box/errcode.c | 11 +-
src/box/error.cc | 83 ++-
src/box/error.h | 86 +--
src/box/execute.c | 98 ++-
src/box/field_def.c | 26 +-
src/box/field_def.h | 2 +-
src/box/field_map.c | 17 +-
src/box/field_map.h | 12 +-
src/box/fk_constraint.h | 5 +-
src/box/func.c | 58 +-
src/box/func.h | 3 +-
src/box/func_def.c | 25 +-
src/box/func_def.h | 4 +-
src/box/gc.c | 38 +-
src/box/identifier.c | 6 +-
src/box/index.cc | 118 ++--
src/box/index.h | 135 ++--
src/box/index_def.c | 46 +-
src/box/index_def.h | 22 +-
src/box/iproto.cc | 220 +++---
src/box/iproto_constants.h | 26 +-
src/box/iterator_type.c | 3 +-
src/box/iterator_type.h | 29 +-
src/box/journal.c | 6 +-
src/box/journal.h | 23 +-
src/box/key_def.c | 114 ++-
src/box/key_def.h | 62 +-
src/box/key_list.c | 8 +-
src/box/key_list.h | 4 +-
src/box/lua/call.c | 104 ++-
src/box/lua/call.h | 8 +-
src/box/lua/cfg.cc | 92 ++-
src/box/lua/console.c | 112 +--
src/box/lua/ctl.c | 14 +-
src/box/lua/error.cc | 35 +-
src/box/lua/execute.c | 55 +-
src/box/lua/index.c | 75 +-
src/box/lua/info.c | 85 +--
src/box/lua/init.c | 110 ++-
src/box/lua/key_def.c | 52 +-
src/box/lua/merger.c | 178 +++--
src/box/lua/misc.cc | 41 +-
src/box/lua/net_box.c | 95 ++-
src/box/lua/sequence.c | 12 +-
src/box/lua/serialize_lua.c | 197 +++---
src/box/lua/session.c | 68 +-
src/box/lua/slab.c | 17 +-
src/box/lua/slab.h | 3 +-
src/box/lua/space.cc | 81 +--
src/box/lua/stat.c | 36 +-
src/box/lua/stat.h | 3 +-
src/box/lua/tuple.c | 102 ++-
src/box/lua/xlog.c | 38 +-
src/box/memtx_bitset.c | 70 +-
src/box/memtx_engine.c | 109 ++-
src/box/memtx_engine.h | 18 +-
src/box/memtx_hash.c | 142 ++--
src/box/memtx_rtree.c | 59 +-
src/box/memtx_space.c | 188 +++--
src/box/memtx_space.h | 4 +-
src/box/memtx_tree.c | 274 ++++----
src/box/memtx_tx.c | 141 ++--
src/box/memtx_tx.h | 4 +-
src/box/merger.c | 9 +-
src/box/mp_error.cc | 80 +--
src/box/msgpack.c | 4 +-
src/box/opt_def.c | 10 +-
src/box/opt_def.h | 72 +-
src/box/port.h | 6 +-
src/box/raft.c | 31 +-
src/box/recovery.cc | 40 +-
src/box/recovery.h | 2 +-
src/box/relay.cc | 71 +-
src/box/relay.h | 2 +-
src/box/replication.cc | 121 ++--
src/box/replication.h | 11 +-
src/box/request.c | 14 +-
src/box/schema.cc | 158 ++---
src/box/schema.h | 17 +-
src/box/schema_def.c | 6 +-
src/box/schema_def.h | 6 +-
src/box/sequence.c | 55 +-
src/box/service_engine.c | 10 +-
src/box/session.cc | 38 +-
src/box/session.h | 9 +-
src/box/session_settings.c | 34 +-
src/box/space.c | 86 +--
src/box/space.h | 87 ++-
src/box/space_def.c | 29 +-
src/box/space_def.h | 16 +-
src/box/sysview.c | 53 +-
src/box/tuple.c | 105 +--
src/box/tuple.h | 32 +-
src/box/tuple_bloom.c | 37 +-
src/box/tuple_bloom.h | 5 +-
src/box/tuple_compare.cc | 394 +++++------
src/box/tuple_convert.c | 28 +-
src/box/tuple_dictionary.c | 37 +-
| 94 +--
src/box/tuple_format.c | 216 +++---
src/box/tuple_format.h | 34 +-
src/box/tuple_hash.cc | 140 ++--
src/box/txn.c | 66 +-
src/box/txn_limbo.c | 34 +-
src/box/user.cc | 101 ++-
src/box/user.h | 5 +-
src/box/user_def.c | 23 +-
src/box/user_def.h | 6 +-
src/box/vclock.c | 151 ++--
src/box/vclock.h | 25 +-
src/box/vinyl.c | 388 +++++-----
src/box/vinyl.h | 12 +-
src/box/vy_cache.c | 65 +-
src/box/vy_cache.h | 10 +-
src/box/vy_history.c | 8 +-
src/box/vy_history.h | 8 +-
src/box/vy_log.c | 224 +++---
src/box/vy_log.h | 52 +-
src/box/vy_lsm.c | 184 ++---
src/box/vy_lsm.h | 18 +-
src/box/vy_mem.c | 91 ++-
src/box/vy_mem.h | 13 +-
src/box/vy_point_lookup.c | 41 +-
src/box/vy_point_lookup.h | 4 +-
src/box/vy_quota.c | 13 +-
src/box/vy_quota.h | 7 +-
src/box/vy_range.c | 22 +-
src/box/vy_range.h | 4 +-
src/box/vy_read_iterator.c | 136 ++--
src/box/vy_read_set.c | 11 +-
src/box/vy_regulator.c | 40 +-
src/box/vy_regulator.h | 7 +-
src/box/vy_run.c | 375 +++++-----
src/box/vy_run.h | 48 +-
src/box/vy_scheduler.c | 182 ++---
src/box/vy_scheduler.h | 6 +-
src/box/vy_stmt.c | 119 ++--
src/box/vy_stmt.h | 84 +--
src/box/vy_stmt_stream.h | 10 +-
src/box/vy_tx.c | 128 ++--
src/box/vy_tx.h | 21 +-
src/box/vy_upsert.c | 39 +-
src/box/vy_upsert.h | 4 +-
src/box/vy_write_iterator.c | 74 +-
src/box/vy_write_iterator.h | 11 +-
src/box/wal.c | 185 +++--
src/box/wal.h | 8 +-
src/box/xlog.c | 289 ++++----
src/box/xlog.h | 36 +-
src/box/xrow.c | 361 +++++-----
src/box/xrow.h | 44 +-
src/box/xrow_io.cc | 16 +-
src/box/xrow_io.h | 1 -
src/box/xrow_update.c | 65 +-
src/box/xrow_update.h | 12 +-
src/box/xrow_update_array.c | 89 +--
src/box/xrow_update_bar.c | 68 +-
src/box/xrow_update_field.c | 104 +--
src/box/xrow_update_field.h | 132 ++--
src/box/xrow_update_map.c | 84 +--
src/box/xrow_update_route.c | 45 +-
183 files changed, 6330 insertions(+), 6493 deletions(-)
diff --git a/src/box/alter.cc b/src/box/alter.cc
index 08957f6..3f2ad9b 100644
--- a/src/box/alter.cc
+++ b/src/box/alter.cc
@@ -45,12 +45,12 @@
#include "fiber.h" /* for gc_pool */
#include "scoped_guard.h"
#include "third_party/base64.h"
-#include <new> /* for placement new */
+#include <new> /* for placement new */
#include <stdio.h> /* snprintf() */
#include <ctype.h>
#include "replication.h" /* for replica_set_id() */
-#include "session.h" /* to fetch the current user. */
-#include "vclock.h" /* VCLOCK_MAX */
+#include "session.h" /* to fetch the current user. */
+#include "vclock.h" /* VCLOCK_MAX */
#include "xrow.h"
#include "iproto_constants.h"
#include "identifier.h"
@@ -68,8 +68,8 @@ access_check_ddl(const char *name, uint32_t object_id, uint32_t owner_uid,
struct credentials *cr = effective_user();
user_access_t has_access = cr->universal_access;
- user_access_t access = ((PRIV_U | (user_access_t) priv_type) &
- ~has_access);
+ user_access_t access =
+ ((PRIV_U | (user_access_t)priv_type) & ~has_access);
bool is_owner = owner_uid == cr->uid || cr->uid == ADMIN;
if (access == 0)
return 0; /* Access granted. */
@@ -135,24 +135,25 @@ index_def_check_sequence(struct index_def *index_def, uint32_t sequence_fieldno,
continue;
if ((part->path == NULL && sequence_path == NULL) ||
(part->path != NULL && sequence_path != NULL &&
- json_path_cmp(part->path, part->path_len,
- sequence_path, sequence_path_len,
- TUPLE_INDEX_BASE) == 0)) {
+ json_path_cmp(part->path, part->path_len, sequence_path,
+ sequence_path_len, TUPLE_INDEX_BASE) == 0)) {
sequence_part = part;
break;
}
}
if (sequence_part == NULL) {
diag_set(ClientError, ER_MODIFY_INDEX, index_def->name,
- space_name, "sequence field must be a part of "
- "the index");
+ space_name,
+ "sequence field must be a part of "
+ "the index");
return -1;
}
enum field_type type = sequence_part->type;
if (type != FIELD_TYPE_UNSIGNED && type != FIELD_TYPE_INTEGER) {
diag_set(ClientError, ER_MODIFY_INDEX, index_def->name,
- space_name, "sequence cannot be used with "
- "a non-integer key");
+ space_name,
+ "sequence cannot be used with "
+ "a non-integer key");
return -1;
}
return 0;
@@ -166,8 +167,8 @@ index_def_check_sequence(struct index_def *index_def, uint32_t sequence_fieldno,
static int
index_def_check_tuple(struct tuple *tuple)
{
- const mp_type common_template[] =
- {MP_UINT, MP_UINT, MP_STR, MP_STR, MP_MAP, MP_ARRAY};
+ const mp_type common_template[] = { MP_UINT, MP_UINT, MP_STR,
+ MP_STR, MP_MAP, MP_ARRAY };
const char *data = tuple_data(tuple);
uint32_t field_count = mp_decode_array(&data);
const char *field_start = data;
@@ -191,8 +192,8 @@ err:
p += snprintf(p, e - p, i ? ", %s" : "%s", mp_type_strs[type]);
}
diag_set(ClientError, ER_WRONG_INDEX_RECORD, got,
- "space id (unsigned), index id (unsigned), name (string), "\
- "type (string), options (map), parts (array)");
+ "space id (unsigned), index id (unsigned), name (string), "
+ "type (string), options (map), parts (array)");
return -1;
}
@@ -210,12 +211,13 @@ index_opts_decode(struct index_opts *opts, const char *map,
return -1;
if (opts->distance == rtree_index_distance_type_MAX) {
diag_set(ClientError, ER_WRONG_INDEX_OPTIONS,
- BOX_INDEX_FIELD_OPTS, "distance must be either "\
- "'euclid' or 'manhattan'");
+ BOX_INDEX_FIELD_OPTS,
+ "distance must be either "
+ "'euclid' or 'manhattan'");
return -1;
}
- if (opts->page_size <= 0 || (opts->range_size > 0 &&
- opts->page_size > opts->range_size)) {
+ if (opts->page_size <= 0 ||
+ (opts->range_size > 0 && opts->page_size > opts->range_size)) {
diag_set(ClientError, ER_WRONG_INDEX_OPTIONS,
BOX_INDEX_FIELD_OPTS,
"page_size must be greater than 0 and "
@@ -250,14 +252,15 @@ index_opts_decode(struct index_opts *opts, const char *map,
* functional index for now.
*/
static int
-func_index_check_func(struct func *func) {
+func_index_check_func(struct func *func)
+{
assert(func != NULL);
if (func->def->language != FUNC_LANGUAGE_LUA ||
func->def->body == NULL || !func->def->is_deterministic ||
!func->def->is_sandboxed) {
diag_set(ClientError, ER_WRONG_INDEX_OPTIONS, 0,
- "referenced function doesn't satisfy "
- "functional index function constraints");
+ "referenced function doesn't satisfy "
+ "functional index function constraints");
return -1;
}
return 0;
@@ -293,12 +296,12 @@ index_def_new_from_tuple(struct tuple *tuple, struct space *space)
return NULL;
enum index_type type = STR2ENUM(index_type, out);
uint32_t name_len;
- const char *name = tuple_field_str(tuple, BOX_INDEX_FIELD_NAME,
- &name_len);
+ const char *name =
+ tuple_field_str(tuple, BOX_INDEX_FIELD_NAME, &name_len);
if (name == NULL)
return NULL;
- const char *opts_field = tuple_field_with_type(tuple,
- BOX_INDEX_FIELD_OPTS, MP_MAP);
+ const char *opts_field =
+ tuple_field_with_type(tuple, BOX_INDEX_FIELD_OPTS, MP_MAP);
if (opts_field == NULL)
return NULL;
if (index_opts_decode(&opts, opts_field, &fiber()->gc) != 0)
@@ -307,18 +310,18 @@ index_def_new_from_tuple(struct tuple *tuple, struct space *space)
uint32_t part_count = mp_decode_array(&parts);
if (name_len > BOX_NAME_MAX) {
diag_set(ClientError, ER_MODIFY_INDEX,
- tt_cstr(name, BOX_INVALID_NAME_MAX),
- space_name(space), "index name is too long");
+ tt_cstr(name, BOX_INVALID_NAME_MAX), space_name(space),
+ "index name is too long");
return NULL;
}
if (identifier_check(name, name_len) != 0)
return NULL;
struct key_def *key_def = NULL;
- struct key_part_def *part_def = (struct key_part_def *)
- malloc(sizeof(*part_def) * part_count);
+ struct key_part_def *part_def =
+ (struct key_part_def *)malloc(sizeof(*part_def) * part_count);
if (part_def == NULL) {
- diag_set(OutOfMemory, sizeof(*part_def) * part_count,
- "malloc", "key_part_def");
+ diag_set(OutOfMemory, sizeof(*part_def) * part_count, "malloc",
+ "key_part_def");
return NULL;
}
auto key_def_guard = make_scoped_guard([&] {
@@ -327,19 +330,20 @@ index_def_new_from_tuple(struct tuple *tuple, struct space *space)
key_def_delete(key_def);
});
if (key_def_decode_parts(part_def, part_count, &parts,
- space->def->fields,
- space->def->field_count, &fiber()->gc) != 0)
+ space->def->fields, space->def->field_count,
+ &fiber()->gc) != 0)
return NULL;
bool for_func_index = opts.func_id > 0;
key_def = key_def_new(part_def, part_count, for_func_index);
if (key_def == NULL)
return NULL;
struct index_def *index_def =
- index_def_new(id, index_id, name, name_len, type,
- &opts, key_def, space_index_key_def(space, 0));
+ index_def_new(id, index_id, name, name_len, type, &opts,
+ key_def, space_index_key_def(space, 0));
if (index_def == NULL)
return NULL;
- auto index_def_guard = make_scoped_guard([=] { index_def_delete(index_def); });
+ auto index_def_guard =
+ make_scoped_guard([=] { index_def_delete(index_def); });
if (!index_def_is_valid(index_def, space_name(space)))
return NULL;
if (space_check_index_def(space, index_def) != 0)
@@ -366,11 +370,13 @@ index_def_new_from_tuple(struct tuple *tuple, struct space *space)
index_def_set_func(index_def, func);
}
if (index_def->iid == 0 && space->sequence != NULL)
- if (index_def_check_sequence(index_def, space->sequence_fieldno,
- space->sequence_path,
- space->sequence_path != NULL ?
- strlen(space->sequence_path) : 0,
- space_name(space)) != 0)
+ if (index_def_check_sequence(
+ index_def, space->sequence_fieldno,
+ space->sequence_path,
+ space->sequence_path != NULL ?
+ strlen(space->sequence_path) :
+ 0,
+ space_name(space)) != 0)
return NULL;
index_def_guard.is_active = false;
return index_def;
@@ -416,8 +422,8 @@ space_opts_decode(struct space_opts *opts, const char *map,
*/
static int
field_def_decode(struct field_def *field, const char **data,
- const char *space_name, uint32_t name_len,
- uint32_t errcode, uint32_t fieldno, struct region *region)
+ const char *space_name, uint32_t name_len, uint32_t errcode,
+ uint32_t fieldno, struct region *region)
{
if (mp_typeof(**data) != MP_MAP) {
diag_set(ClientError, errcode, tt_cstr(space_name, name_len),
@@ -433,8 +439,8 @@ field_def_decode(struct field_def *field, const char **data,
if (mp_typeof(**data) != MP_STR) {
diag_set(ClientError, errcode,
tt_cstr(space_name, name_len),
- tt_sprintf("field %d format is not map"\
- " with string keys",
+ tt_sprintf("field %d format is not map"
+ " with string keys",
fieldno + TUPLE_INDEX_BASE));
return -1;
}
@@ -445,15 +451,14 @@ field_def_decode(struct field_def *field, const char **data,
fieldno + TUPLE_INDEX_BASE, region,
true) != 0)
return -1;
- if (is_action_missing &&
- key_len == action_literal_len &&
+ if (is_action_missing && key_len == action_literal_len &&
memcmp(key, "nullable_action", action_literal_len) == 0)
is_action_missing = false;
}
if (is_action_missing) {
field->nullable_action = field->is_nullable ?
- ON_CONFLICT_ACTION_NONE
- : ON_CONFLICT_ACTION_DEFAULT;
+ ON_CONFLICT_ACTION_NONE :
+ ON_CONFLICT_ACTION_DEFAULT;
}
if (field->name == NULL) {
diag_set(ClientError, errcode, tt_cstr(space_name, name_len),
@@ -483,20 +488,18 @@ field_def_decode(struct field_def *field, const char **data,
fieldno + TUPLE_INDEX_BASE));
return -1;
}
- if (!((field->is_nullable && field->nullable_action ==
- ON_CONFLICT_ACTION_NONE)
- || (!field->is_nullable
- && field->nullable_action != ON_CONFLICT_ACTION_NONE))) {
+ if (!((field->is_nullable &&
+ field->nullable_action == ON_CONFLICT_ACTION_NONE) ||
+ (!field->is_nullable &&
+ field->nullable_action != ON_CONFLICT_ACTION_NONE))) {
diag_set(ClientError, errcode, tt_cstr(space_name, name_len),
tt_sprintf("field %d has conflicting nullability and "
- "nullable action properties", fieldno +
- TUPLE_INDEX_BASE));
+ "nullable action properties",
+ fieldno + TUPLE_INDEX_BASE));
return -1;
}
- if (field->coll_id != COLL_NONE &&
- field->type != FIELD_TYPE_STRING &&
- field->type != FIELD_TYPE_SCALAR &&
- field->type != FIELD_TYPE_ANY) {
+ if (field->coll_id != COLL_NONE && field->type != FIELD_TYPE_STRING &&
+ field->type != FIELD_TYPE_SCALAR && field->type != FIELD_TYPE_ANY) {
diag_set(ClientError, errcode, tt_cstr(space_name, name_len),
tt_sprintf("collation is reasonable only for "
"string, scalar and any fields"));
@@ -505,8 +508,8 @@ field_def_decode(struct field_def *field, const char **data,
const char *dv = field->default_value;
if (dv != NULL) {
- field->default_value_expr = sql_expr_compile(sql_get(), dv,
- strlen(dv));
+ field->default_value_expr =
+ sql_expr_compile(sql_get(), dv, strlen(dv));
if (field->default_value_expr == NULL)
return -1;
}
@@ -526,8 +529,8 @@ field_def_decode(struct field_def *field, const char **data,
*/
static int
space_format_decode(const char *data, uint32_t *out_count,
- const char *space_name, uint32_t name_len,
- uint32_t errcode, struct region *region, struct field_def **fields)
+ const char *space_name, uint32_t name_len, uint32_t errcode,
+ struct region *region, struct field_def **fields)
{
/* Type is checked by _space format. */
assert(mp_typeof(*data) == MP_ARRAY);
@@ -538,9 +541,8 @@ space_format_decode(const char *data, uint32_t *out_count,
return 0;
}
size_t size;
- struct field_def *region_defs =
- region_alloc_array(region, typeof(region_defs[0]), count,
- &size);
+ struct field_def *region_defs = region_alloc_array(
+ region, typeof(region_defs[0]), count, &size);
if (region_defs == NULL) {
diag_set(OutOfMemory, size, "region_alloc_array",
"region_defs");
@@ -552,12 +554,11 @@ space_format_decode(const char *data, uint32_t *out_count,
* work with garbage pointers.
*/
memset(region_defs, 0, size);
- auto fields_guard = make_scoped_guard([=] {
- space_def_destroy_fields(region_defs, count, false);
- });
+ auto fields_guard = make_scoped_guard(
+ [=] { space_def_destroy_fields(region_defs, count, false); });
for (uint32_t i = 0; i < count; ++i) {
- if (field_def_decode(®ion_defs[i], &data, space_name, name_len,
- errcode, i, region) != 0)
+ if (field_def_decode(®ion_defs[i], &data, space_name,
+ name_len, errcode, i, region) != 0)
return -1;
}
fields_guard.is_active = false;
@@ -573,8 +574,8 @@ space_def_new_from_tuple(struct tuple *tuple, uint32_t errcode,
struct region *region)
{
uint32_t name_len;
- const char *name = tuple_field_str(tuple, BOX_SPACE_FIELD_NAME,
- &name_len);
+ const char *name =
+ tuple_field_str(tuple, BOX_SPACE_FIELD_NAME, &name_len);
if (name == NULL)
return NULL;
if (name_len > BOX_NAME_MAX) {
@@ -606,8 +607,8 @@ space_def_new_from_tuple(struct tuple *tuple, uint32_t errcode,
&exact_field_count) != 0)
return NULL;
uint32_t engine_name_len;
- const char *engine_name = tuple_field_str(tuple,
- BOX_SPACE_FIELD_ENGINE, &engine_name_len);
+ const char *engine_name = tuple_field_str(tuple, BOX_SPACE_FIELD_ENGINE,
+ &engine_name_len);
if (engine_name == NULL)
return NULL;
/*
@@ -622,28 +623,26 @@ space_def_new_from_tuple(struct tuple *tuple, uint32_t errcode,
if (identifier_check(engine_name, engine_name_len) != 0)
return NULL;
/* Check space opts. */
- const char *space_opts = tuple_field_with_type(tuple,
- BOX_SPACE_FIELD_OPTS, MP_MAP);
+ const char *space_opts =
+ tuple_field_with_type(tuple, BOX_SPACE_FIELD_OPTS, MP_MAP);
if (space_opts == NULL)
return NULL;
/* Check space format */
- const char *format = tuple_field_with_type(tuple,
- BOX_SPACE_FIELD_FORMAT, MP_ARRAY);
+ const char *format =
+ tuple_field_with_type(tuple, BOX_SPACE_FIELD_FORMAT, MP_ARRAY);
if (format == NULL)
return NULL;
struct field_def *fields = NULL;
uint32_t field_count;
- if (space_format_decode(format, &field_count, name,
- name_len, errcode, region, &fields) != 0)
+ if (space_format_decode(format, &field_count, name, name_len, errcode,
+ region, &fields) != 0)
return NULL;
- auto fields_guard = make_scoped_guard([=] {
- space_def_destroy_fields(fields, field_count, false);
- });
- if (exact_field_count != 0 &&
- exact_field_count < field_count) {
+ auto fields_guard = make_scoped_guard(
+ [=] { space_def_destroy_fields(fields, field_count, false); });
+ if (exact_field_count != 0 && exact_field_count < field_count) {
diag_set(ClientError, errcode, tt_cstr(name, name_len),
- "exact_field_count must be either 0 or >= "\
- "formatted field count");
+ "exact_field_count must be either 0 or >= "
+ "formatted field count");
return NULL;
}
struct space_opts opts;
@@ -653,10 +652,8 @@ space_def_new_from_tuple(struct tuple *tuple, uint32_t errcode,
* Currently, only predefined replication groups
* are supported.
*/
- if (opts.group_id != GROUP_DEFAULT &&
- opts.group_id != GROUP_LOCAL) {
- diag_set(ClientError, ER_NO_SUCH_GROUP,
- int2str(opts.group_id));
+ if (opts.group_id != GROUP_DEFAULT && opts.group_id != GROUP_LOCAL) {
+ diag_set(ClientError, ER_NO_SUCH_GROUP, int2str(opts.group_id));
return NULL;
}
if (opts.is_view && opts.sql == NULL) {
@@ -668,10 +665,10 @@ space_def_new_from_tuple(struct tuple *tuple, uint32_t errcode,
"local space can't be synchronous");
return NULL;
}
- struct space_def *def =
- space_def_new(id, uid, exact_field_count, name, name_len,
- engine_name, engine_name_len, &opts, fields,
- field_count);
+ struct space_def *def = space_def_new(id, uid, exact_field_count, name,
+ name_len, engine_name,
+ engine_name_len, &opts, fields,
+ field_count);
if (def == NULL)
return NULL;
auto def_guard = make_scoped_guard([=] { space_def_delete(def); });
@@ -736,8 +733,8 @@ space_has_data(uint32_t id, uint32_t iid, uint32_t uid, bool *out)
}
if (!space_is_memtx(space)) {
- diag_set(ClientError, ER_UNSUPPORTED,
- space->engine->name, "system data");
+ diag_set(ClientError, ER_UNSUPPORTED, space->engine->name,
+ "system data");
return -1;
}
struct index *index = index_find(space, iid);
@@ -794,7 +791,8 @@ public:
* to WAL. Must not fail.
*/
virtual void commit(struct alter_space * /* alter */,
- int64_t /* signature */) {}
+ int64_t /* signature */)
+ {}
/**
* Called in case a WAL error occurred. It is supposed to undo
* the effect of AlterSpaceOp::prepare and AlterSpaceOp::alter.
@@ -820,9 +818,8 @@ static struct trigger *
txn_alter_trigger_new(trigger_f run, void *data)
{
size_t size = sizeof(struct trigger);
- struct trigger *trigger = (struct trigger *)
- region_aligned_alloc(&in_txn()->region, size,
- alignof(struct trigger));
+ struct trigger *trigger = (struct trigger *)region_aligned_alloc(
+ &in_txn()->region, size, alignof(struct trigger));
if (trigger == NULL) {
diag_set(OutOfMemory, size, "region", "struct trigger");
return NULL;
@@ -868,9 +865,8 @@ alter_space_new(struct space *old_space)
{
struct txn *txn = in_txn();
size_t size = sizeof(struct alter_space);
- struct alter_space *alter = (struct alter_space *)
- region_aligned_alloc(&in_txn()->region, size,
- alignof(struct alter_space));
+ struct alter_space *alter = (struct alter_space *)region_aligned_alloc(
+ &in_txn()->region, size, alignof(struct alter_space));
if (alter == NULL) {
diag_set(OutOfMemory, size, "region", "struct alter_space");
return NULL;
@@ -894,9 +890,9 @@ static void
alter_space_delete(struct alter_space *alter)
{
/* Destroy the ops. */
- while (! rlist_empty(&alter->ops)) {
- AlterSpaceOp *op = rlist_shift_entry(&alter->ops,
- AlterSpaceOp, link);
+ while (!rlist_empty(&alter->ops)) {
+ AlterSpaceOp *op =
+ rlist_shift_entry(&alter->ops, AlterSpaceOp, link);
delete op;
}
/* Delete the new space, if any. */
@@ -946,9 +942,11 @@ class AlterSpaceLock {
static struct mh_i32_t *registry;
/** Identifier of the space this lock is for. */
uint32_t space_id;
+
public:
/** Take a lock for the altered space. */
- AlterSpaceLock(struct alter_space *alter) {
+ AlterSpaceLock(struct alter_space *alter)
+ {
if (registry == NULL) {
registry = mh_i32_new();
if (registry == NULL) {
@@ -966,7 +964,8 @@ public:
if (k == mh_end(registry))
tnt_raise(OutOfMemory, 0, "mh_i32_put", "alter lock");
}
- ~AlterSpaceLock() {
+ ~AlterSpaceLock()
+ {
mh_int_t k = mh_i32_find(registry, space_id, NULL);
assert(k != mh_end(registry));
mh_i32_del(registry, k, NULL);
@@ -986,8 +985,8 @@ struct mh_i32_t *AlterSpaceLock::registry;
static int
alter_space_commit(struct trigger *trigger, void *event)
{
- struct txn *txn = (struct txn *) event;
- struct alter_space *alter = (struct alter_space *) trigger->data;
+ struct txn *txn = (struct txn *)event;
+ struct alter_space *alter = (struct alter_space *)trigger->data;
/*
* The engine (vinyl) expects us to pass the signature of
* the row that performed this operation, not the signature
@@ -1031,7 +1030,7 @@ alter_space_commit(struct trigger *trigger, void *event)
static int
alter_space_rollback(struct trigger *trigger, void * /* event */)
{
- struct alter_space *alter = (struct alter_space *) trigger->data;
+ struct alter_space *alter = (struct alter_space *)trigger->data;
/* Rollback alter ops */
class AlterSpaceOp *op;
try {
@@ -1185,11 +1184,9 @@ alter_space_do(struct txn_stmt *stmt, struct alter_space *alter)
* This operation does not modify the space, it just checks that
* tuples stored in it conform to the new format.
*/
-class CheckSpaceFormat: public AlterSpaceOp
-{
+class CheckSpaceFormat: public AlterSpaceOp {
public:
- CheckSpaceFormat(struct alter_space *alter)
- :AlterSpaceOp(alter) {}
+ CheckSpaceFormat(struct alter_space *alter) : AlterSpaceOp(alter) {}
virtual void prepare(struct alter_space *alter);
};
@@ -1204,16 +1201,16 @@ CheckSpaceFormat::prepare(struct alter_space *alter)
assert(new_format != NULL);
if (!tuple_format1_can_store_format2_tuples(new_format,
old_format))
- space_check_format_xc(old_space, new_format);
+ space_check_format_xc(old_space, new_format);
}
}
/** Change non-essential properties of a space. */
-class ModifySpace: public AlterSpaceOp
-{
+class ModifySpace: public AlterSpaceOp {
public:
ModifySpace(struct alter_space *alter, struct space_def *def)
- :AlterSpaceOp(alter), new_def(def), new_dict(NULL) {}
+ : AlterSpaceOp(alter), new_def(def), new_dict(NULL)
+ {}
/* New space definition. */
struct space_def *new_def;
/**
@@ -1275,11 +1272,11 @@ ModifySpace::~ModifySpace()
/** DropIndex - remove an index from space. */
-class DropIndex: public AlterSpaceOp
-{
+class DropIndex: public AlterSpaceOp {
public:
DropIndex(struct alter_space *alter, struct index *index)
- :AlterSpaceOp(alter), old_index(index) {}
+ : AlterSpaceOp(alter), old_index(index)
+ {}
struct index *old_index;
virtual void alter_def(struct alter_space *alter);
virtual void prepare(struct alter_space *alter);
@@ -1316,11 +1313,11 @@ DropIndex::commit(struct alter_space *alter, int64_t signature)
* Added to the alter specification when the index at hand
* is not affected by alter in any way.
*/
-class MoveIndex: public AlterSpaceOp
-{
+class MoveIndex: public AlterSpaceOp {
public:
MoveIndex(struct alter_space *alter, uint32_t iid_arg)
- :AlterSpaceOp(alter), iid(iid_arg) {}
+ : AlterSpaceOp(alter), iid(iid_arg)
+ {}
/** id of the index on the move. */
uint32_t iid;
virtual void alter(struct alter_space *alter);
@@ -1343,24 +1340,24 @@ MoveIndex::rollback(struct alter_space *alter)
* Change non-essential properties of an index, i.e.
* properties not involving index data or layout on disk.
*/
-class ModifyIndex: public AlterSpaceOp
-{
+class ModifyIndex: public AlterSpaceOp {
public:
- ModifyIndex(struct alter_space *alter,
- struct index *index, struct index_def *def)
- : AlterSpaceOp(alter), old_index(index),
- new_index(NULL), new_index_def(def) {
- if (new_index_def->iid == 0 &&
- key_part_cmp(new_index_def->key_def->parts,
- new_index_def->key_def->part_count,
- old_index->def->key_def->parts,
- old_index->def->key_def->part_count) != 0) {
- /*
+ ModifyIndex(struct alter_space *alter, struct index *index,
+ struct index_def *def)
+ : AlterSpaceOp(alter), old_index(index), new_index(NULL),
+ new_index_def(def)
+ {
+ if (new_index_def->iid == 0 &&
+ key_part_cmp(new_index_def->key_def->parts,
+ new_index_def->key_def->part_count,
+ old_index->def->key_def->parts,
+ old_index->def->key_def->part_count) != 0) {
+ /*
* Primary parts have been changed -
* update secondary indexes.
*/
- alter->pk_def = new_index_def->key_def;
- }
+ alter->pk_def = new_index_def->key_def;
+ }
}
struct index *old_index;
struct index *new_index;
@@ -1422,15 +1419,15 @@ ModifyIndex::~ModifyIndex()
}
/** CreateIndex - add a new index to the space. */
-class CreateIndex: public AlterSpaceOp
-{
+class CreateIndex: public AlterSpaceOp {
/** New index. */
struct index *new_index;
/** New index index_def. */
struct index_def *new_index_def;
+
public:
CreateIndex(struct alter_space *alter, struct index_def *def)
- :AlterSpaceOp(alter), new_index(NULL), new_index_def(def)
+ : AlterSpaceOp(alter), new_index(NULL), new_index_def(def)
{}
virtual void alter_def(struct alter_space *alter);
virtual void prepare(struct alter_space *alter);
@@ -1482,7 +1479,7 @@ CreateIndex::prepare(struct alter_space *alter)
void
CreateIndex::commit(struct alter_space *alter, int64_t signature)
{
- (void) alter;
+ (void)alter;
assert(new_index != NULL);
index_commit_create(new_index, signature);
new_index = NULL;
@@ -1501,15 +1498,14 @@ CreateIndex::~CreateIndex()
* from by reading the primary key. Used when key_def of
* an index is changed.
*/
-class RebuildIndex: public AlterSpaceOp
-{
+class RebuildIndex: public AlterSpaceOp {
public:
RebuildIndex(struct alter_space *alter,
struct index_def *new_index_def_arg,
struct index_def *old_index_def_arg)
- :AlterSpaceOp(alter), new_index(NULL),
- new_index_def(new_index_def_arg),
- old_index_def(old_index_def_arg)
+ : AlterSpaceOp(alter), new_index(NULL),
+ new_index_def(new_index_def_arg),
+ old_index_def(old_index_def_arg)
{
/* We may want to rebuild secondary keys as well. */
if (new_index_def->iid == 0)
@@ -1547,8 +1543,8 @@ RebuildIndex::prepare(struct alter_space *alter)
void
RebuildIndex::commit(struct alter_space *alter, int64_t signature)
{
- struct index *old_index = space_index(alter->old_space,
- old_index_def->iid);
+ struct index *old_index =
+ space_index(alter->old_space, old_index_def->iid);
assert(old_index != NULL);
index_commit_drop(old_index, signature);
assert(new_index != NULL);
@@ -1569,25 +1565,26 @@ RebuildIndex::~RebuildIndex()
* drop the old index data and rebuild index from by reading the
* primary key.
*/
-class RebuildFuncIndex: public RebuildIndex
-{
- struct index_def *
- func_index_def_new(struct index_def *index_def, struct func *func)
+class RebuildFuncIndex: public RebuildIndex {
+ struct index_def *func_index_def_new(struct index_def *index_def,
+ struct func *func)
{
struct index_def *new_index_def = index_def_dup_xc(index_def);
index_def_set_func(new_index_def, func);
return new_index_def;
}
+
public:
RebuildFuncIndex(struct alter_space *alter,
- struct index_def *old_index_def_arg, struct func *func) :
- RebuildIndex(alter, func_index_def_new(old_index_def_arg, func),
- old_index_def_arg) {}
+ struct index_def *old_index_def_arg, struct func *func)
+ : RebuildIndex(alter,
+ func_index_def_new(old_index_def_arg, func),
+ old_index_def_arg)
+ {}
};
/** TruncateIndex - truncate an index. */
-class TruncateIndex: public AlterSpaceOp
-{
+class TruncateIndex: public AlterSpaceOp {
/** id of the index to truncate. */
uint32_t iid;
/**
@@ -1596,10 +1593,12 @@ class TruncateIndex: public AlterSpaceOp
*/
struct index *old_index;
struct index *new_index;
+
public:
TruncateIndex(struct alter_space *alter, uint32_t iid)
- : AlterSpaceOp(alter), iid(iid),
- old_index(NULL), new_index(NULL) {}
+ : AlterSpaceOp(alter), iid(iid), old_index(NULL),
+ new_index(NULL)
+ {}
virtual void prepare(struct alter_space *alter);
virtual void commit(struct alter_space *alter, int64_t signature);
virtual ~TruncateIndex();
@@ -1652,19 +1651,17 @@ TruncateIndex::~TruncateIndex()
* in alter_space_do(), i.e. when creating or dropping
* an index, altering a space.
*/
-class UpdateSchemaVersion: public AlterSpaceOp
-{
+class UpdateSchemaVersion: public AlterSpaceOp {
public:
- UpdateSchemaVersion(struct alter_space * alter)
- :AlterSpaceOp(alter) {}
+ UpdateSchemaVersion(struct alter_space *alter) : AlterSpaceOp(alter) {}
virtual void alter(struct alter_space *alter);
};
void
UpdateSchemaVersion::alter(struct alter_space *alter)
{
- (void)alter;
- ++schema_version;
+ (void)alter;
+ ++schema_version;
}
/**
@@ -1676,13 +1673,15 @@ UpdateSchemaVersion::alter(struct alter_space *alter)
* Finally in ::alter or ::rollback methods we only swap those
* lists securely.
*/
-class RebuildCkConstraints: public AlterSpaceOp
-{
+class RebuildCkConstraints: public AlterSpaceOp {
void space_swap_ck_constraint(struct space *old_space,
struct space *new_space);
+
public:
- RebuildCkConstraints(struct alter_space *alter) : AlterSpaceOp(alter),
- ck_constraint(RLIST_HEAD_INITIALIZER(ck_constraint)) {}
+ RebuildCkConstraints(struct alter_space *alter)
+ : AlterSpaceOp(alter),
+ ck_constraint(RLIST_HEAD_INITIALIZER(ck_constraint))
+ {}
struct rlist ck_constraint;
virtual void prepare(struct alter_space *alter);
virtual void alter(struct alter_space *alter);
@@ -1696,9 +1695,8 @@ RebuildCkConstraints::prepare(struct alter_space *alter)
struct ck_constraint *old_ck_constraint;
rlist_foreach_entry(old_ck_constraint, &alter->old_space->ck_constraint,
link) {
- struct ck_constraint *new_ck_constraint =
- ck_constraint_new(old_ck_constraint->def,
- alter->new_space->def);
+ struct ck_constraint *new_ck_constraint = ck_constraint_new(
+ old_ck_constraint->def, alter->new_space->def);
if (new_ck_constraint == NULL)
diag_raise();
rlist_add_entry(&ck_constraint, new_ck_constraint, link);
@@ -1748,10 +1746,10 @@ RebuildCkConstraints::~RebuildCkConstraints()
* ck constraints rebuild. This may be used in scenarios where
* space format doesn't change i.e. on index alter or space trim.
*/
-class MoveCkConstraints: public AlterSpaceOp
-{
+class MoveCkConstraints: public AlterSpaceOp {
void space_swap_ck_constraint(struct space *old_space,
struct space *new_space);
+
public:
MoveCkConstraints(struct alter_space *alter) : AlterSpaceOp(alter) {}
virtual void alter(struct alter_space *alter);
@@ -1762,8 +1760,7 @@ void
MoveCkConstraints::space_swap_ck_constraint(struct space *old_space,
struct space *new_space)
{
- rlist_swap(&new_space->ck_constraint,
- &old_space->ck_constraint);
+ rlist_swap(&new_space->ck_constraint, &old_space->ck_constraint);
SWAP(new_space->ck_constraint_trigger,
old_space->ck_constraint_trigger);
}
@@ -1822,13 +1819,13 @@ space_delete_constraint_id(struct space *space, const char *name)
}
/** CreateConstraintID - add a new constraint id to a space. */
-class CreateConstraintID: public AlterSpaceOp
-{
+class CreateConstraintID: public AlterSpaceOp {
struct constraint_id *new_id;
+
public:
CreateConstraintID(struct alter_space *alter, enum constraint_type type,
const char *name)
- :AlterSpaceOp(alter), new_id(NULL)
+ : AlterSpaceOp(alter), new_id(NULL)
{
new_id = constraint_id_new(type, name);
if (new_id == NULL)
@@ -1867,8 +1864,8 @@ CreateConstraintID::rollback(struct alter_space *alter)
void
CreateConstraintID::commit(struct alter_space *alter, int64_t signature)
{
- (void) alter;
- (void) signature;
+ (void)alter;
+ (void)signature;
/*
* Constraint id is added to the space, and should not be
* deleted from now on.
@@ -1883,16 +1880,16 @@ CreateConstraintID::~CreateConstraintID()
}
/** DropConstraintID - drop a constraint id from the space. */
-class DropConstraintID: public AlterSpaceOp
-{
+class DropConstraintID: public AlterSpaceOp {
struct constraint_id *old_id;
const char *name;
+
public:
DropConstraintID(struct alter_space *alter, const char *name)
- :AlterSpaceOp(alter), old_id(NULL), name(name)
+ : AlterSpaceOp(alter), old_id(NULL), name(name)
{}
virtual void alter(struct alter_space *alter);
- virtual void commit(struct alter_space *alter , int64_t signature);
+ virtual void commit(struct alter_space *alter, int64_t signature);
virtual void rollback(struct alter_space *alter);
};
@@ -1905,8 +1902,8 @@ DropConstraintID::alter(struct alter_space *alter)
void
DropConstraintID::commit(struct alter_space *alter, int64_t signature)
{
- (void) alter;
- (void) signature;
+ (void)alter;
+ (void)signature;
constraint_id_delete(old_id);
}
@@ -1927,7 +1924,7 @@ DropConstraintID::rollback(struct alter_space *alter)
static int
on_drop_space_commit(struct trigger *trigger, void *event)
{
- (void) event;
+ (void)event;
struct space *space = (struct space *)trigger->data;
space_delete(space);
return 0;
@@ -1941,7 +1938,7 @@ on_drop_space_commit(struct trigger *trigger, void *event)
static int
on_drop_space_rollback(struct trigger *trigger, void *event)
{
- (void) event;
+ (void)event;
struct space *space = (struct space *)trigger->data;
space_cache_replace(NULL, space);
return 0;
@@ -1957,7 +1954,7 @@ on_drop_space_rollback(struct trigger *trigger, void *event)
static int
on_create_space_rollback(struct trigger *trigger, void *event)
{
- (void) event;
+ (void)event;
struct space *space = (struct space *)trigger->data;
space_cache_replace(space, NULL);
space_delete(space);
@@ -1994,13 +1991,15 @@ alter_space_move_indexes(struct alter_space *alter, uint32_t begin,
index_def_update_optionality(new_def,
min_field_count);
try {
- (void) new ModifyIndex(alter, old_index, new_def);
+ (void)new ModifyIndex(alter, old_index,
+ new_def);
} catch (Exception *e) {
return -1;
}
} else {
try {
- (void) new MoveIndex(alter, old_def->iid);
+ (void)new MoveIndex(alter,
+ old_def->iid);
} catch (Exception *e) {
return -1;
}
@@ -2016,16 +2015,18 @@ alter_space_move_indexes(struct alter_space *alter, uint32_t begin,
old_def->type, &old_def->opts,
old_def->key_def, alter->pk_def);
index_def_update_optionality(new_def, min_field_count);
- auto guard = make_scoped_guard([=] { index_def_delete(new_def); });
+ auto guard =
+ make_scoped_guard([=] { index_def_delete(new_def); });
if (!index_def_change_requires_rebuild(old_index, new_def))
try {
- (void) new ModifyIndex(alter, old_index, new_def);
+ (void)new ModifyIndex(alter, old_index,
+ new_def);
} catch (Exception *e) {
return -1;
}
else
try {
- (void) new RebuildIndex(alter, new_def, old_def);
+ (void)new RebuildIndex(alter, new_def, old_def);
} catch (Exception *e) {
return -1;
}
@@ -2071,7 +2072,7 @@ update_view_references(struct Select *select, int update_value,
continue;
struct space *space = space_by_name(space_name);
if (space == NULL) {
- if (! suppress_error) {
+ if (!suppress_error) {
assert(not_found_space != NULL);
*not_found_space = tt_sprintf("%s", space_name);
sqlSrcListDelete(sql_get(), list);
@@ -2093,7 +2094,7 @@ update_view_references(struct Select *select, int update_value,
static int
on_create_view_commit(struct trigger *trigger, void *event)
{
- (void) event;
+ (void)event;
struct Select *select = (struct Select *)trigger->data;
sql_select_delete(sql_get(), select);
return 0;
@@ -2107,7 +2108,7 @@ on_create_view_commit(struct trigger *trigger, void *event)
static int
on_create_view_rollback(struct trigger *trigger, void *event)
{
- (void) event;
+ (void)event;
struct Select *select = (struct Select *)trigger->data;
update_view_references(select, -1, true, NULL);
sql_select_delete(sql_get(), select);
@@ -2122,7 +2123,7 @@ on_create_view_rollback(struct trigger *trigger, void *event)
static int
on_drop_view_commit(struct trigger *trigger, void *event)
{
- (void) event;
+ (void)event;
struct Select *select = (struct Select *)trigger->data;
sql_select_delete(sql_get(), select);
return 0;
@@ -2136,7 +2137,7 @@ on_drop_view_commit(struct trigger *trigger, void *event)
static int
on_drop_view_rollback(struct trigger *trigger, void *event)
{
- (void) event;
+ (void)event;
struct Select *select = (struct Select *)trigger->data;
update_view_references(select, 1, true, NULL);
sql_select_delete(sql_get(), select);
@@ -2196,7 +2197,7 @@ on_drop_view_rollback(struct trigger *trigger, void *event)
static int
on_replace_dd_space(struct trigger * /* trigger */, void *event)
{
- struct txn *txn = (struct txn *) event;
+ struct txn *txn = (struct txn *)event;
struct txn_stmt *stmt = txn_current_stmt(txn);
struct tuple *old_tuple = stmt->old_tuple;
struct tuple *new_tuple = stmt->new_tuple;
@@ -2221,15 +2222,14 @@ on_replace_dd_space(struct trigger * /* trigger */, void *event)
return -1;
struct space *old_space = space_by_id(old_id);
if (new_tuple != NULL && old_space == NULL) { /* INSERT */
- struct space_def *def =
- space_def_new_from_tuple(new_tuple, ER_CREATE_SPACE,
- region);
+ struct space_def *def = space_def_new_from_tuple(
+ new_tuple, ER_CREATE_SPACE, region);
if (def == NULL)
return -1;
auto def_guard =
make_scoped_guard([=] { space_def_delete(def); });
if (access_check_ddl(def->name, def->id, def->uid, SC_SPACE,
- PRIV_C) != 0)
+ PRIV_C) != 0)
return -1;
RLIST_HEAD(empty_list);
struct space *space = space_new(def, &empty_list);
@@ -2262,13 +2262,12 @@ on_replace_dd_space(struct trigger * /* trigger */, void *event)
return -1;
txn_stmt_on_rollback(stmt, on_rollback);
if (def->opts.is_view) {
- struct Select *select = sql_view_compile(sql_get(),
- def->opts.sql);
+ struct Select *select =
+ sql_view_compile(sql_get(), def->opts.sql);
if (select == NULL)
return -1;
- auto select_guard = make_scoped_guard([=] {
- sql_select_delete(sql_get(), select);
- });
+ auto select_guard = make_scoped_guard(
+ [=] { sql_select_delete(sql_get(), select); });
const char *disappeared_space;
if (update_view_references(select, 1, false,
&disappeared_space) != 0) {
@@ -2279,12 +2278,11 @@ on_replace_dd_space(struct trigger * /* trigger */, void *event)
update_view_references(select, -1, false,
&disappeared_space);
diag_set(ClientError, ER_NO_SUCH_SPACE,
- disappeared_space);
+ disappeared_space);
return -1;
}
- struct trigger *on_commit_view =
- txn_alter_trigger_new(on_create_view_commit,
- select);
+ struct trigger *on_commit_view = txn_alter_trigger_new(
+ on_create_view_commit, select);
if (on_commit_view == NULL)
return -1;
txn_stmt_on_commit(stmt, on_commit_view);
@@ -2298,37 +2296,39 @@ on_replace_dd_space(struct trigger * /* trigger */, void *event)
}
} else if (new_tuple == NULL) { /* DELETE */
if (access_check_ddl(old_space->def->name, old_space->def->id,
- old_space->def->uid, SC_SPACE, PRIV_D) != 0)
+ old_space->def->uid, SC_SPACE,
+ PRIV_D) != 0)
return -1;
/* Verify that the space is empty (has no indexes) */
if (old_space->index_count) {
diag_set(ClientError, ER_DROP_SPACE,
- space_name(old_space),
- "the space has indexes");
+ space_name(old_space),
+ "the space has indexes");
return -1;
}
bool out;
- if (schema_find_grants("space", old_space->def->id, &out) != 0) {
+ if (schema_find_grants("space", old_space->def->id, &out) !=
+ 0) {
return -1;
}
if (out) {
diag_set(ClientError, ER_DROP_SPACE,
- space_name(old_space),
- "the space has grants");
+ space_name(old_space), "the space has grants");
return -1;
}
- if (space_has_data(BOX_TRUNCATE_ID, 0, old_space->def->id, &out) != 0)
+ if (space_has_data(BOX_TRUNCATE_ID, 0, old_space->def->id,
+ &out) != 0)
return -1;
if (out) {
diag_set(ClientError, ER_DROP_SPACE,
- space_name(old_space),
- "the space has truncate record");
+ space_name(old_space),
+ "the space has truncate record");
return -1;
}
if (old_space->def->view_ref_count > 0) {
diag_set(ClientError, ER_DROP_SPACE,
- space_name(old_space),
- "other views depend on this space");
+ space_name(old_space),
+ "other views depend on this space");
return -1;
}
/*
@@ -2340,14 +2340,14 @@ on_replace_dd_space(struct trigger * /* trigger */, void *event)
*/
if (!rlist_empty(&old_space->child_fk_constraint)) {
diag_set(ClientError, ER_DROP_SPACE,
- space_name(old_space),
- "the space has foreign key constraints");
+ space_name(old_space),
+ "the space has foreign key constraints");
return -1;
}
if (!rlist_empty(&old_space->ck_constraint)) {
diag_set(ClientError, ER_DROP_SPACE,
- space_name(old_space),
- "the space has check constraints");
+ space_name(old_space),
+ "the space has check constraints");
return -1;
}
/**
@@ -2367,23 +2367,20 @@ on_replace_dd_space(struct trigger * /* trigger */, void *event)
if (on_commit == NULL)
return -1;
txn_stmt_on_commit(stmt, on_commit);
- struct trigger *on_rollback =
- txn_alter_trigger_new(on_drop_space_rollback, old_space);
+ struct trigger *on_rollback = txn_alter_trigger_new(
+ on_drop_space_rollback, old_space);
if (on_rollback == NULL)
return -1;
txn_stmt_on_rollback(stmt, on_rollback);
if (old_space->def->opts.is_view) {
- struct Select *select =
- sql_view_compile(sql_get(),
- old_space->def->opts.sql);
+ struct Select *select = sql_view_compile(
+ sql_get(), old_space->def->opts.sql);
if (select == NULL)
return -1;
- auto select_guard = make_scoped_guard([=] {
- sql_select_delete(sql_get(), select);
- });
- struct trigger *on_commit_view =
- txn_alter_trigger_new(on_drop_view_commit,
- select);
+ auto select_guard = make_scoped_guard(
+ [=] { sql_select_delete(sql_get(), select); });
+ struct trigger *on_commit_view = txn_alter_trigger_new(
+ on_drop_view_commit, select);
if (on_commit_view == NULL)
return -1;
txn_stmt_on_commit(stmt, on_commit_view);
@@ -2404,47 +2401,47 @@ on_replace_dd_space(struct trigger * /* trigger */, void *event)
"view can not be altered");
return -1;
}
- struct space_def *def =
- space_def_new_from_tuple(new_tuple, ER_ALTER_SPACE,
- region);
+ struct space_def *def = space_def_new_from_tuple(
+ new_tuple, ER_ALTER_SPACE, region);
if (def == NULL)
return -1;
auto def_guard =
make_scoped_guard([=] { space_def_delete(def); });
if (access_check_ddl(def->name, def->id, def->uid, SC_SPACE,
- PRIV_A) != 0)
+ PRIV_A) != 0)
return -1;
if (def->id != space_id(old_space)) {
diag_set(ClientError, ER_ALTER_SPACE,
- space_name(old_space),
- "space id is immutable");
+ space_name(old_space),
+ "space id is immutable");
return -1;
}
- if (strcmp(def->engine_name, old_space->def->engine_name) != 0) {
+ if (strcmp(def->engine_name, old_space->def->engine_name) !=
+ 0) {
diag_set(ClientError, ER_ALTER_SPACE,
- space_name(old_space),
- "can not change space engine");
+ space_name(old_space),
+ "can not change space engine");
return -1;
}
if (def->opts.group_id != space_group_id(old_space)) {
diag_set(ClientError, ER_ALTER_SPACE,
- space_name(old_space),
- "replication group is immutable");
+ space_name(old_space),
+ "replication group is immutable");
return -1;
}
if (def->opts.is_view != old_space->def->opts.is_view) {
diag_set(ClientError, ER_ALTER_SPACE,
- space_name(old_space),
- "can not convert a space to "
- "a view and vice versa");
+ space_name(old_space),
+ "can not convert a space to "
+ "a view and vice versa");
return -1;
}
if (strcmp(def->name, old_space->def->name) != 0 &&
old_space->def->view_ref_count > 0) {
diag_set(ClientError, ER_ALTER_SPACE,
- space_name(old_space),
- "can not rename space which is referenced by "
- "view");
+ space_name(old_space),
+ "can not rename space which is referenced by "
+ "view");
return -1;
}
/*
@@ -2455,7 +2452,7 @@ on_replace_dd_space(struct trigger * /* trigger */, void *event)
if (alter == NULL)
return -1;
auto alter_guard =
- make_scoped_guard([=] {alter_space_delete(alter);});
+ make_scoped_guard([=] { alter_space_delete(alter); });
/*
* Calculate a new min_field_count. It can be
* changed by resetting space:format(), if an old
@@ -2475,26 +2472,24 @@ on_replace_dd_space(struct trigger * /* trigger */, void *event)
}
for (uint32_t i = 0; i < old_space->index_count; ++i)
keys[i] = old_space->index[i]->def->key_def;
- alter->new_min_field_count =
- tuple_format_min_field_count(keys,
- old_space->index_count,
- def->fields,
- def->field_count);
+ alter->new_min_field_count = tuple_format_min_field_count(
+ keys, old_space->index_count, def->fields,
+ def->field_count);
try {
- (void) new CheckSpaceFormat(alter);
- (void) new ModifySpace(alter, def);
- (void) new RebuildCkConstraints(alter);
+ (void)new CheckSpaceFormat(alter);
+ (void)new ModifySpace(alter, def);
+ (void)new RebuildCkConstraints(alter);
} catch (Exception *e) {
return -1;
}
def_guard.is_active = false;
/* Create MoveIndex ops for all space indexes. */
if (alter_space_move_indexes(alter, 0,
- old_space->index_id_max + 1) != 0)
+ old_space->index_id_max + 1) != 0)
return -1;
try {
/* Remember to update schema_version. */
- (void) new UpdateSchemaVersion(alter);
+ (void)new UpdateSchemaVersion(alter);
alter_space_do(stmt, alter);
} catch (Exception *e) {
return -1;
@@ -2565,7 +2560,7 @@ index_is_used_by_fk_constraint(struct rlist *fk_list, uint32_t iid)
static int
on_replace_dd_index(struct trigger * /* trigger */, void *event)
{
- struct txn *txn = (struct txn *) event;
+ struct txn *txn = (struct txn *)event;
struct txn_stmt *stmt = txn_current_stmt(txn);
struct tuple *old_tuple = stmt->old_tuple;
struct tuple *new_tuple = stmt->new_tuple;
@@ -2581,14 +2576,14 @@ on_replace_dd_index(struct trigger * /* trigger */, void *event)
return -1;
if (old_space->def->opts.is_view) {
diag_set(ClientError, ER_ALTER_SPACE, space_name(old_space),
- "can not add index on a view");
+ "can not add index on a view");
return -1;
}
enum priv_type priv_type = new_tuple ? PRIV_C : PRIV_D;
if (old_tuple && new_tuple)
priv_type = PRIV_A;
if (access_check_ddl(old_space->def->name, old_space->def->id,
- old_space->def->uid, SC_SPACE, priv_type) != 0)
+ old_space->def->uid, SC_SPACE, priv_type) != 0)
return -1;
struct index *old_index = space_index(old_space, iid);
struct index_def *old_def = old_index != NULL ? old_index->def : NULL;
@@ -2602,7 +2597,7 @@ on_replace_dd_index(struct trigger * /* trigger */, void *event)
*/
if (space_is_system(old_space)) {
diag_set(ClientError, ER_LAST_DROP,
- space_name(old_space));
+ space_name(old_space));
return -1;
}
/*
@@ -2610,7 +2605,7 @@ on_replace_dd_index(struct trigger * /* trigger */, void *event)
*/
if (old_space->index_count > 1) {
diag_set(ClientError, ER_DROP_PRIMARY_KEY,
- space_name(old_space));
+ space_name(old_space));
return -1;
}
/*
@@ -2618,9 +2613,9 @@ on_replace_dd_index(struct trigger * /* trigger */, void *event)
*/
if (old_space->sequence != NULL) {
diag_set(ClientError, ER_ALTER_SPACE,
- space_name(old_space),
- "can not drop primary key while "
- "space sequence exists");
+ space_name(old_space),
+ "can not drop primary key while "
+ "space sequence exists");
return -1;
}
}
@@ -2630,9 +2625,8 @@ on_replace_dd_index(struct trigger * /* trigger */, void *event)
* A secondary index can not be created without
* a primary key.
*/
- diag_set(ClientError, ER_ALTER_SPACE,
- space_name(old_space),
- "can not add a secondary key before primary");
+ diag_set(ClientError, ER_ALTER_SPACE, space_name(old_space),
+ "can not add a secondary key before primary");
return -1;
}
@@ -2655,21 +2649,21 @@ on_replace_dd_index(struct trigger * /* trigger */, void *event)
* Can't drop index if foreign key constraints
* references this index.
*/
- if (index_is_used_by_fk_constraint(&old_space->parent_fk_constraint,
- iid)) {
+ if (index_is_used_by_fk_constraint(
+ &old_space->parent_fk_constraint, iid)) {
diag_set(ClientError, ER_ALTER_SPACE,
- space_name(old_space),
- "can not drop a referenced index");
+ space_name(old_space),
+ "can not drop a referenced index");
return -1;
}
if (alter_space_move_indexes(alter, 0, iid) != 0)
return -1;
try {
if (old_index->def->opts.is_unique) {
- (void) new DropConstraintID(alter,
- old_def->name);
+ (void)new DropConstraintID(alter,
+ old_def->name);
}
- (void) new DropIndex(alter, old_index);
+ (void)new DropIndex(alter, old_index);
} catch (Exception *e) {
return -1;
}
@@ -2685,11 +2679,13 @@ on_replace_dd_index(struct trigger * /* trigger */, void *event)
index_def_update_optionality(def, alter->new_min_field_count);
try {
if (def->opts.is_unique) {
- (void) new CreateConstraintID(
- alter, iid == 0 ? CONSTRAINT_TYPE_PK :
- CONSTRAINT_TYPE_UNIQUE, def->name);
+ (void)new CreateConstraintID(
+ alter,
+ iid == 0 ? CONSTRAINT_TYPE_PK :
+ CONSTRAINT_TYPE_UNIQUE,
+ def->name);
}
- (void) new CreateIndex(alter, def);
+ (void)new CreateIndex(alter, def);
} catch (Exception *e) {
index_def_delete(def);
return -1;
@@ -2708,10 +2704,10 @@ on_replace_dd_index(struct trigger * /* trigger */, void *event)
* becoming unique (i.e. constraint), or when a
* unique index's name is under change.
*/
- bool do_new_constraint_id =
- !old_def->opts.is_unique && index_def->opts.is_unique;
- bool do_drop_constraint_id =
- old_def->opts.is_unique && !index_def->opts.is_unique;
+ bool do_new_constraint_id = !old_def->opts.is_unique &&
+ index_def->opts.is_unique;
+ bool do_drop_constraint_id = old_def->opts.is_unique &&
+ !index_def->opts.is_unique;
if (old_def->opts.is_unique && index_def->opts.is_unique &&
strcmp(index_def->name, old_def->name) != 0) {
@@ -2720,13 +2716,13 @@ on_replace_dd_index(struct trigger * /* trigger */, void *event)
}
try {
if (do_new_constraint_id) {
- (void) new CreateConstraintID(
+ (void)new CreateConstraintID(
alter, CONSTRAINT_TYPE_UNIQUE,
index_def->name);
}
if (do_drop_constraint_id) {
- (void) new DropConstraintID(alter,
- old_def->name);
+ (void)new DropConstraintID(alter,
+ old_def->name);
}
} catch (Exception *e) {
return -1;
@@ -2764,11 +2760,9 @@ on_replace_dd_index(struct trigger * /* trigger */, void *event)
keys[j++] = index_def->key_def;
}
struct space_def *def = old_space->def;
- alter->new_min_field_count =
- tuple_format_min_field_count(keys,
- old_space->index_count,
- def->fields,
- def->field_count);
+ alter->new_min_field_count = tuple_format_min_field_count(
+ keys, old_space->index_count, def->fields,
+ def->field_count);
index_def_update_optionality(index_def,
alter->new_min_field_count);
if (alter_space_move_indexes(alter, 0, iid))
@@ -2776,26 +2770,26 @@ on_replace_dd_index(struct trigger * /* trigger */, void *event)
if (index_def_cmp(index_def, old_index->def) == 0) {
/* Index is not changed so just move it. */
try {
- (void) new MoveIndex(alter, old_index->def->iid);
+ (void)new MoveIndex(alter, old_index->def->iid);
} catch (Exception *e) {
return -1;
}
} else if (index_def_change_requires_rebuild(old_index,
index_def)) {
- if (index_is_used_by_fk_constraint(&old_space->parent_fk_constraint,
- iid)) {
+ if (index_is_used_by_fk_constraint(
+ &old_space->parent_fk_constraint, iid)) {
diag_set(ClientError, ER_ALTER_SPACE,
- space_name(old_space),
- "can not alter a referenced index");
+ space_name(old_space),
+ "can not alter a referenced index");
return -1;
}
/*
* Operation demands an index rebuild.
*/
try {
- (void) new RebuildIndex(alter, index_def,
- old_index->def);
+ (void)new RebuildIndex(alter, index_def,
+ old_index->def);
} catch (Exception *e) {
return -1;
}
@@ -2807,8 +2801,9 @@ on_replace_dd_index(struct trigger * /* trigger */, void *event)
* in the space conform to the new format.
*/
try {
- (void) new CheckSpaceFormat(alter);
- (void) new ModifyIndex(alter, old_index, index_def);
+ (void)new CheckSpaceFormat(alter);
+ (void)new ModifyIndex(alter, old_index,
+ index_def);
} catch (Exception *e) {
return -1;
}
@@ -2819,12 +2814,13 @@ on_replace_dd_index(struct trigger * /* trigger */, void *event)
* Create MoveIndex ops for the remaining indexes in the
* old space.
*/
- if (alter_space_move_indexes(alter, iid + 1, old_space->index_id_max + 1) != 0)
+ if (alter_space_move_indexes(alter, iid + 1,
+ old_space->index_id_max + 1) != 0)
return -1;
try {
- (void) new MoveCkConstraints(alter);
+ (void)new MoveCkConstraints(alter);
/* Add an op to update schema_version on commit. */
- (void) new UpdateSchemaVersion(alter);
+ (void)new UpdateSchemaVersion(alter);
alter_space_do(stmt, alter);
} catch (Exception *e) {
return -1;
@@ -2847,7 +2843,7 @@ on_replace_dd_index(struct trigger * /* trigger */, void *event)
static int
on_replace_dd_truncate(struct trigger * /* trigger */, void *event)
{
- struct txn *txn = (struct txn *) event;
+ struct txn *txn = (struct txn *)event;
struct txn_stmt *stmt = txn_current_stmt(txn);
struct tuple *new_tuple = stmt->new_tuple;
@@ -2857,7 +2853,8 @@ on_replace_dd_truncate(struct trigger * /* trigger */, void *event)
}
uint32_t space_id;
- if (tuple_field_u32(new_tuple, BOX_TRUNCATE_FIELD_SPACE_ID, &space_id) != 0)
+ if (tuple_field_u32(new_tuple, BOX_TRUNCATE_FIELD_SPACE_ID,
+ &space_id) != 0)
return -1;
struct space *old_space = space_cache_find(space_id);
if (old_space == NULL)
@@ -2878,7 +2875,7 @@ on_replace_dd_truncate(struct trigger * /* trigger */, void *event)
*/
if (space_is_system(old_space)) {
diag_set(ClientError, ER_TRUNCATE_SYSTEM_SPACE,
- space_name(old_space));
+ space_name(old_space));
return -1;
}
@@ -2910,10 +2907,10 @@ on_replace_dd_truncate(struct trigger * /* trigger */, void *event)
*/
for (uint32_t i = 0; i < old_space->index_count; i++) {
struct index *old_index = old_space->index[i];
- (void) new TruncateIndex(alter, old_index->def->iid);
+ (void)new TruncateIndex(alter, old_index->def->iid);
}
- (void) new MoveCkConstraints(alter);
+ (void)new MoveCkConstraints(alter);
alter_space_do(stmt, alter);
} catch (Exception *e) {
return -1;
@@ -2935,7 +2932,7 @@ user_has_data(struct user *user, bool *has_data)
* For _priv also check that the user has no grants.
*/
uint32_t indexes[] = { 1, 1, 1, 1, 0 };
- uint32_t count = sizeof(spaces)/sizeof(*spaces);
+ uint32_t count = sizeof(spaces) / sizeof(*spaces);
bool out;
for (uint32_t i = 0; i < count; i++) {
if (space_has_data(spaces[i], indexes[i], uid, &out) != 0)
@@ -2945,7 +2942,7 @@ user_has_data(struct user *user, bool *has_data)
return 0;
}
}
- if (! user_map_is_empty(&user->users)) {
+ if (!user_map_is_empty(&user->users)) {
*has_data = true;
return 0;
}
@@ -2979,9 +2976,9 @@ user_def_fill_auth_data(struct user_def *user, const char *auth_data)
}
if (mp_typeof(*auth_data) != MP_MAP) {
/** Prevent users from making silly mistakes */
- diag_set(ClientError, ER_CREATE_USER,
- user->name, "invalid password format, "
- "use box.schema.user.passwd() to reset password");
+ diag_set(ClientError, ER_CREATE_USER, user->name,
+ "invalid password format, "
+ "use box.schema.user.passwd() to reset password");
return -1;
}
uint32_t mech_count = mp_decode_map(&auth_data);
@@ -2999,13 +2996,14 @@ user_def_fill_auth_data(struct user_def *user, const char *auth_data)
}
const char *hash2_base64 = mp_decode_str(&auth_data, &len);
if (len != 0 && len != SCRAMBLE_BASE64_SIZE) {
- diag_set(ClientError, ER_CREATE_USER,
- user->name, "invalid user password");
+ diag_set(ClientError, ER_CREATE_USER, user->name,
+ "invalid user password");
return -1;
}
if (user->uid == GUEST) {
/** Guest user is permitted to have empty password */
- if (strncmp(hash2_base64, CHAP_SHA1_EMPTY_PASSWORD, len)) {
+ if (strncmp(hash2_base64, CHAP_SHA1_EMPTY_PASSWORD,
+ len)) {
diag_set(ClientError, ER_GUEST_USER_PASSWORD);
return -1;
}
@@ -3022,19 +3020,19 @@ static struct user_def *
user_def_new_from_tuple(struct tuple *tuple)
{
uint32_t name_len;
- const char *name = tuple_field_str(tuple, BOX_USER_FIELD_NAME,
- &name_len);
+ const char *name =
+ tuple_field_str(tuple, BOX_USER_FIELD_NAME, &name_len);
if (name == NULL)
return NULL;
if (name_len > BOX_NAME_MAX) {
diag_set(ClientError, ER_CREATE_USER,
- tt_cstr(name, BOX_INVALID_NAME_MAX),
- "user name is too long");
+ tt_cstr(name, BOX_INVALID_NAME_MAX),
+ "user name is too long");
return NULL;
}
size_t size = user_def_sizeof(name_len);
/* Use calloc: in case user password is empty, fill it with \0 */
- struct user_def *user = (struct user_def *) malloc(size);
+ struct user_def *user = (struct user_def *)malloc(size);
if (user == NULL) {
diag_set(OutOfMemory, size, "malloc", "user");
return NULL;
@@ -3051,8 +3049,8 @@ user_def_new_from_tuple(struct tuple *tuple)
memcpy(user->name, name, name_len);
user->name[name_len] = 0;
if (user->type != SC_ROLE && user->type != SC_USER) {
- diag_set(ClientError, ER_CREATE_USER,
- user->name, "unknown user type");
+ diag_set(ClientError, ER_CREATE_USER, user->name,
+ "unknown user type");
return NULL;
}
if (identifier_check(user->name, name_len) != 0)
@@ -3079,8 +3077,8 @@ user_def_new_from_tuple(struct tuple *tuple)
}
if (!is_auth_empty && user->type == SC_ROLE) {
diag_set(ClientError, ER_CREATE_ROLE, user->name,
- "authentication data can not be set for a "\
- "role");
+ "authentication data can not be set for a "
+ "role");
return NULL;
}
if (user_def_fill_auth_data(user, auth_data) != 0)
@@ -3125,7 +3123,7 @@ user_cache_alter_user(struct trigger *trigger, void * /* event */)
static int
on_replace_dd_user(struct trigger * /* trigger */, void *event)
{
- struct txn *txn = (struct txn *) event;
+ struct txn *txn = (struct txn *)event;
struct txn_stmt *stmt = txn_current_stmt(txn);
struct tuple *old_tuple = stmt->old_tuple;
struct tuple *new_tuple = stmt->new_tuple;
@@ -3139,31 +3137,30 @@ on_replace_dd_user(struct trigger * /* trigger */, void *event)
struct user_def *user = user_def_new_from_tuple(new_tuple);
if (user == NULL)
return -1;
- if (access_check_ddl(user->name, user->uid, user->owner, user->type,
- PRIV_C) != 0)
+ if (access_check_ddl(user->name, user->uid, user->owner,
+ user->type, PRIV_C) != 0)
return -1;
auto def_guard = make_scoped_guard([=] { free(user); });
try {
- (void) user_cache_replace(user);
+ (void)user_cache_replace(user);
} catch (Exception *e) {
return -1;
}
def_guard.is_active = false;
- struct trigger *on_rollback =
- txn_alter_trigger_new(user_cache_remove_user, new_tuple);
+ struct trigger *on_rollback = txn_alter_trigger_new(
+ user_cache_remove_user, new_tuple);
if (on_rollback == NULL)
return -1;
txn_stmt_on_rollback(stmt, on_rollback);
} else if (new_tuple == NULL) { /* DELETE */
if (access_check_ddl(old_user->def->name, old_user->def->uid,
- old_user->def->owner, old_user->def->type,
- PRIV_D) != 0)
+ old_user->def->owner, old_user->def->type,
+ PRIV_D) != 0)
return -1;
/* Can't drop guest or super user */
- if (uid <= (uint32_t) BOX_SYSTEM_USER_ID_MAX || uid == SUPER) {
- diag_set(ClientError, ER_DROP_USER,
- old_user->def->name,
- "the user or the role is a system");
+ if (uid <= (uint32_t)BOX_SYSTEM_USER_ID_MAX || uid == SUPER) {
+ diag_set(ClientError, ER_DROP_USER, old_user->def->name,
+ "the user or the role is a system");
return -1;
}
/*
@@ -3175,8 +3172,8 @@ on_replace_dd_user(struct trigger * /* trigger */, void *event)
return -1;
}
if (has_data) {
- diag_set(ClientError, ER_DROP_USER,
- old_user->def->name, "the user has objects");
+ diag_set(ClientError, ER_DROP_USER, old_user->def->name,
+ "the user has objects");
return -1;
}
user_cache_delete(uid);
@@ -3196,7 +3193,7 @@ on_replace_dd_user(struct trigger * /* trigger */, void *event)
if (user == NULL)
return -1;
if (access_check_ddl(user->name, user->uid, user->uid,
- old_user->def->type, PRIV_A) != 0)
+ old_user->def->type, PRIV_A) != 0)
return -1;
auto def_guard = make_scoped_guard([=] { free(user); });
try {
@@ -3241,8 +3238,8 @@ func_def_new_from_tuple(struct tuple *tuple)
return NULL;
if (name_len > BOX_NAME_MAX) {
diag_set(ClientError, ER_CREATE_FUNCTION,
- tt_cstr(name, BOX_INVALID_NAME_MAX),
- "function name is too long");
+ tt_cstr(name, BOX_INVALID_NAME_MAX),
+ "function name is too long");
return NULL;
}
if (identifier_check(name, name_len) != 0)
@@ -3256,24 +3253,24 @@ func_def_new_from_tuple(struct tuple *tuple)
if (comment == NULL)
return NULL;
uint32_t len;
- const char *routine_type = tuple_field_str(tuple,
- BOX_FUNC_FIELD_ROUTINE_TYPE, &len);
+ const char *routine_type = tuple_field_str(
+ tuple, BOX_FUNC_FIELD_ROUTINE_TYPE, &len);
if (routine_type == NULL)
return NULL;
if (len != strlen("function") ||
strncasecmp(routine_type, "function", len) != 0) {
diag_set(ClientError, ER_CREATE_FUNCTION, name,
- "unsupported routine_type value");
+ "unsupported routine_type value");
return NULL;
}
- const char *sql_data_access = tuple_field_str(tuple,
- BOX_FUNC_FIELD_SQL_DATA_ACCESS, &len);
+ const char *sql_data_access = tuple_field_str(
+ tuple, BOX_FUNC_FIELD_SQL_DATA_ACCESS, &len);
if (sql_data_access == NULL)
return NULL;
if (len != strlen("none") ||
strncasecmp(sql_data_access, "none", len) != 0) {
diag_set(ClientError, ER_CREATE_FUNCTION, name,
- "unsupported sql_data_access value");
+ "unsupported sql_data_access value");
return NULL;
}
bool is_null_call;
@@ -3282,7 +3279,7 @@ func_def_new_from_tuple(struct tuple *tuple)
return NULL;
if (is_null_call != true) {
diag_set(ClientError, ER_CREATE_FUNCTION, name,
- "unsupported is_null_call value");
+ "unsupported is_null_call value");
return NULL;
}
} else {
@@ -3294,7 +3291,7 @@ func_def_new_from_tuple(struct tuple *tuple)
uint32_t body_offset, comment_offset;
uint32_t def_sz = func_def_sizeof(name_len, body_len, comment_len,
&body_offset, &comment_offset);
- struct func_def *def = (struct func_def *) malloc(def_sz);
+ struct func_def *def = (struct func_def *)malloc(def_sz);
if (def == NULL) {
diag_set(OutOfMemory, def_sz, "malloc", "def");
return NULL;
@@ -3304,7 +3301,7 @@ func_def_new_from_tuple(struct tuple *tuple)
return NULL;
if (def->fid > BOX_FUNCTION_MAX) {
diag_set(ClientError, ER_CREATE_FUNCTION,
- tt_cstr(name, name_len), "function id is too big");
+ tt_cstr(name, name_len), "function id is too big");
return NULL;
}
func_opts_create(&def->opts);
@@ -3341,8 +3338,8 @@ func_def_new_from_tuple(struct tuple *tuple)
def->language = STR2ENUM(func_language, language);
if (def->language == func_language_MAX ||
def->language == FUNC_LANGUAGE_SQL) {
- diag_set(ClientError, ER_FUNCTION_LANGUAGE,
- language, def->name);
+ diag_set(ClientError, ER_FUNCTION_LANGUAGE, language,
+ def->name);
return NULL;
}
} else {
@@ -3362,20 +3359,20 @@ func_def_new_from_tuple(struct tuple *tuple)
return NULL;
def->returns = STR2ENUM(field_type, returns);
if (def->returns == field_type_MAX) {
- diag_set(ClientError, ER_CREATE_FUNCTION,
- def->name, "invalid returns value");
+ diag_set(ClientError, ER_CREATE_FUNCTION, def->name,
+ "invalid returns value");
return NULL;
}
def->exports.all = 0;
- const char *exports = tuple_field_with_type(tuple,
- BOX_FUNC_FIELD_EXPORTS, MP_ARRAY);
+ const char *exports = tuple_field_with_type(
+ tuple, BOX_FUNC_FIELD_EXPORTS, MP_ARRAY);
if (exports == NULL)
return NULL;
uint32_t cnt = mp_decode_array(&exports);
for (uint32_t i = 0; i < cnt; i++) {
if (mp_typeof(*exports) != MP_STR) {
diag_set(ClientError, ER_FIELD_TYPE,
- int2str(BOX_FUNC_FIELD_EXPORTS + 1),
+ int2str(BOX_FUNC_FIELD_EXPORTS + 1),
mp_type_strs[MP_STR]);
return NULL;
}
@@ -3390,7 +3387,7 @@ func_def_new_from_tuple(struct tuple *tuple)
break;
default:
diag_set(ClientError, ER_CREATE_FUNCTION,
- def->name, "invalid exports value");
+ def->name, "invalid exports value");
return NULL;
}
}
@@ -3400,19 +3397,19 @@ func_def_new_from_tuple(struct tuple *tuple)
return NULL;
def->aggregate = STR2ENUM(func_aggregate, aggregate);
if (def->aggregate == func_aggregate_MAX) {
- diag_set(ClientError, ER_CREATE_FUNCTION,
- def->name, "invalid aggregate value");
+ diag_set(ClientError, ER_CREATE_FUNCTION, def->name,
+ "invalid aggregate value");
return NULL;
}
- const char *param_list = tuple_field_with_type(tuple,
- BOX_FUNC_FIELD_PARAM_LIST, MP_ARRAY);
+ const char *param_list = tuple_field_with_type(
+ tuple, BOX_FUNC_FIELD_PARAM_LIST, MP_ARRAY);
if (param_list == NULL)
return NULL;
uint32_t argc = mp_decode_array(¶m_list);
for (uint32_t i = 0; i < argc; i++) {
- if (mp_typeof(*param_list) != MP_STR) {
+ if (mp_typeof(*param_list) != MP_STR) {
diag_set(ClientError, ER_FIELD_TYPE,
- int2str(BOX_FUNC_FIELD_PARAM_LIST + 1),
+ int2str(BOX_FUNC_FIELD_PARAM_LIST + 1),
mp_type_strs[MP_STR]);
return NULL;
}
@@ -3420,7 +3417,7 @@ func_def_new_from_tuple(struct tuple *tuple)
const char *str = mp_decode_str(¶m_list, &len);
if (STRN2ENUM(field_type, str, len) == field_type_MAX) {
diag_set(ClientError, ER_CREATE_FUNCTION,
- def->name, "invalid argument type");
+ def->name, "invalid argument type");
return NULL;
}
}
@@ -3485,7 +3482,7 @@ on_drop_func_rollback(struct trigger *trigger, void * /* event */)
static int
on_replace_dd_func(struct trigger * /* trigger */, void *event)
{
- struct txn *txn = (struct txn *) event;
+ struct txn *txn = (struct txn *)event;
struct txn_stmt *stmt = txn_current_stmt(txn);
struct tuple *old_tuple = stmt->old_tuple;
struct tuple *new_tuple = stmt->new_tuple;
@@ -3501,7 +3498,7 @@ on_replace_dd_func(struct trigger * /* trigger */, void *event)
return -1;
auto def_guard = make_scoped_guard([=] { free(def); });
if (access_check_ddl(def->name, def->fid, def->uid, SC_FUNCTION,
- PRIV_C) != 0)
+ PRIV_C) != 0)
return -1;
struct trigger *on_rollback =
txn_alter_trigger_new(on_create_func_rollback, NULL);
@@ -3516,7 +3513,7 @@ on_replace_dd_func(struct trigger * /* trigger */, void *event)
txn_stmt_on_rollback(stmt, on_rollback);
if (trigger_run(&on_alter_func, func) != 0)
return -1;
- } else if (new_tuple == NULL) { /* DELETE */
+ } else if (new_tuple == NULL) { /* DELETE */
uint32_t uid;
if (func_def_get_ids_from_tuple(old_tuple, &fid, &uid) != 0)
return -1;
@@ -3525,32 +3522,34 @@ on_replace_dd_func(struct trigger * /* trigger */, void *event)
* who created it or a superuser.
*/
if (access_check_ddl(old_func->def->name, fid, uid, SC_FUNCTION,
- PRIV_D) != 0)
+ PRIV_D) != 0)
return -1;
/* Can only delete func if it has no grants. */
bool out;
- if (schema_find_grants("function", old_func->def->fid, &out) != 0) {
+ if (schema_find_grants("function", old_func->def->fid, &out) !=
+ 0) {
return -1;
}
if (out) {
diag_set(ClientError, ER_DROP_FUNCTION,
- (unsigned) old_func->def->uid,
- "function has grants");
+ (unsigned)old_func->def->uid,
+ "function has grants");
return -1;
}
- if (space_has_data(BOX_FUNC_INDEX_ID, 1, old_func->def->fid, &out) != 0)
+ if (space_has_data(BOX_FUNC_INDEX_ID, 1, old_func->def->fid,
+ &out) != 0)
return -1;
if (old_func != NULL && out) {
diag_set(ClientError, ER_DROP_FUNCTION,
- (unsigned) old_func->def->uid,
- "function has references");
+ (unsigned)old_func->def->uid,
+ "function has references");
return -1;
}
/* Can't' drop a builtin function. */
if (old_func->def->language == FUNC_LANGUAGE_SQL_BUILTIN) {
diag_set(ClientError, ER_DROP_FUNCTION,
- (unsigned) old_func->def->uid,
- "function is SQL built-in");
+ (unsigned)old_func->def->uid,
+ "function is SQL built-in");
return -1;
}
struct trigger *on_commit =
@@ -3564,7 +3563,7 @@ on_replace_dd_func(struct trigger * /* trigger */, void *event)
txn_stmt_on_rollback(stmt, on_rollback);
if (trigger_run(&on_alter_func, old_func) != 0)
return -1;
- } else { /* UPDATE, REPLACE */
+ } else { /* UPDATE, REPLACE */
assert(new_tuple != NULL && old_tuple != NULL);
/**
* Allow an alter that doesn't change the
@@ -3581,7 +3580,7 @@ on_replace_dd_func(struct trigger * /* trigger */, void *event)
return -1;
if (func_def_cmp(new_def, old_def) != 0) {
diag_set(ClientError, ER_UNSUPPORTED, "function",
- "alter");
+ "alter");
return -1;
}
}
@@ -3602,31 +3601,32 @@ coll_id_def_new_from_tuple(struct tuple *tuple, struct coll_id_def *def)
def->name_len = name_len;
if (name_len > BOX_NAME_MAX) {
diag_set(ClientError, ER_CANT_CREATE_COLLATION,
- "collation name is too long");
+ "collation name is too long");
return -1;
}
if (identifier_check(def->name, name_len) != 0)
return -1;
- if (tuple_field_u32(tuple, BOX_COLLATION_FIELD_UID, &(def->owner_id)) != 0)
+ if (tuple_field_u32(tuple, BOX_COLLATION_FIELD_UID, &(def->owner_id)) !=
+ 0)
return -1;
- const char *type = tuple_field_str(tuple, BOX_COLLATION_FIELD_TYPE,
- &type_len);
+ const char *type =
+ tuple_field_str(tuple, BOX_COLLATION_FIELD_TYPE, &type_len);
if (type == NULL)
return -1;
struct coll_def *base = &def->base;
base->type = STRN2ENUM(coll_type, type, type_len);
if (base->type == coll_type_MAX) {
diag_set(ClientError, ER_CANT_CREATE_COLLATION,
- "unknown collation type");
+ "unknown collation type");
return -1;
}
- const char *locale = tuple_field_str(tuple, BOX_COLLATION_FIELD_LOCALE,
- &locale_len);
+ const char *locale =
+ tuple_field_str(tuple, BOX_COLLATION_FIELD_LOCALE, &locale_len);
if (locale == NULL)
return -1;
if (locale_len > COLL_LOCALE_LEN_MAX) {
diag_set(ClientError, ER_CANT_CREATE_COLLATION,
- "collation locale is too long");
+ "collation locale is too long");
return -1;
}
if (locale_len > 0)
@@ -3634,62 +3634,62 @@ coll_id_def_new_from_tuple(struct tuple *tuple, struct coll_id_def *def)
return -1;
snprintf(base->locale, sizeof(base->locale), "%.*s", locale_len,
locale);
- const char *options = tuple_field_with_type(tuple,
- BOX_COLLATION_FIELD_OPTIONS, MP_MAP);
+ const char *options = tuple_field_with_type(
+ tuple, BOX_COLLATION_FIELD_OPTIONS, MP_MAP);
if (options == NULL)
return -1;
if (opts_decode(&base->icu, coll_icu_opts_reg, &options,
- ER_WRONG_COLLATION_OPTIONS,
- BOX_COLLATION_FIELD_OPTIONS, NULL) != 0)
+ ER_WRONG_COLLATION_OPTIONS, BOX_COLLATION_FIELD_OPTIONS,
+ NULL) != 0)
return -1;
if (base->icu.french_collation == coll_icu_on_off_MAX) {
diag_set(ClientError, ER_CANT_CREATE_COLLATION,
- "ICU wrong french_collation option setting, "
- "expected ON | OFF");
+ "ICU wrong french_collation option setting, "
+ "expected ON | OFF");
return -1;
}
if (base->icu.alternate_handling == coll_icu_alternate_handling_MAX) {
diag_set(ClientError, ER_CANT_CREATE_COLLATION,
- "ICU wrong alternate_handling option setting, "
- "expected NON_IGNORABLE | SHIFTED");
+ "ICU wrong alternate_handling option setting, "
+ "expected NON_IGNORABLE | SHIFTED");
return -1;
}
if (base->icu.case_first == coll_icu_case_first_MAX) {
diag_set(ClientError, ER_CANT_CREATE_COLLATION,
- "ICU wrong case_first option setting, "
- "expected OFF | UPPER_FIRST | LOWER_FIRST");
+ "ICU wrong case_first option setting, "
+ "expected OFF | UPPER_FIRST | LOWER_FIRST");
return -1;
}
if (base->icu.case_level == coll_icu_on_off_MAX) {
diag_set(ClientError, ER_CANT_CREATE_COLLATION,
- "ICU wrong case_level option setting, "
- "expected ON | OFF");
+ "ICU wrong case_level option setting, "
+ "expected ON | OFF");
return -1;
}
if (base->icu.normalization_mode == coll_icu_on_off_MAX) {
diag_set(ClientError, ER_CANT_CREATE_COLLATION,
- "ICU wrong normalization_mode option setting, "
- "expected ON | OFF");
+ "ICU wrong normalization_mode option setting, "
+ "expected ON | OFF");
return -1;
}
if (base->icu.strength == coll_icu_strength_MAX) {
diag_set(ClientError, ER_CANT_CREATE_COLLATION,
- "ICU wrong strength option setting, "
- "expected PRIMARY | SECONDARY | "
- "TERTIARY | QUATERNARY | IDENTICAL");
+ "ICU wrong strength option setting, "
+ "expected PRIMARY | SECONDARY | "
+ "TERTIARY | QUATERNARY | IDENTICAL");
return -1;
}
if (base->icu.numeric_collation == coll_icu_on_off_MAX) {
diag_set(ClientError, ER_CANT_CREATE_COLLATION,
- "ICU wrong numeric_collation option setting, "
- "expected ON | OFF");
+ "ICU wrong numeric_collation option setting, "
+ "expected ON | OFF");
return -1;
}
return 0;
@@ -3699,20 +3699,19 @@ coll_id_def_new_from_tuple(struct tuple *tuple, struct coll_id_def *def)
static int
on_create_collation_rollback(struct trigger *trigger, void *event)
{
- (void) event;
- struct coll_id *coll_id = (struct coll_id *) trigger->data;
+ (void)event;
+ struct coll_id *coll_id = (struct coll_id *)trigger->data;
coll_id_cache_delete(coll_id);
coll_id_delete(coll_id);
return 0;
}
-
/** Free a deleted collation identifier on commit. */
static int
on_drop_collation_commit(struct trigger *trigger, void *event)
{
- (void) event;
- struct coll_id *coll_id = (struct coll_id *) trigger->data;
+ (void)event;
+ struct coll_id *coll_id = (struct coll_id *)trigger->data;
coll_id_delete(coll_id);
return 0;
}
@@ -3721,8 +3720,8 @@ on_drop_collation_commit(struct trigger *trigger, void *event)
static int
on_drop_collation_rollback(struct trigger *trigger, void *event)
{
- (void) event;
- struct coll_id *coll_id = (struct coll_id *) trigger->data;
+ (void)event;
+ struct coll_id *coll_id = (struct coll_id *)trigger->data;
struct coll_id *replaced_id;
if (coll_id_cache_replace(coll_id, &replaced_id) != 0)
panic("Out of memory on insertion into collation cache");
@@ -3737,7 +3736,7 @@ on_drop_collation_rollback(struct trigger *trigger, void *event)
static int
on_replace_dd_collation(struct trigger * /* trigger */, void *event)
{
- struct txn *txn = (struct txn *) event;
+ struct txn *txn = (struct txn *)event;
struct txn_stmt *stmt = txn_current_stmt(txn);
struct tuple *old_tuple = stmt->old_tuple;
struct tuple *new_tuple = stmt->new_tuple;
@@ -3754,7 +3753,8 @@ on_replace_dd_collation(struct trigger * /* trigger */, void *event)
* identifier.
*/
uint32_t out;
- if (tuple_field_u32(old_tuple, BOX_COLLATION_FIELD_ID, &out) != 0)
+ if (tuple_field_u32(old_tuple, BOX_COLLATION_FIELD_ID, &out) !=
+ 0)
return -1;
int32_t old_id = out;
/*
@@ -3765,14 +3765,14 @@ on_replace_dd_collation(struct trigger * /* trigger */, void *event)
*/
if (old_id == COLL_NONE) {
diag_set(ClientError, ER_DROP_COLLATION, "none",
- "system collation");
+ "system collation");
return -1;
}
struct coll_id *old_coll_id = coll_by_id(old_id);
assert(old_coll_id != NULL);
if (access_check_ddl(old_coll_id->name, old_coll_id->id,
- old_coll_id->owner_id, SC_COLLATION,
- PRIV_D) != 0)
+ old_coll_id->owner_id, SC_COLLATION,
+ PRIV_D) != 0)
return -1;
/*
* Set on_commit/on_rollback triggers after
@@ -3786,15 +3786,15 @@ on_replace_dd_collation(struct trigger * /* trigger */, void *event)
txn_stmt_on_commit(stmt, on_commit);
} else if (new_tuple != NULL && old_tuple == NULL) {
/* INSERT */
- struct trigger *on_rollback =
- txn_alter_trigger_new(on_create_collation_rollback, NULL);
+ struct trigger *on_rollback = txn_alter_trigger_new(
+ on_create_collation_rollback, NULL);
if (on_rollback == NULL)
return -1;
struct coll_id_def new_def;
if (coll_id_def_new_from_tuple(new_tuple, &new_def) != 0)
return -1;
if (access_check_ddl(new_def.name, new_def.id, new_def.owner_id,
- SC_COLLATION, PRIV_C) != 0)
+ SC_COLLATION, PRIV_C) != 0)
return -1;
struct coll_id *new_coll_id = coll_id_new(&new_def);
if (new_coll_id == NULL)
@@ -3822,8 +3822,10 @@ on_replace_dd_collation(struct trigger * /* trigger */, void *event)
int
priv_def_create_from_tuple(struct priv_def *priv, struct tuple *tuple)
{
- if (tuple_field_u32(tuple, BOX_PRIV_FIELD_ID, &(priv->grantor_id)) != 0 ||
- tuple_field_u32(tuple, BOX_PRIV_FIELD_UID, &(priv->grantee_id)) != 0)
+ if (tuple_field_u32(tuple, BOX_PRIV_FIELD_ID, &(priv->grantor_id)) !=
+ 0 ||
+ tuple_field_u32(tuple, BOX_PRIV_FIELD_UID, &(priv->grantee_id)) !=
+ 0)
return -1;
const char *object_type =
@@ -3835,7 +3837,7 @@ priv_def_create_from_tuple(struct priv_def *priv, struct tuple *tuple)
const char *data = tuple_field(tuple, BOX_PRIV_FIELD_OBJECT_ID);
if (data == NULL) {
diag_set(ClientError, ER_NO_SUCH_FIELD_NO,
- BOX_PRIV_FIELD_OBJECT_ID + TUPLE_INDEX_BASE);
+ BOX_PRIV_FIELD_OBJECT_ID + TUPLE_INDEX_BASE);
return -1;
}
/*
@@ -3849,18 +3851,18 @@ priv_def_create_from_tuple(struct priv_def *priv, struct tuple *tuple)
if (mp_decode_strl(&data) == 0) {
/* Entity-wide privilege. */
priv->object_id = 0;
- priv->object_type = schema_entity_type(priv->object_type);
+ priv->object_type =
+ schema_entity_type(priv->object_type);
break;
}
FALLTHROUGH;
default:
- if (tuple_field_u32(tuple,
- BOX_PRIV_FIELD_OBJECT_ID, &(priv->object_id)) != 0)
+ if (tuple_field_u32(tuple, BOX_PRIV_FIELD_OBJECT_ID,
+ &(priv->object_id)) != 0)
return -1;
}
if (priv->object_type == SC_UNKNOWN) {
- diag_set(ClientError, ER_UNKNOWN_SCHEMA_OBJECT,
- object_type);
+ diag_set(ClientError, ER_UNKNOWN_SCHEMA_OBJECT, object_type);
return -1;
}
uint32_t out;
@@ -3890,7 +3892,7 @@ priv_def_check(struct priv_def *priv, enum priv_type priv_type)
struct user *grantee = user_by_id(priv->grantee_id);
if (grantee == NULL) {
diag_set(ClientError, ER_NO_SUCH_USER,
- int2str(priv->grantee_id));
+ int2str(priv->grantee_id));
return -1;
}
const char *name = schema_find_name(priv->object_type, priv->object_id);
@@ -3900,70 +3902,63 @@ priv_def_check(struct priv_def *priv, enum priv_type priv_type)
switch (priv->object_type) {
case SC_UNIVERSE:
if (grantor->def->uid != ADMIN) {
- diag_set(AccessDeniedError,
- priv_name(priv_type),
- schema_object_name(SC_UNIVERSE),
- name,
- grantor->def->name);
+ diag_set(AccessDeniedError, priv_name(priv_type),
+ schema_object_name(SC_UNIVERSE), name,
+ grantor->def->name);
return -1;
}
break;
- case SC_SPACE:
- {
+ case SC_SPACE: {
struct space *space = space_cache_find(priv->object_id);
if (space == NULL)
return -1;
if (space->def->uid != grantor->def->uid &&
grantor->def->uid != ADMIN) {
- diag_set(AccessDeniedError,
- priv_name(priv_type),
- schema_object_name(SC_SPACE), name,
- grantor->def->name);
+ diag_set(AccessDeniedError, priv_name(priv_type),
+ schema_object_name(SC_SPACE), name,
+ grantor->def->name);
return -1;
}
break;
}
- case SC_FUNCTION:
- {
+ case SC_FUNCTION: {
struct func *func = func_by_id(priv->object_id);
if (func == NULL) {
- diag_set(ClientError, ER_NO_SUCH_FUNCTION, int2str(priv->object_id));
+ diag_set(ClientError, ER_NO_SUCH_FUNCTION,
+ int2str(priv->object_id));
return -1;
}
if (func->def->uid != grantor->def->uid &&
grantor->def->uid != ADMIN) {
- diag_set(AccessDeniedError,
- priv_name(priv_type),
- schema_object_name(SC_FUNCTION), name,
- grantor->def->name);
+ diag_set(AccessDeniedError, priv_name(priv_type),
+ schema_object_name(SC_FUNCTION), name,
+ grantor->def->name);
return -1;
}
break;
}
- case SC_SEQUENCE:
- {
+ case SC_SEQUENCE: {
struct sequence *seq = sequence_by_id(priv->object_id);
if (seq == NULL) {
- diag_set(ClientError, ER_NO_SUCH_SEQUENCE, int2str(priv->object_id));
+ diag_set(ClientError, ER_NO_SUCH_SEQUENCE,
+ int2str(priv->object_id));
return -1;
}
if (seq->def->uid != grantor->def->uid &&
grantor->def->uid != ADMIN) {
- diag_set(AccessDeniedError,
- priv_name(priv_type),
- schema_object_name(SC_SEQUENCE), name,
- grantor->def->name);
+ diag_set(AccessDeniedError, priv_name(priv_type),
+ schema_object_name(SC_SEQUENCE), name,
+ grantor->def->name);
return -1;
}
break;
}
- case SC_ROLE:
- {
+ case SC_ROLE: {
struct user *role = user_by_id(priv->object_id);
if (role == NULL || role->def->type != SC_ROLE) {
diag_set(ClientError, ER_NO_SUCH_ROLE,
- role ? role->def->name :
- int2str(priv->object_id));
+ role ? role->def->name :
+ int2str(priv->object_id));
return -1;
}
/*
@@ -3973,10 +3968,9 @@ priv_def_check(struct priv_def *priv, enum priv_type priv_type)
if (role->def->owner != grantor->def->uid &&
grantor->def->uid != ADMIN &&
(role->def->uid != PUBLIC || priv->access != PRIV_X)) {
- diag_set(AccessDeniedError,
- priv_name(priv_type),
- schema_object_name(SC_ROLE), name,
- grantor->def->name);
+ diag_set(AccessDeniedError, priv_name(priv_type),
+ schema_object_name(SC_ROLE), name,
+ grantor->def->name);
return -1;
}
/* Not necessary to do during revoke, but who cares. */
@@ -3984,21 +3978,19 @@ priv_def_check(struct priv_def *priv, enum priv_type priv_type)
return -1;
break;
}
- case SC_USER:
- {
+ case SC_USER: {
struct user *user = user_by_id(priv->object_id);
if (user == NULL || user->def->type != SC_USER) {
diag_set(ClientError, ER_NO_SUCH_USER,
- user ? user->def->name :
- int2str(priv->object_id));
+ user ? user->def->name :
+ int2str(priv->object_id));
return -1;
}
if (user->def->owner != grantor->def->uid &&
grantor->def->uid != ADMIN) {
- diag_set(AccessDeniedError,
- priv_name(priv_type),
- schema_object_name(SC_USER), name,
- grantor->def->name);
+ diag_set(AccessDeniedError, priv_name(priv_type),
+ schema_object_name(SC_USER), name,
+ grantor->def->name);
return -1;
}
break;
@@ -4007,13 +3999,12 @@ priv_def_check(struct priv_def *priv, enum priv_type priv_type)
case SC_ENTITY_FUNCTION:
case SC_ENTITY_SEQUENCE:
case SC_ENTITY_ROLE:
- case SC_ENTITY_USER:
- {
+ case SC_ENTITY_USER: {
/* Only admin may grant privileges on an entire entity. */
if (grantor->def->uid != ADMIN) {
diag_set(AccessDeniedError, priv_name(priv_type),
- schema_object_name(priv->object_type), name,
- grantor->def->name);
+ schema_object_name(priv->object_type), name,
+ grantor->def->name);
return -1;
}
}
@@ -4022,7 +4013,7 @@ priv_def_check(struct priv_def *priv, enum priv_type priv_type)
}
if (priv->access == 0) {
diag_set(ClientError, ER_GRANT,
- "the grant tuple has no privileges");
+ "the grant tuple has no privileges");
return -1;
}
return 0;
@@ -4064,7 +4055,7 @@ grant_or_revoke(struct priv_def *priv)
static int
revoke_priv(struct trigger *trigger, void *event)
{
- (void) event;
+ (void)event;
struct tuple *tuple = (struct tuple *)trigger->data;
struct priv_def priv;
if (priv_def_create_from_tuple(&priv, tuple) != 0)
@@ -4079,7 +4070,7 @@ revoke_priv(struct trigger *trigger, void *event)
static int
modify_priv(struct trigger *trigger, void *event)
{
- (void) event;
+ (void)event;
struct tuple *tuple = (struct tuple *)trigger->data;
struct priv_def priv;
if (priv_def_create_from_tuple(&priv, tuple) != 0 ||
@@ -4095,13 +4086,13 @@ modify_priv(struct trigger *trigger, void *event)
static int
on_replace_dd_priv(struct trigger * /* trigger */, void *event)
{
- struct txn *txn = (struct txn *) event;
+ struct txn *txn = (struct txn *)event;
struct txn_stmt *stmt = txn_current_stmt(txn);
struct tuple *old_tuple = stmt->old_tuple;
struct tuple *new_tuple = stmt->new_tuple;
struct priv_def priv;
- if (new_tuple != NULL && old_tuple == NULL) { /* grant */
+ if (new_tuple != NULL && old_tuple == NULL) { /* grant */
if (priv_def_create_from_tuple(&priv, new_tuple) != 0 ||
priv_def_check(&priv, PRIV_GRANT) != 0 ||
grant_or_revoke(&priv) != 0)
@@ -4111,7 +4102,7 @@ on_replace_dd_priv(struct trigger * /* trigger */, void *event)
if (on_rollback == NULL)
return -1;
txn_stmt_on_rollback(stmt, on_rollback);
- } else if (new_tuple == NULL) { /* revoke */
+ } else if (new_tuple == NULL) { /* revoke */
assert(old_tuple);
if (priv_def_create_from_tuple(&priv, old_tuple) != 0 ||
priv_def_check(&priv, PRIV_REVOKE) != 0)
@@ -4124,7 +4115,7 @@ on_replace_dd_priv(struct trigger * /* trigger */, void *event)
if (on_rollback == NULL)
return -1;
txn_stmt_on_rollback(stmt, on_rollback);
- } else { /* modify */
+ } else { /* modify */
if (priv_def_create_from_tuple(&priv, new_tuple) != 0 ||
priv_def_check(&priv, PRIV_GRANT) != 0 ||
grant_or_revoke(&priv) != 0)
@@ -4154,12 +4145,12 @@ on_replace_dd_priv(struct trigger * /* trigger */, void *event)
static int
on_replace_dd_schema(struct trigger * /* trigger */, void *event)
{
- struct txn *txn = (struct txn *) event;
+ struct txn *txn = (struct txn *)event;
struct txn_stmt *stmt = txn_current_stmt(txn);
struct tuple *old_tuple = stmt->old_tuple;
struct tuple *new_tuple = stmt->new_tuple;
const char *key = tuple_field_cstr(new_tuple ? new_tuple : old_tuple,
- BOX_SCHEMA_FIELD_KEY);
+ BOX_SCHEMA_FIELD_KEY);
if (key == NULL)
return -1;
if (strcmp(key, "cluster") == 0) {
@@ -4168,7 +4159,8 @@ on_replace_dd_schema(struct trigger * /* trigger */, void *event)
return -1;
}
tt_uuid uu;
- if (tuple_field_uuid(new_tuple, BOX_CLUSTER_FIELD_UUID, &uu) != 0)
+ if (tuple_field_uuid(new_tuple, BOX_CLUSTER_FIELD_UUID, &uu) !=
+ 0)
return -1;
REPLICASET_UUID = uu;
say_info("cluster uuid %s", tt_uuid_str(&uu));
@@ -4198,7 +4190,7 @@ register_replica(struct trigger *trigger, void * /* event */)
try {
replica = replicaset_add(id, &uuid);
/* Can't throw exceptions from on_commit trigger */
- } catch(Exception *e) {
+ } catch (Exception *e) {
panic("Can't register replica: %s", e->errmsg);
}
}
@@ -4241,25 +4233,26 @@ unregister_replica(struct trigger *trigger, void * /* event */)
static int
on_replace_dd_cluster(struct trigger *trigger, void *event)
{
- (void) trigger;
- struct txn *txn = (struct txn *) event;
+ (void)trigger;
+ struct txn *txn = (struct txn *)event;
struct txn_stmt *stmt = txn_current_stmt(txn);
struct tuple *old_tuple = stmt->old_tuple;
struct tuple *new_tuple = stmt->new_tuple;
if (new_tuple != NULL) { /* Insert or replace */
/* Check fields */
uint32_t replica_id;
- if (tuple_field_u32(new_tuple, BOX_CLUSTER_FIELD_ID, &replica_id) != 0)
+ if (tuple_field_u32(new_tuple, BOX_CLUSTER_FIELD_ID,
+ &replica_id) != 0)
return -1;
if (replica_check_id(replica_id) != 0)
return -1;
tt_uuid replica_uuid;
if (tuple_field_uuid(new_tuple, BOX_CLUSTER_FIELD_UUID,
- &replica_uuid) != 0)
+ &replica_uuid) != 0)
return -1;
if (tt_uuid_is_nil(&replica_uuid)) {
diag_set(ClientError, ER_INVALID_UUID,
- tt_uuid_str(&replica_uuid));
+ tt_uuid_str(&replica_uuid));
return -1;
}
if (old_tuple != NULL) {
@@ -4270,12 +4263,12 @@ on_replace_dd_cluster(struct trigger *trigger, void *event)
*/
tt_uuid old_uuid;
if (tuple_field_uuid(old_tuple, BOX_CLUSTER_FIELD_UUID,
- &old_uuid) != 0)
+ &old_uuid) != 0)
return -1;
if (!tt_uuid_is_equal(&replica_uuid, &old_uuid)) {
diag_set(ClientError, ER_UNSUPPORTED,
- "Space _cluster",
- "updates of instance uuid");
+ "Space _cluster",
+ "updates of instance uuid");
return -1;
}
} else {
@@ -4293,14 +4286,15 @@ on_replace_dd_cluster(struct trigger *trigger, void *event)
*/
assert(old_tuple != NULL);
uint32_t replica_id;
- if (tuple_field_u32(old_tuple, BOX_CLUSTER_FIELD_ID, &replica_id) != 0)
+ if (tuple_field_u32(old_tuple, BOX_CLUSTER_FIELD_ID,
+ &replica_id) != 0)
return -1;
if (replica_check_id(replica_id) != 0)
return -1;
struct trigger *on_commit;
- on_commit = txn_alter_trigger_new(unregister_replica,
- old_tuple);
+ on_commit =
+ txn_alter_trigger_new(unregister_replica, old_tuple);
if (on_commit == NULL)
return -1;
txn_stmt_on_commit(stmt, on_commit);
@@ -4317,20 +4311,20 @@ static struct sequence_def *
sequence_def_new_from_tuple(struct tuple *tuple, uint32_t errcode)
{
uint32_t name_len;
- const char *name = tuple_field_str(tuple, BOX_USER_FIELD_NAME,
- &name_len);
+ const char *name =
+ tuple_field_str(tuple, BOX_USER_FIELD_NAME, &name_len);
if (name == NULL)
return NULL;
if (name_len > BOX_NAME_MAX) {
diag_set(ClientError, errcode,
- tt_cstr(name, BOX_INVALID_NAME_MAX),
- "sequence name is too long");
+ tt_cstr(name, BOX_INVALID_NAME_MAX),
+ "sequence name is too long");
return NULL;
}
if (identifier_check(name, name_len) != 0)
return NULL;
size_t sz = sequence_def_sizeof(name_len);
- struct sequence_def *def = (struct sequence_def *) malloc(sz);
+ struct sequence_def *def = (struct sequence_def *)malloc(sz);
if (def == NULL) {
diag_set(OutOfMemory, sz, "malloc", "sequence");
return NULL;
@@ -4348,11 +4342,14 @@ sequence_def_new_from_tuple(struct tuple *tuple, uint32_t errcode)
return NULL;
if (tuple_field_i64(tuple, BOX_SEQUENCE_FIELD_MAX, &(def->max)) != 0)
return NULL;
- if (tuple_field_i64(tuple, BOX_SEQUENCE_FIELD_START, &(def->start)) != 0)
+ if (tuple_field_i64(tuple, BOX_SEQUENCE_FIELD_START, &(def->start)) !=
+ 0)
return NULL;
- if (tuple_field_i64(tuple, BOX_SEQUENCE_FIELD_CACHE, &(def->cache)) != 0)
+ if (tuple_field_i64(tuple, BOX_SEQUENCE_FIELD_CACHE, &(def->cache)) !=
+ 0)
return NULL;
- if (tuple_field_bool(tuple, BOX_SEQUENCE_FIELD_CYCLE, &(def->cycle)) != 0)
+ if (tuple_field_bool(tuple, BOX_SEQUENCE_FIELD_CYCLE, &(def->cycle)) !=
+ 0)
return NULL;
if (def->step == 0) {
diag_set(ClientError, errcode, def->name,
@@ -4405,7 +4402,6 @@ on_drop_sequence_rollback(struct trigger *trigger, void * /* event */)
return 0;
}
-
static int
on_alter_sequence_commit(struct trigger *trigger, void * /* event */)
{
@@ -4436,7 +4432,7 @@ on_alter_sequence_rollback(struct trigger *trigger, void * /* event */)
static int
on_replace_dd_sequence(struct trigger * /* trigger */, void *event)
{
- struct txn *txn = (struct txn *) event;
+ struct txn *txn = (struct txn *)event;
struct txn_stmt *stmt = txn_current_stmt(txn);
struct tuple *old_tuple = stmt->old_tuple;
struct tuple *new_tuple = stmt->new_tuple;
@@ -4445,16 +4441,16 @@ on_replace_dd_sequence(struct trigger * /* trigger */, void *event)
auto def_guard = make_scoped_guard([&new_def] { free(new_def); });
struct sequence *seq;
- if (old_tuple == NULL && new_tuple != NULL) { /* INSERT */
+ if (old_tuple == NULL && new_tuple != NULL) { /* INSERT */
new_def = sequence_def_new_from_tuple(new_tuple,
ER_CREATE_SEQUENCE);
if (new_def == NULL)
return -1;
if (access_check_ddl(new_def->name, new_def->id, new_def->uid,
- SC_SEQUENCE, PRIV_C) != 0)
+ SC_SEQUENCE, PRIV_C) != 0)
return -1;
- struct trigger *on_rollback =
- txn_alter_trigger_new(on_create_sequence_rollback, NULL);
+ struct trigger *on_rollback = txn_alter_trigger_new(
+ on_create_sequence_rollback, NULL);
if (on_rollback == NULL)
return -1;
seq = sequence_new(new_def);
@@ -4463,36 +4459,37 @@ on_replace_dd_sequence(struct trigger * /* trigger */, void *event)
sequence_cache_insert(seq);
on_rollback->data = seq;
txn_stmt_on_rollback(stmt, on_rollback);
- } else if (old_tuple != NULL && new_tuple == NULL) { /* DELETE */
+ } else if (old_tuple != NULL && new_tuple == NULL) { /* DELETE */
uint32_t id;
- if (tuple_field_u32(old_tuple, BOX_SEQUENCE_DATA_FIELD_ID, &id) != 0)
+ if (tuple_field_u32(old_tuple, BOX_SEQUENCE_DATA_FIELD_ID,
+ &id) != 0)
return -1;
seq = sequence_by_id(id);
assert(seq != NULL);
- if (access_check_ddl(seq->def->name, seq->def->id, seq->def->uid,
- SC_SEQUENCE, PRIV_D) != 0)
+ if (access_check_ddl(seq->def->name, seq->def->id,
+ seq->def->uid, SC_SEQUENCE, PRIV_D) != 0)
return -1;
bool out;
if (space_has_data(BOX_SEQUENCE_DATA_ID, 0, id, &out) != 0)
return -1;
if (out) {
- diag_set(ClientError, ER_DROP_SEQUENCE,
- seq->def->name, "the sequence has data");
+ diag_set(ClientError, ER_DROP_SEQUENCE, seq->def->name,
+ "the sequence has data");
return -1;
}
if (space_has_data(BOX_SPACE_SEQUENCE_ID, 1, id, &out) != 0)
return -1;
if (out) {
- diag_set(ClientError, ER_DROP_SEQUENCE,
- seq->def->name, "the sequence is in use");
+ diag_set(ClientError, ER_DROP_SEQUENCE, seq->def->name,
+ "the sequence is in use");
return -1;
}
if (schema_find_grants("sequence", seq->def->id, &out) != 0) {
return -1;
}
if (out) {
- diag_set(ClientError, ER_DROP_SEQUENCE,
- seq->def->name, "the sequence has grants");
+ diag_set(ClientError, ER_DROP_SEQUENCE, seq->def->name,
+ "the sequence has grants");
return -1;
}
struct trigger *on_commit =
@@ -4504,20 +4501,20 @@ on_replace_dd_sequence(struct trigger * /* trigger */, void *event)
sequence_cache_delete(seq->def->id);
txn_stmt_on_commit(stmt, on_commit);
txn_stmt_on_rollback(stmt, on_rollback);
- } else { /* UPDATE */
+ } else { /* UPDATE */
new_def = sequence_def_new_from_tuple(new_tuple,
ER_ALTER_SEQUENCE);
if (new_def == NULL)
return -1;
seq = sequence_by_id(new_def->id);
assert(seq != NULL);
- if (access_check_ddl(seq->def->name, seq->def->id, seq->def->uid,
- SC_SEQUENCE, PRIV_A) != 0)
+ if (access_check_ddl(seq->def->name, seq->def->id,
+ seq->def->uid, SC_SEQUENCE, PRIV_A) != 0)
return -1;
- struct trigger *on_commit =
- txn_alter_trigger_new(on_alter_sequence_commit, seq->def);
- struct trigger *on_rollback =
- txn_alter_trigger_new(on_alter_sequence_rollback, seq->def);
+ struct trigger *on_commit = txn_alter_trigger_new(
+ on_alter_sequence_commit, seq->def);
+ struct trigger *on_rollback = txn_alter_trigger_new(
+ on_alter_sequence_rollback, seq->def);
if (on_commit == NULL || on_rollback == NULL)
return -1;
seq->def = new_def;
@@ -4556,7 +4553,7 @@ on_drop_sequence_data_rollback(struct trigger *trigger, void * /* event */)
static int
on_replace_dd_sequence_data(struct trigger * /* trigger */, void *event)
{
- struct txn *txn = (struct txn *) event;
+ struct txn *txn = (struct txn *)event;
struct txn_stmt *stmt = txn_current_stmt(txn);
struct tuple *old_tuple = stmt->old_tuple;
struct tuple *new_tuple = stmt->new_tuple;
@@ -4570,14 +4567,14 @@ on_replace_dd_sequence_data(struct trigger * /* trigger */, void *event)
diag_set(ClientError, ER_NO_SUCH_SEQUENCE, int2str(id));
return -1;
}
- if (new_tuple != NULL) { /* INSERT, UPDATE */
+ if (new_tuple != NULL) { /* INSERT, UPDATE */
int64_t value;
if (tuple_field_i64(new_tuple, BOX_SEQUENCE_DATA_FIELD_VALUE,
&value) != 0)
return -1;
if (sequence_set(seq, value) != 0)
return -1;
- } else { /* DELETE */
+ } else { /* DELETE */
/*
* A sequence isn't supposed to roll back to the old
* value if the transaction it was used in is aborted
@@ -4586,7 +4583,7 @@ on_replace_dd_sequence_data(struct trigger * /* trigger */, void *event)
* on rollback.
*/
struct trigger *on_rollback = txn_alter_trigger_new(
- on_drop_sequence_data_rollback, old_tuple);
+ on_drop_sequence_data_rollback, old_tuple);
if (on_rollback == NULL)
return -1;
txn_stmt_on_rollback(stmt, on_rollback);
@@ -4631,8 +4628,8 @@ sequence_field_from_tuple(struct space *space, struct tuple *tuple,
if (path_raw != NULL) {
path = (char *)malloc(path_len + 1);
if (path == NULL) {
- diag_set(OutOfMemory, path_len + 1,
- "malloc", "sequence path");
+ diag_set(OutOfMemory, path_len + 1, "malloc",
+ "sequence path");
return -1;
}
memcpy(path, path_raw, path_len);
@@ -4657,7 +4654,7 @@ set_space_sequence(struct trigger *trigger, void * /* event */)
return -1;
bool is_generated;
if (tuple_field_bool(tuple, BOX_SPACE_SEQUENCE_FIELD_IS_GENERATED,
- &is_generated) != 0)
+ &is_generated) != 0)
return -1;
struct space *space = space_by_id(space_id);
assert(space != NULL);
@@ -4705,9 +4702,10 @@ clear_space_sequence(struct trigger *trigger, void * /* event */)
static int
on_replace_dd_space_sequence(struct trigger * /* trigger */, void *event)
{
- struct txn *txn = (struct txn *) event;
+ struct txn *txn = (struct txn *)event;
struct txn_stmt *stmt = txn_current_stmt(txn);
- struct tuple *tuple = stmt->new_tuple ? stmt->new_tuple : stmt->old_tuple;
+ struct tuple *tuple = stmt->new_tuple ? stmt->new_tuple :
+ stmt->old_tuple;
uint32_t space_id;
if (tuple_field_u32(tuple, BOX_SPACE_SEQUENCE_FIELD_ID, &space_id) != 0)
return -1;
@@ -4724,7 +4722,8 @@ on_replace_dd_space_sequence(struct trigger * /* trigger */, void *event)
return -1;
struct sequence *seq = sequence_by_id(sequence_id);
if (seq == NULL) {
- diag_set(ClientError, ER_NO_SUCH_SEQUENCE, int2str(sequence_id));
+ diag_set(ClientError, ER_NO_SUCH_SEQUENCE,
+ int2str(sequence_id));
return -1;
}
if (stmt->new_tuple != NULL && stmt->old_tuple != NULL) {
@@ -4741,39 +4740,38 @@ on_replace_dd_space_sequence(struct trigger * /* trigger */, void *event)
/* Check we have the correct access type on the sequence. * */
if (is_generated || !stmt->new_tuple) {
- if (access_check_ddl(seq->def->name, seq->def->id, seq->def->uid,
- SC_SEQUENCE, priv_type) != 0)
+ if (access_check_ddl(seq->def->name, seq->def->id,
+ seq->def->uid, SC_SEQUENCE,
+ priv_type) != 0)
return -1;
} else {
/*
* In case user wants to attach an existing sequence,
* check that it has read and write access.
*/
- if (access_check_ddl(seq->def->name, seq->def->id, seq->def->uid,
- SC_SEQUENCE, PRIV_R) != 0)
+ if (access_check_ddl(seq->def->name, seq->def->id,
+ seq->def->uid, SC_SEQUENCE, PRIV_R) != 0)
return -1;
- if (access_check_ddl(seq->def->name, seq->def->id, seq->def->uid,
- SC_SEQUENCE, PRIV_W) != 0)
+ if (access_check_ddl(seq->def->name, seq->def->id,
+ seq->def->uid, SC_SEQUENCE, PRIV_W) != 0)
return -1;
}
/** Check we have alter access on space. */
if (access_check_ddl(space->def->name, space->def->id, space->def->uid,
- SC_SPACE, PRIV_A) != 0)
+ SC_SPACE, PRIV_A) != 0)
return -1;
- if (stmt->new_tuple != NULL) { /* INSERT, UPDATE */
+ if (stmt->new_tuple != NULL) { /* INSERT, UPDATE */
char *sequence_path;
uint32_t sequence_fieldno;
if (sequence_field_from_tuple(space, tuple, &sequence_path,
&sequence_fieldno) != 0)
return -1;
- auto sequence_path_guard = make_scoped_guard([=] {
- free(sequence_path);
- });
+ auto sequence_path_guard =
+ make_scoped_guard([=] { free(sequence_path); });
if (seq->is_generated) {
- diag_set(ClientError, ER_ALTER_SPACE,
- space_name(space),
- "can not attach generated sequence");
+ diag_set(ClientError, ER_ALTER_SPACE, space_name(space),
+ "can not attach generated sequence");
return -1;
}
struct trigger *on_rollback;
@@ -4781,8 +4779,8 @@ on_replace_dd_space_sequence(struct trigger * /* trigger */, void *event)
on_rollback = txn_alter_trigger_new(set_space_sequence,
stmt->old_tuple);
else
- on_rollback = txn_alter_trigger_new(clear_space_sequence,
- stmt->new_tuple);
+ on_rollback = txn_alter_trigger_new(
+ clear_space_sequence, stmt->new_tuple);
if (on_rollback == NULL)
return -1;
seq->is_generated = is_generated;
@@ -4792,7 +4790,7 @@ on_replace_dd_space_sequence(struct trigger * /* trigger */, void *event)
space->sequence_path = sequence_path;
sequence_path_guard.is_active = false;
txn_stmt_on_rollback(stmt, on_rollback);
- } else { /* DELETE */
+ } else { /* DELETE */
struct trigger *on_rollback;
on_rollback = txn_alter_trigger_new(set_space_sequence,
stmt->old_tuple);
@@ -4820,8 +4818,8 @@ on_create_trigger_rollback(struct trigger *trigger, void * /* event */)
struct sql_trigger *old_trigger = (struct sql_trigger *)trigger->data;
struct sql_trigger *new_trigger;
int rc = sql_trigger_replace(sql_trigger_name(old_trigger),
- sql_trigger_space_id(old_trigger),
- NULL, &new_trigger);
+ sql_trigger_space_id(old_trigger), NULL,
+ &new_trigger);
(void)rc;
assert(rc == 0);
assert(new_trigger == old_trigger);
@@ -4838,8 +4836,8 @@ on_drop_trigger_rollback(struct trigger *trigger, void * /* event */)
if (old_trigger == NULL)
return 0;
if (sql_trigger_replace(sql_trigger_name(old_trigger),
- sql_trigger_space_id(old_trigger),
- old_trigger, &new_trigger) != 0)
+ sql_trigger_space_id(old_trigger), old_trigger,
+ &new_trigger) != 0)
panic("Out of memory on insertion into trigger hash");
assert(new_trigger == NULL);
return 0;
@@ -4855,8 +4853,8 @@ on_replace_trigger_rollback(struct trigger *trigger, void * /* event */)
struct sql_trigger *old_trigger = (struct sql_trigger *)trigger->data;
struct sql_trigger *new_trigger;
if (sql_trigger_replace(sql_trigger_name(old_trigger),
- sql_trigger_space_id(old_trigger),
- old_trigger, &new_trigger) != 0)
+ sql_trigger_space_id(old_trigger), old_trigger,
+ &new_trigger) != 0)
panic("Out of memory on insertion into trigger hash");
sql_trigger_delete(sql_get(), new_trigger);
return 0;
@@ -4881,7 +4879,7 @@ on_replace_trigger_commit(struct trigger *trigger, void * /* event */)
static int
on_replace_dd_trigger(struct trigger * /* trigger */, void *event)
{
- struct txn *txn = (struct txn *) event;
+ struct txn *txn = (struct txn *)event;
struct txn_stmt *stmt = txn_current_stmt(txn);
struct tuple *old_tuple = stmt->old_tuple;
struct tuple *new_tuple = stmt->new_tuple;
@@ -4895,8 +4893,8 @@ on_replace_dd_trigger(struct trigger * /* trigger */, void *event)
if (old_tuple != NULL && new_tuple == NULL) {
/* DROP trigger. */
uint32_t trigger_name_len;
- const char *trigger_name_src = tuple_field_str(old_tuple,
- BOX_TRIGGER_FIELD_NAME, &trigger_name_len);
+ const char *trigger_name_src = tuple_field_str(
+ old_tuple, BOX_TRIGGER_FIELD_NAME, &trigger_name_len);
if (trigger_name_src == NULL)
return -1;
uint32_t space_id;
@@ -4922,12 +4920,12 @@ on_replace_dd_trigger(struct trigger * /* trigger */, void *event)
} else {
/* INSERT, REPLACE trigger. */
uint32_t trigger_name_len;
- const char *trigger_name_src = tuple_field_str(new_tuple,
- BOX_TRIGGER_FIELD_NAME, &trigger_name_len);
+ const char *trigger_name_src = tuple_field_str(
+ new_tuple, BOX_TRIGGER_FIELD_NAME, &trigger_name_len);
if (trigger_name_src == NULL)
return -1;
- const char *space_opts = tuple_field_with_type(new_tuple,
- BOX_TRIGGER_FIELD_OPTS,MP_MAP);
+ const char *space_opts = tuple_field_with_type(
+ new_tuple, BOX_TRIGGER_FIELD_OPTS, MP_MAP);
if (space_opts == NULL)
return -1;
struct space_opts opts;
@@ -4939,17 +4937,16 @@ on_replace_dd_trigger(struct trigger * /* trigger */, void *event)
if (new_trigger == NULL)
return -1;
- auto new_trigger_guard = make_scoped_guard([=] {
- sql_trigger_delete(sql_get(), new_trigger);
- });
+ auto new_trigger_guard = make_scoped_guard(
+ [=] { sql_trigger_delete(sql_get(), new_trigger); });
const char *trigger_name = sql_trigger_name(new_trigger);
if (strlen(trigger_name) != trigger_name_len ||
- memcmp(trigger_name_src, trigger_name,
- trigger_name_len) != 0) {
+ memcmp(trigger_name_src, trigger_name, trigger_name_len) !=
+ 0) {
diag_set(ClientError, ER_SQL_EXECUTE,
- "trigger name does not match extracted "
- "from SQL");
+ "trigger name does not match extracted "
+ "from SQL");
return -1;
}
uint32_t space_id;
@@ -4958,8 +4955,8 @@ on_replace_dd_trigger(struct trigger * /* trigger */, void *event)
return -1;
if (space_id != sql_trigger_space_id(new_trigger)) {
diag_set(ClientError, ER_SQL_EXECUTE,
- "trigger space_id does not match the value "
- "resolved on AST building from SQL");
+ "trigger space_id does not match the value "
+ "resolved on AST building from SQL");
return -1;
}
@@ -5003,33 +5000,32 @@ decode_fk_links(struct tuple *tuple, uint32_t *out_count,
const char *constraint_name, uint32_t constraint_len,
uint32_t errcode)
{
- const char *parent_cols = tuple_field_with_type(tuple,
- BOX_FK_CONSTRAINT_FIELD_PARENT_COLS, MP_ARRAY);
+ const char *parent_cols = tuple_field_with_type(
+ tuple, BOX_FK_CONSTRAINT_FIELD_PARENT_COLS, MP_ARRAY);
if (parent_cols == NULL)
return NULL;
uint32_t count = mp_decode_array(&parent_cols);
if (count == 0) {
diag_set(ClientError, errcode,
- tt_cstr(constraint_name, constraint_len),
- "at least one link must be specified");
+ tt_cstr(constraint_name, constraint_len),
+ "at least one link must be specified");
return NULL;
}
- const char *child_cols = tuple_field_with_type(tuple,
- BOX_FK_CONSTRAINT_FIELD_CHILD_COLS, MP_ARRAY);
+ const char *child_cols = tuple_field_with_type(
+ tuple, BOX_FK_CONSTRAINT_FIELD_CHILD_COLS, MP_ARRAY);
if (child_cols == NULL)
return NULL;
if (mp_decode_array(&child_cols) != count) {
diag_set(ClientError, errcode,
- tt_cstr(constraint_name, constraint_len),
- "number of referenced and referencing fields "
- "must be the same");
+ tt_cstr(constraint_name, constraint_len),
+ "number of referenced and referencing fields "
+ "must be the same");
return NULL;
}
*out_count = count;
size_t size;
- struct field_link *region_links =
- region_alloc_array(&fiber()->gc, typeof(region_links[0]), count,
- &size);
+ struct field_link *region_links = region_alloc_array(
+ &fiber()->gc, typeof(region_links[0]), count, &size);
if (region_links == NULL) {
diag_set(OutOfMemory, size, "region_alloc_array",
"region_links");
@@ -5040,9 +5036,9 @@ decode_fk_links(struct tuple *tuple, uint32_t *out_count,
if (mp_typeof(*parent_cols) != MP_UINT ||
mp_typeof(*child_cols) != MP_UINT) {
diag_set(ClientError, errcode,
- tt_cstr(constraint_name, constraint_len),
- tt_sprintf("value of %d link is not unsigned",
- i));
+ tt_cstr(constraint_name, constraint_len),
+ tt_sprintf("value of %d link is not unsigned",
+ i));
return NULL;
}
region_links[i].parent_field = mp_decode_uint(&parent_cols);
@@ -5056,31 +5052,31 @@ static struct fk_constraint_def *
fk_constraint_def_new_from_tuple(struct tuple *tuple, uint32_t errcode)
{
uint32_t name_len;
- const char *name = tuple_field_str(tuple,
- BOX_FK_CONSTRAINT_FIELD_NAME, &name_len);
+ const char *name =
+ tuple_field_str(tuple, BOX_FK_CONSTRAINT_FIELD_NAME, &name_len);
if (name == NULL)
return NULL;
if (name_len > BOX_NAME_MAX) {
diag_set(ClientError, errcode,
- tt_cstr(name, BOX_INVALID_NAME_MAX),
- "constraint name is too long");
+ tt_cstr(name, BOX_INVALID_NAME_MAX),
+ "constraint name is too long");
return NULL;
}
if (identifier_check(name, name_len) != 0)
return NULL;
uint32_t link_count;
- struct field_link *links = decode_fk_links(tuple, &link_count, name,
- name_len, errcode);
+ struct field_link *links =
+ decode_fk_links(tuple, &link_count, name, name_len, errcode);
if (links == NULL)
return NULL;
uint32_t links_offset;
- size_t fk_def_sz = fk_constraint_def_sizeof(link_count, name_len,
- &links_offset);
+ size_t fk_def_sz =
+ fk_constraint_def_sizeof(link_count, name_len, &links_offset);
struct fk_constraint_def *fk_def =
- (struct fk_constraint_def *) malloc(fk_def_sz);
+ (struct fk_constraint_def *)malloc(fk_def_sz);
if (fk_def == NULL) {
diag_set(OutOfMemory, fk_def_sz, "malloc",
- "struct fk_constraint_def");
+ "struct fk_constraint_def");
return NULL;
}
auto def_guard = make_scoped_guard([=] { free(fk_def); });
@@ -5090,7 +5086,7 @@ fk_constraint_def_new_from_tuple(struct tuple *tuple, uint32_t errcode)
memcpy(fk_def->links, links, link_count * sizeof(struct field_link));
fk_def->field_count = link_count;
if (tuple_field_u32(tuple, BOX_FK_CONSTRAINT_FIELD_CHILD_ID,
- &(fk_def->child_id )) != 0)
+ &(fk_def->child_id)) != 0)
return NULL;
if (tuple_field_u32(tuple, BOX_FK_CONSTRAINT_FIELD_PARENT_ID,
&(fk_def->parent_id)) != 0)
@@ -5098,36 +5094,36 @@ fk_constraint_def_new_from_tuple(struct tuple *tuple, uint32_t errcode)
if (tuple_field_bool(tuple, BOX_FK_CONSTRAINT_FIELD_DEFERRED,
&(fk_def->is_deferred)) != 0)
return NULL;
- const char *match = tuple_field_str(tuple,
- BOX_FK_CONSTRAINT_FIELD_MATCH, &name_len);
+ const char *match = tuple_field_str(
+ tuple, BOX_FK_CONSTRAINT_FIELD_MATCH, &name_len);
if (match == NULL)
return NULL;
fk_def->match = STRN2ENUM(fk_constraint_match, match, name_len);
if (fk_def->match == fk_constraint_match_MAX) {
diag_set(ClientError, errcode, fk_def->name,
- "unknown MATCH clause");
+ "unknown MATCH clause");
return NULL;
}
- const char *on_delete_action = tuple_field_str(tuple,
- BOX_FK_CONSTRAINT_FIELD_ON_DELETE, &name_len);
+ const char *on_delete_action = tuple_field_str(
+ tuple, BOX_FK_CONSTRAINT_FIELD_ON_DELETE, &name_len);
if (on_delete_action == NULL)
return NULL;
- fk_def->on_delete = STRN2ENUM(fk_constraint_action,
- on_delete_action, name_len);
+ fk_def->on_delete =
+ STRN2ENUM(fk_constraint_action, on_delete_action, name_len);
if (fk_def->on_delete == fk_constraint_action_MAX) {
diag_set(ClientError, errcode, fk_def->name,
- "unknown ON DELETE action");
+ "unknown ON DELETE action");
return NULL;
}
- const char *on_update_action = tuple_field_str(tuple,
- BOX_FK_CONSTRAINT_FIELD_ON_UPDATE, &name_len);
+ const char *on_update_action = tuple_field_str(
+ tuple, BOX_FK_CONSTRAINT_FIELD_ON_UPDATE, &name_len);
if (on_update_action == NULL)
return NULL;
- fk_def->on_update = STRN2ENUM(fk_constraint_action,
- on_update_action, name_len);
+ fk_def->on_update =
+ STRN2ENUM(fk_constraint_action, on_update_action, name_len);
if (fk_def->on_update == fk_constraint_action_MAX) {
diag_set(ClientError, errcode, fk_def->name,
- "unknown ON UPDATE action");
+ "unknown ON UPDATE action");
return NULL;
}
def_guard.is_active = false;
@@ -5186,12 +5182,10 @@ space_reset_fk_constraint_mask(struct space *space)
space->fk_constraint_mask = 0;
struct fk_constraint *fk;
rlist_foreach_entry(fk, &space->child_fk_constraint, in_child_space) {
-
fk_constraint_set_mask(fk, &space->fk_constraint_mask,
FIELD_LINK_CHILD);
}
rlist_foreach_entry(fk, &space->parent_fk_constraint, in_parent_space) {
-
fk_constraint_set_mask(fk, &space->fk_constraint_mask,
FIELD_LINK_PARENT);
}
@@ -5205,7 +5199,7 @@ space_reset_fk_constraint_mask(struct space *space)
static int
on_create_fk_constraint_rollback(struct trigger *trigger, void *event)
{
- (void) event;
+ (void)event;
struct fk_constraint *fk = (struct fk_constraint *)trigger->data;
rlist_del_entry(fk, in_parent_space);
rlist_del_entry(fk, in_child_space);
@@ -5222,13 +5216,12 @@ on_create_fk_constraint_rollback(struct trigger *trigger, void *event)
static int
on_replace_fk_constraint_rollback(struct trigger *trigger, void *event)
{
- (void) event;
+ (void)event;
struct fk_constraint *old_fk = (struct fk_constraint *)trigger->data;
struct space *parent = space_by_id(old_fk->def->parent_id);
struct space *child = space_by_id(old_fk->def->child_id);
- struct fk_constraint *new_fk =
- fk_constraint_remove(&child->child_fk_constraint,
- old_fk->def->name);
+ struct fk_constraint *new_fk = fk_constraint_remove(
+ &child->child_fk_constraint, old_fk->def->name);
fk_constraint_delete(new_fk);
rlist_add_entry(&child->child_fk_constraint, old_fk, in_child_space);
rlist_add_entry(&parent->parent_fk_constraint, old_fk, in_parent_space);
@@ -5241,7 +5234,7 @@ on_replace_fk_constraint_rollback(struct trigger *trigger, void *event)
static int
on_drop_fk_constraint_rollback(struct trigger *trigger, void *event)
{
- (void) event;
+ (void)event;
struct fk_constraint *old_fk = (struct fk_constraint *)trigger->data;
struct space *parent = space_by_id(old_fk->def->parent_id);
struct space *child = space_by_id(old_fk->def->child_id);
@@ -5267,8 +5260,8 @@ on_drop_fk_constraint_rollback(struct trigger *trigger, void *event)
static int
on_drop_or_replace_fk_constraint_commit(struct trigger *trigger, void *event)
{
- (void) event;
- fk_constraint_delete((struct fk_constraint *) trigger->data);
+ (void)event;
+ fk_constraint_delete((struct fk_constraint *)trigger->data);
return 0;
}
@@ -5287,7 +5280,7 @@ fk_constraint_check_dup_links(struct fk_constraint_def *fk_def)
uint32_t parent_field = fk_def->links[i].parent_field;
if (parent_field > 63)
goto slow_check;
- parent_field = ((uint64_t) 1) << parent_field;
+ parent_field = ((uint64_t)1) << parent_field;
if ((field_mask & parent_field) != 0)
goto error;
field_mask |= parent_field;
@@ -5304,7 +5297,7 @@ slow_check:
return 0;
error:
diag_set(ClientError, ER_CREATE_FK_CONSTRAINT, fk_def->name,
- "referenced fields can not contain duplicates");
+ "referenced fields can not contain duplicates");
return -1;
}
@@ -5312,15 +5305,15 @@ error:
static int
on_replace_dd_fk_constraint(struct trigger * /* trigger*/, void *event)
{
- struct txn *txn = (struct txn *) event;
+ struct txn *txn = (struct txn *)event;
struct txn_stmt *stmt = txn_current_stmt(txn);
struct tuple *old_tuple = stmt->old_tuple;
struct tuple *new_tuple = stmt->new_tuple;
if (new_tuple != NULL) {
/* Create or replace foreign key. */
struct fk_constraint_def *fk_def =
- fk_constraint_def_new_from_tuple(new_tuple,
- ER_CREATE_FK_CONSTRAINT);
+ fk_constraint_def_new_from_tuple(
+ new_tuple, ER_CREATE_FK_CONSTRAINT);
if (fk_def == NULL)
return -1;
auto fk_def_guard = make_scoped_guard([=] { free(fk_def); });
@@ -5329,17 +5322,18 @@ on_replace_dd_fk_constraint(struct trigger * /* trigger*/, void *event)
return -1;
if (child_space->def->opts.is_view) {
diag_set(ClientError, ER_CREATE_FK_CONSTRAINT,
- fk_def->name,
- "referencing space can't be VIEW");
+ fk_def->name,
+ "referencing space can't be VIEW");
return -1;
}
- struct space *parent_space = space_cache_find(fk_def->parent_id);
+ struct space *parent_space =
+ space_cache_find(fk_def->parent_id);
if (parent_space == NULL)
return -1;
if (parent_space->def->opts.is_view) {
diag_set(ClientError, ER_CREATE_FK_CONSTRAINT,
- fk_def->name,
- "referenced space can't be VIEW");
+ fk_def->name,
+ "referenced space can't be VIEW");
return -1;
}
/*
@@ -5352,8 +5346,8 @@ on_replace_dd_fk_constraint(struct trigger * /* trigger*/, void *event)
struct index *pk = space_index(child_space, 0);
if (pk != NULL && index_size(pk) > 0) {
diag_set(ClientError, ER_CREATE_FK_CONSTRAINT,
- fk_def->name,
- "referencing space must be empty");
+ fk_def->name,
+ "referencing space must be empty");
return -1;
}
/* Check types of referenced fields. */
@@ -5363,24 +5357,25 @@ on_replace_dd_fk_constraint(struct trigger * /* trigger*/, void *event)
if (child_fieldno >= child_space->def->field_count ||
parent_fieldno >= parent_space->def->field_count) {
diag_set(ClientError, ER_CREATE_FK_CONSTRAINT,
- fk_def->name, "foreign key refers to "
- "nonexistent field");
+ fk_def->name,
+ "foreign key refers to "
+ "nonexistent field");
return -1;
}
struct field_def *child_field =
&child_space->def->fields[child_fieldno];
struct field_def *parent_field =
&parent_space->def->fields[parent_fieldno];
- if (! field_type1_contains_type2(parent_field->type,
- child_field->type)) {
+ if (!field_type1_contains_type2(parent_field->type,
+ child_field->type)) {
diag_set(ClientError, ER_CREATE_FK_CONSTRAINT,
- fk_def->name, "field type mismatch");
+ fk_def->name, "field type mismatch");
return -1;
}
if (child_field->coll_id != parent_field->coll_id) {
diag_set(ClientError, ER_CREATE_FK_CONSTRAINT,
- fk_def->name,
- "field collation mismatch");
+ fk_def->name,
+ "field collation mismatch");
return -1;
}
}
@@ -5402,10 +5397,10 @@ on_replace_dd_fk_constraint(struct trigger * /* trigger*/, void *event)
continue;
uint32_t j;
for (j = 0; j < fk_def->field_count; ++j) {
- if (key_def_find_by_fieldno(idx->def->key_def,
- fk_def->links[j].
- parent_field) ==
- NULL)
+ if (key_def_find_by_fieldno(
+ idx->def->key_def,
+ fk_def->links[j].parent_field) ==
+ NULL)
break;
}
if (j != fk_def->field_count)
@@ -5415,15 +5410,16 @@ on_replace_dd_fk_constraint(struct trigger * /* trigger*/, void *event)
}
if (fk_index == NULL) {
diag_set(ClientError, ER_CREATE_FK_CONSTRAINT,
- fk_def->name, "referenced fields don't "
- "compose unique index");
+ fk_def->name,
+ "referenced fields don't "
+ "compose unique index");
return -1;
}
struct fk_constraint *fk =
- (struct fk_constraint *) malloc(sizeof(*fk));
+ (struct fk_constraint *)malloc(sizeof(*fk));
if (fk == NULL) {
- diag_set(OutOfMemory, sizeof(*fk),
- "malloc", "struct fk_constraint");
+ diag_set(OutOfMemory, sizeof(*fk), "malloc",
+ "struct fk_constraint");
return -1;
}
auto fk_guard = make_scoped_guard([=] { free(fk); });
@@ -5431,43 +5427,41 @@ on_replace_dd_fk_constraint(struct trigger * /* trigger*/, void *event)
fk->def = fk_def;
fk->index_id = fk_index->def->iid;
if (old_tuple == NULL) {
- struct trigger *on_rollback =
- txn_alter_trigger_new(on_create_fk_constraint_rollback,
- fk);
+ struct trigger *on_rollback = txn_alter_trigger_new(
+ on_create_fk_constraint_rollback, fk);
if (on_rollback == NULL)
return -1;
if (space_insert_constraint_id(child_space,
CONSTRAINT_TYPE_FK,
fk_def->name) != 0)
return -1;
- rlist_add_entry(&child_space->child_fk_constraint,
- fk, in_child_space);
- rlist_add_entry(&parent_space->parent_fk_constraint,
- fk, in_parent_space);
+ rlist_add_entry(&child_space->child_fk_constraint, fk,
+ in_child_space);
+ rlist_add_entry(&parent_space->parent_fk_constraint, fk,
+ in_parent_space);
txn_stmt_on_rollback(stmt, on_rollback);
- fk_constraint_set_mask(fk,
- &parent_space->fk_constraint_mask,
- FIELD_LINK_PARENT);
+ fk_constraint_set_mask(
+ fk, &parent_space->fk_constraint_mask,
+ FIELD_LINK_PARENT);
fk_constraint_set_mask(fk,
&child_space->fk_constraint_mask,
FIELD_LINK_CHILD);
} else {
- struct fk_constraint *old_fk =
- fk_constraint_remove(&child_space->child_fk_constraint,
- fk_def->name);
+ struct fk_constraint *old_fk = fk_constraint_remove(
+ &child_space->child_fk_constraint,
+ fk_def->name);
rlist_add_entry(&child_space->child_fk_constraint, fk,
in_child_space);
rlist_add_entry(&parent_space->parent_fk_constraint, fk,
in_parent_space);
- struct trigger *on_rollback =
- txn_alter_trigger_new(on_replace_fk_constraint_rollback,
- old_fk);
+ struct trigger *on_rollback = txn_alter_trigger_new(
+ on_replace_fk_constraint_rollback, old_fk);
if (on_rollback == NULL)
return -1;
txn_stmt_on_rollback(stmt, on_rollback);
- struct trigger *on_commit =
- txn_alter_trigger_new(on_drop_or_replace_fk_constraint_commit,
- old_fk);
+ struct trigger *on_commit = txn_alter_trigger_new(
+ on_drop_or_replace_fk_constraint_commit,
+ old_fk);
if (on_commit == NULL)
return -1;
txn_stmt_on_commit(stmt, on_commit);
@@ -5480,28 +5474,26 @@ on_replace_dd_fk_constraint(struct trigger * /* trigger*/, void *event)
/* Drop foreign key. */
struct fk_constraint_def *fk_def =
fk_constraint_def_new_from_tuple(old_tuple,
- ER_DROP_FK_CONSTRAINT);
+ ER_DROP_FK_CONSTRAINT);
if (fk_def == NULL)
return -1;
auto fk_def_guard = make_scoped_guard([=] { free(fk_def); });
struct space *child_space = space_cache_find(fk_def->child_id);
if (child_space == NULL)
return -1;
- struct space *parent_space = space_cache_find(fk_def->parent_id);
+ struct space *parent_space =
+ space_cache_find(fk_def->parent_id);
if (parent_space == NULL)
return -1;
- struct fk_constraint *old_fk=
- fk_constraint_remove(&child_space->child_fk_constraint,
- fk_def->name);
- struct trigger *on_commit =
- txn_alter_trigger_new(on_drop_or_replace_fk_constraint_commit,
- old_fk);
+ struct fk_constraint *old_fk = fk_constraint_remove(
+ &child_space->child_fk_constraint, fk_def->name);
+ struct trigger *on_commit = txn_alter_trigger_new(
+ on_drop_or_replace_fk_constraint_commit, old_fk);
if (on_commit == NULL)
return -1;
txn_stmt_on_commit(stmt, on_commit);
- struct trigger *on_rollback =
- txn_alter_trigger_new(on_drop_fk_constraint_rollback,
- old_fk);
+ struct trigger *on_rollback = txn_alter_trigger_new(
+ on_drop_fk_constraint_rollback, old_fk);
if (on_rollback == NULL)
return -1;
space_delete_constraint_id(child_space, fk_def->name);
@@ -5518,14 +5510,14 @@ static struct ck_constraint_def *
ck_constraint_def_new_from_tuple(struct tuple *tuple)
{
uint32_t name_len;
- const char *name = tuple_field_str(tuple, BOX_CK_CONSTRAINT_FIELD_NAME,
- &name_len);
+ const char *name =
+ tuple_field_str(tuple, BOX_CK_CONSTRAINT_FIELD_NAME, &name_len);
if (name == NULL)
return NULL;
if (name_len > BOX_NAME_MAX) {
diag_set(ClientError, ER_CREATE_CK_CONSTRAINT,
- tt_cstr(name, BOX_INVALID_NAME_MAX),
- "check constraint name is too long");
+ tt_cstr(name, BOX_INVALID_NAME_MAX),
+ "check constraint name is too long");
return NULL;
}
if (identifier_check(name, name_len) != 0)
@@ -5534,26 +5526,25 @@ ck_constraint_def_new_from_tuple(struct tuple *tuple)
if (tuple_field_u32(tuple, BOX_CK_CONSTRAINT_FIELD_SPACE_ID,
&space_id) != 0)
return NULL;
- const char *language_str = tuple_field_cstr(tuple,
- BOX_CK_CONSTRAINT_FIELD_LANGUAGE);
+ const char *language_str =
+ tuple_field_cstr(tuple, BOX_CK_CONSTRAINT_FIELD_LANGUAGE);
if (language_str == NULL)
return NULL;
enum ck_constraint_language language =
STR2ENUM(ck_constraint_language, language_str);
if (language == ck_constraint_language_MAX) {
diag_set(ClientError, ER_FUNCTION_LANGUAGE, language_str,
- tt_cstr(name, name_len));
+ tt_cstr(name, name_len));
return NULL;
}
uint32_t expr_str_len;
- const char *expr_str = tuple_field_str(tuple,
- BOX_CK_CONSTRAINT_FIELD_CODE, &expr_str_len);
+ const char *expr_str = tuple_field_str(
+ tuple, BOX_CK_CONSTRAINT_FIELD_CODE, &expr_str_len);
if (expr_str == NULL)
return NULL;
bool is_enabled = true;
if (tuple_field_count(tuple) > BOX_CK_CONSTRAINT_FIELD_IS_ENABLED) {
- if (tuple_field_bool(tuple,
- BOX_CK_CONSTRAINT_FIELD_IS_ENABLED,
+ if (tuple_field_bool(tuple, BOX_CK_CONSTRAINT_FIELD_IS_ENABLED,
&is_enabled) != 0)
return NULL;
}
@@ -5627,8 +5618,8 @@ on_replace_ck_constraint_rollback(struct trigger *trigger, void * /* event */)
assert(ck != NULL);
struct space *space = space_by_id(ck->def->space_id);
assert(space != NULL);
- struct ck_constraint *new_ck = space_ck_constraint_by_name(space,
- ck->def->name, strlen(ck->def->name));
+ struct ck_constraint *new_ck = space_ck_constraint_by_name(
+ space, ck->def->name, strlen(ck->def->name));
assert(new_ck != NULL);
rlist_del_entry(new_ck, link);
rlist_add_entry(&space->ck_constraint, ck, link);
@@ -5642,7 +5633,7 @@ on_replace_ck_constraint_rollback(struct trigger *trigger, void * /* event */)
static int
on_replace_dd_ck_constraint(struct trigger * /* trigger*/, void *event)
{
- struct txn *txn = (struct txn *) event;
+ struct txn *txn = (struct txn *)event;
struct txn_stmt *stmt = txn_current_stmt(txn);
struct tuple *old_tuple = stmt->old_tuple;
struct tuple *new_tuple = stmt->new_tuple;
@@ -5661,11 +5652,12 @@ on_replace_dd_ck_constraint(struct trigger * /* trigger*/, void *event)
if (new_tuple != NULL) {
bool is_deferred;
if (tuple_field_bool(new_tuple,
- BOX_CK_CONSTRAINT_FIELD_DEFERRED, &is_deferred) != 0)
+ BOX_CK_CONSTRAINT_FIELD_DEFERRED,
+ &is_deferred) != 0)
return -1;
if (is_deferred) {
diag_set(ClientError, ER_UNSUPPORTED, "Tarantool",
- "deferred ck constraints");
+ "deferred ck constraints");
return -1;
}
/* Create or replace check constraint. */
@@ -5673,9 +5665,8 @@ on_replace_dd_ck_constraint(struct trigger * /* trigger*/, void *event)
ck_constraint_def_new_from_tuple(new_tuple);
if (ck_def == NULL)
return -1;
- auto ck_def_guard = make_scoped_guard([=] {
- ck_constraint_def_delete(ck_def);
- });
+ auto ck_def_guard = make_scoped_guard(
+ [=] { ck_constraint_def_delete(ck_def); });
/*
* A corner case: enabling/disabling an existent
* ck constraint doesn't require the object
@@ -5691,7 +5682,7 @@ on_replace_dd_ck_constraint(struct trigger * /* trigger*/, void *event)
bool is_insert = old_ck_constraint == NULL;
if (!is_insert) {
struct ck_constraint_def *old_def =
- old_ck_constraint->def;
+ old_ck_constraint->def;
assert(old_def->space_id == ck_def->space_id);
assert(strcmp(old_def->name, ck_def->name) == 0);
if (old_def->language == ck_def->language &&
@@ -5708,9 +5699,8 @@ on_replace_dd_ck_constraint(struct trigger * /* trigger*/, void *event)
*/
struct index *pk = space_index(space, 0);
if (pk != NULL && index_size(pk) > 0) {
- diag_set(ClientError, ER_CREATE_CK_CONSTRAINT,
- name,
- "referencing space must be empty");
+ diag_set(ClientError, ER_CREATE_CK_CONSTRAINT, name,
+ "referencing space must be empty");
return -1;
}
struct ck_constraint *new_ck_constraint =
@@ -5718,9 +5708,8 @@ on_replace_dd_ck_constraint(struct trigger * /* trigger*/, void *event)
if (new_ck_constraint == NULL)
return -1;
ck_def_guard.is_active = false;
- auto ck_guard = make_scoped_guard([=] {
- ck_constraint_delete(new_ck_constraint);
- });
+ auto ck_guard = make_scoped_guard(
+ [=] { ck_constraint_delete(new_ck_constraint); });
if (space_add_ck_constraint(space, new_ck_constraint) != 0)
return -1;
if (!is_insert) {
@@ -5728,9 +5717,8 @@ on_replace_dd_ck_constraint(struct trigger * /* trigger*/, void *event)
on_rollback->data = old_ck_constraint;
on_rollback->run = on_replace_ck_constraint_rollback;
} else {
- if (space_insert_constraint_id(space,
- CONSTRAINT_TYPE_CK,
- name) != 0) {
+ if (space_insert_constraint_id(
+ space, CONSTRAINT_TYPE_CK, name) != 0) {
space_remove_ck_constraint(space,
new_ck_constraint);
return -1;
@@ -5745,8 +5733,8 @@ on_replace_dd_ck_constraint(struct trigger * /* trigger*/, void *event)
assert(new_tuple == NULL && old_tuple != NULL);
/* Drop check constraint. */
uint32_t name_len;
- const char *name = tuple_field_str(old_tuple,
- BOX_CK_CONSTRAINT_FIELD_NAME, &name_len);
+ const char *name = tuple_field_str(
+ old_tuple, BOX_CK_CONSTRAINT_FIELD_NAME, &name_len);
if (name == NULL)
return -1;
struct ck_constraint *old_ck_constraint =
@@ -5773,8 +5761,8 @@ on_replace_dd_ck_constraint(struct trigger * /* trigger*/, void *event)
static int
on_replace_dd_func_index(struct trigger *trigger, void *event)
{
- (void) trigger;
- struct txn *txn = (struct txn *) event;
+ (void)trigger;
+ struct txn *txn = (struct txn *)event;
struct txn_stmt *stmt = txn_current_stmt(txn);
struct tuple *old_tuple = stmt->old_tuple;
struct tuple *new_tuple = stmt->new_tuple;
@@ -5794,7 +5782,7 @@ on_replace_dd_func_index(struct trigger *trigger, void *event)
&index_id) != 0)
return -1;
if (tuple_field_u32(new_tuple, BOX_FUNC_INDEX_FUNCTION_ID,
- &fid) != 0)
+ &fid) != 0)
return -1;
space = space_cache_find(space_id);
if (space == NULL)
@@ -5804,15 +5792,16 @@ on_replace_dd_func_index(struct trigger *trigger, void *event)
return -1;
func = func_by_id(fid);
if (func == NULL) {
- diag_set(ClientError, ER_NO_SUCH_FUNCTION, int2str(fid));
+ diag_set(ClientError, ER_NO_SUCH_FUNCTION,
+ int2str(fid));
return -1;
}
if (func_index_check_func(func) != 0)
return -1;
if (index->def->opts.func_id != func->def->fid) {
diag_set(ClientError, ER_WRONG_INDEX_OPTIONS, 0,
- "Function ids defined in _index and "
- "_func_index don't match");
+ "Function ids defined in _index and "
+ "_func_index don't match");
return -1;
}
} else if (old_tuple != NULL && new_tuple == NULL) {
@@ -5833,7 +5822,8 @@ on_replace_dd_func_index(struct trigger *trigger, void *event)
func = NULL;
} else {
assert(old_tuple != NULL && new_tuple != NULL);
- diag_set(ClientError, ER_UNSUPPORTED, "functional index", "alter");
+ diag_set(ClientError, ER_UNSUPPORTED, "functional index",
+ "alter");
return -1;
}
@@ -5847,11 +5837,12 @@ on_replace_dd_func_index(struct trigger *trigger, void *event)
alter = alter_space_new(space);
if (alter == NULL)
return -1;
- auto scoped_guard = make_scoped_guard([=] {alter_space_delete(alter);});
+ auto scoped_guard =
+ make_scoped_guard([=] { alter_space_delete(alter); });
if (alter_space_move_indexes(alter, 0, index->def->iid) != 0)
return -1;
try {
- (void) new RebuildFuncIndex(alter, index->def, func);
+ (void)new RebuildFuncIndex(alter, index->def, func);
} catch (Exception *e) {
return -1;
}
@@ -5859,8 +5850,8 @@ on_replace_dd_func_index(struct trigger *trigger, void *event)
space->index_id_max + 1) != 0)
return -1;
try {
- (void) new MoveCkConstraints(alter);
- (void) new UpdateSchemaVersion(alter);
+ (void)new MoveCkConstraints(alter);
+ (void)new UpdateSchemaVersion(alter);
alter_space_do(stmt, alter);
} catch (Exception *e) {
return -1;
@@ -5870,68 +5861,58 @@ on_replace_dd_func_index(struct trigger *trigger, void *event)
return 0;
}
-struct trigger alter_space_on_replace_space = {
- RLIST_LINK_INITIALIZER, on_replace_dd_space, NULL, NULL
-};
+struct trigger alter_space_on_replace_space = { RLIST_LINK_INITIALIZER,
+ on_replace_dd_space, NULL,
+ NULL };
-struct trigger alter_space_on_replace_index = {
- RLIST_LINK_INITIALIZER, on_replace_dd_index, NULL, NULL
-};
+struct trigger alter_space_on_replace_index = { RLIST_LINK_INITIALIZER,
+ on_replace_dd_index, NULL,
+ NULL };
-struct trigger on_replace_truncate = {
- RLIST_LINK_INITIALIZER, on_replace_dd_truncate, NULL, NULL
-};
+struct trigger on_replace_truncate = { RLIST_LINK_INITIALIZER,
+ on_replace_dd_truncate, NULL, NULL };
-struct trigger on_replace_schema = {
- RLIST_LINK_INITIALIZER, on_replace_dd_schema, NULL, NULL
-};
+struct trigger on_replace_schema = { RLIST_LINK_INITIALIZER,
+ on_replace_dd_schema, NULL, NULL };
-struct trigger on_replace_user = {
- RLIST_LINK_INITIALIZER, on_replace_dd_user, NULL, NULL
-};
+struct trigger on_replace_user = { RLIST_LINK_INITIALIZER, on_replace_dd_user,
+ NULL, NULL };
-struct trigger on_replace_func = {
- RLIST_LINK_INITIALIZER, on_replace_dd_func, NULL, NULL
-};
+struct trigger on_replace_func = { RLIST_LINK_INITIALIZER, on_replace_dd_func,
+ NULL, NULL };
-struct trigger on_replace_collation = {
- RLIST_LINK_INITIALIZER, on_replace_dd_collation, NULL, NULL
-};
+struct trigger on_replace_collation = { RLIST_LINK_INITIALIZER,
+ on_replace_dd_collation, NULL, NULL };
-struct trigger on_replace_priv = {
- RLIST_LINK_INITIALIZER, on_replace_dd_priv, NULL, NULL
-};
+struct trigger on_replace_priv = { RLIST_LINK_INITIALIZER, on_replace_dd_priv,
+ NULL, NULL };
-struct trigger on_replace_cluster = {
- RLIST_LINK_INITIALIZER, on_replace_dd_cluster, NULL, NULL
-};
+struct trigger on_replace_cluster = { RLIST_LINK_INITIALIZER,
+ on_replace_dd_cluster, NULL, NULL };
-struct trigger on_replace_sequence = {
- RLIST_LINK_INITIALIZER, on_replace_dd_sequence, NULL, NULL
-};
+struct trigger on_replace_sequence = { RLIST_LINK_INITIALIZER,
+ on_replace_dd_sequence, NULL, NULL };
-struct trigger on_replace_sequence_data = {
- RLIST_LINK_INITIALIZER, on_replace_dd_sequence_data, NULL, NULL
-};
+struct trigger on_replace_sequence_data = { RLIST_LINK_INITIALIZER,
+ on_replace_dd_sequence_data, NULL,
+ NULL };
-struct trigger on_replace_space_sequence = {
- RLIST_LINK_INITIALIZER, on_replace_dd_space_sequence, NULL, NULL
-};
+struct trigger on_replace_space_sequence = { RLIST_LINK_INITIALIZER,
+ on_replace_dd_space_sequence, NULL,
+ NULL };
-struct trigger on_replace_trigger = {
- RLIST_LINK_INITIALIZER, on_replace_dd_trigger, NULL, NULL
-};
+struct trigger on_replace_trigger = { RLIST_LINK_INITIALIZER,
+ on_replace_dd_trigger, NULL, NULL };
-struct trigger on_replace_fk_constraint = {
- RLIST_LINK_INITIALIZER, on_replace_dd_fk_constraint, NULL, NULL
-};
+struct trigger on_replace_fk_constraint = { RLIST_LINK_INITIALIZER,
+ on_replace_dd_fk_constraint, NULL,
+ NULL };
-struct trigger on_replace_ck_constraint = {
- RLIST_LINK_INITIALIZER, on_replace_dd_ck_constraint, NULL, NULL
-};
+struct trigger on_replace_ck_constraint = { RLIST_LINK_INITIALIZER,
+ on_replace_dd_ck_constraint, NULL,
+ NULL };
-struct trigger on_replace_func_index = {
- RLIST_LINK_INITIALIZER, on_replace_dd_func_index, NULL, NULL
-};
+struct trigger on_replace_func_index = { RLIST_LINK_INITIALIZER,
+ on_replace_dd_func_index, NULL, NULL };
/* vim: set foldmethod=marker */
diff --git a/src/box/applier.cc b/src/box/applier.cc
index 7686d6c..06e55e6 100644
--- a/src/box/applier.cc
+++ b/src/box/applier.cc
@@ -63,8 +63,7 @@ static inline void
applier_set_state(struct applier *applier, enum applier_state state)
{
applier->state = state;
- say_debug("=> %s", applier_state_strs[state] +
- strlen("APPLIER_"));
+ say_debug("=> %s", applier_state_strs[state] + strlen("APPLIER_"));
trigger_run_xc(&applier->on_state, applier);
}
@@ -185,9 +184,8 @@ applier_writer_f(va_list ap)
struct xrow_header xrow;
xrow_encode_vclock(&xrow, &replicaset.vclock);
coio_write_xrow(&io, &xrow);
- ERROR_INJECT(ERRINJ_APPLIER_SLOW_ACK, {
- fiber_sleep(0.01);
- });
+ ERROR_INJECT(ERRINJ_APPLIER_SLOW_ACK,
+ { fiber_sleep(0.01); });
/*
* Even if new ACK is requested during the
* write, don't send it again right away.
@@ -385,16 +383,17 @@ applier_connect(struct applier *applier)
coio_read_xrow(coio, ibuf, &row);
if (row.type == IPROTO_OK) {
xrow_decode_ballot_xc(&row, &applier->ballot);
- } else try {
- xrow_decode_error_xc(&row);
- } catch (ClientError *e) {
- if (e->errcode() != ER_UNKNOWN_REQUEST_TYPE)
- e->raise();
- /*
+ } else
+ try {
+ xrow_decode_error_xc(&row);
+ } catch (ClientError *e) {
+ if (e->errcode() != ER_UNKNOWN_REQUEST_TYPE)
+ e->raise();
+ /*
* Master isn't aware of IPROTO_VOTE request.
* It's OK - we can proceed without it.
*/
- }
+ }
applier_set_state(applier, APPLIER_CONNECTED);
@@ -442,7 +441,7 @@ applier_wait_snapshot(struct applier *applier)
xrow_decode_error_xc(&row); /* re-throw error */
} else if (row.type != IPROTO_OK) {
tnt_raise(ClientError, ER_UNKNOWN_REQUEST_TYPE,
- (uint32_t) row.type);
+ (uint32_t)row.type);
}
/*
* Start vclock. The vclock of the checkpoint
@@ -464,7 +463,8 @@ applier_wait_snapshot(struct applier *applier)
if (apply_snapshot_row(&row) != 0)
diag_raise();
if (++row_count % 100000 == 0)
- say_info("%.1fM rows received", row_count / 1e6);
+ say_info("%.1fM rows received",
+ row_count / 1e6);
} else if (row.type == IPROTO_OK) {
if (applier->version_id < version_id(1, 7, 0)) {
/*
@@ -478,10 +478,10 @@ applier_wait_snapshot(struct applier *applier)
}
break; /* end of stream */
} else if (iproto_type_is_error(row.type)) {
- xrow_decode_error_xc(&row); /* rethrow error */
+ xrow_decode_error_xc(&row); /* rethrow error */
} else {
tnt_raise(ClientError, ER_UNKNOWN_REQUEST_TYPE,
- (uint32_t) row.type);
+ (uint32_t)row.type);
}
}
@@ -531,7 +531,8 @@ applier_wait_register(struct applier *applier, uint64_t row_count)
if (apply_final_join_row(&row) != 0)
diag_raise();
if (++row_count % 100000 == 0)
- say_info("%.1fM rows received", row_count / 1e6);
+ say_info("%.1fM rows received",
+ row_count / 1e6);
} else if (row.type == IPROTO_OK) {
/*
* Current vclock. This is not used now,
@@ -540,10 +541,10 @@ applier_wait_register(struct applier *applier, uint64_t row_count)
++row_count;
break; /* end of stream */
} else if (iproto_type_is_error(row.type)) {
- xrow_decode_error_xc(&row); /* rethrow error */
+ xrow_decode_error_xc(&row); /* rethrow error */
} else {
tnt_raise(ClientError, ER_UNKNOWN_REQUEST_TYPE,
- (uint32_t) row.type);
+ (uint32_t)row.type);
}
}
@@ -695,8 +696,7 @@ applier_read_tx(struct applier *applier, struct stailq *rows)
"transaction.");
}
if (tsn != row->tsn)
- tnt_raise(ClientError, ER_UNSUPPORTED,
- "replication",
+ tnt_raise(ClientError, ER_UNSUPPORTED, "replication",
"interleaving transactions");
assert(row->bodycnt <= 1);
@@ -708,8 +708,8 @@ applier_read_tx(struct applier *applier, struct stailq *rows)
* buffer will not be used while the
* transaction is applied.
*/
- void *new_base = region_alloc(&fiber()->gc,
- row->body->iov_len);
+ void *new_base =
+ region_alloc(&fiber()->gc, row->body->iov_len);
if (new_base == NULL)
tnt_raise(OutOfMemory, row->body->iov_len,
"region", "xrow body");
@@ -720,8 +720,8 @@ applier_read_tx(struct applier *applier, struct stailq *rows)
}
stailq_add_tail(rows, &tx_row->next);
- } while (!stailq_last_entry(rows, struct applier_tx_row,
- next)->row.is_commit);
+ } while (!stailq_last_entry(rows, struct applier_tx_row, next)
+ ->row.is_commit);
}
static void
@@ -740,8 +740,7 @@ applier_rollback_by_wal_io(void)
* the journal engine.
*/
diag_set(ClientError, ER_WAL_IO);
- diag_set_error(&replicaset.applier.diag,
- diag_last_error(diag_get()));
+ diag_set_error(&replicaset.applier.diag, diag_last_error(diag_get()));
/* Broadcast the rollback across all appliers. */
trigger_run(&replicaset.applier.on_rollback, NULL);
@@ -753,8 +752,8 @@ applier_rollback_by_wal_io(void)
static int
applier_txn_rollback_cb(struct trigger *trigger, void *event)
{
- (void) trigger;
- struct txn *txn = (struct txn *) event;
+ (void)trigger;
+ struct txn *txn = (struct txn *)event;
/*
* Synchronous transaction rollback due to receiving a
* ROLLBACK entry is a normal event and requires no
@@ -768,8 +767,8 @@ applier_txn_rollback_cb(struct trigger *trigger, void *event)
static int
applier_txn_wal_write_cb(struct trigger *trigger, void *event)
{
- (void) trigger;
- (void) event;
+ (void)trigger;
+ (void)event;
/* Broadcast the WAL write across all appliers. */
trigger_run(&replicaset.applier.on_wal_write, NULL);
return 0;
@@ -777,7 +776,7 @@ applier_txn_wal_write_cb(struct trigger *trigger, void *event)
struct synchro_entry {
/** Encoded form of a synchro record. */
- struct synchro_body_bin body_bin;
+ struct synchro_body_bin body_bin;
/** xrow to write, used by the journal engine. */
struct xrow_header row;
@@ -818,8 +817,7 @@ apply_synchro_row_cb(struct journal_entry *entry)
* the journal engine in async write way.
*/
static struct synchro_entry *
-synchro_entry_new(struct xrow_header *applier_row,
- struct synchro_request *req)
+synchro_entry_new(struct xrow_header *applier_row, struct synchro_request *req)
{
struct synchro_entry *entry;
size_t size = sizeof(*entry) + sizeof(struct xrow_header *);
@@ -884,7 +882,8 @@ applier_handle_raft(struct applier *applier, struct xrow_header *row)
{
assert(iproto_type_is_raft_request(row->type));
if (applier->instance_id == 0) {
- diag_set(ClientError, ER_PROTOCOL, "Can't apply a Raft request "
+ diag_set(ClientError, ER_PROTOCOL,
+ "Can't apply a Raft request "
"from an instance without an ID");
return -1;
}
@@ -917,8 +916,8 @@ applier_apply_tx(struct applier *applier, struct stailq *rows)
*/
if (!raft_is_source_allowed(applier->instance_id))
return 0;
- struct xrow_header *first_row = &stailq_first_entry(rows,
- struct applier_tx_row, next)->row;
+ struct xrow_header *first_row =
+ &stailq_first_entry(rows, struct applier_tx_row, next)->row;
struct xrow_header *last_row;
last_row = &stailq_last_entry(rows, struct applier_tx_row, next)->row;
struct replica *replica = replica_by_id(first_row->replica_id);
@@ -929,10 +928,10 @@ applier_apply_tx(struct applier *applier, struct stailq *rows)
* that belong to the same server id.
*/
struct latch *latch = (replica ? &replica->order_latch :
- &replicaset.applier.order_latch);
+ &replicaset.applier.order_latch);
latch_lock(latch);
- if (vclock_get(&replicaset.applier.vclock,
- last_row->replica_id) >= last_row->lsn) {
+ if (vclock_get(&replicaset.applier.vclock, last_row->replica_id) >=
+ last_row->lsn) {
latch_unlock(latch);
return 0;
} else if (vclock_get(&replicaset.applier.vclock,
@@ -944,9 +943,9 @@ applier_apply_tx(struct applier *applier, struct stailq *rows)
*/
struct xrow_header *tmp;
while (true) {
- tmp = &stailq_first_entry(rows,
- struct applier_tx_row,
- next)->row;
+ tmp = &stailq_first_entry(rows, struct applier_tx_row,
+ next)
+ ->row;
if (tmp->lsn <= vclock_get(&replicaset.applier.vclock,
tmp->replica_id)) {
stailq_shift(rows);
@@ -981,7 +980,8 @@ applier_apply_tx(struct applier *applier, struct stailq *rows)
latch_unlock(latch);
return -1;
}
- stailq_foreach_entry(item, rows, next) {
+ stailq_foreach_entry(item, rows, next)
+ {
struct xrow_header *row = &item->row;
int res = apply_row(row);
if (res != 0) {
@@ -1017,18 +1017,18 @@ applier_apply_tx(struct applier *applier, struct stailq *rows)
* new changes which local rows may overwrite.
* Raise an error.
*/
- diag_set(ClientError, ER_UNSUPPORTED,
- "Replication", "distributed transactions");
+ diag_set(ClientError, ER_UNSUPPORTED, "Replication",
+ "distributed transactions");
goto rollback;
}
/* We are ready to submit txn to wal. */
struct trigger *on_rollback, *on_wal_write;
size_t size;
- on_rollback = region_alloc_object(&txn->region, typeof(*on_rollback),
- &size);
- on_wal_write = region_alloc_object(&txn->region, typeof(*on_wal_write),
- &size);
+ on_rollback =
+ region_alloc_object(&txn->region, typeof(*on_rollback), &size);
+ on_wal_write =
+ region_alloc_object(&txn->region, typeof(*on_wal_write), &size);
if (on_rollback == NULL || on_wal_write == NULL) {
diag_set(OutOfMemory, size, "region_alloc_object",
"on_rollback/on_wal_write");
@@ -1081,7 +1081,7 @@ applier_signal_ack(struct applier *applier)
static int
applier_on_wal_write(struct trigger *trigger, void *event)
{
- (void) event;
+ (void)event;
struct applier *applier = (struct applier *)trigger->data;
applier_signal_ack(applier);
return 0;
@@ -1093,7 +1093,7 @@ applier_on_wal_write(struct trigger *trigger, void *event)
static int
applier_on_rollback(struct trigger *trigger, void *event)
{
- (void) event;
+ (void)event;
struct applier *applier = (struct applier *)trigger->data;
/* Setup a shared error. */
if (!diag_is_empty(&replicaset.applier.diag)) {
@@ -1133,7 +1133,7 @@ applier_subscribe(struct applier *applier)
if (applier->version_id >= version_id(1, 6, 7)) {
coio_read_xrow(coio, ibuf, &row);
if (iproto_type_is_error(row.type)) {
- xrow_decode_error_xc(&row); /* error */
+ xrow_decode_error_xc(&row); /* error */
} else if (row.type != IPROTO_OK) {
tnt_raise(ClientError, ER_PROTOCOL,
"Invalid response to SUBSCRIBE");
@@ -1147,8 +1147,9 @@ applier_subscribe(struct applier *applier)
* its and master's cluster ids match.
*/
vclock_create(&applier->remote_vclock_at_subscribe);
- xrow_decode_subscribe_response_xc(&row, &cluster_id,
- &applier->remote_vclock_at_subscribe);
+ xrow_decode_subscribe_response_xc(
+ &row, &cluster_id,
+ &applier->remote_vclock_at_subscribe);
applier->instance_id = row.replica_id;
/*
* If master didn't send us its cluster id
@@ -1204,7 +1205,8 @@ applier_subscribe(struct applier *applier)
char name[FIBER_NAME_MAX];
int pos = snprintf(name, sizeof(name), "applierw/");
- uri_format(name + pos, sizeof(name) - pos, &applier->uri, false);
+ uri_format(name + pos, sizeof(name) - pos, &applier->uri,
+ false);
applier->writer = fiber_new_xc(name, applier_writer_f);
fiber_set_joinable(applier->writer, true);
@@ -1254,14 +1256,14 @@ applier_subscribe(struct applier *applier)
* and check applier state.
*/
struct xrow_header *first_row =
- &stailq_first_entry(&rows, struct applier_tx_row,
- next)->row;
+ &stailq_first_entry(&rows, struct applier_tx_row, next)
+ ->row;
raft_process_heartbeat(applier->instance_id);
if (first_row->lsn == 0) {
if (unlikely(iproto_type_is_raft_request(
- first_row->type))) {
- if (applier_handle_raft(applier,
- first_row) != 0)
+ first_row->type))) {
+ if (applier_handle_raft(applier, first_row) !=
+ 0)
diag_raise();
}
applier_signal_ack(applier);
@@ -1381,7 +1383,8 @@ applier_f(va_list ap)
} else if (e->errcode() == ER_SYSTEM) {
/* System error from master instance. */
applier_log_error(applier, e);
- applier_disconnect(applier, APPLIER_DISCONNECTED);
+ applier_disconnect(applier,
+ APPLIER_DISCONNECTED);
goto reconnect;
} else {
/* Unrecoverable errors */
@@ -1448,7 +1451,7 @@ applier_f(va_list ap)
*
* See: https://github.com/tarantool/tarantool/issues/136
*/
-reconnect:
+ reconnect:
fiber_sleep(replication_reconnect_interval());
}
return 0;
@@ -1488,8 +1491,8 @@ applier_stop(struct applier *applier)
struct applier *
applier_new(const char *uri)
{
- struct applier *applier = (struct applier *)
- calloc(1, sizeof(struct applier));
+ struct applier *applier =
+ (struct applier *)calloc(1, sizeof(struct applier));
if (applier == NULL) {
diag_set(OutOfMemory, sizeof(*applier), "malloc",
"struct applier");
@@ -1503,7 +1506,7 @@ applier_new(const char *uri)
int rc = uri_parse(&applier->uri, applier->source);
/* URI checked by box_check_replication() */
assert(rc == 0 && applier->uri.service != NULL);
- (void) rc;
+ (void)rc;
applier->last_row_time = ev_monotonic_now(loop());
rlist_create(&applier->on_state);
@@ -1554,7 +1557,7 @@ struct applier_on_state {
static int
applier_on_state_f(struct trigger *trigger, void *event)
{
- (void) event;
+ (void)event;
struct applier_on_state *on_state =
container_of(trigger, struct applier_on_state, base);
@@ -1573,8 +1576,7 @@ applier_on_state_f(struct trigger *trigger, void *event)
}
static inline void
-applier_add_on_state(struct applier *applier,
- struct applier_on_state *trigger,
+applier_add_on_state(struct applier *applier, struct applier_on_state *trigger,
enum applier_state desired_state)
{
trigger_create(&trigger->base, applier_on_state_f, NULL, NULL);
diff --git a/src/box/applier.h b/src/box/applier.h
index 15ca1fc..d519cee 100644
--- a/src/box/applier.h
+++ b/src/box/applier.h
@@ -47,24 +47,24 @@
enum { APPLIER_SOURCE_MAXLEN = 1024 }; /* enough to fit URI with passwords */
-#define applier_STATE(_) \
- _(APPLIER_OFF, 0) \
- _(APPLIER_CONNECT, 1) \
- _(APPLIER_CONNECTED, 2) \
- _(APPLIER_AUTH, 3) \
- _(APPLIER_READY, 4) \
- _(APPLIER_INITIAL_JOIN, 5) \
- _(APPLIER_FINAL_JOIN, 6) \
- _(APPLIER_JOINED, 7) \
- _(APPLIER_SYNC, 8) \
- _(APPLIER_FOLLOW, 9) \
- _(APPLIER_STOPPED, 10) \
- _(APPLIER_DISCONNECTED, 11) \
- _(APPLIER_LOADING, 12) \
- _(APPLIER_FETCH_SNAPSHOT, 13) \
- _(APPLIER_FETCHED_SNAPSHOT, 14) \
- _(APPLIER_REGISTER, 15) \
- _(APPLIER_REGISTERED, 16) \
+#define applier_STATE(_) \
+ _(APPLIER_OFF, 0) \
+ _(APPLIER_CONNECT, 1) \
+ _(APPLIER_CONNECTED, 2) \
+ _(APPLIER_AUTH, 3) \
+ _(APPLIER_READY, 4) \
+ _(APPLIER_INITIAL_JOIN, 5) \
+ _(APPLIER_FINAL_JOIN, 6) \
+ _(APPLIER_JOINED, 7) \
+ _(APPLIER_SYNC, 8) \
+ _(APPLIER_FOLLOW, 9) \
+ _(APPLIER_STOPPED, 10) \
+ _(APPLIER_DISCONNECTED, 11) \
+ _(APPLIER_LOADING, 12) \
+ _(APPLIER_FETCH_SNAPSHOT, 13) \
+ _(APPLIER_FETCHED_SNAPSHOT, 14) \
+ _(APPLIER_REGISTER, 15) \
+ _(APPLIER_REGISTERED, 16)
/** States for the applier */
ENUM(applier_state, applier_STATE);
diff --git a/src/box/authentication.cc b/src/box/authentication.cc
index a7a3587..e62723a 100644
--- a/src/box/authentication.cc
+++ b/src/box/authentication.cc
@@ -68,7 +68,7 @@ authenticate(const char *user_name, uint32_t len, const char *salt,
if (part_count < 2) {
/* Expected at least: authentication mechanism and data. */
tnt_raise(ClientError, ER_INVALID_MSGPACK,
- "authentication request body");
+ "authentication request body");
}
mp_next(&tuple); /* Skip authentication mechanism. */
if (mp_typeof(*tuple) == MP_STR) {
@@ -81,12 +81,12 @@ authenticate(const char *user_name, uint32_t len, const char *salt,
scramble = mp_decode_bin(&tuple, &scramble_len);
} else {
tnt_raise(ClientError, ER_INVALID_MSGPACK,
- "authentication scramble");
+ "authentication scramble");
}
if (scramble_len != SCRAMBLE_SIZE) {
/* Authentication mechanism, data. */
tnt_raise(ClientError, ER_INVALID_MSGPACK,
- "invalid scramble size");
+ "invalid scramble size");
}
if (scramble_check(scramble, salt, user->def->hash2)) {
@@ -97,7 +97,7 @@ authenticate(const char *user_name, uint32_t len, const char *salt,
}
ok:
/* check and run auth triggers on success */
- if (! rlist_empty(&session_on_auth) &&
+ if (!rlist_empty(&session_on_auth) &&
session_run_on_auth_triggers(&auth_res) != 0)
diag_raise();
credentials_reset(&session->credentials, user);
diff --git a/src/box/authentication.h b/src/box/authentication.h
index 9935e35..0f37c7d 100644
--- a/src/box/authentication.h
+++ b/src/box/authentication.h
@@ -43,7 +43,6 @@ struct on_auth_trigger_ctx {
bool is_authenticated;
};
-
void
authenticate(const char *user_name, uint32_t len, const char *salt,
const char *tuple);
diff --git a/src/box/bind.c b/src/box/bind.c
index d45a0f9..c871e5d 100644
--- a/src/box/bind.c
+++ b/src/box/bind.c
@@ -41,7 +41,7 @@ sql_bind_name(const struct sql_bind *bind)
if (bind->name)
return tt_sprintf("'%.*s'", bind->name_len, bind->name);
else
- return tt_sprintf("%d", (int) bind->pos);
+ return tt_sprintf("%d", (int)bind->pos);
}
int
@@ -132,14 +132,14 @@ sql_bind_list_decode(const char *data, struct sql_bind **out_bind)
return 0;
if (bind_count > SQL_BIND_PARAMETER_MAX) {
diag_set(ClientError, ER_SQL_BIND_PARAMETER_MAX,
- (int) bind_count);
+ (int)bind_count);
return -1;
}
struct region *region = &fiber()->gc;
uint32_t used = region_used(region);
size_t size;
- struct sql_bind *bind = region_alloc_array(region, typeof(bind[0]),
- bind_count, &size);
+ struct sql_bind *bind =
+ region_alloc_array(region, typeof(bind[0]), bind_count, &size);
if (bind == NULL) {
diag_set(OutOfMemory, size, "region_alloc_array", "bind");
return -1;
@@ -155,8 +155,7 @@ sql_bind_list_decode(const char *data, struct sql_bind **out_bind)
}
int
-sql_bind_column(struct sql_stmt *stmt, const struct sql_bind *p,
- uint32_t pos)
+sql_bind_column(struct sql_stmt *stmt, const struct sql_bind *p, uint32_t pos)
{
if (p->name != NULL) {
pos = sql_bind_parameter_lindex(stmt, p->name, p->name_len);
@@ -189,7 +188,7 @@ sql_bind_column(struct sql_stmt *stmt, const struct sql_bind *p,
case MP_NIL:
return sql_bind_null(stmt, pos);
case MP_BIN:
- return sql_bind_blob64(stmt, pos, (const void *) p->s, p->bytes,
+ return sql_bind_blob64(stmt, pos, (const void *)p->s, p->bytes,
SQL_STATIC);
default:
unreachable();
diff --git a/src/box/bind.h b/src/box/bind.h
index 568c558..58fabd3 100644
--- a/src/box/bind.h
+++ b/src/box/bind.h
@@ -116,8 +116,7 @@ sql_bind_decode(struct sql_bind *bind, int i, const char **packet);
* @retval -1 SQL error.
*/
int
-sql_bind_column(struct sql_stmt *stmt, const struct sql_bind *p,
- uint32_t pos);
+sql_bind_column(struct sql_stmt *stmt, const struct sql_bind *p, uint32_t pos);
/**
* Bind parameter values to the prepared statement.
diff --git a/src/box/blackhole.c b/src/box/blackhole.c
index 69f1deb..46d449c 100644
--- a/src/box/blackhole.c
+++ b/src/box/blackhole.c
@@ -52,8 +52,8 @@ blackhole_space_execute_replace(struct space *space, struct txn *txn,
struct request *request, struct tuple **result)
{
struct txn_stmt *stmt = txn_current_stmt(txn);
- stmt->new_tuple = tuple_new(space->format, request->tuple,
- request->tuple_end);
+ stmt->new_tuple =
+ tuple_new(space->format, request->tuple, request->tuple_end);
if (stmt->new_tuple == NULL)
return -1;
tuple_ref(stmt->new_tuple);
@@ -146,8 +146,7 @@ blackhole_engine_create_space(struct engine *engine, struct space_def *def,
struct space *space = (struct space *)calloc(1, sizeof(*space));
if (space == NULL) {
- diag_set(OutOfMemory, sizeof(*space),
- "malloc", "struct space");
+ diag_set(OutOfMemory, sizeof(*space), "malloc", "struct space");
return NULL;
}
@@ -163,8 +162,8 @@ blackhole_engine_create_space(struct engine *engine, struct space_def *def,
}
tuple_format_ref(format);
- if (space_create(space, engine, &blackhole_space_vtab,
- def, key_list, format) != 0) {
+ if (space_create(space, engine, &blackhole_space_vtab, def, key_list,
+ format) != 0) {
tuple_format_unref(format);
free(space);
return NULL;
@@ -205,8 +204,8 @@ blackhole_engine_new(void)
{
struct engine *engine = calloc(1, sizeof(*engine));
if (engine == NULL) {
- diag_set(OutOfMemory, sizeof(*engine),
- "malloc", "struct engine");
+ diag_set(OutOfMemory, sizeof(*engine), "malloc",
+ "struct engine");
return NULL;
}
diff --git a/src/box/box.cc b/src/box/box.cc
index 6ec813c..10658d9 100644
--- a/src/box/box.cc
+++ b/src/box/box.cc
@@ -88,7 +88,8 @@ struct rmean *rmean_box;
struct rlist box_on_shutdown = RLIST_HEAD_INITIALIZER(box_on_shutdown);
-static void title(const char *new_status)
+static void
+title(const char *new_status)
{
snprintf(status, sizeof(status), "%s", new_status);
title_set_status(new_status);
@@ -192,10 +193,9 @@ box_check_writable_xc(void)
static void
box_check_memtx_min_tuple_size(ssize_t memtx_min_tuple_size)
{
-
if (memtx_min_tuple_size < 8 || memtx_min_tuple_size > 1048280)
- tnt_raise(ClientError, ER_CFG, "memtx_min_tuple_size",
- "specified value is out of bounds");
+ tnt_raise(ClientError, ER_CFG, "memtx_min_tuple_size",
+ "specified value is out of bounds");
}
int
@@ -254,7 +254,7 @@ box_process_rw(struct request *request, struct space *space,
}
if (res < 0)
goto error;
- fiber_gc();
+ fiber_gc();
}
if (return_tuple) {
tuple_bless(tuple);
@@ -354,10 +354,9 @@ struct recovery_journal {
* min/max LSN of created LSM levels.
*/
static int
-recovery_journal_write(struct journal *base,
- struct journal_entry *entry)
+recovery_journal_write(struct journal *base, struct journal_entry *entry)
{
- struct recovery_journal *journal = (struct recovery_journal *) base;
+ struct recovery_journal *journal = (struct recovery_journal *)base;
entry->res = vclock_sum(journal->vclock);
/*
* Since there're no actual writes, fire a
@@ -401,7 +400,8 @@ apply_wal_row(struct xstream *stream, struct xrow_header *row)
if (request.type != IPROTO_NOP) {
struct space *space = space_cache_find_xc(request.space_id);
if (box_process_rw(&request, space, NULL) != 0) {
- say_error("error applying row: %s", request_str(&request));
+ say_error("error applying row: %s",
+ request_str(&request));
diag_raise();
}
}
@@ -450,7 +450,7 @@ box_check_say(void)
enum say_format format = say_format_by_name(log_format);
if (format == say_format_MAX)
tnt_raise(ClientError, ER_CFG, "log_format",
- "expected 'plain' or 'json'");
+ "expected 'plain' or 'json'");
if (type == SAY_LOGGER_SYSLOG && format == SF_JSON) {
tnt_raise(ClientError, ER_CFG, "log_format",
"'json' can't be used with syslog logger");
@@ -662,15 +662,14 @@ box_check_wal_mode(const char *mode_name)
int mode = strindex(wal_mode_STRS, mode_name, WAL_MODE_MAX);
if (mode == WAL_MODE_MAX)
tnt_raise(ClientError, ER_CFG, "wal_mode", mode_name);
- return (enum wal_mode) mode;
+ return (enum wal_mode)mode;
}
static void
box_check_readahead(int readahead)
{
enum { READAHEAD_MIN = 128, READAHEAD_MAX = 2147483647 };
- if (readahead < (int) READAHEAD_MIN ||
- readahead > (int) READAHEAD_MAX) {
+ if (readahead < (int)READAHEAD_MIN || readahead > (int)READAHEAD_MAX) {
tnt_raise(ClientError, ER_CFG, "readahead",
"specified value is out of bounds");
}
@@ -700,11 +699,11 @@ static ssize_t
box_check_memory_quota(const char *quota_name)
{
int64_t size = cfg_geti64(quota_name);
- if (size >= 0 && (size_t) size <= QUOTA_MAX)
+ if (size >= 0 && (size_t)size <= QUOTA_MAX)
return size;
diag_set(ClientError, ER_CFG, quota_name,
tt_sprintf("must be >= 0 and <= %zu, but it is %lld",
- QUOTA_MAX, size));
+ QUOTA_MAX, size));
return -1;
}
@@ -832,14 +831,13 @@ box_set_election_timeout(void)
static struct applier **
cfg_get_replication(int *p_count)
{
-
/* Use static buffer for result */
static struct applier *appliers[VCLOCK_MAX];
int count = cfg_getarr_size("replication");
if (count >= VCLOCK_MAX) {
tnt_raise(ClientError, ER_CFG, "replication",
- "too many replicas");
+ "too many replicas");
}
for (int i = 0; i < count; i++) {
@@ -871,7 +869,7 @@ box_sync_replication(bool connect_quorum)
if (appliers == NULL)
diag_raise();
- auto guard = make_scoped_guard([=]{
+ auto guard = make_scoped_guard([=] {
for (int i = 0; i < count; i++)
applier_delete(appliers[i]); /* doesn't affect diag */
});
@@ -976,9 +974,8 @@ box_set_replication_anon(void)
return;
if (!anon) {
- auto guard = make_scoped_guard([&]{
- replication_anon = !anon;
- });
+ auto guard =
+ make_scoped_guard([&] { replication_anon = !anon; });
/* Turn anonymous instance into a normal one. */
replication_anon = anon;
/*
@@ -1021,7 +1018,6 @@ box_set_replication_anon(void)
"cannot be turned on after bootstrap"
" has finished");
}
-
}
void
@@ -1046,16 +1042,17 @@ box_clear_synchro_queue(void)
if (!txn_limbo_is_empty(&txn_limbo)) {
int64_t lsns[VCLOCK_MAX];
int len = 0;
- const struct vclock *vclock;
- replicaset_foreach(replica) {
+ const struct vclock *vclock;
+ replicaset_foreach(replica)
+ {
if (replica->relay != NULL &&
relay_get_state(replica->relay) != RELAY_OFF &&
!replica->anon) {
assert(!tt_uuid_is_equal(&INSTANCE_UUID,
&replica->uuid));
vclock = relay_vclock(replica->relay);
- int64_t lsn = vclock_get(vclock,
- former_leader_id);
+ int64_t lsn =
+ vclock_get(vclock, former_leader_id);
lsns[len++] = lsn;
}
}
@@ -1108,11 +1105,11 @@ box_set_snap_io_rate_limit(void)
memtx = (struct memtx_engine *)engine_by_name("memtx");
assert(memtx != NULL);
memtx_engine_set_snap_io_rate_limit(memtx,
- cfg_getd("snap_io_rate_limit"));
+ cfg_getd("snap_io_rate_limit"));
struct engine *vinyl = engine_by_name("vinyl");
assert(vinyl != NULL);
vinyl_engine_set_snap_io_rate_limit(vinyl,
- cfg_getd("snap_io_rate_limit"));
+ cfg_getd("snap_io_rate_limit"));
}
void
@@ -1134,7 +1131,7 @@ box_set_memtx_max_tuple_size(void)
memtx = (struct memtx_engine *)engine_by_name("memtx");
assert(memtx != NULL);
memtx_engine_set_max_tuple_size(memtx,
- cfg_geti("memtx_max_tuple_size"));
+ cfg_geti("memtx_max_tuple_size"));
}
void
@@ -1194,7 +1191,7 @@ box_set_vinyl_max_tuple_size(void)
struct engine *vinyl = engine_by_name("vinyl");
assert(vinyl != NULL);
vinyl_engine_set_max_tuple_size(vinyl,
- cfg_geti("vinyl_max_tuple_size"));
+ cfg_geti("vinyl_max_tuple_size"));
}
void
@@ -1210,7 +1207,7 @@ box_set_vinyl_timeout(void)
{
struct engine *vinyl = engine_by_name("vinyl");
assert(vinyl != NULL);
- vinyl_engine_set_timeout(vinyl, cfg_getd("vinyl_timeout"));
+ vinyl_engine_set_timeout(vinyl, cfg_getd("vinyl_timeout"));
}
void
@@ -1220,7 +1217,7 @@ box_set_net_msg_max(void)
iproto_set_msg_max(new_iproto_msg_max);
fiber_pool_set_max_size(&tx_fiber_pool,
new_iproto_msg_max *
- IPROTO_FIBER_POOL_SIZE_FACTOR);
+ IPROTO_FIBER_POOL_SIZE_FACTOR);
}
int
@@ -1317,7 +1314,7 @@ box_space_id_by_name(const char *name, uint32_t len)
if (len > BOX_NAME_MAX)
return BOX_ID_NIL;
uint32_t size = mp_sizeof_array(1) + mp_sizeof_str(len);
- char *begin = (char *) region_alloc(&fiber()->gc, size);
+ char *begin = (char *)region_alloc(&fiber()->gc, size);
if (begin == NULL) {
diag_set(OutOfMemory, size, "region_alloc", "begin");
return BOX_ID_NIL;
@@ -1332,7 +1329,7 @@ box_space_id_by_name(const char *name, uint32_t len)
if (tuple == NULL)
return BOX_ID_NIL;
uint32_t result = BOX_ID_NIL;
- (void) tuple_field_u32(tuple, BOX_SPACE_FIELD_ID, &result);
+ (void)tuple_field_u32(tuple, BOX_SPACE_FIELD_ID, &result);
return result;
}
@@ -1343,7 +1340,7 @@ box_index_id_by_name(uint32_t space_id, const char *name, uint32_t len)
return BOX_ID_NIL;
uint32_t size = mp_sizeof_array(2) + mp_sizeof_uint(space_id) +
mp_sizeof_str(len);
- char *begin = (char *) region_alloc(&fiber()->gc, size);
+ char *begin = (char *)region_alloc(&fiber()->gc, size);
if (begin == NULL) {
diag_set(OutOfMemory, size, "region_alloc", "begin");
return BOX_ID_NIL;
@@ -1359,7 +1356,7 @@ box_index_id_by_name(uint32_t space_id, const char *name, uint32_t len)
if (tuple == NULL)
return BOX_ID_NIL;
uint32_t result = BOX_ID_NIL;
- (void) tuple_field_u32(tuple, BOX_INDEX_FIELD_ID, &result);
+ (void)tuple_field_u32(tuple, BOX_INDEX_FIELD_ID, &result);
return result;
}
/** \endcond public */
@@ -1372,16 +1369,14 @@ box_process1(struct request *request, box_tuple_t **result)
if (space == NULL)
return -1;
if (!space_is_temporary(space) &&
- space_group_id(space) != GROUP_LOCAL &&
- box_check_writable() != 0)
+ space_group_id(space) != GROUP_LOCAL && box_check_writable() != 0)
return -1;
return box_process_rw(request, space, result);
}
API_EXPORT int
-box_select(uint32_t space_id, uint32_t index_id,
- int iterator, uint32_t offset, uint32_t limit,
- const char *key, const char *key_end,
+box_select(uint32_t space_id, uint32_t index_id, int iterator, uint32_t offset,
+ uint32_t limit, const char *key, const char *key_end,
struct port *port)
{
(void)key_end;
@@ -1404,7 +1399,7 @@ box_select(uint32_t space_id, uint32_t index_id,
if (index == NULL)
return -1;
- enum iterator_type type = (enum iterator_type) iterator;
+ enum iterator_type type = (enum iterator_type)iterator;
uint32_t part_count = key ? mp_decode_array(&key) : 0;
if (key_validate(index->def, type, key, part_count))
return -1;
@@ -1418,8 +1413,8 @@ box_select(uint32_t space_id, uint32_t index_id,
if (txn_begin_ro_stmt(space, &txn) != 0)
return -1;
- struct iterator *it = index_create_iterator(index, type,
- key, part_count);
+ struct iterator *it =
+ index_create_iterator(index, type, key, part_count);
if (it == NULL) {
txn_rollback_stmt(txn);
return -1;
@@ -1564,8 +1559,8 @@ space_truncate(struct space *space)
ops_buf_end = mp_encode_uint(ops_buf_end, 1);
assert(ops_buf_end < buf + buf_size);
- if (box_upsert(BOX_TRUNCATE_ID, 0, tuple_buf, tuple_buf_end,
- ops_buf, ops_buf_end, 0, NULL) != 0)
+ if (box_upsert(BOX_TRUNCATE_ID, 0, tuple_buf, tuple_buf_end, ops_buf,
+ ops_buf_end, 0, NULL) != 0)
diag_raise();
}
@@ -1585,9 +1580,9 @@ box_truncate(uint32_t space_id)
static int
sequence_data_update(uint32_t seq_id, int64_t value)
{
- size_t tuple_buf_size = (mp_sizeof_array(2) +
- 2 * mp_sizeof_uint(UINT64_MAX));
- char *tuple_buf = (char *) region_alloc(&fiber()->gc, tuple_buf_size);
+ size_t tuple_buf_size =
+ (mp_sizeof_array(2) + 2 * mp_sizeof_uint(UINT64_MAX));
+ char *tuple_buf = (char *)region_alloc(&fiber()->gc, tuple_buf_size);
if (tuple_buf == NULL) {
diag_set(OutOfMemory, tuple_buf_size, "region", "tuple");
return -1;
@@ -1595,16 +1590,15 @@ sequence_data_update(uint32_t seq_id, int64_t value)
char *tuple_buf_end = tuple_buf;
tuple_buf_end = mp_encode_array(tuple_buf_end, 2);
tuple_buf_end = mp_encode_uint(tuple_buf_end, seq_id);
- tuple_buf_end = (value < 0 ?
- mp_encode_int(tuple_buf_end, value) :
- mp_encode_uint(tuple_buf_end, value));
+ tuple_buf_end = (value < 0 ? mp_encode_int(tuple_buf_end, value) :
+ mp_encode_uint(tuple_buf_end, value));
assert(tuple_buf_end < tuple_buf + tuple_buf_size);
struct credentials *orig_credentials = effective_user();
fiber_set_user(fiber(), &admin_credentials);
- int rc = box_replace(BOX_SEQUENCE_DATA_ID,
- tuple_buf, tuple_buf_end, NULL);
+ int rc = box_replace(BOX_SEQUENCE_DATA_ID, tuple_buf, tuple_buf_end,
+ NULL);
fiber_set_user(fiber(), orig_credentials);
return rc;
@@ -1615,7 +1609,7 @@ static int
sequence_data_delete(uint32_t seq_id)
{
size_t key_buf_size = mp_sizeof_array(1) + mp_sizeof_uint(UINT64_MAX);
- char *key_buf = (char *) region_alloc(&fiber()->gc, key_buf_size);
+ char *key_buf = (char *)region_alloc(&fiber()->gc, key_buf_size);
if (key_buf == NULL) {
diag_set(OutOfMemory, key_buf_size, "region", "key");
return -1;
@@ -1628,8 +1622,8 @@ sequence_data_delete(uint32_t seq_id)
struct credentials *orig_credentials = effective_user();
fiber_set_user(fiber(), &admin_credentials);
- int rc = box_delete(BOX_SEQUENCE_DATA_ID, 0,
- key_buf, key_buf_end, NULL);
+ int rc =
+ box_delete(BOX_SEQUENCE_DATA_ID, 0, key_buf, key_buf_end, NULL);
fiber_set_user(fiber(), orig_credentials);
return rc;
@@ -1707,8 +1701,8 @@ box_session_push(const char *data, const char *data_end)
static inline void
box_register_replica(uint32_t id, const struct tt_uuid *uuid)
{
- if (boxk(IPROTO_INSERT, BOX_CLUSTER_ID, "[%u%s]",
- (unsigned) id, tt_uuid_str(uuid)) != 0)
+ if (boxk(IPROTO_INSERT, BOX_CLUSTER_ID, "[%u%s]", (unsigned)id,
+ tt_uuid_str(uuid)) != 0)
diag_raise();
assert(replica_by_uuid(uuid)->id == id);
}
@@ -1732,15 +1726,15 @@ box_on_join(const tt_uuid *instance_uuid)
/** Find the largest existing replica id. */
struct space *space = space_cache_find_xc(BOX_CLUSTER_ID);
struct index *index = index_find_system_xc(space, 0);
- struct iterator *it = index_create_iterator_xc(index, ITER_ALL,
- NULL, 0);
+ struct iterator *it =
+ index_create_iterator_xc(index, ITER_ALL, NULL, 0);
IteratorGuard iter_guard(it);
struct tuple *tuple;
/** Assign a new replica id. */
uint32_t replica_id = 1;
while ((tuple = iterator_next_xc(it)) != NULL) {
- if (tuple_field_u32_xc(tuple,
- BOX_CLUSTER_FIELD_ID) != replica_id)
+ if (tuple_field_u32_xc(tuple, BOX_CLUSTER_FIELD_ID) !=
+ replica_id)
break;
replica_id++;
}
@@ -1779,7 +1773,8 @@ box_process_fetch_snapshot(struct ev_io *io, struct xrow_header *header)
"wal_mode = 'none'");
}
- say_info("sending current read-view to replica at %s", sio_socketname(io->fd));
+ say_info("sending current read-view to replica at %s",
+ sio_socketname(io->fd));
/* Send the snapshot data to the instance. */
struct vclock start_vclock;
@@ -1830,14 +1825,14 @@ box_process_register(struct ev_io *io, struct xrow_header *header)
"wal_mode = 'none'");
}
- struct gc_consumer *gc = gc_consumer_register(&replicaset.vclock,
- "replica %s", tt_uuid_str(&instance_uuid));
+ struct gc_consumer *gc = gc_consumer_register(
+ &replicaset.vclock, "replica %s", tt_uuid_str(&instance_uuid));
if (gc == NULL)
diag_raise();
auto gc_guard = make_scoped_guard([&] { gc_consumer_unregister(gc); });
- say_info("registering replica %s at %s",
- tt_uuid_str(&instance_uuid), sio_socketname(io->fd));
+ say_info("registering replica %s at %s", tt_uuid_str(&instance_uuid),
+ sio_socketname(io->fd));
/* See box_process_join() */
int64_t limbo_rollback_count = txn_limbo.rollback_count;
@@ -1974,14 +1969,14 @@ box_process_join(struct ev_io *io, struct xrow_header *header)
* Register the replica as a WAL consumer so that
* it can resume FINAL JOIN where INITIAL JOIN ends.
*/
- struct gc_consumer *gc = gc_consumer_register(&replicaset.vclock,
- "replica %s", tt_uuid_str(&instance_uuid));
+ struct gc_consumer *gc = gc_consumer_register(
+ &replicaset.vclock, "replica %s", tt_uuid_str(&instance_uuid));
if (gc == NULL)
diag_raise();
auto gc_guard = make_scoped_guard([&] { gc_consumer_unregister(gc); });
- say_info("joining replica %s at %s",
- tt_uuid_str(&instance_uuid), sio_socketname(io->fd));
+ say_info("joining replica %s at %s", tt_uuid_str(&instance_uuid),
+ sio_socketname(io->fd));
/*
* In order to join a replica, master has to make sure it
@@ -2102,7 +2097,8 @@ box_process_subscribe(struct ev_io *io, struct xrow_header *header)
tt_uuid_str(&replica_uuid));
}
if (anon && replica != NULL && replica->id != REPLICA_ID_NIL) {
- tnt_raise(ClientError, ER_PROTOCOL, "Can't subscribe an "
+ tnt_raise(ClientError, ER_PROTOCOL,
+ "Can't subscribe an "
"anonymous replica having an ID assigned");
}
if (replica == NULL)
@@ -2147,8 +2143,8 @@ box_process_subscribe(struct ev_io *io, struct xrow_header *header)
row.sync = header->sync;
coio_write_xrow(io, &row);
- say_info("subscribed replica %s at %s",
- tt_uuid_str(&replica_uuid), sio_socketname(io->fd));
+ say_info("subscribed replica %s at %s", tt_uuid_str(&replica_uuid),
+ sio_socketname(io->fd));
say_info("remote vclock %s local vclock %s",
vclock_to_string(&replica_clock), vclock_to_string(&vclock));
if (raft_is_enabled()) {
@@ -2265,12 +2261,10 @@ engine_init()
* so it must be registered first.
*/
struct memtx_engine *memtx;
- memtx = memtx_engine_new_xc(cfg_gets("memtx_dir"),
- cfg_geti("force_recovery"),
- cfg_getd("memtx_memory"),
- cfg_geti("memtx_min_tuple_size"),
- cfg_geti("strip_core"),
- cfg_getd("slab_alloc_factor"));
+ memtx = memtx_engine_new_xc(
+ cfg_gets("memtx_dir"), cfg_geti("force_recovery"),
+ cfg_getd("memtx_memory"), cfg_geti("memtx_min_tuple_size"),
+ cfg_geti("strip_core"), cfg_getd("slab_alloc_factor"));
engine_register((struct engine *)memtx);
box_set_memtx_max_tuple_size();
@@ -2368,15 +2362,15 @@ bootstrap_from_master(struct replica *master)
assert(!tt_uuid_is_nil(&INSTANCE_UUID));
enum applier_state wait_state = replication_anon ?
- APPLIER_FETCH_SNAPSHOT :
- APPLIER_INITIAL_JOIN;
+ APPLIER_FETCH_SNAPSHOT :
+ APPLIER_INITIAL_JOIN;
applier_resume_to_state(applier, wait_state, TIMEOUT_INFINITY);
/*
* Process initial data (snapshot or dirty disk data).
*/
engine_begin_initial_recovery_xc(NULL);
wait_state = replication_anon ? APPLIER_FETCHED_SNAPSHOT :
- APPLIER_FINAL_JOIN;
+ APPLIER_FINAL_JOIN;
applier_resume_to_state(applier, wait_state, TIMEOUT_INFINITY);
/*
@@ -2423,8 +2417,7 @@ bootstrap_from_master(struct replica *master)
*/
static void
bootstrap(const struct tt_uuid *instance_uuid,
- const struct tt_uuid *replicaset_uuid,
- bool *is_bootstrap_leader)
+ const struct tt_uuid *replicaset_uuid, bool *is_bootstrap_leader)
{
/* Initialize instance UUID. */
assert(tt_uuid_is_nil(&INSTANCE_UUID));
@@ -2456,7 +2449,8 @@ bootstrap(const struct tt_uuid *instance_uuid,
struct replica *master = replicaset_leader();
assert(master == NULL || master->applier != NULL);
- if (master != NULL && !tt_uuid_is_equal(&master->uuid, &INSTANCE_UUID)) {
+ if (master != NULL &&
+ !tt_uuid_is_equal(&master->uuid, &INSTANCE_UUID)) {
bootstrap_from_master(master);
/* Check replica set UUID */
if (!tt_uuid_is_nil(replicaset_uuid) &&
@@ -2506,7 +2500,7 @@ local_recovery(const struct tt_uuid *instance_uuid,
* in box.info while local recovery is in progress.
*/
box_vclock = &recovery->vclock;
- auto guard = make_scoped_guard([&]{
+ auto guard = make_scoped_guard([&] {
box_vclock = &replicaset.vclock;
recovery_delete(recovery);
});
@@ -2611,8 +2605,8 @@ local_recovery(const struct tt_uuid *instance_uuid,
static void
tx_prio_cb(struct ev_loop *loop, ev_watcher *watcher, int events)
{
- (void) loop;
- (void) events;
+ (void)loop;
+ (void)events;
struct cbus_endpoint *endpoint = (struct cbus_endpoint *)watcher->data;
cbus_process(endpoint);
}
@@ -2669,7 +2663,8 @@ box_cfg_xc(void)
IPROTO_MSG_MAX_MIN * IPROTO_FIBER_POOL_SIZE_FACTOR,
FIBER_POOL_IDLE_TIMEOUT);
/* Add an extra endpoint for WAL wake up/rollback messages. */
- cbus_endpoint_create(&tx_prio_endpoint, "tx_prio", tx_prio_cb, &tx_prio_endpoint);
+ cbus_endpoint_create(&tx_prio_endpoint, "tx_prio", tx_prio_cb,
+ &tx_prio_endpoint);
rmean_box = rmean_new(iproto_type_strs, IPROTO_TYPE_STAT_MAX);
rmean_error = rmean_new(rmean_error_strings, RMEAN_ERROR_LAST);
@@ -2682,7 +2677,8 @@ box_cfg_xc(void)
iproto_init();
sql_init();
- int64_t wal_max_size = box_check_wal_max_size(cfg_geti64("wal_max_size"));
+ int64_t wal_max_size =
+ box_check_wal_max_size(cfg_geti64("wal_max_size"));
enum wal_mode wal_mode = box_check_wal_mode(cfg_gets("wal_mode"));
if (wal_init(wal_mode, cfg_gets("wal_dir"), wal_max_size,
&INSTANCE_UUID, on_wal_garbage_collection,
@@ -2734,9 +2730,8 @@ box_cfg_xc(void)
struct journal bootstrap_journal;
journal_create(&bootstrap_journal, NULL, bootstrap_journal_write);
journal_set(&bootstrap_journal);
- auto bootstrap_journal_guard = make_scoped_guard([] {
- journal_set(NULL);
- });
+ auto bootstrap_journal_guard =
+ make_scoped_guard([] { journal_set(NULL); });
bool is_bootstrap_leader = false;
if (checkpoint != NULL) {
@@ -2855,7 +2850,7 @@ int
box_checkpoint(void)
{
/* Signal arrived before box.cfg{} */
- if (! is_box_configured)
+ if (!is_box_configured)
return 0;
return gc_checkpoint();
@@ -2870,7 +2865,8 @@ box_backup_start(int checkpoint_idx, box_backup_cb cb, void *cb_arg)
return -1;
}
struct gc_checkpoint *checkpoint;
- gc_foreach_checkpoint_reverse(checkpoint) {
+ gc_foreach_checkpoint_reverse(checkpoint)
+ {
if (checkpoint_idx-- == 0)
break;
}
@@ -2900,7 +2896,7 @@ box_backup_stop(void)
const char *
box_status(void)
{
- return status;
+ return status;
}
static int
diff --git a/src/box/box.h b/src/box/box.h
index 45ff8bb..448f931 100644
--- a/src/box/box.h
+++ b/src/box/box.h
@@ -147,7 +147,8 @@ box_update_ro_summary(void);
* Iterate over all spaces and save them to the
* snapshot file.
*/
-int box_checkpoint(void);
+int
+box_checkpoint(void);
typedef int (*box_backup_cb)(const char *path, void *arg);
@@ -174,7 +175,8 @@ box_backup_stop(void);
/**
* Spit out some basic module status (master/slave, etc.
*/
-const char *box_status(void);
+const char *
+box_status(void);
/**
* Reset box statistics.
@@ -228,36 +230,66 @@ box_process_vote(struct ballot *ballot);
void
box_check_config(void);
-void box_listen(void);
-void box_set_replication(void);
-void box_set_log_level(void);
-void box_set_log_format(void);
-void box_set_io_collect_interval(void);
-void box_set_snap_io_rate_limit(void);
-void box_set_too_long_threshold(void);
-void box_set_readahead(void);
-void box_set_checkpoint_count(void);
-void box_set_checkpoint_interval(void);
-void box_set_checkpoint_wal_threshold(void);
-void box_set_memtx_memory(void);
-void box_set_memtx_max_tuple_size(void);
-void box_set_vinyl_memory(void);
-void box_set_vinyl_max_tuple_size(void);
-void box_set_vinyl_cache(void);
-void box_set_vinyl_timeout(void);
-int box_set_election_is_enabled(void);
-int box_set_election_is_candidate(void);
-int box_set_election_timeout(void);
-void box_set_replication_timeout(void);
-void box_set_replication_connect_timeout(void);
-void box_set_replication_connect_quorum(void);
-void box_set_replication_sync_lag(void);
-int box_set_replication_synchro_quorum(void);
-int box_set_replication_synchro_timeout(void);
-void box_set_replication_sync_timeout(void);
-void box_set_replication_skip_conflict(void);
-void box_set_replication_anon(void);
-void box_set_net_msg_max(void);
+void
+box_listen(void);
+void
+box_set_replication(void);
+void
+box_set_log_level(void);
+void
+box_set_log_format(void);
+void
+box_set_io_collect_interval(void);
+void
+box_set_snap_io_rate_limit(void);
+void
+box_set_too_long_threshold(void);
+void
+box_set_readahead(void);
+void
+box_set_checkpoint_count(void);
+void
+box_set_checkpoint_interval(void);
+void
+box_set_checkpoint_wal_threshold(void);
+void
+box_set_memtx_memory(void);
+void
+box_set_memtx_max_tuple_size(void);
+void
+box_set_vinyl_memory(void);
+void
+box_set_vinyl_max_tuple_size(void);
+void
+box_set_vinyl_cache(void);
+void
+box_set_vinyl_timeout(void);
+int
+box_set_election_is_enabled(void);
+int
+box_set_election_is_candidate(void);
+int
+box_set_election_timeout(void);
+void
+box_set_replication_timeout(void);
+void
+box_set_replication_connect_timeout(void);
+void
+box_set_replication_connect_quorum(void);
+void
+box_set_replication_sync_lag(void);
+int
+box_set_replication_synchro_quorum(void);
+int
+box_set_replication_synchro_timeout(void);
+void
+box_set_replication_sync_timeout(void);
+void
+box_set_replication_skip_conflict(void);
+void
+box_set_replication_anon(void);
+void
+box_set_net_msg_max(void);
int
box_set_prepared_stmt_cache_size(void);
@@ -267,13 +299,13 @@ extern "C" {
typedef struct tuple box_tuple_t;
-void box_clear_synchro_queue(void);
+void
+box_clear_synchro_queue(void);
/* box_select is private and used only by FFI */
API_EXPORT int
-box_select(uint32_t space_id, uint32_t index_id,
- int iterator, uint32_t offset, uint32_t limit,
- const char *key, const char *key_end,
+box_select(uint32_t space_id, uint32_t index_id, int iterator, uint32_t offset,
+ uint32_t limit, const char *key, const char *key_end,
struct port *port);
/** \cond public */
diff --git a/src/box/call.c b/src/box/call.c
index 9c29126..bdc83d0 100644
--- a/src/box/call.c
+++ b/src/box/call.c
@@ -48,7 +48,7 @@ static const struct port_vtab port_msgpack_vtab;
void
port_msgpack_create(struct port *base, const char *data, uint32_t data_sz)
{
- struct port_msgpack *port_msgpack = (struct port_msgpack *) base;
+ struct port_msgpack *port_msgpack = (struct port_msgpack *)base;
memset(port_msgpack, 0, sizeof(*port_msgpack));
port_msgpack->vtab = &port_msgpack_vtab;
port_msgpack->data = data;
@@ -58,7 +58,7 @@ port_msgpack_create(struct port *base, const char *data, uint32_t data_sz)
static const char *
port_msgpack_get_msgpack(struct port *base, uint32_t *size)
{
- struct port_msgpack *port = (struct port_msgpack *) base;
+ struct port_msgpack *port = (struct port_msgpack *)base;
assert(port->vtab == &port_msgpack_vtab);
*size = port->data_sz;
return port->data;
@@ -158,8 +158,9 @@ box_process_call(struct call_request *request, struct port *port)
struct func *func = func_by_name(name, name_len);
if (func != NULL) {
rc = func_call(func, &args, port);
- } else if ((rc = access_check_universe_object(PRIV_X | PRIV_U,
- SC_FUNCTION, tt_cstr(name, name_len))) == 0) {
+ } else if ((rc = access_check_universe_object(
+ PRIV_X | PRIV_U, SC_FUNCTION,
+ tt_cstr(name, name_len))) == 0) {
rc = box_lua_call(name, name_len, &args, port);
}
if (rc != 0)
diff --git a/src/box/checkpoint_schedule.c b/src/box/checkpoint_schedule.c
index d37eba7..4e77f87 100644
--- a/src/box/checkpoint_schedule.c
+++ b/src/box/checkpoint_schedule.c
@@ -35,8 +35,8 @@
#include <stdlib.h>
void
-checkpoint_schedule_cfg(struct checkpoint_schedule *sched,
- double now, double interval)
+checkpoint_schedule_cfg(struct checkpoint_schedule *sched, double now,
+ double interval)
{
sched->interval = interval;
sched->start_time = now + interval;
diff --git a/src/box/checkpoint_schedule.h b/src/box/checkpoint_schedule.h
index 7fbbfe2..dc3b185 100644
--- a/src/box/checkpoint_schedule.h
+++ b/src/box/checkpoint_schedule.h
@@ -55,8 +55,8 @@ struct checkpoint_schedule {
* @interval is the configured interval between checkpoints.
*/
void
-checkpoint_schedule_cfg(struct checkpoint_schedule *sched,
- double now, double interval);
+checkpoint_schedule_cfg(struct checkpoint_schedule *sched, double now,
+ double interval);
/**
* Reset a checkpoint schedule.
diff --git a/src/box/ck_constraint.c b/src/box/ck_constraint.c
index b629a73..ab606ee 100644
--- a/src/box/ck_constraint.c
+++ b/src/box/ck_constraint.c
@@ -40,7 +40,7 @@
#include "sql/vdbeInt.h"
#include "tuple.h"
-const char *ck_constraint_language_strs[] = {"SQL"};
+const char *ck_constraint_language_strs[] = { "SQL" };
struct ck_constraint_def *
ck_constraint_def_new(const char *name, uint32_t name_len, const char *expr_str,
@@ -51,7 +51,7 @@ ck_constraint_def_new(const char *name, uint32_t name_len, const char *expr_str,
uint32_t ck_def_sz = ck_constraint_def_sizeof(name_len, expr_str_len,
&expr_str_offset);
struct ck_constraint_def *ck_def =
- (struct ck_constraint_def *) malloc(ck_def_sz);
+ (struct ck_constraint_def *)malloc(ck_def_sz);
if (ck_def == NULL) {
diag_set(OutOfMemory, ck_def_sz, "malloc", "ck_def");
return NULL;
@@ -131,7 +131,8 @@ ck_constraint_program_compile(struct ck_constraint_def *ck_constraint_def,
sqlVdbeAddOp2(v, OP_Variable, ++parser.nVar, vdbe_field_ref_reg);
/* Generate ck constraint test code. */
vdbe_emit_ck_constraint(&parser, expr, ck_constraint_def->name,
- ck_constraint_def->expr_str, vdbe_field_ref_reg);
+ ck_constraint_def->expr_str,
+ vdbe_field_ref_reg);
/* Clean-up and restore user-defined sql context. */
bool is_error = parser.is_aborted;
@@ -142,10 +143,10 @@ ck_constraint_program_compile(struct ck_constraint_def *ck_constraint_def,
diag_set(ClientError, ER_CREATE_CK_CONSTRAINT,
ck_constraint_def->name,
box_error_message(box_error_last()));
- sql_stmt_finalize((struct sql_stmt *) v);
+ sql_stmt_finalize((struct sql_stmt *)v);
return NULL;
}
- return (struct sql_stmt *) v;
+ return (struct sql_stmt *)v;
}
/**
@@ -167,7 +168,7 @@ ck_constraint_program_run(struct ck_constraint *ck_constraint,
return -1;
}
/* Checks VDBE can't expire, reset expired flag and go. */
- struct Vdbe *v = (struct Vdbe *) ck_constraint->stmt;
+ struct Vdbe *v = (struct Vdbe *)ck_constraint->stmt;
v->expired = 0;
sql_step(ck_constraint->stmt);
/*
@@ -180,8 +181,8 @@ ck_constraint_program_run(struct ck_constraint *ck_constraint,
int
ck_constraint_on_replace_trigger(struct trigger *trigger, void *event)
{
- (void) trigger;
- struct txn *txn = (struct txn *) event;
+ (void)trigger;
+ struct txn *txn = (struct txn *)event;
struct txn_stmt *stmt = txn_current_stmt(txn);
assert(stmt != NULL);
struct tuple *new_tuple = stmt->new_tuple;
@@ -193,8 +194,8 @@ ck_constraint_on_replace_trigger(struct trigger *trigger, void *event)
struct vdbe_field_ref *field_ref;
size_t size = sizeof(field_ref->slots[0]) * space->def->field_count +
sizeof(*field_ref);
- field_ref = (struct vdbe_field_ref *)
- region_aligned_alloc(&fiber()->gc, size, alignof(*field_ref));
+ field_ref = (struct vdbe_field_ref *)region_aligned_alloc(
+ &fiber()->gc, size, alignof(*field_ref));
if (field_ref == NULL) {
diag_set(OutOfMemory, size, "region_aligned_alloc",
"field_ref");
diff --git a/src/box/ck_constraint.h b/src/box/ck_constraint.h
index f8f2465..abf313a 100644
--- a/src/box/ck_constraint.h
+++ b/src/box/ck_constraint.h
@@ -46,8 +46,8 @@ struct trigger;
/** Supported languages of ck constraint. */
enum ck_constraint_language {
- CK_CONSTRAINT_LANGUAGE_SQL,
- ck_constraint_language_MAX,
+ CK_CONSTRAINT_LANGUAGE_SQL,
+ ck_constraint_language_MAX,
};
/** The supported languages strings. */
diff --git a/src/box/coll_id.c b/src/box/coll_id.c
index 5abeaed..fddac49 100644
--- a/src/box/coll_id.c
+++ b/src/box/coll_id.c
@@ -38,7 +38,7 @@ struct coll_id *
coll_id_new(const struct coll_id_def *def)
{
size_t total_len = sizeof(struct coll_id) + def->name_len + 1;
- struct coll_id *coll_id = (struct coll_id *) malloc(total_len);
+ struct coll_id *coll_id = (struct coll_id *)malloc(total_len);
if (coll_id == NULL) {
diag_set(OutOfMemory, total_len, "malloc", "coll_id");
return NULL;
diff --git a/src/box/coll_id_cache.c b/src/box/coll_id_cache.c
index 22673ef..a5c43d6 100644
--- a/src/box/coll_id_cache.c
+++ b/src/box/coll_id_cache.c
@@ -67,8 +67,8 @@ coll_id_cache_destroy(void)
int
coll_id_cache_replace(struct coll_id *coll_id, struct coll_id **replaced_id)
{
- const struct mh_i32ptr_node_t id_node = {coll_id->id, coll_id};
- struct mh_i32ptr_node_t repl_id_node = {0, NULL};
+ const struct mh_i32ptr_node_t id_node = { coll_id->id, coll_id };
+ struct mh_i32ptr_node_t repl_id_node = { 0, NULL };
struct mh_i32ptr_node_t *prepl_id_node = &repl_id_node;
mh_int_t i =
mh_i32ptr_put(coll_id_cache, &id_node, &prepl_id_node, NULL);
@@ -79,8 +79,9 @@ coll_id_cache_replace(struct coll_id *coll_id, struct coll_id **replaced_id)
}
uint32_t hash = mh_strn_hash(coll_id->name, coll_id->name_len);
- const struct mh_strnptr_node_t name_node =
- { coll_id->name, coll_id->name_len, hash, coll_id };
+ const struct mh_strnptr_node_t name_node = { coll_id->name,
+ coll_id->name_len, hash,
+ coll_id };
struct mh_strnptr_node_t repl_name_node = { NULL, 0, 0, NULL };
struct mh_strnptr_node_t *prepl_node_name = &repl_name_node;
if (mh_strnptr_put(coll_cache_name, &name_node, &prepl_node_name,
diff --git a/src/box/coll_id_def.c b/src/box/coll_id_def.c
index 9fe0cda..d518ead 100644
--- a/src/box/coll_id_def.c
+++ b/src/box/coll_id_def.c
@@ -35,35 +35,40 @@ static int64_t
icu_on_off_from_str(const char *str, uint32_t len)
{
return strnindex(coll_icu_on_off_strs + 1, str, len,
- coll_icu_on_off_MAX - 1) + 1;
+ coll_icu_on_off_MAX - 1) +
+ 1;
}
static int64_t
icu_alternate_handling_from_str(const char *str, uint32_t len)
{
return strnindex(coll_icu_alternate_handling_strs + 1, str, len,
- coll_icu_alternate_handling_MAX - 1) + 1;
+ coll_icu_alternate_handling_MAX - 1) +
+ 1;
}
static int64_t
icu_case_first_from_str(const char *str, uint32_t len)
{
return strnindex(coll_icu_case_first_strs + 1, str, len,
- coll_icu_case_first_MAX - 1) + 1;
+ coll_icu_case_first_MAX - 1) +
+ 1;
}
static int64_t
icu_strength_from_str(const char *str, uint32_t len)
{
return strnindex(coll_icu_strength_strs + 1, str, len,
- coll_icu_strength_MAX - 1) + 1;
+ coll_icu_strength_MAX - 1) +
+ 1;
}
const struct opt_def coll_icu_opts_reg[] = {
OPT_DEF_ENUM("french_collation", coll_icu_on_off, struct coll_icu_def,
french_collation, icu_on_off_from_str),
- OPT_DEF_ENUM("alternate_handling", coll_icu_alternate_handling, struct coll_icu_def,
- alternate_handling, icu_alternate_handling_from_str),
+ OPT_DEF_ENUM("alternate_handling", coll_icu_alternate_handling,
+ struct coll_icu_def, alternate_handling,
+ icu_alternate_handling_from_str),
OPT_DEF_ENUM("case_first", coll_icu_case_first, struct coll_icu_def,
case_first, icu_case_first_from_str),
OPT_DEF_ENUM("case_level", coll_icu_on_off, struct coll_icu_def,
diff --git a/src/box/column_mask.h b/src/box/column_mask.h
index 9470fe1..9731b44 100644
--- a/src/box/column_mask.h
+++ b/src/box/column_mask.h
@@ -66,9 +66,9 @@ column_mask_set_fieldno(uint64_t *column_mask, uint32_t fieldno)
* @sa column_mask key_def declaration for
* details.
*/
- *column_mask |= ((uint64_t) 1) << 63;
+ *column_mask |= ((uint64_t)1) << 63;
else
- *column_mask |= ((uint64_t) 1) << fieldno;
+ *column_mask |= ((uint64_t)1) << fieldno;
}
/**
@@ -90,7 +90,7 @@ column_mask_set_range(uint64_t *column_mask, uint32_t first_fieldno_in_range)
*column_mask |= COLUMN_MASK_FULL << first_fieldno_in_range;
} else {
/* A range outside "short" range. */
- *column_mask |= ((uint64_t) 1) << 63;
+ *column_mask |= ((uint64_t)1) << 63;
}
}
@@ -119,7 +119,7 @@ key_update_can_be_skipped(uint64_t key_mask, uint64_t update_mask)
static inline bool
column_mask_fieldno_is_set(uint64_t column_mask, uint32_t fieldno)
{
- uint64_t mask = (uint64_t) 1 << (fieldno < 63 ? fieldno : 63);
+ uint64_t mask = (uint64_t)1 << (fieldno < 63 ? fieldno : 63);
return (column_mask & mask) != 0;
}
diff --git a/src/box/constraint_id.c b/src/box/constraint_id.c
index ba6ed85..1047c7e 100644
--- a/src/box/constraint_id.c
+++ b/src/box/constraint_id.c
@@ -35,10 +35,10 @@
#include "diag.h"
const char *constraint_type_strs[] = {
- [CONSTRAINT_TYPE_PK] = "PRIMARY KEY",
- [CONSTRAINT_TYPE_UNIQUE] = "UNIQUE",
- [CONSTRAINT_TYPE_FK] = "FOREIGN KEY",
- [CONSTRAINT_TYPE_CK] = "CHECK",
+ [CONSTRAINT_TYPE_PK] = "PRIMARY KEY",
+ [CONSTRAINT_TYPE_UNIQUE] = "UNIQUE",
+ [CONSTRAINT_TYPE_FK] = "FOREIGN KEY",
+ [CONSTRAINT_TYPE_CK] = "CHECK",
};
struct constraint_id *
diff --git a/src/box/engine.c b/src/box/engine.c
index 88ed928..63ab517 100644
--- a/src/box/engine.c
+++ b/src/box/engine.c
@@ -43,7 +43,8 @@ RLIST_HEAD(engines);
enum { MAX_ENGINE_COUNT = 10 };
/** Register engine instance. */
-void engine_register(struct engine *engine)
+void
+engine_register(struct engine *engine)
{
static int n_engines;
rlist_add_tail_entry(&engines, engine, link);
@@ -55,7 +56,8 @@ struct engine *
engine_by_name(const char *name)
{
struct engine *e;
- engine_foreach(e) {
+ engine_foreach(e)
+ {
if (strcmp(e->name, name) == 0)
return e;
}
@@ -75,15 +77,15 @@ void
engine_switch_to_ro(void)
{
struct engine *engine;
- engine_foreach(engine)
- engine->vtab->switch_to_ro(engine);
+ engine_foreach(engine) engine->vtab->switch_to_ro(engine);
}
int
engine_bootstrap(void)
{
struct engine *engine;
- engine_foreach(engine) {
+ engine_foreach(engine)
+ {
if (engine->vtab->bootstrap(engine) != 0)
return -1;
}
@@ -94,9 +96,10 @@ int
engine_begin_initial_recovery(const struct vclock *recovery_vclock)
{
struct engine *engine;
- engine_foreach(engine) {
+ engine_foreach(engine)
+ {
if (engine->vtab->begin_initial_recovery(engine,
- recovery_vclock) != 0)
+ recovery_vclock) != 0)
return -1;
}
return 0;
@@ -106,7 +109,8 @@ int
engine_begin_final_recovery(void)
{
struct engine *engine;
- engine_foreach(engine) {
+ engine_foreach(engine)
+ {
if (engine->vtab->begin_final_recovery(engine) != 0)
return -1;
}
@@ -121,7 +125,8 @@ engine_end_recovery(void)
* when the primary key is added, enable all keys.
*/
struct engine *engine;
- engine_foreach(engine) {
+ engine_foreach(engine)
+ {
if (engine->vtab->end_recovery(engine) != 0)
return -1;
}
@@ -132,7 +137,8 @@ int
engine_begin_checkpoint(bool is_scheduled)
{
struct engine *engine;
- engine_foreach(engine) {
+ engine_foreach(engine)
+ {
if (engine->vtab->begin_checkpoint(engine, is_scheduled) < 0)
return -1;
}
@@ -143,11 +149,13 @@ int
engine_commit_checkpoint(const struct vclock *vclock)
{
struct engine *engine;
- engine_foreach(engine) {
+ engine_foreach(engine)
+ {
if (engine->vtab->wait_checkpoint(engine, vclock) < 0)
return -1;
}
- engine_foreach(engine) {
+ engine_foreach(engine)
+ {
engine->vtab->commit_checkpoint(engine, vclock);
}
return 0;
@@ -157,23 +165,22 @@ void
engine_abort_checkpoint(void)
{
struct engine *engine;
- engine_foreach(engine)
- engine->vtab->abort_checkpoint(engine);
+ engine_foreach(engine) engine->vtab->abort_checkpoint(engine);
}
void
engine_collect_garbage(const struct vclock *vclock)
{
struct engine *engine;
- engine_foreach(engine)
- engine->vtab->collect_garbage(engine, vclock);
+ engine_foreach(engine) engine->vtab->collect_garbage(engine, vclock);
}
int
engine_backup(const struct vclock *vclock, engine_backup_cb cb, void *cb_arg)
{
struct engine *engine;
- engine_foreach(engine) {
+ engine_foreach(engine)
+ {
if (engine->vtab->backup(engine, vclock, cb, cb_arg) < 0)
return -1;
}
@@ -191,7 +198,8 @@ engine_prepare_join(struct engine_join_ctx *ctx)
}
int i = 0;
struct engine *engine;
- engine_foreach(engine) {
+ engine_foreach(engine)
+ {
assert(i < MAX_ENGINE_COUNT);
if (engine->vtab->prepare_join(engine, &ctx->array[i]) != 0)
goto fail;
@@ -208,7 +216,8 @@ engine_join(struct engine_join_ctx *ctx, struct xstream *stream)
{
int i = 0;
struct engine *engine;
- engine_foreach(engine) {
+ engine_foreach(engine)
+ {
if (engine->vtab->join(engine, ctx->array[i], stream) != 0)
return -1;
i++;
@@ -221,7 +230,8 @@ engine_complete_join(struct engine_join_ctx *ctx)
{
int i = 0;
struct engine *engine;
- engine_foreach(engine) {
+ engine_foreach(engine)
+ {
if (ctx->array[i] != NULL)
engine->vtab->complete_join(engine, ctx->array[i]);
i++;
@@ -234,16 +244,14 @@ engine_memory_stat(struct engine_memory_stat *stat)
{
memset(stat, 0, sizeof(*stat));
struct engine *engine;
- engine_foreach(engine)
- engine->vtab->memory_stat(engine, stat);
+ engine_foreach(engine) engine->vtab->memory_stat(engine, stat);
}
void
engine_reset_stat(void)
{
struct engine *engine;
- engine_foreach(engine)
- engine->vtab->reset_stat(engine);
+ engine_foreach(engine) engine->vtab->reset_stat(engine);
}
/* {{{ Virtual method stubs */
diff --git a/src/box/engine.h b/src/box/engine.h
index c4da01e..2270c9f 100644
--- a/src/box/engine.h
+++ b/src/box/engine.h
@@ -72,7 +72,8 @@ struct engine_vtab {
void (*shutdown)(struct engine *);
/** Allocate a new space instance. */
struct space *(*create_space)(struct engine *engine,
- struct space_def *def, struct rlist *key_list);
+ struct space_def *def,
+ struct rlist *key_list);
/**
* Freeze a read view to feed to a new replica.
* Setup and return a context that will be used
@@ -140,7 +141,7 @@ struct engine_vtab {
* On remote recovery, it is set to NULL.
*/
int (*begin_initial_recovery)(struct engine *engine,
- const struct vclock *recovery_vclock);
+ const struct vclock *recovery_vclock);
/**
* Notify engine about a start of recovering from WALs
* that could be local WALs during local recovery
@@ -237,7 +238,8 @@ struct engine_join_ctx {
};
/** Register engine engine instance. */
-void engine_register(struct engine *engine);
+void
+engine_register(struct engine *engine);
/** Call a visitor function on every registered engine. */
#define engine_foreach(engine) rlist_foreach_entry(engine, &engines, link)
@@ -380,32 +382,54 @@ engine_reset_stat(void);
/*
* Virtual method stubs.
*/
-int generic_engine_prepare_join(struct engine *, void **);
-int generic_engine_join(struct engine *, void *, struct xstream *);
-void generic_engine_complete_join(struct engine *, void *);
-int generic_engine_begin(struct engine *, struct txn *);
-int generic_engine_begin_statement(struct engine *, struct txn *);
-int generic_engine_prepare(struct engine *, struct txn *);
-void generic_engine_commit(struct engine *, struct txn *);
-void generic_engine_rollback_statement(struct engine *, struct txn *,
- struct txn_stmt *);
-void generic_engine_rollback(struct engine *, struct txn *);
-void generic_engine_switch_to_ro(struct engine *);
-int generic_engine_bootstrap(struct engine *);
-int generic_engine_begin_initial_recovery(struct engine *,
- const struct vclock *);
-int generic_engine_begin_final_recovery(struct engine *);
-int generic_engine_end_recovery(struct engine *);
-int generic_engine_begin_checkpoint(struct engine *, bool);
-int generic_engine_wait_checkpoint(struct engine *, const struct vclock *);
-void generic_engine_commit_checkpoint(struct engine *, const struct vclock *);
-void generic_engine_abort_checkpoint(struct engine *);
-void generic_engine_collect_garbage(struct engine *, const struct vclock *);
-int generic_engine_backup(struct engine *, const struct vclock *,
- engine_backup_cb, void *);
-void generic_engine_memory_stat(struct engine *, struct engine_memory_stat *);
-void generic_engine_reset_stat(struct engine *);
-int generic_engine_check_space_def(struct space_def *);
+int
+generic_engine_prepare_join(struct engine *, void **);
+int
+generic_engine_join(struct engine *, void *, struct xstream *);
+void
+generic_engine_complete_join(struct engine *, void *);
+int
+generic_engine_begin(struct engine *, struct txn *);
+int
+generic_engine_begin_statement(struct engine *, struct txn *);
+int
+generic_engine_prepare(struct engine *, struct txn *);
+void
+generic_engine_commit(struct engine *, struct txn *);
+void
+generic_engine_rollback_statement(struct engine *, struct txn *,
+ struct txn_stmt *);
+void
+generic_engine_rollback(struct engine *, struct txn *);
+void
+generic_engine_switch_to_ro(struct engine *);
+int
+generic_engine_bootstrap(struct engine *);
+int
+generic_engine_begin_initial_recovery(struct engine *, const struct vclock *);
+int
+generic_engine_begin_final_recovery(struct engine *);
+int
+generic_engine_end_recovery(struct engine *);
+int
+generic_engine_begin_checkpoint(struct engine *, bool);
+int
+generic_engine_wait_checkpoint(struct engine *, const struct vclock *);
+void
+generic_engine_commit_checkpoint(struct engine *, const struct vclock *);
+void
+generic_engine_abort_checkpoint(struct engine *);
+void
+generic_engine_collect_garbage(struct engine *, const struct vclock *);
+int
+generic_engine_backup(struct engine *, const struct vclock *, engine_backup_cb,
+ void *);
+void
+generic_engine_memory_stat(struct engine *, struct engine_memory_stat *);
+void
+generic_engine_reset_stat(struct engine *);
+int
+generic_engine_check_space_def(struct space_def *);
#if defined(__cplusplus)
} /* extern "C" */
@@ -421,7 +445,7 @@ engine_find_xc(const char *name)
static inline struct space *
engine_create_space_xc(struct engine *engine, struct space_def *def,
- struct rlist *key_list)
+ struct rlist *key_list)
{
struct space *space = engine_create_space(engine, def, key_list);
if (space == NULL)
diff --git a/src/box/errcode.c b/src/box/errcode.c
index c1cb594..778054f 100644
--- a/src/box/errcode.c
+++ b/src/box/errcode.c
@@ -31,12 +31,7 @@
*/
#include "errcode.h"
-#define ERRCODE_RECORD_MEMBER(s, d) { \
- .errstr = #s, \
- .errdesc = d \
-},
-
-struct errcode_record box_error_codes[box_error_code_MAX] = {
- ERROR_CODES(ERRCODE_RECORD_MEMBER)
-};
+#define ERRCODE_RECORD_MEMBER(s, d) { .errstr = #s, .errdesc = d },
+struct errcode_record box_error_codes[box_error_code_MAX] = { ERROR_CODES(
+ ERRCODE_RECORD_MEMBER) };
diff --git a/src/box/error.cc b/src/box/error.cc
index ca1d73e..9305a2f 100644
--- a/src/box/error.cc
+++ b/src/box/error.cc
@@ -70,8 +70,8 @@ box_error_clear(void)
}
int
-box_error_set(const char *file, unsigned line, uint32_t code,
- const char *fmt, ...)
+box_error_set(const char *file, unsigned line, uint32_t code, const char *fmt,
+ ...)
{
struct error *e = BuildClientError(file, line, ER_UNKNOWN);
ClientError *client_error = type_cast(ClientError, e);
@@ -99,8 +99,8 @@ box_error_new_va(const char *file, unsigned line, uint32_t code,
}
return e;
} else {
- struct error *e = BuildCustomError(file, line, custom_type,
- code);
+ struct error *e =
+ BuildCustomError(file, line, custom_type, code);
CustomError *custom_error = type_cast(CustomError, e);
if (custom_error != NULL) {
error_vformat_msg(e, fmt, ap);
@@ -116,8 +116,8 @@ box_error_new(const char *file, unsigned line, uint32_t code,
{
va_list ap;
va_start(ap, fmt);
- struct error *e = box_error_new_va(file, line, code, custom_type,
- fmt, ap);
+ struct error *e =
+ box_error_new_va(file, line, code, custom_type, fmt, ap);
va_end(ap);
return e;
}
@@ -128,8 +128,8 @@ box_error_add(const char *file, unsigned line, uint32_t code,
{
va_list ap;
va_start(ap, fmt);
- struct error *e = box_error_new_va(file, line, code, custom_type,
- fmt, ap);
+ struct error *e =
+ box_error_new_va(file, line, code, custom_type, fmt, ap);
va_end(ap);
struct diag *d = &fiber()->diag;
@@ -154,9 +154,7 @@ box_error_custom_type(const struct error *e)
struct rmean *rmean_error = NULL;
-const char *rmean_error_strings[RMEAN_ERROR_LAST] = {
- "ERROR"
-};
+const char *rmean_error_strings[RMEAN_ERROR_LAST] = { "ERROR" };
static struct method_info clienterror_methods[] = {
make_method(&type_ClientError, "code", &ClientError::errcode),
@@ -168,16 +166,15 @@ const struct type_info type_ClientError =
ClientError::ClientError(const type_info *type, const char *file, unsigned line,
uint32_t errcode)
- :Exception(type, file, line)
+ : Exception(type, file, line)
{
m_errcode = errcode;
if (rmean_error)
rmean_collect(rmean_error, RMEAN_ERROR, 1);
}
-ClientError::ClientError(const char *file, unsigned line,
- uint32_t errcode, ...)
- :ClientError(&type_ClientError, file, line, errcode)
+ClientError::ClientError(const char *file, unsigned line, uint32_t errcode, ...)
+ : ClientError(&type_ClientError, file, line, errcode)
{
va_list ap;
va_start(ap, errcode);
@@ -208,7 +205,6 @@ ClientError::log() const
tnt_errcode_str(m_errcode));
}
-
uint32_t
ClientError::get_errcode(const struct error *e)
{
@@ -247,26 +243,25 @@ const struct type_info type_XlogGapError =
XlogGapError::XlogGapError(const char *file, unsigned line,
const struct vclock *from, const struct vclock *to)
- : XlogError(&type_XlogGapError, file, line)
+ : XlogError(&type_XlogGapError, file, line)
{
const char *s_from = vclock_to_string(from);
const char *s_to = vclock_to_string(to);
snprintf(errmsg, sizeof(errmsg),
"Missing .xlog file between LSN %lld %s and %lld %s",
- (long long) vclock_sum(from), s_from ? s_from : "",
- (long long) vclock_sum(to), s_to ? s_to : "");
+ (long long)vclock_sum(from), s_from ? s_from : "",
+ (long long)vclock_sum(to), s_to ? s_to : "");
}
-XlogGapError::XlogGapError(const char *file, unsigned line,
- const char *msg)
- : XlogError(&type_XlogGapError, file, line)
+XlogGapError::XlogGapError(const char *file, unsigned line, const char *msg)
+ : XlogError(&type_XlogGapError, file, line)
{
error_format_msg(this, "%s", msg);
}
struct error *
-BuildXlogGapError(const char *file, unsigned line,
- const struct vclock *from, const struct vclock *to)
+BuildXlogGapError(const char *file, unsigned line, const struct vclock *from,
+ const struct vclock *to)
{
try {
return new XlogGapError(file, line, from, to);
@@ -278,34 +273,36 @@ BuildXlogGapError(const char *file, unsigned line,
struct rlist on_access_denied = RLIST_HEAD_INITIALIZER(on_access_denied);
static struct method_info accessdeniederror_methods[] = {
- make_method(&type_AccessDeniedError, "access_type", &AccessDeniedError::access_type),
- make_method(&type_AccessDeniedError, "object_type", &AccessDeniedError::object_type),
- make_method(&type_AccessDeniedError, "object_name", &AccessDeniedError::object_name),
+ make_method(&type_AccessDeniedError, "access_type",
+ &AccessDeniedError::access_type),
+ make_method(&type_AccessDeniedError, "object_type",
+ &AccessDeniedError::object_type),
+ make_method(&type_AccessDeniedError, "object_name",
+ &AccessDeniedError::object_name),
METHODS_SENTINEL
};
-const struct type_info type_AccessDeniedError =
- make_type("AccessDeniedError", &type_ClientError,
- accessdeniederror_methods);
+const struct type_info type_AccessDeniedError = make_type(
+ "AccessDeniedError", &type_ClientError, accessdeniederror_methods);
AccessDeniedError::AccessDeniedError(const char *file, unsigned int line,
const char *access_type,
const char *object_type,
const char *object_name,
- const char *user_name,
- bool run_trigers)
- :ClientError(&type_AccessDeniedError, file, line, ER_ACCESS_DENIED)
+ const char *user_name, bool run_trigers)
+ : ClientError(&type_AccessDeniedError, file, line, ER_ACCESS_DENIED)
{
- error_format_msg(this, tnt_errcode_desc(m_errcode),
- access_type, object_type, object_name, user_name);
+ error_format_msg(this, tnt_errcode_desc(m_errcode), access_type,
+ object_type, object_name, user_name);
- struct on_access_denied_ctx ctx = {access_type, object_type, object_name};
+ struct on_access_denied_ctx ctx = { access_type, object_type,
+ object_name };
/*
* Don't run the triggers when create after marshaling
* through network.
*/
if (run_trigers)
- trigger_run(&on_access_denied, (void *) &ctx);
+ trigger_run(&on_access_denied, (void *)&ctx);
m_object_type = strdup(object_type);
m_access_type = strdup(access_type);
m_object_name = strdup(object_name);
@@ -314,8 +311,7 @@ AccessDeniedError::AccessDeniedError(const char *file, unsigned int line,
struct error *
BuildAccessDeniedError(const char *file, unsigned int line,
const char *access_type, const char *object_type,
- const char *object_name,
- const char *user_name)
+ const char *object_name, const char *user_name)
{
try {
return new AccessDeniedError(file, line, access_type,
@@ -327,7 +323,8 @@ BuildAccessDeniedError(const char *file, unsigned int line,
}
static struct method_info customerror_methods[] = {
- make_method(&type_CustomError, "custom_type", &CustomError::custom_type),
+ make_method(&type_CustomError, "custom_type",
+ &CustomError::custom_type),
METHODS_SENTINEL
};
@@ -336,7 +333,7 @@ const struct type_info type_CustomError =
CustomError::CustomError(const char *file, unsigned int line,
const char *custom_type, uint32_t errcode)
- :ClientError(&type_CustomError, file, line, errcode)
+ : ClientError(&type_CustomError, file, line, errcode)
{
strncpy(m_custom_type, custom_type, sizeof(m_custom_type) - 1);
m_custom_type[sizeof(m_custom_type) - 1] = '\0';
@@ -345,8 +342,8 @@ CustomError::CustomError(const char *file, unsigned int line,
void
CustomError::log() const
{
- say_file_line(S_ERROR, file, line, errmsg, "%s",
- "Custom type %s", m_custom_type);
+ say_file_line(S_ERROR, file, line, errmsg, "%s", "Custom type %s",
+ m_custom_type);
}
struct error *
diff --git a/src/box/error.h b/src/box/error.h
index 338121d..7bab7e8 100644
--- a/src/box/error.h
+++ b/src/box/error.h
@@ -50,8 +50,8 @@ struct error *
BuildXlogError(const char *file, unsigned line, const char *format, ...);
struct error *
-BuildXlogGapError(const char *file, unsigned line,
- const struct vclock *from, const struct vclock *to);
+BuildXlogGapError(const char *file, unsigned line, const struct vclock *from,
+ const struct vclock *to);
struct error *
BuildCustomError(const char *file, unsigned int line, const char *custom_type,
@@ -189,43 +189,33 @@ extern const struct type_info type_CustomError;
struct rmean;
extern "C" struct rmean *rmean_error;
-enum rmean_error_name {
- RMEAN_ERROR,
- RMEAN_ERROR_LAST
-};
+enum rmean_error_name { RMEAN_ERROR, RMEAN_ERROR_LAST };
extern const char *rmean_error_strings[RMEAN_ERROR_LAST];
-class ClientError: public Exception
-{
+class ClientError: public Exception {
public:
- virtual void raise()
- {
- throw this;
- }
+ virtual void raise() { throw this; }
virtual void log() const;
- int
- errcode() const
- {
- return m_errcode;
- }
+ int errcode() const { return m_errcode; }
ClientError(const char *file, unsigned line, uint32_t errcode, ...);
static uint32_t get_errcode(const struct error *e);
/* client errno code */
int m_errcode;
+
protected:
ClientError(const type_info *type, const char *file, unsigned line,
uint32_t errcode);
};
-class LoggedError: public ClientError
-{
+class LoggedError: public ClientError {
public:
- template <typename ... Args>
- LoggedError(const char *file, unsigned line, uint32_t errcode, Args ... args)
+ template <typename... Args>
+ LoggedError(const char *file, unsigned line, uint32_t errcode,
+ Args... args)
: ClientError(file, line, errcode, args...)
{
/* TODO: actually calls ClientError::log */
@@ -237,8 +227,7 @@ public:
* A special type of exception which must be used
* for all access denied errors, since it invokes audit triggers.
*/
-class AccessDeniedError: public ClientError
-{
+class AccessDeniedError: public ClientError {
public:
AccessDeniedError(const char *file, unsigned int line,
const char *access_type, const char *object_type,
@@ -252,23 +241,11 @@ public:
free(m_access_type);
}
- const char *
- object_type()
- {
- return m_object_type;
- }
+ const char *object_type() { return m_object_type; }
- const char *
- object_name()
- {
- return m_object_name?:"(nil)";
- }
+ const char *object_name() { return m_object_name ?: "(nil)"; }
- const char *
- access_type()
- {
- return m_access_type;
- }
+ const char *access_type() { return m_access_type; }
private:
/** Type of object the required access was denied to */
@@ -285,46 +262,37 @@ private:
* of exception is introduced to gracefully skip such errors
* in force_recovery = true mode.
*/
-struct XlogError: public Exception
-{
+struct XlogError: public Exception {
XlogError(const char *file, unsigned line, const char *format,
va_list ap)
- :Exception(&type_XlogError, file, line)
+ : Exception(&type_XlogError, file, line)
{
error_vformat_msg(this, format, ap);
}
- XlogError(const struct type_info *type, const char *file,
- unsigned line)
- :Exception(type, file, line)
- {
- }
+ XlogError(const struct type_info *type, const char *file, unsigned line)
+ : Exception(type, file, line)
+ {}
virtual void raise() { throw this; }
};
-struct XlogGapError: public XlogError
-{
- XlogGapError(const char *file, unsigned line,
- const struct vclock *from, const struct vclock *to);
- XlogGapError(const char *file, unsigned line,
- const char *msg);
+struct XlogGapError: public XlogError {
+ XlogGapError(const char *file, unsigned line, const struct vclock *from,
+ const struct vclock *to);
+ XlogGapError(const char *file, unsigned line, const char *msg);
virtual void raise() { throw this; }
};
-class CustomError: public ClientError
-{
+class CustomError: public ClientError {
public:
CustomError(const char *file, unsigned int line,
const char *custom_type, uint32_t errcode);
virtual void log() const;
- const char*
- custom_type()
- {
- return m_custom_type;
- }
+ const char *custom_type() { return m_custom_type; }
+
private:
/** Custom type name. */
char m_custom_type[64];
diff --git a/src/box/execute.c b/src/box/execute.c
index e14da20..11736c7 100644
--- a/src/box/execute.c
+++ b/src/box/execute.c
@@ -104,7 +104,7 @@ static void
port_sql_destroy(struct port *base)
{
port_c_vtab.destroy(base);
- struct port_sql *port_sql = (struct port_sql *) base;
+ struct port_sql *port_sql = (struct port_sql *)base;
if (port_sql->do_finalize)
sql_stmt_finalize(((struct port_sql *)base)->stmt);
}
@@ -125,7 +125,7 @@ port_sql_create(struct port *port, struct sql_stmt *stmt,
{
port_c_create(port);
port->vtab = &port_sql_vtab;
- struct port_sql *port_sql = (struct port_sql *) port;
+ struct port_sql *port_sql = (struct port_sql *)port;
port_sql->stmt = stmt;
port_sql->serialization_format = format;
port_sql->do_finalize = do_finalize;
@@ -142,8 +142,7 @@ port_sql_create(struct port *port, struct sql_stmt *stmt,
* @retval -1 Out of memory when resizing the output buffer.
*/
static inline int
-sql_column_to_messagepack(struct sql_stmt *stmt, int i,
- struct region *region)
+sql_column_to_messagepack(struct sql_stmt *stmt, int i, struct region *region)
{
size_t size;
enum mp_type type = sql_column_type(stmt, i);
@@ -151,7 +150,7 @@ sql_column_to_messagepack(struct sql_stmt *stmt, int i,
case MP_INT: {
int64_t n = sql_column_int64(stmt, i);
size = mp_sizeof_int(n);
- char *pos = (char *) region_alloc(region, size);
+ char *pos = (char *)region_alloc(region, size);
if (pos == NULL)
goto oom;
mp_encode_int(pos, n);
@@ -160,7 +159,7 @@ sql_column_to_messagepack(struct sql_stmt *stmt, int i,
case MP_UINT: {
uint64_t n = sql_column_uint64(stmt, i);
size = mp_sizeof_uint(n);
- char *pos = (char *) region_alloc(region, size);
+ char *pos = (char *)region_alloc(region, size);
if (pos == NULL)
goto oom;
mp_encode_uint(pos, n);
@@ -169,7 +168,7 @@ sql_column_to_messagepack(struct sql_stmt *stmt, int i,
case MP_DOUBLE: {
double d = sql_column_double(stmt, i);
size = mp_sizeof_double(d);
- char *pos = (char *) region_alloc(region, size);
+ char *pos = (char *)region_alloc(region, size);
if (pos == NULL)
goto oom;
mp_encode_double(pos, d);
@@ -178,7 +177,7 @@ sql_column_to_messagepack(struct sql_stmt *stmt, int i,
case MP_STR: {
uint32_t len = sql_column_bytes(stmt, i);
size = mp_sizeof_str(len);
- char *pos = (char *) region_alloc(region, size);
+ char *pos = (char *)region_alloc(region, size);
if (pos == NULL)
goto oom;
const char *s;
@@ -190,8 +189,7 @@ sql_column_to_messagepack(struct sql_stmt *stmt, int i,
case MP_MAP:
case MP_ARRAY: {
uint32_t len = sql_column_bytes(stmt, i);
- const char *s =
- (const char *)sql_column_blob(stmt, i);
+ const char *s = (const char *)sql_column_blob(stmt, i);
if (sql_column_subtype(stmt, i) == SQL_SUBTYPE_MSGPACK) {
size = len;
char *pos = (char *)region_alloc(region, size);
@@ -210,7 +208,7 @@ sql_column_to_messagepack(struct sql_stmt *stmt, int i,
case MP_BOOL: {
bool b = sql_column_boolean(stmt, i);
size = mp_sizeof_bool(b);
- char *pos = (char *) region_alloc(region, size);
+ char *pos = (char *)region_alloc(region, size);
if (pos == NULL)
goto oom;
mp_encode_bool(pos, b);
@@ -218,7 +216,7 @@ sql_column_to_messagepack(struct sql_stmt *stmt, int i,
}
case MP_NIL: {
size = mp_sizeof_nil();
- char *pos = (char *) region_alloc(region, size);
+ char *pos = (char *)region_alloc(region, size);
if (pos == NULL)
goto oom;
mp_encode_nil(pos);
@@ -245,13 +243,13 @@ oom:
* @retval -1 Memory error.
*/
static inline int
-sql_row_to_port(struct sql_stmt *stmt, int column_count,
- struct region *region, struct port *port)
+sql_row_to_port(struct sql_stmt *stmt, int column_count, struct region *region,
+ struct port *port)
{
assert(column_count > 0);
size_t size = mp_sizeof_array(column_count);
size_t svp = region_used(region);
- char *pos = (char *) region_alloc(region, size);
+ char *pos = (char *)region_alloc(region, size);
if (pos == NULL) {
diag_set(OutOfMemory, size, "region_alloc", "SQL row");
return -1;
@@ -263,7 +261,7 @@ sql_row_to_port(struct sql_stmt *stmt, int column_count,
goto error;
}
size = region_used(region) - svp;
- pos = (char *) region_join(region, size);
+ pos = (char *)region_join(region, size);
if (pos == NULL) {
diag_set(OutOfMemory, size, "region_join", "pos");
goto error;
@@ -305,7 +303,7 @@ metadata_map_sizeof(const char *name, const char *type, const char *coll,
members_count++;
map_size += mp_sizeof_uint(IPROTO_FIELD_SPAN);
map_size += span != NULL ? mp_sizeof_str(strlen(span)) :
- mp_sizeof_nil();
+ mp_sizeof_nil();
}
map_size += mp_sizeof_uint(IPROTO_FIELD_NAME);
map_size += mp_sizeof_uint(IPROTO_FIELD_TYPE);
@@ -340,7 +338,7 @@ metadata_map_encode(char *buf, const char *name, const char *type,
buf = mp_encode_uint(buf, IPROTO_FIELD_IS_AUTOINCREMENT);
buf = mp_encode_bool(buf, true);
}
- if (! is_full)
+ if (!is_full)
return;
/*
* Span is an original expression that forms
@@ -370,9 +368,9 @@ static inline int
sql_get_metadata(struct sql_stmt *stmt, struct obuf *out, int column_count)
{
assert(column_count > 0);
- int size = mp_sizeof_uint(IPROTO_METADATA) +
- mp_sizeof_array(column_count);
- char *pos = (char *) obuf_alloc(out, size);
+ int size =
+ mp_sizeof_uint(IPROTO_METADATA) + mp_sizeof_array(column_count);
+ char *pos = (char *)obuf_alloc(out, size);
if (pos == NULL) {
diag_set(OutOfMemory, size, "obuf_alloc", "pos");
return -1;
@@ -395,7 +393,7 @@ sql_get_metadata(struct sql_stmt *stmt, struct obuf *out, int column_count)
assert(type != NULL);
size = metadata_map_sizeof(name, type, coll, span, nullable,
is_autoincrement);
- char *pos = (char *) obuf_alloc(out, size);
+ char *pos = (char *)obuf_alloc(out, size);
if (pos == NULL) {
diag_set(OutOfMemory, size, "obuf_alloc", "pos");
return -1;
@@ -412,7 +410,7 @@ sql_get_params_metadata(struct sql_stmt *stmt, struct obuf *out)
int bind_count = sql_bind_parameter_count(stmt);
int size = mp_sizeof_uint(IPROTO_BIND_METADATA) +
mp_sizeof_array(bind_count);
- char *pos = (char *) obuf_alloc(out, size);
+ char *pos = (char *)obuf_alloc(out, size);
if (pos == NULL) {
diag_set(OutOfMemory, size, "obuf_alloc", "pos");
return -1;
@@ -429,7 +427,7 @@ sql_get_params_metadata(struct sql_stmt *stmt, struct obuf *out)
const char *type = "ANY";
size += mp_sizeof_str(strlen(name));
size += mp_sizeof_str(strlen(type));
- char *pos = (char *) obuf_alloc(out, size);
+ char *pos = (char *)obuf_alloc(out, size);
if (pos == NULL) {
diag_set(OutOfMemory, size, "obuf_alloc", "pos");
return -1;
@@ -448,12 +446,10 @@ sql_get_prepare_common_keys(struct sql_stmt *stmt, struct obuf *out, int keys)
{
const char *sql_str = sql_stmt_query_str(stmt);
uint32_t stmt_id = sql_stmt_calculate_id(sql_str, strlen(sql_str));
- int size = mp_sizeof_map(keys) +
- mp_sizeof_uint(IPROTO_STMT_ID) +
- mp_sizeof_uint(stmt_id) +
- mp_sizeof_uint(IPROTO_BIND_COUNT) +
+ int size = mp_sizeof_map(keys) + mp_sizeof_uint(IPROTO_STMT_ID) +
+ mp_sizeof_uint(stmt_id) + mp_sizeof_uint(IPROTO_BIND_COUNT) +
mp_sizeof_uint(sql_bind_parameter_count(stmt));
- char *pos = (char *) obuf_alloc(out, size);
+ char *pos = (char *)obuf_alloc(out, size);
if (pos == NULL) {
diag_set(OutOfMemory, size, "obuf_alloc", "pos");
return -1;
@@ -479,7 +475,7 @@ port_sql_dump_msgpack(struct port *port, struct obuf *out)
case DQL_EXECUTE: {
int keys = 2;
int size = mp_sizeof_map(keys);
- char *pos = (char *) obuf_alloc(out, size);
+ char *pos = (char *)obuf_alloc(out, size);
if (pos == NULL) {
diag_set(OutOfMemory, size, "obuf_alloc", "pos");
return -1;
@@ -488,7 +484,7 @@ port_sql_dump_msgpack(struct port *port, struct obuf *out)
if (sql_get_metadata(stmt, out, sql_column_count(stmt)) != 0)
return -1;
size = mp_sizeof_uint(IPROTO_DATA);
- pos = (char *) obuf_alloc(out, size);
+ pos = (char *)obuf_alloc(out, size);
if (pos == NULL) {
diag_set(OutOfMemory, size, "obuf_alloc", "pos");
return -1;
@@ -507,7 +503,7 @@ port_sql_dump_msgpack(struct port *port, struct obuf *out)
int size = mp_sizeof_map(keys) +
mp_sizeof_uint(IPROTO_SQL_INFO) +
mp_sizeof_map(map_size);
- char *pos = (char *) obuf_alloc(out, size);
+ char *pos = (char *)obuf_alloc(out, size);
if (pos == NULL) {
diag_set(OutOfMemory, size, "obuf_alloc", "pos");
return -1;
@@ -521,10 +517,11 @@ port_sql_dump_msgpack(struct port *port, struct obuf *out)
mp_sizeof_uint(changes);
if (!stailq_empty(autoinc_id_list)) {
struct autoinc_id_entry *id_entry;
- stailq_foreach_entry(id_entry, autoinc_id_list, link) {
+ stailq_foreach_entry(id_entry, autoinc_id_list, link)
+ {
size += id_entry->id >= 0 ?
- mp_sizeof_uint(id_entry->id) :
- mp_sizeof_int(id_entry->id);
+ mp_sizeof_uint(id_entry->id) :
+ mp_sizeof_int(id_entry->id);
id_count++;
}
size += mp_sizeof_uint(SQL_INFO_AUTOINCREMENT_IDS) +
@@ -541,10 +538,12 @@ port_sql_dump_msgpack(struct port *port, struct obuf *out)
buf = mp_encode_uint(buf, SQL_INFO_AUTOINCREMENT_IDS);
buf = mp_encode_array(buf, id_count);
struct autoinc_id_entry *id_entry;
- stailq_foreach_entry(id_entry, autoinc_id_list, link) {
+ stailq_foreach_entry(id_entry, autoinc_id_list, link)
+ {
buf = id_entry->id >= 0 ?
- mp_encode_uint(buf, id_entry->id) :
- mp_encode_int(buf, id_entry->id);
+ mp_encode_uint(buf,
+ id_entry->id) :
+ mp_encode_int(buf, id_entry->id);
}
}
break;
@@ -569,7 +568,7 @@ port_sql_dump_msgpack(struct port *port, struct obuf *out)
*/
int keys = 3;
return sql_get_prepare_common_keys(stmt, out, keys);
- }
+ }
default: {
unreachable();
}
@@ -592,8 +591,8 @@ sql_reprepare(struct sql_stmt **stmt)
{
const char *sql_str = sql_stmt_query_str(*stmt);
struct sql_stmt *new_stmt;
- if (sql_stmt_compile(sql_str, strlen(sql_str), NULL,
- &new_stmt, NULL) != 0)
+ if (sql_stmt_compile(sql_str, strlen(sql_str), NULL, &new_stmt, NULL) !=
+ 0)
return -1;
if (sql_stmt_cache_update(*stmt, new_stmt) != 0)
return -1;
@@ -630,8 +629,8 @@ sql_prepare(const char *sql, int len, struct port *port)
/* Add id to the list of available statements in session. */
if (!session_check_stmt_id(current_session(), stmt_id))
session_add_stmt_id(current_session(), stmt_id);
- enum sql_serialization_format format = sql_column_count(stmt) > 0 ?
- DQL_PREPARE : DML_PREPARE;
+ enum sql_serialization_format format =
+ sql_column_count(stmt) > 0 ? DQL_PREPARE : DML_PREPARE;
port_sql_create(port, stmt, format, false);
return 0;
@@ -677,8 +676,8 @@ sql_execute(struct sql_stmt *stmt, struct port *port, struct region *region)
if (column_count > 0) {
/* Either ROW or DONE or ERROR. */
while ((rc = sql_step(stmt)) == SQL_ROW) {
- if (sql_row_to_port(stmt, column_count, region,
- port) != 0)
+ if (sql_row_to_port(stmt, column_count, region, port) !=
+ 0)
return -1;
}
assert(rc == SQL_DONE || rc != 0);
@@ -697,7 +696,6 @@ sql_execute_prepared(uint32_t stmt_id, const struct sql_bind *bind,
uint32_t bind_count, struct port *port,
struct region *region)
{
-
if (!session_check_stmt_id(current_session(), stmt_id)) {
diag_set(ClientError, ER_WRONG_QUERY_ID, stmt_id);
return -1;
@@ -720,8 +718,8 @@ sql_execute_prepared(uint32_t stmt_id, const struct sql_bind *bind,
sql_unbind(stmt);
if (sql_bind(stmt, bind, bind_count) != 0)
return -1;
- enum sql_serialization_format format = sql_column_count(stmt) > 0 ?
- DQL_EXECUTE : DML_EXECUTE;
+ enum sql_serialization_format format =
+ sql_column_count(stmt) > 0 ? DQL_EXECUTE : DML_EXECUTE;
port_sql_create(port, stmt, format, false);
if (sql_execute(stmt, port, region) != 0) {
port_destroy(port);
@@ -742,8 +740,8 @@ sql_prepare_and_execute(const char *sql, int len, const struct sql_bind *bind,
if (sql_stmt_compile(sql, len, NULL, &stmt, NULL) != 0)
return -1;
assert(stmt != NULL);
- enum sql_serialization_format format = sql_column_count(stmt) > 0 ?
- DQL_EXECUTE : DML_EXECUTE;
+ enum sql_serialization_format format =
+ sql_column_count(stmt) > 0 ? DQL_EXECUTE : DML_EXECUTE;
port_sql_create(port, stmt, format, true);
if (sql_bind(stmt, bind, bind_count) == 0 &&
sql_execute(stmt, port, region) == 0)
diff --git a/src/box/field_def.c b/src/box/field_def.c
index 34cecfa..e893641 100644
--- a/src/box/field_def.c
+++ b/src/box/field_def.c
@@ -75,7 +75,8 @@ const uint32_t field_mp_type[] = {
};
const uint32_t field_ext_type[] = {
- /* [FIELD_TYPE_ANY] = */ UINT32_MAX ^ (1U << MP_UNKNOWN_EXTENSION),
+ /* [FIELD_TYPE_ANY] = */ UINT32_MAX ^
+ (1U << MP_UNKNOWN_EXTENSION),
/* [FIELD_TYPE_UNSIGNED] = */ 0,
/* [FIELD_TYPE_STRING] = */ 0,
/* [FIELD_TYPE_NUMBER] = */ 1U << MP_DECIMAL,
@@ -98,7 +99,7 @@ const char *field_type_strs[] = {
/* [FIELD_TYPE_DOUBLE] = */ "double",
/* [FIELD_TYPE_INTEGER] = */ "integer",
/* [FIELD_TYPE_BOOLEAN] = */ "boolean",
- /* [FIELD_TYPE_VARBINARY] = */"varbinary",
+ /* [FIELD_TYPE_VARBINARY] = */ "varbinary",
/* [FIELD_TYPE_SCALAR] = */ "scalar",
/* [FIELD_TYPE_DECIMAL] = */ "decimal",
/* [FIELD_TYPE_UUID] = */ "uuid",
@@ -165,21 +166,20 @@ const struct opt_def field_def_reg[] = {
OPT_END,
};
-const struct field_def field_def_default = {
- .type = FIELD_TYPE_ANY,
- .name = NULL,
- .is_nullable = false,
- .nullable_action = ON_CONFLICT_ACTION_DEFAULT,
- .coll_id = COLL_NONE,
- .default_value = NULL,
- .default_value_expr = NULL
-};
+const struct field_def field_def_default = { .type = FIELD_TYPE_ANY,
+ .name = NULL,
+ .is_nullable = false,
+ .nullable_action =
+ ON_CONFLICT_ACTION_DEFAULT,
+ .coll_id = COLL_NONE,
+ .default_value = NULL,
+ .default_value_expr = NULL };
enum field_type
field_type_by_name(const char *name, size_t len)
{
- enum field_type field_type = strnindex(field_type_strs, name, len,
- field_type_MAX);
+ enum field_type field_type =
+ strnindex(field_type_strs, name, len, field_type_MAX);
if (field_type != field_type_MAX)
return field_type;
/* 'num' and 'str' in _index are deprecated since Tarantool 1.7 */
diff --git a/src/box/field_def.h b/src/box/field_def.h
index c5cfe5e..e839180 100644
--- a/src/box/field_def.h
+++ b/src/box/field_def.h
@@ -92,7 +92,7 @@ enum {
* For detailed explanation see context of OP_Eq, OP_Lt etc
* opcodes in vdbe.c.
*/
-static_assert((int) field_type_MAX <= (int) FIELD_TYPE_MASK,
+static_assert((int)field_type_MAX <= (int)FIELD_TYPE_MASK,
"values of enum field_type should fit into 4 bits of VdbeOp.p5");
extern const char *field_type_strs[];
diff --git a/src/box/field_map.c b/src/box/field_map.c
index dc90311..6b4433c 100644
--- a/src/box/field_map.c
+++ b/src/box/field_map.c
@@ -34,8 +34,7 @@
int
field_map_builder_create(struct field_map_builder *builder,
- uint32_t minimal_field_map_size,
- struct region *region)
+ uint32_t minimal_field_map_size, struct region *region)
{
builder->extents_size = 0;
builder->slot_count = minimal_field_map_size / sizeof(uint32_t);
@@ -63,10 +62,10 @@ field_map_builder_slot_extent_new(struct field_map_builder *builder,
{
struct field_map_builder_slot_extent *extent;
assert(!builder->slots[offset_slot].has_extent);
- uint32_t sz = sizeof(*extent) +
- multikey_count * sizeof(extent->offset[0]);
- extent = (struct field_map_builder_slot_extent *)
- region_aligned_alloc(region, sz, alignof(*extent));
+ uint32_t sz =
+ sizeof(*extent) + multikey_count * sizeof(extent->offset[0]);
+ extent = (struct field_map_builder_slot_extent *)region_aligned_alloc(
+ region, sz, alignof(*extent));
if (extent == NULL) {
diag_set(OutOfMemory, sz, "region_aligned_alloc", "extent");
return NULL;
@@ -112,13 +111,13 @@ field_map_build(struct field_map_builder *builder, char *buffer)
continue;
}
struct field_map_builder_slot_extent *extent =
- builder->slots[i].extent;
+ builder->slots[i].extent;
/** Retrive memory for the extent. */
store_u32(&field_map[i], extent_wptr - (char *)field_map);
store_u32(extent_wptr, extent->size);
uint32_t extent_offset_sz = extent->size * sizeof(uint32_t);
- memcpy(&((uint32_t *) extent_wptr)[1], extent->offset,
- extent_offset_sz);
+ memcpy(&((uint32_t *)extent_wptr)[1], extent->offset,
+ extent_offset_sz);
extent_wptr += sizeof(uint32_t) + extent_offset_sz;
}
assert(extent_wptr == buffer + builder->extents_size);
diff --git a/src/box/field_map.h b/src/box/field_map.h
index d8ef726..f96663b 100644
--- a/src/box/field_map.h
+++ b/src/box/field_map.h
@@ -163,8 +163,9 @@ field_map_get_offset(const uint32_t *field_map, int32_t offset_slot,
* The field_map extent has the following
* structure: [size=N|slot1|slot2|..|slotN]
*/
- const uint32_t *extent = (const uint32_t *)
- ((const char *)field_map + (int32_t)offset);
+ const uint32_t *extent =
+ (const uint32_t *)((const char *)field_map +
+ (int32_t)offset);
if ((uint32_t)multikey_idx >= load_u32(&extent[0]))
return 0;
offset = load_u32(&extent[multikey_idx + 1]);
@@ -229,8 +230,8 @@ field_map_builder_set_slot(struct field_map_builder *builder,
assert(extent != NULL);
assert(extent->size == multikey_count);
} else {
- extent = field_map_builder_slot_extent_new(builder,
- offset_slot, multikey_count, region);
+ extent = field_map_builder_slot_extent_new(
+ builder, offset_slot, multikey_count, region);
if (extent == NULL)
return -1;
}
@@ -245,8 +246,7 @@ field_map_builder_set_slot(struct field_map_builder *builder,
static inline uint32_t
field_map_build_size(struct field_map_builder *builder)
{
- return builder->slot_count * sizeof(uint32_t) +
- builder->extents_size;
+ return builder->slot_count * sizeof(uint32_t) + builder->extents_size;
}
/**
diff --git a/src/box/fk_constraint.h b/src/box/fk_constraint.h
index dcc5363..c1d90a6 100644
--- a/src/box/fk_constraint.h
+++ b/src/box/fk_constraint.h
@@ -135,8 +135,9 @@ static inline size_t
fk_constraint_def_sizeof(uint32_t link_count, uint32_t name_len,
uint32_t *links_offset)
{
- *links_offset = small_align(sizeof(struct fk_constraint_def) +
- name_len + 1, alignof(struct field_link));
+ *links_offset =
+ small_align(sizeof(struct fk_constraint_def) + name_len + 1,
+ alignof(struct field_link));
return *links_offset + link_count * sizeof(struct field_link);
}
diff --git a/src/box/func.c b/src/box/func.c
index 8087c95..6adbb71 100644
--- a/src/box/func.c
+++ b/src/box/func.c
@@ -111,8 +111,8 @@ struct module_find_ctx {
static int
luaT_module_find(lua_State *L)
{
- struct module_find_ctx *ctx = (struct module_find_ctx *)
- lua_topointer(L, 1);
+ struct module_find_ctx *ctx =
+ (struct module_find_ctx *)lua_topointer(L, 1);
/*
* Call package.searchpath(name, package.cpath) and use
@@ -156,7 +156,7 @@ module_find(const char *package, const char *package_end, char *path,
lua_State *L = tarantool_L;
int top = lua_gettop(L);
if (luaT_cpcall(L, luaT_module_find, &ctx) != 0) {
- int package_len = (int) (package_end - package);
+ int package_len = (int)(package_end - package);
diag_set(ClientError, ER_LOAD_MODULE, package_len, package,
lua_tostring(L, -1));
lua_settop(L, top);
@@ -177,7 +177,7 @@ module_init(void)
modules = mh_strnptr_new();
if (modules == NULL) {
diag_set(OutOfMemory, sizeof(*modules), "malloc",
- "modules hash table");
+ "modules hash table");
return -1;
}
return 0;
@@ -189,7 +189,7 @@ module_free(void)
while (mh_size(modules) > 0) {
mh_int_t i = mh_first(modules);
struct module *module =
- (struct module *) mh_strnptr_node(modules, i)->val;
+ (struct module *)mh_strnptr_node(modules, i)->val;
/* Can't delete modules if they have active calls */
module_gc(module);
}
@@ -216,8 +216,8 @@ module_cache_put(struct module *module)
{
size_t package_len = strlen(module->package);
uint32_t name_hash = mh_strn_hash(module->package, package_len);
- const struct mh_strnptr_node_t strnode = {
- module->package, package_len, name_hash, module};
+ const struct mh_strnptr_node_t strnode = { module->package, package_len,
+ name_hash, module };
if (mh_strnptr_put(modules, &strnode, NULL, NULL) == mh_end(modules)) {
diag_set(OutOfMemory, sizeof(strnode), "malloc", "modules");
@@ -252,8 +252,8 @@ module_load(const char *package, const char *package_end)
return NULL;
int package_len = package_end - package;
- struct module *module = (struct module *)
- malloc(sizeof(*module) + package_len + 1);
+ struct module *module =
+ (struct module *)malloc(sizeof(*module) + package_len + 1);
if (module == NULL) {
diag_set(OutOfMemory, sizeof(struct module) + package_len + 1,
"malloc", "struct module");
@@ -269,7 +269,7 @@ module_load(const char *package, const char *package_end)
tmpdir = "/tmp";
char dir_name[PATH_MAX];
int rc = snprintf(dir_name, sizeof(dir_name), "%s/tntXXXXXX", tmpdir);
- if (rc < 0 || (size_t) rc >= sizeof(dir_name)) {
+ if (rc < 0 || (size_t)rc >= sizeof(dir_name)) {
diag_set(SystemError, "failed to generate path to tmp dir");
goto error;
}
@@ -282,7 +282,7 @@ module_load(const char *package, const char *package_end)
char load_name[PATH_MAX];
rc = snprintf(load_name, sizeof(load_name), "%s/%.*s." TARANTOOL_LIBEXT,
dir_name, package_len, package);
- if (rc < 0 || (size_t) rc >= sizeof(dir_name)) {
+ if (rc < 0 || (size_t)rc >= sizeof(dir_name)) {
diag_set(SystemError, "failed to generate path to DSO");
goto error;
}
@@ -295,11 +295,13 @@ module_load(const char *package, const char *package_end)
int source_fd = open(path, O_RDONLY);
if (source_fd < 0) {
- diag_set(SystemError, "failed to open module %s file for" \
- " reading", path);
+ diag_set(SystemError,
+ "failed to open module %s file for"
+ " reading",
+ path);
goto error;
}
- int dest_fd = open(load_name, O_WRONLY|O_CREAT|O_TRUNC,
+ int dest_fd = open(load_name, O_WRONLY | O_CREAT | O_TRUNC,
st.st_mode & 0777);
if (dest_fd < 0) {
diag_set(SystemError, "failed to open file %s for writing ",
@@ -312,8 +314,8 @@ module_load(const char *package, const char *package_end)
close(source_fd);
close(dest_fd);
if (ret != st.st_size) {
- diag_set(SystemError, "failed to copy DSO %s to %s",
- path, load_name);
+ diag_set(SystemError, "failed to copy DSO %s to %s", path,
+ load_name);
goto error;
}
@@ -323,8 +325,8 @@ module_load(const char *package, const char *package_end)
if (rmdir(dir_name) != 0)
say_warn("failed to delete temporary dir %s", dir_name);
if (module->handle == NULL) {
- diag_set(ClientError, ER_LOAD_MODULE, package_len,
- package, dlerror());
+ diag_set(ClientError, ER_LOAD_MODULE, package_len, package,
+ dlerror());
goto error;
}
struct errinj *e = errinj(ERRINJ_DYN_MODULE_COUNT, ERRINJ_INT);
@@ -372,7 +374,8 @@ module_sym(struct module *module, const char *name)
}
int
-module_reload(const char *package, const char *package_end, struct module **module)
+module_reload(const char *package, const char *package_end,
+ struct module **module)
{
struct module *old_module = module_cache_find(package, package_end);
if (old_module == NULL) {
@@ -420,8 +423,8 @@ restore:
}
func->module = old_module;
rlist_move(&old_module->funcs, &func->item);
- } while (func != rlist_first_entry(&old_module->funcs,
- struct func_c, item));
+ } while (func !=
+ rlist_first_entry(&old_module->funcs, struct func_c, item));
assert(rlist_empty(&new_module->funcs));
module_delete(new_module);
return -1;
@@ -478,7 +481,7 @@ func_c_new(MAYBE_UNUSED struct func_def *def)
{
assert(def->language == FUNC_LANGUAGE_C);
assert(def->body == NULL && !def->is_sandboxed);
- struct func_c *func = (struct func_c *) malloc(sizeof(struct func_c));
+ struct func_c *func = (struct func_c *)malloc(sizeof(struct func_c));
if (func == NULL) {
diag_set(OutOfMemory, sizeof(*func), "malloc", "func");
return NULL;
@@ -510,7 +513,7 @@ func_c_destroy(struct func *base)
{
assert(base->vtab == &func_c_vtab);
assert(base != NULL && base->def->language == FUNC_LANGUAGE_C);
- struct func_c *func = (struct func_c *) base;
+ struct func_c *func = (struct func_c *)base;
func_c_unload(func);
TRASH(base);
free(func);
@@ -528,8 +531,8 @@ func_c_load(struct func_c *func)
struct func_name name;
func_split_name(func->base.def->name, &name);
- struct module *module = module_cache_find(name.package,
- name.package_end);
+ struct module *module =
+ module_cache_find(name.package, name.package_end);
if (module == NULL) {
/* Try to find loaded module in the cache */
module = module_load(name.package, name.package_end);
@@ -554,7 +557,7 @@ func_c_call(struct func *base, struct port *args, struct port *ret)
{
assert(base->vtab == &func_c_vtab);
assert(base != NULL && base->def->language == FUNC_LANGUAGE_C);
- struct func_c *func = (struct func_c *) base;
+ struct func_c *func = (struct func_c *)base;
if (func->func == NULL) {
if (func_c_load(func) != 0)
return -1;
@@ -618,7 +621,8 @@ func_access_check(struct func *func)
return 0;
user_access_t access = PRIV_X | PRIV_U;
/* Check access for all functions. */
- access &= ~entity_access_get(SC_FUNCTION)[credentials->auth_token].effective;
+ access &= ~entity_access_get(SC_FUNCTION)[credentials->auth_token]
+ .effective;
user_access_t func_access = access & ~credentials->universal_access;
if ((func_access & PRIV_U) != 0 ||
(func->def->uid != credentials->uid &&
diff --git a/src/box/func.h b/src/box/func.h
index 581e468..4e9d22d 100644
--- a/src/box/func.h
+++ b/src/box/func.h
@@ -118,7 +118,8 @@ func_call(struct func *func, struct port *args, struct port *ret);
* @retval 0 on success.
*/
int
-module_reload(const char *package, const char *package_end, struct module **module);
+module_reload(const char *package, const char *package_end,
+ struct module **module);
#if defined(__cplusplus)
} /* extern "C" */
diff --git a/src/box/func_def.c b/src/box/func_def.c
index 11d2bdb..47ead9e 100644
--- a/src/box/func_def.c
+++ b/src/box/func_def.c
@@ -34,9 +34,9 @@
#include "diag.h"
#include "error.h"
-const char *func_language_strs[] = {"LUA", "C", "SQL", "SQL_BUILTIN"};
+const char *func_language_strs[] = { "LUA", "C", "SQL", "SQL_BUILTIN" };
-const char *func_aggregate_strs[] = {"none", "group"};
+const char *func_aggregate_strs[] = { "none", "group" };
const struct func_opts func_opts_default = {
/* .is_multikey = */ false,
@@ -102,25 +102,28 @@ func_def_check(struct func_def *def)
switch (def->language) {
case FUNC_LANGUAGE_C:
if (def->body != NULL || def->is_sandboxed) {
- diag_set(ClientError, ER_CREATE_FUNCTION, def->name,
- "body and is_sandboxed options are not compatible "
- "with C language");
+ diag_set(
+ ClientError, ER_CREATE_FUNCTION, def->name,
+ "body and is_sandboxed options are not compatible "
+ "with C language");
return -1;
}
break;
case FUNC_LANGUAGE_LUA:
if (def->is_sandboxed && def->body == NULL) {
- diag_set(ClientError, ER_CREATE_FUNCTION, def->name,
- "is_sandboxed option may be set only for a persistent "
- "Lua function (one with a non-empty body)");
+ diag_set(
+ ClientError, ER_CREATE_FUNCTION, def->name,
+ "is_sandboxed option may be set only for a persistent "
+ "Lua function (one with a non-empty body)");
return -1;
}
break;
case FUNC_LANGUAGE_SQL_BUILTIN:
if (def->body != NULL || def->is_sandboxed) {
- diag_set(ClientError, ER_CREATE_FUNCTION, def->name,
- "body and is_sandboxed options are not compatible "
- "with SQL language");
+ diag_set(
+ ClientError, ER_CREATE_FUNCTION, def->name,
+ "body and is_sandboxed options are not compatible "
+ "with SQL language");
return -1;
}
break;
diff --git a/src/box/func_def.h b/src/box/func_def.h
index d99d891..13db751 100644
--- a/src/box/func_def.h
+++ b/src/box/func_def.h
@@ -179,8 +179,8 @@ struct box_function_ctx {
};
typedef struct box_function_ctx box_function_ctx_t;
-typedef int (*box_function_f)(box_function_ctx_t *ctx,
- const char *args, const char *args_end);
+typedef int (*box_function_f)(box_function_ctx_t *ctx, const char *args,
+ const char *args_end);
#ifdef __cplusplus
}
diff --git a/src/box/gc.c b/src/box/gc.c
index 76f7c63..0763a9b 100644
--- a/src/box/gc.c
+++ b/src/box/gc.c
@@ -54,17 +54,15 @@
#include "say.h"
#include "vclock.h"
#include "cbus.h"
-#include "engine.h" /* engine_collect_garbage() */
-#include "wal.h" /* wal_collect_garbage() */
+#include "engine.h" /* engine_collect_garbage() */
+#include "wal.h" /* wal_collect_garbage() */
#include "checkpoint_schedule.h"
#include "txn_limbo.h"
struct gc_state gc;
-static int
-gc_cleanup_fiber_f(va_list);
-static int
-gc_checkpoint_fiber_f(va_list);
+static int gc_cleanup_fiber_f(va_list);
+static int gc_checkpoint_fiber_f(va_list);
/**
* Comparator used for ordering gc_consumer objects
@@ -83,8 +81,8 @@ gc_consumer_cmp(const struct gc_consumer *a, const struct gc_consumer *b)
return 0;
}
-rb_gen(MAYBE_UNUSED static inline, gc_tree_, gc_tree_t,
- struct gc_consumer, node, gc_consumer_cmp);
+rb_gen(MAYBE_UNUSED static inline, gc_tree_, gc_tree_t, struct gc_consumer,
+ node, gc_consumer_cmp);
/** Free a consumer object. */
static void
@@ -119,8 +117,8 @@ gc_init(void)
if (gc.cleanup_fiber == NULL)
panic("failed to start garbage collection fiber");
- gc.checkpoint_fiber = fiber_new("checkpoint_daemon",
- gc_checkpoint_fiber_f);
+ gc.checkpoint_fiber =
+ fiber_new("checkpoint_daemon", gc_checkpoint_fiber_f);
if (gc.checkpoint_fiber == NULL)
panic("failed to start checkpoint daemon fiber");
@@ -145,8 +143,8 @@ gc_free(void)
/* Free all registered consumers. */
struct gc_consumer *consumer = gc_tree_first(&gc.consumers);
while (consumer != NULL) {
- struct gc_consumer *next = gc_tree_next(&gc.consumers,
- consumer);
+ struct gc_consumer *next =
+ gc_tree_next(&gc.consumers, consumer);
gc_tree_remove(&gc.consumers, consumer);
gc_consumer_delete(consumer);
consumer = next;
@@ -172,8 +170,8 @@ gc_run_cleanup(void)
*/
struct gc_checkpoint *checkpoint = NULL;
while (true) {
- checkpoint = rlist_first_entry(&gc.checkpoints,
- struct gc_checkpoint, in_checkpoints);
+ checkpoint = rlist_first_entry(
+ &gc.checkpoints, struct gc_checkpoint, in_checkpoints);
if (gc.checkpoint_count <= gc.min_checkpoint_count)
break;
if (!rlist_empty(&checkpoint->refs))
@@ -297,8 +295,8 @@ gc_advance(const struct vclock *vclock)
struct gc_consumer *consumer = gc_tree_first(&gc.consumers);
while (consumer != NULL) {
- struct gc_consumer *next = gc_tree_next(&gc.consumers,
- consumer);
+ struct gc_consumer *next =
+ gc_tree_next(&gc.consumers, consumer);
/*
* Remove all the consumers whose vclocks are
* either less than or incomparable with the wal
@@ -496,8 +494,8 @@ gc_checkpoint_fiber_f(va_list ap)
struct checkpoint_schedule *sched = &gc.checkpoint_schedule;
while (!fiber_is_cancelled()) {
- double timeout = checkpoint_schedule_timeout(sched,
- ev_monotonic_now(loop()));
+ double timeout = checkpoint_schedule_timeout(
+ sched, ev_monotonic_now(loop()));
if (timeout > 0) {
char buf[128];
struct tm tm;
@@ -556,8 +554,8 @@ gc_consumer_register(const struct vclock *vclock, const char *format, ...)
{
struct gc_consumer *consumer = calloc(1, sizeof(*consumer));
if (consumer == NULL) {
- diag_set(OutOfMemory, sizeof(*consumer),
- "malloc", "struct gc_consumer");
+ diag_set(OutOfMemory, sizeof(*consumer), "malloc",
+ "struct gc_consumer");
return NULL;
}
diff --git a/src/box/identifier.c b/src/box/identifier.c
index b1c56bd..4432cc8 100644
--- a/src/box/identifier.c
+++ b/src/box/identifier.c
@@ -57,10 +57,8 @@ identifier_check(const char *str, int str_len)
* Here the `c` symbol printability is determined by comparison
* with unicode category types explicitly.
*/
- if (type == U_UNASSIGNED ||
- type == U_LINE_SEPARATOR ||
- type == U_CONTROL_CHAR ||
- type == U_PARAGRAPH_SEPARATOR)
+ if (type == U_UNASSIGNED || type == U_LINE_SEPARATOR ||
+ type == U_CONTROL_CHAR || type == U_PARAGRAPH_SEPARATOR)
goto error;
}
return 0;
diff --git a/src/box/index.cc b/src/box/index.cc
index c2fc008..0d88298 100644
--- a/src/box/index.cc
+++ b/src/box/index.cc
@@ -42,14 +42,16 @@
/* {{{ Utilities. **********************************************/
UnsupportedIndexFeature::UnsupportedIndexFeature(const char *file,
- unsigned line, struct index_def *index_def, const char *what)
+ unsigned line,
+ struct index_def *index_def,
+ const char *what)
: ClientError(file, line, ER_UNKNOWN)
{
struct space *space = space_cache_find_xc(index_def->space_id);
m_errcode = ER_UNSUPPORTED_INDEX_FEATURE;
error_format_msg(this, tnt_errcode_desc(m_errcode), index_def->name,
- index_type_strs[index_def->type],
- space->def->name, space->def->engine_name, what);
+ index_type_strs[index_def->type], space->def->name,
+ space->def->engine_name, what);
}
struct error *
@@ -84,7 +86,7 @@ key_validate(const struct index_def *index_def, enum iterator_type type,
if (index_def->type == RTREE) {
unsigned d = index_def->opts.dimension;
if (part_count != 1 && part_count != d && part_count != d * 2) {
- diag_set(ClientError, ER_KEY_PART_COUNT, d * 2,
+ diag_set(ClientError, ER_KEY_PART_COUNT, d * 2,
part_count);
return -1;
}
@@ -98,8 +100,8 @@ key_validate(const struct index_def *index_def, enum iterator_type type,
return -1;
}
for (uint32_t part = 0; part < array_size; part++) {
- if (key_part_validate(FIELD_TYPE_NUMBER, key,
- 0, false))
+ if (key_part_validate(FIELD_TYPE_NUMBER, key, 0,
+ false))
return -1;
mp_next(&key);
}
@@ -119,16 +121,16 @@ key_validate(const struct index_def *index_def, enum iterator_type type,
}
/* Partial keys are allowed only for TREE index type. */
- if (index_def->type != TREE && part_count < index_def->key_def->part_count) {
+ if (index_def->type != TREE &&
+ part_count < index_def->key_def->part_count) {
diag_set(ClientError, ER_PARTIAL_KEY,
index_type_strs[index_def->type],
- index_def->key_def->part_count,
- part_count);
+ index_def->key_def->part_count, part_count);
return -1;
}
const char *key_end;
- if (key_validate_parts(index_def->key_def, key,
- part_count, true, &key_end) != 0)
+ if (key_validate_parts(index_def->key_def, key, part_count,
+ true, &key_end) != 0)
return -1;
}
return 0;
@@ -158,13 +160,13 @@ box_tuple_extract_key(box_tuple_t *tuple, uint32_t space_id, uint32_t index_id,
struct index *index = index_find(space, index_id);
if (index == NULL)
return NULL;
- return tuple_extract_key(tuple, index->def->key_def,
- MULTIKEY_NONE, key_size);
+ return tuple_extract_key(tuple, index->def->key_def, MULTIKEY_NONE,
+ key_size);
}
static inline int
-check_index(uint32_t space_id, uint32_t index_id,
- struct space **space, struct index **index)
+check_index(uint32_t space_id, uint32_t index_id, struct space **space,
+ struct index **index)
{
*space = space_cache_find(space_id);
if (*space == NULL)
@@ -205,7 +207,7 @@ box_index_bsize(uint32_t space_id, uint32_t index_id)
int
box_index_random(uint32_t space_id, uint32_t index_id, uint32_t rnd,
- box_tuple_t **result)
+ box_tuple_t **result)
{
assert(result != NULL);
struct space *space;
@@ -318,8 +320,8 @@ box_index_max(uint32_t space_id, uint32_t index_id, const char *key,
}
ssize_t
-box_index_count(uint32_t space_id, uint32_t index_id, int type,
- const char *key, const char *key_end)
+box_index_count(uint32_t space_id, uint32_t index_id, int type, const char *key,
+ const char *key_end)
{
assert(key != NULL && key_end != NULL);
mp_tuple_assert(key, key_end);
@@ -328,7 +330,7 @@ box_index_count(uint32_t space_id, uint32_t index_id, int type,
"Invalid iterator type");
return -1;
}
- enum iterator_type itype = (enum iterator_type) type;
+ enum iterator_type itype = (enum iterator_type)type;
struct space *space;
struct index *index;
if (check_index(space_id, index_id, &space, &index) != 0)
@@ -355,7 +357,7 @@ box_index_count(uint32_t space_id, uint32_t index_id, int type,
box_iterator_t *
box_index_iterator(uint32_t space_id, uint32_t index_id, int type,
- const char *key, const char *key_end)
+ const char *key, const char *key_end)
{
assert(key != NULL && key_end != NULL);
mp_tuple_assert(key, key_end);
@@ -364,7 +366,7 @@ box_index_iterator(uint32_t space_id, uint32_t index_id, int type,
"Invalid iterator type");
return NULL;
}
- enum iterator_type itype = (enum iterator_type) type;
+ enum iterator_type itype = (enum iterator_type)type;
struct space *space;
struct index *index;
if (check_index(space_id, index_id, &space, &index) != 0)
@@ -376,8 +378,8 @@ box_index_iterator(uint32_t space_id, uint32_t index_id, int type,
struct txn *txn;
if (txn_begin_ro_stmt(space, &txn) != 0)
return NULL;
- struct iterator *it = index_create_iterator(index, itype,
- key, part_count);
+ struct iterator *it =
+ index_create_iterator(index, itype, key, part_count);
if (it == NULL) {
txn_rollback_stmt(txn);
return NULL;
@@ -409,8 +411,7 @@ box_iterator_free(box_iterator_t *it)
/* {{{ Other index functions */
int
-box_index_stat(uint32_t space_id, uint32_t index_id,
- struct info_handler *info)
+box_index_stat(uint32_t space_id, uint32_t index_id, struct info_handler *info)
{
struct space *space;
struct index *index;
@@ -520,9 +521,8 @@ index_build(struct index *index, struct index *pk)
return -1;
if (n_tuples > 0) {
- say_info("Adding %zd keys to %s index '%s' ...",
- n_tuples, index_type_strs[index->def->type],
- index->def->name);
+ say_info("Adding %zd keys to %s index '%s' ...", n_tuples,
+ index_type_strs[index->def->type], index->def->name);
}
struct iterator *it = index_create_iterator(pk, ITER_ALL, NULL, 0);
@@ -555,30 +555,26 @@ index_build(struct index *index, struct index *pk)
void
generic_index_commit_create(struct index *, int64_t)
-{
-}
+{}
void
generic_index_abort_create(struct index *)
-{
-}
+{}
void
generic_index_commit_modify(struct index *, int64_t)
-{
-}
+{}
void
generic_index_commit_drop(struct index *, int64_t)
-{
-}
+{}
void
generic_index_update_def(struct index *)
-{
-}
+{}
-bool generic_index_depends_on_pk(struct index *)
+bool
+generic_index_depends_on_pk(struct index *)
{
return false;
}
@@ -604,11 +600,11 @@ generic_index_bsize(struct index *)
}
int
-generic_index_min(struct index *index, const char *key,
- uint32_t part_count, struct tuple **result)
+generic_index_min(struct index *index, const char *key, uint32_t part_count,
+ struct tuple **result)
{
- struct iterator *it = index_create_iterator(index, ITER_EQ,
- key, part_count);
+ struct iterator *it =
+ index_create_iterator(index, ITER_EQ, key, part_count);
if (it == NULL)
return -1;
int rc = iterator_next(it, result);
@@ -617,11 +613,11 @@ generic_index_min(struct index *index, const char *key,
}
int
-generic_index_max(struct index *index, const char *key,
- uint32_t part_count, struct tuple **result)
+generic_index_max(struct index *index, const char *key, uint32_t part_count,
+ struct tuple **result)
{
- struct iterator *it = index_create_iterator(index, ITER_REQ,
- key, part_count);
+ struct iterator *it =
+ index_create_iterator(index, ITER_REQ, key, part_count);
if (it == NULL)
return -1;
int rc = iterator_next(it, result);
@@ -642,8 +638,8 @@ ssize_t
generic_index_count(struct index *index, enum iterator_type type,
const char *key, uint32_t part_count)
{
- struct iterator *it = index_create_iterator(index, type,
- key, part_count);
+ struct iterator *it =
+ index_create_iterator(index, type, key, part_count);
if (it == NULL)
return -1;
int rc = 0;
@@ -658,8 +654,8 @@ generic_index_count(struct index *index, enum iterator_type type,
}
int
-generic_index_get(struct index *index, const char *key,
- uint32_t part_count, struct tuple **result)
+generic_index_get(struct index *index, const char *key, uint32_t part_count,
+ struct tuple **result)
{
(void)key;
(void)part_count;
@@ -685,12 +681,13 @@ struct iterator *
generic_index_create_iterator(struct index *base, enum iterator_type type,
const char *key, uint32_t part_count)
{
- (void) type; (void) key; (void) part_count;
+ (void)type;
+ (void)key;
+ (void)part_count;
diag_set(UnsupportedIndexFeature, base->def, "read view");
return NULL;
}
-
struct snapshot_iterator *
generic_index_create_snapshot_iterator(struct index *index)
{
@@ -720,8 +717,7 @@ generic_index_reset_stat(struct index *index)
void
generic_index_begin_build(struct index *)
-{
-}
+{}
int
generic_index_reserve(struct index *, uint32_t)
@@ -745,13 +741,13 @@ generic_index_build_next(struct index *index, struct tuple *tuple)
void
generic_index_end_build(struct index *)
-{
-}
+{}
int
disabled_index_build_next(struct index *index, struct tuple *tuple)
{
- (void) index; (void) tuple;
+ (void)index;
+ (void)tuple;
return 0;
}
@@ -760,8 +756,10 @@ disabled_index_replace(struct index *index, struct tuple *old_tuple,
struct tuple *new_tuple, enum dup_replace_mode mode,
struct tuple **result)
{
- (void) old_tuple; (void) new_tuple; (void) mode;
- (void) index;
+ (void)old_tuple;
+ (void)new_tuple;
+ (void)mode;
+ (void)index;
*result = NULL;
return 0;
}
diff --git a/src/box/index.h b/src/box/index.h
index 8614802..70bf2e5 100644
--- a/src/box/index.h
+++ b/src/box/index.h
@@ -124,7 +124,7 @@ box_index_bsize(uint32_t space_id, uint32_t index_id);
*/
int
box_index_random(uint32_t space_id, uint32_t index_id, uint32_t rnd,
- box_tuple_t **result);
+ box_tuple_t **result);
/**
* Get a tuple from index by the key.
@@ -192,8 +192,8 @@ box_index_max(uint32_t space_id, uint32_t index_id, const char *key,
* { iterator = type }) \endcode
*/
ssize_t
-box_index_count(uint32_t space_id, uint32_t index_id, int type,
- const char *key, const char *key_end);
+box_index_count(uint32_t space_id, uint32_t index_id, int type, const char *key,
+ const char *key_end);
/**
* Extract key from tuple according to key definition of given
@@ -206,8 +206,8 @@ box_index_count(uint32_t space_id, uint32_t index_id, int type,
* @retval NULL Memory Allocation error
*/
char *
-box_tuple_extract_key(box_tuple_t *tuple, uint32_t space_id,
- uint32_t index_id, uint32_t *key_size);
+box_tuple_extract_key(box_tuple_t *tuple, uint32_t space_id, uint32_t index_id,
+ uint32_t *key_size);
/** \endcond public */
@@ -221,8 +221,7 @@ box_tuple_extract_key(box_tuple_t *tuple, uint32_t space_id,
* \retval >=0 on success
*/
int
-box_index_stat(uint32_t space_id, uint32_t index_id,
- struct info_handler *info);
+box_index_stat(uint32_t space_id, uint32_t index_id, struct info_handler *info);
/**
* Trigger index compaction (index:compact())
@@ -293,8 +292,8 @@ struct snapshot_iterator {
* Returns a pointer to the tuple data and its
* size or NULL if EOF.
*/
- int (*next)(struct snapshot_iterator *,
- const char **data, uint32_t *size);
+ int (*next)(struct snapshot_iterator *, const char **data,
+ uint32_t *size);
/**
* Destroy the iterator.
*/
@@ -403,22 +402,23 @@ struct index_vtab {
ssize_t (*size)(struct index *);
ssize_t (*bsize)(struct index *);
- int (*min)(struct index *index, const char *key,
- uint32_t part_count, struct tuple **result);
- int (*max)(struct index *index, const char *key,
- uint32_t part_count, struct tuple **result);
+ int (*min)(struct index *index, const char *key, uint32_t part_count,
+ struct tuple **result);
+ int (*max)(struct index *index, const char *key, uint32_t part_count,
+ struct tuple **result);
int (*random)(struct index *index, uint32_t rnd, struct tuple **result);
ssize_t (*count)(struct index *index, enum iterator_type type,
const char *key, uint32_t part_count);
- int (*get)(struct index *index, const char *key,
- uint32_t part_count, struct tuple **result);
+ int (*get)(struct index *index, const char *key, uint32_t part_count,
+ struct tuple **result);
int (*replace)(struct index *index, struct tuple *old_tuple,
struct tuple *new_tuple, enum dup_replace_mode mode,
struct tuple **result);
/** Create an index iterator. */
struct iterator *(*create_iterator)(struct index *index,
- enum iterator_type type,
- const char *key, uint32_t part_count);
+ enum iterator_type type,
+ const char *key,
+ uint32_t part_count);
/**
* Create an ALL iterator with personal read view so further
* index modifications will not affect the iteration results.
@@ -475,8 +475,8 @@ replace_check_dup(struct tuple *old_tuple, struct tuple *dup_tuple,
* dup_replace_mode is DUP_REPLACE, and
* a tuple with the same key is not found.
*/
- return old_tuple ?
- ER_CANT_UPDATE_PRIMARY_KEY : ER_TUPLE_NOT_FOUND;
+ return old_tuple ? ER_CANT_UPDATE_PRIMARY_KEY :
+ ER_TUPLE_NOT_FOUND;
}
} else { /* dup_tuple != NULL */
if (dup_tuple != old_tuple &&
@@ -589,15 +589,15 @@ index_bsize(struct index *index)
}
static inline int
-index_min(struct index *index, const char *key,
- uint32_t part_count, struct tuple **result)
+index_min(struct index *index, const char *key, uint32_t part_count,
+ struct tuple **result)
{
return index->vtab->min(index, key, part_count, result);
}
static inline int
-index_max(struct index *index, const char *key,
- uint32_t part_count, struct tuple **result)
+index_max(struct index *index, const char *key, uint32_t part_count,
+ struct tuple **result)
{
return index->vtab->max(index, key, part_count, result);
}
@@ -609,15 +609,15 @@ index_random(struct index *index, uint32_t rnd, struct tuple **result)
}
static inline ssize_t
-index_count(struct index *index, enum iterator_type type,
- const char *key, uint32_t part_count)
+index_count(struct index *index, enum iterator_type type, const char *key,
+ uint32_t part_count)
{
return index->vtab->count(index, type, key, part_count);
}
static inline int
-index_get(struct index *index, const char *key,
- uint32_t part_count, struct tuple **result)
+index_get(struct index *index, const char *key, uint32_t part_count,
+ struct tuple **result)
{
return index->vtab->get(index, key, part_count, result);
}
@@ -688,35 +688,57 @@ index_end_build(struct index *index)
/*
* Virtual method stubs.
*/
-void generic_index_commit_create(struct index *, int64_t);
-void generic_index_abort_create(struct index *);
-void generic_index_commit_modify(struct index *, int64_t);
-void generic_index_commit_drop(struct index *, int64_t);
-void generic_index_update_def(struct index *);
-bool generic_index_depends_on_pk(struct index *);
-bool generic_index_def_change_requires_rebuild(struct index *,
- const struct index_def *);
-ssize_t generic_index_bsize(struct index *);
-ssize_t generic_index_size(struct index *);
-int generic_index_min(struct index *, const char *, uint32_t, struct tuple **);
-int generic_index_max(struct index *, const char *, uint32_t, struct tuple **);
-int generic_index_random(struct index *, uint32_t, struct tuple **);
-ssize_t generic_index_count(struct index *, enum iterator_type,
- const char *, uint32_t);
-int generic_index_get(struct index *, const char *, uint32_t, struct tuple **);
-int generic_index_replace(struct index *, struct tuple *, struct tuple *,
- enum dup_replace_mode, struct tuple **);
-struct snapshot_iterator *generic_index_create_snapshot_iterator(struct index *);
-void generic_index_stat(struct index *, struct info_handler *);
-void generic_index_compact(struct index *);
-void generic_index_reset_stat(struct index *);
-void generic_index_begin_build(struct index *);
-int generic_index_reserve(struct index *, uint32_t);
+void
+generic_index_commit_create(struct index *, int64_t);
+void
+generic_index_abort_create(struct index *);
+void
+generic_index_commit_modify(struct index *, int64_t);
+void
+generic_index_commit_drop(struct index *, int64_t);
+void
+generic_index_update_def(struct index *);
+bool
+generic_index_depends_on_pk(struct index *);
+bool
+generic_index_def_change_requires_rebuild(struct index *,
+ const struct index_def *);
+ssize_t
+generic_index_bsize(struct index *);
+ssize_t
+generic_index_size(struct index *);
+int
+generic_index_min(struct index *, const char *, uint32_t, struct tuple **);
+int
+generic_index_max(struct index *, const char *, uint32_t, struct tuple **);
+int
+generic_index_random(struct index *, uint32_t, struct tuple **);
+ssize_t
+generic_index_count(struct index *, enum iterator_type, const char *, uint32_t);
+int
+generic_index_get(struct index *, const char *, uint32_t, struct tuple **);
+int
+generic_index_replace(struct index *, struct tuple *, struct tuple *,
+ enum dup_replace_mode, struct tuple **);
+struct snapshot_iterator *
+generic_index_create_snapshot_iterator(struct index *);
+void
+generic_index_stat(struct index *, struct info_handler *);
+void
+generic_index_compact(struct index *);
+void
+generic_index_reset_stat(struct index *);
+void
+generic_index_begin_build(struct index *);
+int
+generic_index_reserve(struct index *, uint32_t);
struct iterator *
generic_index_create_iterator(struct index *base, enum iterator_type type,
const char *key, uint32_t part_count);
-int generic_index_build_next(struct index *, struct tuple *);
-void generic_index_end_build(struct index *);
+int
+generic_index_build_next(struct index *, struct tuple *);
+void
+generic_index_end_build(struct index *);
int
disabled_index_build_next(struct index *index, struct tuple *tuple);
int
@@ -739,8 +761,7 @@ public:
struct index_def *index_def, const char *what);
};
-struct IteratorGuard
-{
+struct IteratorGuard {
struct iterator *it;
IteratorGuard(struct iterator *it_arg) : it(it_arg) {}
~IteratorGuard() { iterator_delete(it); }
@@ -755,8 +776,8 @@ static inline struct iterator *
index_create_iterator_xc(struct index *index, enum iterator_type type,
const char *key, uint32_t part_count)
{
- struct iterator *it = index_create_iterator(index, type,
- key, part_count);
+ struct iterator *it =
+ index_create_iterator(index, type, key, part_count);
if (it == NULL)
diag_raise();
return it;
diff --git a/src/box/index_def.c b/src/box/index_def.c
index 9802961..5b1c538 100644
--- a/src/box/index_def.c
+++ b/src/box/index_def.c
@@ -60,7 +60,8 @@ const struct opt_def index_opts_reg[] = {
distance, NULL),
OPT_DEF("range_size", OPT_INT64, struct index_opts, range_size),
OPT_DEF("page_size", OPT_INT64, struct index_opts, page_size),
- OPT_DEF("run_count_per_level", OPT_INT64, struct index_opts, run_count_per_level),
+ OPT_DEF("run_count_per_level", OPT_INT64, struct index_opts,
+ run_count_per_level),
OPT_DEF("run_size_ratio", OPT_FLOAT, struct index_opts, run_size_ratio),
OPT_DEF("bloom_fpr", OPT_FLOAT, struct index_opts, bloom_fpr),
OPT_DEF("lsn", OPT_INT64, struct index_opts, lsn),
@@ -72,14 +73,15 @@ const struct opt_def index_opts_reg[] = {
struct index_def *
index_def_new(uint32_t space_id, uint32_t iid, const char *name,
uint32_t name_len, enum index_type type,
- const struct index_opts *opts,
- struct key_def *key_def, struct key_def *pk_def)
+ const struct index_opts *opts, struct key_def *key_def,
+ struct key_def *pk_def)
{
assert(name_len <= BOX_NAME_MAX);
/* Use calloc to make index_def_delete() safe at all times. */
- struct index_def *def = (struct index_def *) calloc(1, sizeof(*def));
+ struct index_def *def = (struct index_def *)calloc(1, sizeof(*def));
if (def == NULL) {
- diag_set(OutOfMemory, sizeof(*def), "malloc", "struct index_def");
+ diag_set(OutOfMemory, sizeof(*def), "malloc",
+ "struct index_def");
return NULL;
}
def->name = strndup(name, name_len);
@@ -95,7 +97,7 @@ index_def_new(uint32_t space_id, uint32_t iid, const char *name,
def->key_def = key_def_dup(key_def);
if (iid != 0) {
def->cmp_def = key_def_merge(key_def, pk_def);
- if (! opts->is_unique) {
+ if (!opts->is_unique) {
def->cmp_def->unique_part_count =
def->cmp_def->part_count;
} else {
@@ -121,7 +123,7 @@ index_def_new(uint32_t space_id, uint32_t iid, const char *name,
struct index_def *
index_def_dup(const struct index_def *def)
{
- struct index_def *dup = (struct index_def *) malloc(sizeof(*dup));
+ struct index_def *dup = (struct index_def *)malloc(sizeof(*dup));
if (dup == NULL) {
diag_set(OutOfMemory, sizeof(*dup), "malloc",
"struct index_def");
@@ -180,7 +182,7 @@ index_stat_dup(const struct index_stat *src)
{
size_t size = index_stat_sizeof(src->samples, src->sample_count,
src->sample_field_count);
- struct index_stat *dup = (struct index_stat *) malloc(size);
+ struct index_stat *dup = (struct index_stat *)malloc(size);
if (dup == NULL) {
diag_set(OutOfMemory, size, "malloc", "index stat");
return NULL;
@@ -188,21 +190,21 @@ index_stat_dup(const struct index_stat *src)
memcpy(dup, src, size);
uint32_t array_size = src->sample_field_count * sizeof(uint32_t);
uint32_t stat1_offset = sizeof(struct index_stat);
- char *pos = (char *) dup + stat1_offset;
- dup->tuple_stat1 = (uint32_t *) pos;
+ char *pos = (char *)dup + stat1_offset;
+ dup->tuple_stat1 = (uint32_t *)pos;
pos += array_size + sizeof(uint32_t);
- dup->tuple_log_est = (log_est_t *) pos;
+ dup->tuple_log_est = (log_est_t *)pos;
pos += array_size + sizeof(uint32_t);
- dup->avg_eq = (uint32_t *) pos;
+ dup->avg_eq = (uint32_t *)pos;
pos += array_size;
- dup->samples = (struct index_sample *) pos;
+ dup->samples = (struct index_sample *)pos;
pos += src->sample_count * sizeof(struct index_sample);
for (uint32_t i = 0; i < src->sample_count; ++i) {
- dup->samples[i].eq = (uint32_t *) pos;
+ dup->samples[i].eq = (uint32_t *)pos;
pos += array_size;
- dup->samples[i].lt = (uint32_t *) pos;
+ dup->samples[i].lt = (uint32_t *)pos;
pos += array_size;
- dup->samples[i].dlt = (uint32_t *) pos;
+ dup->samples[i].dlt = (uint32_t *)pos;
pos += array_size;
dup->samples[i].sample_key = pos;
pos += dup->samples[i].key_size;
@@ -240,7 +242,7 @@ index_def_cmp(const struct index_def *key1, const struct index_def *key2)
if (strcmp(key1->name, key2->name))
return strcmp(key1->name, key2->name);
if (key1->type != key2->type)
- return (int) key1->type < (int) key2->type ? -1 : 1;
+ return (int)key1->type < (int)key2->type ? -1 : 1;
if (index_opts_cmp(&key1->opts, &key2->opts))
return index_opts_cmp(&key1->opts, &key2->opts);
@@ -256,9 +258,8 @@ index_def_to_key_def(struct rlist *index_defs, int *size)
rlist_foreach_entry(index_def, index_defs, link)
key_count++;
size_t bsize;
- struct key_def **keys =
- region_alloc_array(&fiber()->gc, typeof(keys[0]), key_count,
- &bsize);
+ struct key_def **keys = region_alloc_array(
+ &fiber()->gc, typeof(keys[0]), key_count, &bsize);
if (keys == NULL) {
diag_set(OutOfMemory, bsize, "region_alloc_array", "keys");
return NULL;
@@ -301,12 +302,13 @@ index_def_is_valid(struct index_def *index_def, const char *space_name)
}
if (index_def->iid == 0 && index_def->key_def->for_func_index) {
diag_set(ClientError, ER_MODIFY_INDEX, index_def->name,
- space_name, "primary key can not use a function");
+ space_name, "primary key can not use a function");
return false;
}
for (uint32_t i = 0; i < index_def->key_def->part_count; i++) {
assert(index_def->key_def->parts[i].type < field_type_MAX);
- if (index_def->key_def->parts[i].fieldno > BOX_INDEX_FIELD_MAX) {
+ if (index_def->key_def->parts[i].fieldno >
+ BOX_INDEX_FIELD_MAX) {
diag_set(ClientError, ER_MODIFY_INDEX, index_def->name,
space_name, "field no is too big");
return false;
diff --git a/src/box/index_def.h b/src/box/index_def.h
index d928b23..a0088b9 100644
--- a/src/box/index_def.h
+++ b/src/box/index_def.h
@@ -41,16 +41,16 @@ extern "C" {
enum index_type {
HASH = 0, /* HASH Index */
- TREE, /* TREE Index */
- BITSET, /* BITSET Index */
- RTREE, /* R-Tree Index */
+ TREE, /* TREE Index */
+ BITSET, /* BITSET Index */
+ RTREE, /* R-Tree Index */
index_type_MAX,
};
extern const char *index_type_strs[];
enum rtree_index_distance_type {
- /* Euclid distance, sqrt(dx*dx + dy*dy) */
+ /* Euclid distance, sqrt(dx*dx + dy*dy) */
RTREE_INDEX_DISTANCE_TYPE_EUCLID,
/* Manhattan distance, fabs(dx) + fabs(dy) */
RTREE_INDEX_DISTANCE_TYPE_MANHATTAN,
@@ -203,8 +203,8 @@ index_opts_cmp(const struct index_opts *o1, const struct index_opts *o2)
if (o1->page_size != o2->page_size)
return o1->page_size < o2->page_size ? -1 : 1;
if (o1->run_count_per_level != o2->run_count_per_level)
- return o1->run_count_per_level < o2->run_count_per_level ?
- -1 : 1;
+ return o1->run_count_per_level < o2->run_count_per_level ? -1 :
+ 1;
if (o1->run_size_ratio != o2->run_size_ratio)
return o1->run_size_ratio < o2->run_size_ratio ? -1 : 1;
if (o1->bloom_fpr != o2->bloom_fpr)
@@ -310,8 +310,8 @@ index_def_update_optionality(struct index_def *def, uint32_t min_field_count)
static inline void
index_def_set_func(struct index_def *def, struct func *func)
{
- assert(def->opts.func_id > 0 &&
- def->key_def->for_func_index && def->cmp_def->for_func_index);
+ assert(def->opts.func_id > 0 && def->key_def->for_func_index &&
+ def->cmp_def->for_func_index);
/*
* def->key_def is used in key_list module to build a key
* a key for given tuple.
@@ -368,8 +368,8 @@ index_def_list_add(struct rlist *index_def_list, struct index_def *index_def)
struct index_def *
index_def_new(uint32_t space_id, uint32_t iid, const char *name,
uint32_t name_len, enum index_type type,
- const struct index_opts *opts,
- struct key_def *key_def, struct key_def *pk_def);
+ const struct index_opts *opts, struct key_def *key_def,
+ struct key_def *pk_def);
/**
* Create an array (on a region) of key_defs from list of index
@@ -415,7 +415,7 @@ index_def_dup_xc(const struct index_def *def)
static inline void
index_def_check_xc(struct index_def *index_def, const char *space_name)
{
- if (! index_def_is_valid(index_def, space_name))
+ if (!index_def_is_valid(index_def, space_name))
diag_raise();
}
diff --git a/src/box/iproto.cc b/src/box/iproto.cc
index b8f65e5..d31c58c 100644
--- a/src/box/iproto.cc
+++ b/src/box/iproto.cc
@@ -56,7 +56,7 @@
#include "tuple_convert.h"
#include "session.h"
#include "xrow.h"
-#include "schema.h" /* schema_version */
+#include "schema.h" /* schema_version */
#include "replication.h" /* instance_uuid */
#include "iproto_constants.h"
#include "rmean.h"
@@ -142,7 +142,7 @@ iproto_bound_address(void)
{
if (iproto_bound_address_len == 0)
return NULL;
- return sio_strfaddr((struct sockaddr *) &iproto_bound_address_storage,
+ return sio_strfaddr((struct sockaddr *)&iproto_bound_address_storage,
iproto_bound_address_len);
}
@@ -180,8 +180,7 @@ iproto_reset_input(struct ibuf *ibuf)
* from all connections are queued into a single queue
* and processed in FIFO order.
*/
-struct iproto_msg
-{
+struct iproto_msg {
struct cmsg base;
struct iproto_connection *connection;
@@ -339,11 +338,8 @@ iproto_process_push(struct cmsg *m);
static void
tx_end_push(struct cmsg *m);
-static const struct cmsg_hop push_route[] = {
- { iproto_process_push, &tx_pipe },
- { tx_end_push, NULL }
-};
-
+static const struct cmsg_hop push_route[] = { { iproto_process_push, &tx_pipe },
+ { tx_end_push, NULL } };
/* }}} */
@@ -408,8 +404,7 @@ enum iproto_connection_state {
* messages are
* discarded
*/
-struct iproto_connection
-{
+struct iproto_connection {
/**
* Two rotating buffers for input. Input is first read into
* ibuf[0]. As soon as it buffer becomes full, the buffers are
@@ -522,8 +517,8 @@ struct iproto_connection
*/
struct {
alignas(CACHELINE_SIZE)
- /** Pointer to the current output buffer. */
- struct obuf *p_obuf;
+ /** Pointer to the current output buffer. */
+ struct obuf *p_obuf;
/** True if Kharon is in use/travelling. */
bool is_push_sent;
/**
@@ -547,14 +542,14 @@ static inline bool
iproto_check_msg_max(void)
{
size_t request_count = mempool_count(&iproto_msg_pool);
- return request_count > (size_t) iproto_msg_max;
+ return request_count > (size_t)iproto_msg_max;
}
static struct iproto_msg *
iproto_msg_new(struct iproto_connection *con)
{
struct iproto_msg *msg =
- (struct iproto_msg *) mempool_alloc(&iproto_msg_pool);
+ (struct iproto_msg *)mempool_alloc(&iproto_msg_pool);
ERROR_INJECT(ERRINJ_TESTING, {
mempool_free(&iproto_msg_pool, msg);
msg = NULL;
@@ -562,7 +557,8 @@ iproto_msg_new(struct iproto_connection *con)
if (msg == NULL) {
diag_set(OutOfMemory, sizeof(*msg), "mempool_alloc", "msg");
say_warn("can not allocate memory for a new message, "
- "connection %s", sio_socketname(con->input.fd));
+ "connection %s",
+ sio_socketname(con->input.fd));
return NULL;
}
msg->connection = con;
@@ -588,8 +584,7 @@ iproto_msg_new(struct iproto_connection *con)
static inline bool
iproto_connection_is_idle(struct iproto_connection *con)
{
- return con->long_poll_count == 0 &&
- ibuf_used(&con->ibuf[0]) == 0 &&
+ return con->long_poll_count == 0 && ibuf_used(&con->ibuf[0]) == 0 &&
ibuf_used(&con->ibuf[1]) == 0;
}
@@ -821,19 +816,18 @@ iproto_enqueue_batch(struct iproto_connection *con, struct ibuf *in)
/* Read request length. */
if (mp_typeof(*pos) != MP_UINT) {
errmsg = "packet length";
-err_msgpack:
+ err_msgpack:
cpipe_flush_input(&tx_pipe);
- diag_set(ClientError, ER_INVALID_MSGPACK,
- errmsg);
+ diag_set(ClientError, ER_INVALID_MSGPACK, errmsg);
return -1;
}
if (mp_check_uint(pos, in->wpos) >= 0)
break;
uint64_t len = mp_decode_uint(&pos);
if (len > IPROTO_PACKET_SIZE_MAX) {
- errmsg = tt_sprintf("too big packet size in the "\
+ errmsg = tt_sprintf("too big packet size in the "
"header: %llu",
- (unsigned long long) len);
+ (unsigned long long)len);
goto err_msgpack;
}
const char *reqend = pos + len;
@@ -862,7 +856,7 @@ err_msgpack:
n_requests++;
/* Request is parsed */
assert(reqend > reqstart);
- assert(con->parse_size >= (size_t) (reqend - reqstart));
+ assert(con->parse_size >= (size_t)(reqend - reqstart));
con->parse_size -= reqend - reqstart;
}
if (stop_input) {
@@ -908,7 +902,7 @@ err_msgpack:
static void
iproto_connection_resume(struct iproto_connection *con)
{
- assert(! iproto_check_msg_max());
+ assert(!iproto_check_msg_max());
rlist_del(&con->in_stop_list);
/*
* Enqueue_batch() stops the connection again, if the
@@ -939,10 +933,9 @@ iproto_resume(void)
* Shift from list head to ensure strict FIFO
* (fairness) for resumed connections.
*/
- struct iproto_connection *con =
- rlist_first_entry(&stopped_connections,
- struct iproto_connection,
- in_stop_list);
+ struct iproto_connection *con = rlist_first_entry(
+ &stopped_connections, struct iproto_connection,
+ in_stop_list);
iproto_connection_resume(con);
}
}
@@ -952,7 +945,7 @@ iproto_connection_on_input(ev_loop *loop, struct ev_io *watcher,
int /* revents */)
{
struct iproto_connection *con =
- (struct iproto_connection *) watcher->data;
+ (struct iproto_connection *)watcher->data;
int fd = con->input.fd;
assert(fd >= 0);
assert(rlist_empty(&con->in_stop_list));
@@ -976,13 +969,13 @@ iproto_connection_on_input(ev_loop *loop, struct ev_io *watcher,
}
/* Read input. */
int nrd = sio_read(fd, in->wpos, ibuf_unused(in));
- if (nrd < 0) { /* Socket is not ready. */
- if (! sio_wouldblock(errno))
+ if (nrd < 0) { /* Socket is not ready. */
+ if (!sio_wouldblock(errno))
diag_raise();
ev_io_start(loop, &con->input);
return;
}
- if (nrd == 0) { /* EOF */
+ if (nrd == 0) { /* EOF */
iproto_connection_close(con);
return;
}
@@ -1030,7 +1023,7 @@ iproto_flush(struct iproto_connection *con)
return 1;
}
assert(begin->used < end->used);
- struct iovec iov[SMALL_OBUF_IOV_MAX+1];
+ struct iovec iov[SMALL_OBUF_IOV_MAX + 1];
struct iovec *src = obuf->iov;
int iovcnt = end->pos - begin->pos + 1;
/*
@@ -1040,7 +1033,7 @@ iproto_flush(struct iproto_connection *con)
memcpy(iov, src + begin->pos, iovcnt * sizeof(struct iovec));
sio_add_to_iov(iov, -begin->iov_len);
/* *Overwrite* iov_len of the last pos as it may be garbage. */
- iov[iovcnt-1].iov_len = end->iov_len - begin->iov_len * (iovcnt == 1);
+ iov[iovcnt - 1].iov_len = end->iov_len - begin->iov_len * (iovcnt == 1);
ssize_t nwr = sio_writev(fd, iov, iovcnt);
@@ -1054,11 +1047,12 @@ iproto_flush(struct iproto_connection *con)
size_t offset = 0;
int advance = 0;
advance = sio_move_iov(iov, nwr, &offset);
- begin->used += nwr; /* advance write position */
- begin->iov_len = advance == 0 ? begin->iov_len + offset: offset;
+ begin->used += nwr; /* advance write position */
+ begin->iov_len = advance == 0 ? begin->iov_len + offset :
+ offset;
begin->pos += advance;
assert(begin->pos <= end->pos);
- } else if (nwr < 0 && ! sio_wouldblock(errno)) {
+ } else if (nwr < 0 && !sio_wouldblock(errno)) {
diag_raise();
}
return -1;
@@ -1068,7 +1062,8 @@ static void
iproto_connection_on_output(ev_loop *loop, struct ev_io *watcher,
int /* revents */)
{
- struct iproto_connection *con = (struct iproto_connection *) watcher->data;
+ struct iproto_connection *con =
+ (struct iproto_connection *)watcher->data;
try {
int rc;
@@ -1077,7 +1072,7 @@ iproto_connection_on_output(ev_loop *loop, struct ev_io *watcher,
ev_io_start(loop, &con->output);
return;
}
- if (! ev_is_active(&con->input) &&
+ if (!ev_is_active(&con->input) &&
rlist_empty(&con->in_stop_list)) {
ev_feed_event(loop, &con->input, EV_READ);
}
@@ -1093,8 +1088,9 @@ iproto_connection_on_output(ev_loop *loop, struct ev_io *watcher,
static struct iproto_connection *
iproto_connection_new(int fd)
{
- struct iproto_connection *con = (struct iproto_connection *)
- mempool_alloc(&iproto_connection_pool);
+ struct iproto_connection *con =
+ (struct iproto_connection *)mempool_alloc(
+ &iproto_connection_pool);
if (con == NULL) {
diag_set(OutOfMemory, sizeof(*con), "mempool_alloc", "con");
return NULL;
@@ -1140,10 +1136,8 @@ iproto_connection_delete(struct iproto_connection *con)
*/
ibuf_destroy(&con->ibuf[0]);
ibuf_destroy(&con->ibuf[1]);
- assert(con->obuf[0].pos == 0 &&
- con->obuf[0].iov[0].iov_base == NULL);
- assert(con->obuf[1].pos == 0 &&
- con->obuf[1].iov[0].iov_base == NULL);
+ assert(con->obuf[0].pos == 0 && con->obuf[0].iov[0].iov_base == NULL);
+ assert(con->obuf[1].pos == 0 && con->obuf[1].iov[0].iov_base == NULL);
mempool_free(&iproto_connection_pool, con);
}
@@ -1213,20 +1207,20 @@ static const struct cmsg_hop sql_route[] = {
};
static const struct cmsg_hop *dml_route[IPROTO_TYPE_STAT_MAX] = {
- NULL, /* IPROTO_OK */
- select_route, /* IPROTO_SELECT */
- process1_route, /* IPROTO_INSERT */
- process1_route, /* IPROTO_REPLACE */
- process1_route, /* IPROTO_UPDATE */
- process1_route, /* IPROTO_DELETE */
- call_route, /* IPROTO_CALL_16 */
- misc_route, /* IPROTO_AUTH */
- call_route, /* IPROTO_EVAL */
- process1_route, /* IPROTO_UPSERT */
- call_route, /* IPROTO_CALL */
- sql_route, /* IPROTO_EXECUTE */
- NULL, /* IPROTO_NOP */
- sql_route, /* IPROTO_PREPARE */
+ NULL, /* IPROTO_OK */
+ select_route, /* IPROTO_SELECT */
+ process1_route, /* IPROTO_INSERT */
+ process1_route, /* IPROTO_REPLACE */
+ process1_route, /* IPROTO_UPDATE */
+ process1_route, /* IPROTO_DELETE */
+ call_route, /* IPROTO_CALL_16 */
+ misc_route, /* IPROTO_AUTH */
+ call_route, /* IPROTO_EVAL */
+ process1_route, /* IPROTO_UPSERT */
+ call_route, /* IPROTO_CALL */
+ sql_route, /* IPROTO_EXECUTE */
+ NULL, /* IPROTO_NOP */
+ sql_route, /* IPROTO_PREPARE */
};
static const struct cmsg_hop join_route[] = {
@@ -1271,7 +1265,7 @@ iproto_msg_decode(struct iproto_msg *msg, const char **pos, const char *reqend,
if (xrow_decode_dml(&msg->header, &msg->dml,
dml_request_key_map(type)))
goto error;
- assert(type < sizeof(dml_route)/sizeof(*dml_route));
+ assert(type < sizeof(dml_route) / sizeof(*dml_route));
cmsg_init(&msg->base, dml_route[type]);
break;
case IPROTO_CALL_16:
@@ -1310,8 +1304,7 @@ iproto_msg_decode(struct iproto_msg *msg, const char **pos, const char *reqend,
cmsg_init(&msg->base, misc_route);
break;
default:
- diag_set(ClientError, ER_UNKNOWN_REQUEST_TYPE,
- (uint32_t) type);
+ diag_set(ClientError, ER_UNKNOWN_REQUEST_TYPE, (uint32_t)type);
goto error;
}
return;
@@ -1354,7 +1347,7 @@ tx_process_disconnect(struct cmsg *m)
container_of(m, struct iproto_connection, disconnect_msg);
if (con->session != NULL) {
session_close(con->session);
- if (! rlist_empty(&session_on_disconnect)) {
+ if (!rlist_empty(&session_on_disconnect)) {
tx_fiber_init(con->session, 0);
session_run_on_disconnect_triggers(con->session);
}
@@ -1403,7 +1396,6 @@ net_finish_destroy(struct cmsg *m)
iproto_connection_delete(con);
}
-
static int
tx_check_schema(uint32_t new_schema_version)
{
@@ -1418,8 +1410,8 @@ tx_check_schema(uint32_t new_schema_version)
static void
net_discard_input(struct cmsg *m)
{
- struct iproto_msg *msg = container_of(m, struct iproto_msg,
- discard_input);
+ struct iproto_msg *msg =
+ container_of(m, struct iproto_msg, discard_input);
struct iproto_connection *con = msg->connection;
msg->p_ibuf->rpos += msg->len;
msg->len = 0;
@@ -1481,7 +1473,7 @@ tx_accept_wpos(struct iproto_connection *con, const struct iproto_wpos *wpos)
static inline struct iproto_msg *
tx_accept_msg(struct cmsg *m)
{
- struct iproto_msg *msg = (struct iproto_msg *) m;
+ struct iproto_msg *msg = (struct iproto_msg *)m;
tx_accept_wpos(msg->connection, &msg->wpos);
tx_fiber_init(msg->connection->session, msg->header.sync);
return msg;
@@ -1509,8 +1501,8 @@ tx_reply_iproto_error(struct cmsg *m)
{
struct iproto_msg *msg = tx_accept_msg(m);
struct obuf *out = msg->connection->tx.p_obuf;
- iproto_reply_error(out, diag_last_error(&msg->diag),
- msg->header.sync, ::schema_version);
+ iproto_reply_error(out, diag_last_error(&msg->diag), msg->header.sync,
+ ::schema_version);
iproto_wpos_create(&msg->wpos, out);
}
@@ -1564,9 +1556,8 @@ tx_process_select(struct cmsg *m)
goto error;
tx_inject_delay();
- rc = box_select(req->space_id, req->index_id,
- req->iterator, req->offset, req->limit,
- req->key, req->key_end, &port);
+ rc = box_select(req->space_id, req->index_id, req->iterator,
+ req->offset, req->limit, req->key, req->key_end, &port);
if (rc < 0)
goto error;
@@ -1585,8 +1576,8 @@ tx_process_select(struct cmsg *m)
obuf_rollback_to_svp(out, &svp);
goto error;
}
- iproto_reply_select(out, &svp, msg->header.sync,
- ::schema_version, count);
+ iproto_reply_select(out, &svp, msg->header.sync, ::schema_version,
+ count);
iproto_wpos_create(&msg->wpos, out);
return;
error:
@@ -1675,8 +1666,8 @@ tx_process_call(struct cmsg *m)
goto error;
}
- iproto_reply_select(out, &svp, msg->header.sync,
- ::schema_version, count);
+ iproto_reply_select(out, &svp, msg->header.sync, ::schema_version,
+ count);
iproto_wpos_create(&msg->wpos, out);
return;
error:
@@ -1868,7 +1859,7 @@ tx_process_replication(struct cmsg *m)
static void
net_send_msg(struct cmsg *m)
{
- struct iproto_msg *msg = (struct iproto_msg *) m;
+ struct iproto_msg *msg = (struct iproto_msg *)m;
struct iproto_connection *con = msg->connection;
if (msg->len != 0) {
@@ -1882,7 +1873,7 @@ net_send_msg(struct cmsg *m)
con->wend = msg->wpos;
if (evio_has_fd(&con->output)) {
- if (! ev_is_active(&con->output))
+ if (!ev_is_active(&con->output))
ev_feed_event(con->loop, &con->output, EV_WRITE);
} else if (iproto_connection_is_idle(con)) {
iproto_connection_close(con);
@@ -1897,7 +1888,7 @@ net_send_msg(struct cmsg *m)
static void
net_send_error(struct cmsg *m)
{
- struct iproto_msg *msg = (struct iproto_msg *) m;
+ struct iproto_msg *msg = (struct iproto_msg *)m;
/* Recycle the exception. */
diag_move(&msg->diag, &fiber()->diag);
net_send_msg(m);
@@ -1906,13 +1897,13 @@ net_send_error(struct cmsg *m)
static void
net_end_join(struct cmsg *m)
{
- struct iproto_msg *msg = (struct iproto_msg *) m;
+ struct iproto_msg *msg = (struct iproto_msg *)m;
struct iproto_connection *con = msg->connection;
msg->p_ibuf->rpos += msg->len;
iproto_msg_delete(msg);
- assert(! ev_is_active(&con->input));
+ assert(!ev_is_active(&con->input));
/*
* Enqueue any messages if they are in the readahead
* queue. Will simply start input otherwise.
@@ -1924,13 +1915,13 @@ net_end_join(struct cmsg *m)
static void
net_end_subscribe(struct cmsg *m)
{
- struct iproto_msg *msg = (struct iproto_msg *) m;
+ struct iproto_msg *msg = (struct iproto_msg *)m;
struct iproto_connection *con = msg->connection;
msg->p_ibuf->rpos += msg->len;
iproto_msg_delete(msg);
- assert(! ev_is_active(&con->input));
+ assert(!ev_is_active(&con->input));
iproto_connection_close(con);
}
@@ -1943,23 +1934,23 @@ net_end_subscribe(struct cmsg *m)
static void
tx_process_connect(struct cmsg *m)
{
- struct iproto_msg *msg = (struct iproto_msg *) m;
+ struct iproto_msg *msg = (struct iproto_msg *)m;
struct iproto_connection *con = msg->connection;
struct obuf *out = msg->connection->tx.p_obuf;
- try { /* connect. */
+ try { /* connect. */
con->session = session_create(SESSION_TYPE_BINARY);
if (con->session == NULL)
diag_raise();
con->session->meta.connection = con;
tx_fiber_init(con->session, 0);
- char *greeting = (char *) static_alloc(IPROTO_GREETING_SIZE);
+ char *greeting = (char *)static_alloc(IPROTO_GREETING_SIZE);
/* TODO: dirty read from tx thread */
struct tt_uuid uuid = INSTANCE_UUID;
random_bytes(con->salt, IPROTO_SALT_SIZE);
greeting_encode(greeting, tarantool_version_id(), &uuid,
con->salt, IPROTO_SALT_SIZE);
obuf_dup_xc(out, greeting, IPROTO_GREETING_SIZE);
- if (! rlist_empty(&session_on_connect)) {
+ if (!rlist_empty(&session_on_connect)) {
if (session_run_on_connect_triggers(con->session) != 0)
diag_raise();
}
@@ -1977,17 +1968,17 @@ tx_process_connect(struct cmsg *m)
static void
net_send_greeting(struct cmsg *m)
{
- struct iproto_msg *msg = (struct iproto_msg *) m;
+ struct iproto_msg *msg = (struct iproto_msg *)m;
struct iproto_connection *con = msg->connection;
if (msg->close_connection) {
struct obuf *out = msg->wpos.obuf;
- int64_t nwr = sio_writev(con->output.fd, out->iov,
- obuf_iovcnt(out));
+ int64_t nwr =
+ sio_writev(con->output.fd, out->iov, obuf_iovcnt(out));
if (nwr > 0) {
/* Count statistics. */
rmean_collect(rmean_net, IPROTO_SENT, nwr);
- } else if (nwr < 0 && ! sio_wouldblock(errno)) {
+ } else if (nwr < 0 && !sio_wouldblock(errno)) {
diag_log();
}
assert(iproto_connection_is_idle(con));
@@ -2021,8 +2012,8 @@ static int
iproto_on_accept(struct evio_service * /* service */, int fd,
struct sockaddr *addr, socklen_t addrlen)
{
- (void) addr;
- (void) addrlen;
+ (void)addr;
+ (void)addrlen;
struct iproto_msg *msg;
struct iproto_connection *con = iproto_connection_new(fd);
if (con == NULL)
@@ -2051,24 +2042,21 @@ static struct evio_service binary; /* iproto binary listener */
* The network io thread main function:
* begin serving the message bus.
*/
-static int
-net_cord_f(va_list /* ap */)
+static int net_cord_f(va_list /* ap */)
{
mempool_create(&iproto_msg_pool, &cord()->slabc,
sizeof(struct iproto_msg));
mempool_create(&iproto_connection_pool, &cord()->slabc,
sizeof(struct iproto_connection));
- evio_service_init(loop(), &binary, "binary",
- iproto_on_accept, NULL);
-
+ evio_service_init(loop(), &binary, "binary", iproto_on_accept, NULL);
/* Init statistics counter */
rmean_net = rmean_new(rmean_net_strings, IPROTO_LAST);
if (rmean_net == NULL) {
- tnt_raise(OutOfMemory, sizeof(struct rmean),
- "rmean", "struct rmean");
+ tnt_raise(OutOfMemory, sizeof(struct rmean), "rmean",
+ "struct rmean");
}
struct cbus_endpoint endpoint;
@@ -2097,14 +2085,14 @@ int
iproto_session_fd(struct session *session)
{
struct iproto_connection *con =
- (struct iproto_connection *) session->meta.connection;
+ (struct iproto_connection *)session->meta.connection;
return con->output.fd;
}
int64_t
iproto_session_sync(struct session *session)
{
- (void) session;
+ (void)session;
assert(session == fiber()->storage.session);
return fiber()->storage.net.sync;
}
@@ -2114,7 +2102,7 @@ iproto_session_sync(struct session *session)
static void
iproto_process_push(struct cmsg *m)
{
- struct iproto_kharon *kharon = (struct iproto_kharon *) m;
+ struct iproto_kharon *kharon = (struct iproto_kharon *)m;
struct iproto_connection *con =
container_of(kharon, struct iproto_connection, kharon);
con->wend = kharon->wpos;
@@ -2130,18 +2118,18 @@ iproto_process_push(struct cmsg *m)
static void
tx_begin_push(struct iproto_connection *con)
{
- assert(! con->tx.is_push_sent);
+ assert(!con->tx.is_push_sent);
cmsg_init(&con->kharon.base, push_route);
iproto_wpos_create(&con->kharon.wpos, con->tx.p_obuf);
con->tx.is_push_pending = false;
con->tx.is_push_sent = true;
- cpipe_push(&net_pipe, (struct cmsg *) &con->kharon);
+ cpipe_push(&net_pipe, (struct cmsg *)&con->kharon);
}
static void
tx_end_push(struct cmsg *m)
{
- struct iproto_kharon *kharon = (struct iproto_kharon *) m;
+ struct iproto_kharon *kharon = (struct iproto_kharon *)m;
struct iproto_connection *con =
container_of(kharon, struct iproto_connection, kharon);
tx_accept_wpos(con, &kharon->wpos);
@@ -2164,7 +2152,7 @@ static int
iproto_session_push(struct session *session, struct port *port)
{
struct iproto_connection *con =
- (struct iproto_connection *) session->meta.connection;
+ (struct iproto_connection *)session->meta.connection;
struct obuf_svp svp;
if (iproto_prepare_select(con->tx.p_obuf, &svp) != 0)
return -1;
@@ -2174,7 +2162,7 @@ iproto_session_push(struct session *session, struct port *port)
}
iproto_reply_chunk(con->tx.p_obuf, &svp, iproto_session_sync(session),
::schema_version);
- if (! con->tx.is_push_sent)
+ if (!con->tx.is_push_sent)
tx_begin_push(con);
else
con->tx.is_push_pending = true;
@@ -2204,10 +2192,7 @@ iproto_init(void)
}
/** Available iproto configuration changes. */
-enum iproto_cfg_op {
- IPROTO_CFG_MSG_MAX,
- IPROTO_CFG_LISTEN
-};
+enum iproto_cfg_op { IPROTO_CFG_MSG_MAX, IPROTO_CFG_LISTEN };
/**
* Since there is no way to "synchronously" change the
@@ -2215,8 +2200,7 @@ enum iproto_cfg_op {
* message count in flight send a special message to iproto
* thread.
*/
-struct iproto_cfg_msg: public cbus_call_msg
-{
+struct iproto_cfg_msg: public cbus_call_msg {
/** Operation to execute in iproto thread. */
enum iproto_cfg_op op;
union {
@@ -2244,7 +2228,7 @@ iproto_cfg_msg_create(struct iproto_cfg_msg *msg, enum iproto_cfg_op op)
static int
iproto_do_cfg_f(struct cbus_call_msg *m)
{
- struct iproto_cfg_msg *cfg_msg = (struct iproto_cfg_msg *) m;
+ struct iproto_cfg_msg *cfg_msg = (struct iproto_cfg_msg *)m;
int old;
try {
switch (cfg_msg->op) {
@@ -2278,8 +2262,8 @@ iproto_do_cfg_f(struct cbus_call_msg *m)
static inline void
iproto_do_cfg(struct iproto_cfg_msg *msg)
{
- if (cbus_call(&net_pipe, &tx_pipe, msg, iproto_do_cfg_f,
- NULL, TIMEOUT_INFINITY) != 0)
+ if (cbus_call(&net_pipe, &tx_pipe, msg, iproto_do_cfg_f, NULL,
+ TIMEOUT_INFINITY) != 0)
diag_raise();
}
diff --git a/src/box/iproto_constants.h b/src/box/iproto_constants.h
index d3738c7..eb0926b 100644
--- a/src/box/iproto_constants.h
+++ b/src/box/iproto_constants.h
@@ -94,7 +94,7 @@ enum iproto_key {
/* Also request keys. See the comment above. */
IPROTO_EXPR = 0x27, /* EVAL */
- IPROTO_OPS = 0x28, /* UPSERT but not UPDATE ops, because of legacy */
+ IPROTO_OPS = 0x28, /* UPSERT but not UPDATE ops, because of legacy */
IPROTO_BALLOT = 0x29,
IPROTO_TUPLE_META = 0x2a,
IPROTO_OPTIONS = 0x2b,
@@ -152,26 +152,28 @@ enum iproto_ballot_key {
IPROTO_BALLOT_IS_ANON = 0x05,
};
-#define bit(c) (1ULL<<IPROTO_##c)
+#define bit(c) (1ULL << IPROTO_##c)
-#define IPROTO_HEAD_BMAP (bit(REQUEST_TYPE) | bit(SYNC) | bit(REPLICA_ID) |\
- bit(LSN) | bit(SCHEMA_VERSION))
-#define IPROTO_DML_BODY_BMAP (bit(SPACE_ID) | bit(INDEX_ID) | bit(LIMIT) |\
- bit(OFFSET) | bit(ITERATOR) | bit(INDEX_BASE) |\
- bit(KEY) | bit(TUPLE) | bit(OPS) | bit(TUPLE_META))
+#define IPROTO_HEAD_BMAP \
+ (bit(REQUEST_TYPE) | bit(SYNC) | bit(REPLICA_ID) | bit(LSN) | \
+ bit(SCHEMA_VERSION))
+#define IPROTO_DML_BODY_BMAP \
+ (bit(SPACE_ID) | bit(INDEX_ID) | bit(LIMIT) | bit(OFFSET) | \
+ bit(ITERATOR) | bit(INDEX_BASE) | bit(KEY) | bit(TUPLE) | bit(OPS) | \
+ bit(TUPLE_META))
static inline bool
xrow_header_has_key(const char *pos, const char *end)
{
- unsigned char key = pos < end ? *pos : (unsigned char) IPROTO_KEY_MAX;
- return key < IPROTO_KEY_MAX && IPROTO_HEAD_BMAP & (1ULL<<key);
+ unsigned char key = pos < end ? *pos : (unsigned char)IPROTO_KEY_MAX;
+ return key < IPROTO_KEY_MAX && IPROTO_HEAD_BMAP & (1ULL << key);
}
static inline bool
iproto_dml_body_has_key(const char *pos, const char *end)
{
- unsigned char key = pos < end ? *pos : (unsigned char) IPROTO_KEY_MAX;
- return key < IPROTO_KEY_MAX && IPROTO_DML_BODY_BMAP & (1ULL<<key);
+ unsigned char key = pos < end ? *pos : (unsigned char)IPROTO_KEY_MAX;
+ return key < IPROTO_KEY_MAX && IPROTO_DML_BODY_BMAP & (1ULL << key);
}
#undef bit
@@ -319,7 +321,7 @@ static inline bool
iproto_type_is_dml(uint32_t type)
{
return (type >= IPROTO_SELECT && type <= IPROTO_DELETE) ||
- type == IPROTO_UPSERT || type == IPROTO_NOP;
+ type == IPROTO_UPSERT || type == IPROTO_NOP;
}
/**
diff --git a/src/box/iterator_type.c b/src/box/iterator_type.c
index 5d6b55f..c49f021 100644
--- a/src/box/iterator_type.c
+++ b/src/box/iterator_type.c
@@ -47,4 +47,5 @@ const char *iterator_type_strs[] = {
};
static_assert(sizeof(iterator_type_strs) / sizeof(const char *) ==
- iterator_type_MAX, "iterator_type_str constants");
+ iterator_type_MAX,
+ "iterator_type_str constants");
diff --git a/src/box/iterator_type.h b/src/box/iterator_type.h
index c57e614..5a66701 100644
--- a/src/box/iterator_type.h
+++ b/src/box/iterator_type.h
@@ -61,18 +61,19 @@ extern "C" {
*/
enum iterator_type {
/* ITER_EQ must be the first member for request_create */
- ITER_EQ = 0, /* key == x ASC order */
- ITER_REQ = 1, /* key == x DESC order */
- ITER_ALL = 2, /* all tuples */
- ITER_LT = 3, /* key < x */
- ITER_LE = 4, /* key <= x */
- ITER_GE = 5, /* key >= x */
- ITER_GT = 6, /* key > x */
- ITER_BITS_ALL_SET = 7, /* all bits from x are set in key */
- ITER_BITS_ANY_SET = 8, /* at least one x's bit is set */
- ITER_BITS_ALL_NOT_SET = 9, /* all bits are not set */
- ITER_OVERLAPS = 10, /* key overlaps x */
- ITER_NEIGHBOR = 11, /* tuples in distance ascending order from specified point */
+ ITER_EQ = 0, /* key == x ASC order */
+ ITER_REQ = 1, /* key == x DESC order */
+ ITER_ALL = 2, /* all tuples */
+ ITER_LT = 3, /* key < x */
+ ITER_LE = 4, /* key <= x */
+ ITER_GE = 5, /* key >= x */
+ ITER_GT = 6, /* key > x */
+ ITER_BITS_ALL_SET = 7, /* all bits from x are set in key */
+ ITER_BITS_ANY_SET = 8, /* at least one x's bit is set */
+ ITER_BITS_ALL_NOT_SET = 9, /* all bits are not set */
+ ITER_OVERLAPS = 10, /* key overlaps x */
+ ITER_NEIGHBOR =
+ 11, /* tuples in distance ascending order from specified point */
iterator_type_MAX
};
@@ -87,8 +88,8 @@ extern const char *iterator_type_strs[];
static inline int
iterator_direction(enum iterator_type type)
{
- const unsigned reverse =
- (1u << ITER_REQ) | (1u << ITER_LT) | (1u << ITER_LE);
+ const unsigned reverse = (1u << ITER_REQ) | (1u << ITER_LT) |
+ (1u << ITER_LE);
return (reverse & (1u << type)) ? -1 : 1;
}
diff --git a/src/box/journal.c b/src/box/journal.c
index cb320b5..afaf30f 100644
--- a/src/box/journal.c
+++ b/src/box/journal.c
@@ -36,8 +36,7 @@ struct journal *current_journal = NULL;
struct journal_entry *
journal_entry_new(size_t n_rows, struct region *region,
- journal_write_async_f write_async_cb,
- void *complete_data)
+ journal_write_async_f write_async_cb, void *complete_data)
{
struct journal_entry *entry;
@@ -51,7 +50,6 @@ journal_entry_new(size_t n_rows, struct region *region,
return NULL;
}
- journal_entry_create(entry, n_rows, 0, write_async_cb,
- complete_data);
+ journal_entry_create(entry, n_rows, 0, write_async_cb, complete_data);
return entry;
}
diff --git a/src/box/journal.h b/src/box/journal.h
index 5d8d5a7..8d9f347 100644
--- a/src/box/journal.h
+++ b/src/box/journal.h
@@ -88,15 +88,14 @@ struct region;
*/
static inline void
journal_entry_create(struct journal_entry *entry, size_t n_rows,
- size_t approx_len,
- journal_write_async_f write_async_cb,
+ size_t approx_len, journal_write_async_f write_async_cb,
void *complete_data)
{
- entry->write_async_cb = write_async_cb;
- entry->complete_data = complete_data;
- entry->approx_len = approx_len;
- entry->n_rows = n_rows;
- entry->res = -1;
+ entry->write_async_cb = write_async_cb;
+ entry->complete_data = complete_data;
+ entry->approx_len = approx_len;
+ entry->n_rows = n_rows;
+ entry->res = -1;
}
/**
@@ -106,8 +105,7 @@ journal_entry_create(struct journal_entry *entry, size_t n_rows,
*/
struct journal_entry *
journal_entry_new(size_t n_rows, struct region *region,
- journal_write_async_f write_async_cb,
- void *complete_data);
+ journal_write_async_f write_async_cb, void *complete_data);
/**
* An API for an abstract journal for all transactions of this
@@ -120,8 +118,7 @@ struct journal {
struct journal_entry *entry);
/** Synchronous write */
- int (*write)(struct journal *journal,
- struct journal_entry *entry);
+ int (*write)(struct journal *journal, struct journal_entry *entry);
};
/**
@@ -196,8 +193,8 @@ journal_create(struct journal *journal,
int (*write)(struct journal *journal,
struct journal_entry *entry))
{
- journal->write_async = write_async;
- journal->write = write;
+ journal->write_async = write_async;
+ journal->write = write;
}
static inline bool
diff --git a/src/box/key_def.c b/src/box/key_def.c
index a035372..8cc0a29 100644
--- a/src/box/key_def.c
+++ b/src/box/key_def.c
@@ -42,15 +42,13 @@
const char *sort_order_strs[] = { "asc", "desc", "undef" };
-const struct key_part_def key_part_def_default = {
- 0,
- field_type_MAX,
- COLL_NONE,
- false,
- ON_CONFLICT_ACTION_DEFAULT,
- SORT_ORDER_ASC,
- NULL
-};
+const struct key_part_def key_part_def_default = { 0,
+ field_type_MAX,
+ COLL_NONE,
+ false,
+ ON_CONFLICT_ACTION_DEFAULT,
+ SORT_ORDER_ASC,
+ NULL };
static int64_t
part_type_by_name_wrapper(const char *str, uint32_t len)
@@ -58,13 +56,13 @@ part_type_by_name_wrapper(const char *str, uint32_t len)
return field_type_by_name(str, len);
}
-#define PART_OPT_TYPE "type"
-#define PART_OPT_FIELD "field"
-#define PART_OPT_COLLATION "collation"
-#define PART_OPT_NULLABILITY "is_nullable"
+#define PART_OPT_TYPE "type"
+#define PART_OPT_FIELD "field"
+#define PART_OPT_COLLATION "collation"
+#define PART_OPT_NULLABILITY "is_nullable"
#define PART_OPT_NULLABLE_ACTION "nullable_action"
-#define PART_OPT_SORT_ORDER "sort_order"
-#define PART_OPT_PATH "path"
+#define PART_OPT_SORT_ORDER "sort_order"
+#define PART_OPT_PATH "path"
const struct opt_def part_def_reg[] = {
OPT_DEF_ENUM(PART_OPT_TYPE, field_type, struct key_part_def, type,
@@ -177,7 +175,7 @@ key_def_set_part_path(struct key_def *def, uint32_t part_no, const char *path,
*/
int multikey_path_len =
json_path_multikey_offset(path, path_len, TUPLE_INDEX_BASE);
- if ((uint32_t) multikey_path_len == path_len)
+ if ((uint32_t)multikey_path_len == path_len)
return 0;
/*
@@ -192,7 +190,7 @@ key_def_set_part_path(struct key_def *def, uint32_t part_no, const char *path,
*/
def->multikey_path = part->path;
def->multikey_fieldno = part->fieldno;
- def->multikey_path_len = (uint32_t) multikey_path_len;
+ def->multikey_path_len = (uint32_t)multikey_path_len;
def->is_multikey = true;
} else if (def->multikey_fieldno != part->fieldno ||
json_path_cmp(path, multikey_path_len, def->multikey_path,
@@ -278,7 +276,8 @@ key_def_new(const struct key_part_def *parts, uint32_t part_count,
struct coll_id *coll_id = coll_by_id(part->coll_id);
if (coll_id == NULL) {
diag_set(ClientError, ER_WRONG_INDEX_OPTIONS,
- i + 1, "collation was not found by ID");
+ i + 1,
+ "collation was not found by ID");
goto error;
}
coll = coll_id->coll;
@@ -287,8 +286,7 @@ key_def_new(const struct key_part_def *parts, uint32_t part_count,
if (key_def_set_part(def, i, part->fieldno, part->type,
part->nullable_action, coll, part->coll_id,
part->sort_order, part->path, path_len,
- &path_pool, TUPLE_OFFSET_SLOT_NIL,
- 0) != 0)
+ &path_pool, TUPLE_OFFSET_SLOT_NIL, 0) != 0)
goto error;
}
if (for_func_index) {
@@ -297,7 +295,7 @@ key_def_new(const struct key_part_def *parts, uint32_t part_count,
"Functional index", "json paths");
goto error;
}
- if(!key_def_is_sequential(def) || parts->fieldno != 0) {
+ if (!key_def_is_sequential(def) || parts->fieldno != 0) {
diag_set(ClientError, ER_FUNC_INDEX_PARTS,
"key part numbers must be sequential and "
"first part number must be 1");
@@ -386,9 +384,8 @@ box_tuple_compare_with_key(box_tuple_t *tuple_a, const char *key_b,
box_key_def_t *key_def)
{
uint32_t part_count = mp_decode_array(&key_b);
- return tuple_compare_with_key(tuple_a, HINT_NONE, key_b,
- part_count, HINT_NONE, key_def);
-
+ return tuple_compare_with_key(tuple_a, HINT_NONE, key_b, part_count,
+ HINT_NONE, key_def);
}
int
@@ -402,16 +399,19 @@ key_part_cmp(const struct key_part *parts1, uint32_t part_count1,
for (; part1 != end; part1++, part2++) {
if (part1->fieldno != part2->fieldno)
return part1->fieldno < part2->fieldno ? -1 : 1;
- if ((int) part1->type != (int) part2->type)
- return (int) part1->type < (int) part2->type ? -1 : 1;
+ if ((int)part1->type != (int)part2->type)
+ return (int)part1->type < (int)part2->type ? -1 : 1;
if (part1->coll != part2->coll)
- return (uintptr_t) part1->coll <
- (uintptr_t) part2->coll ? -1 : 1;
+ return (uintptr_t)part1->coll < (uintptr_t)part2->coll ?
+ -1 :
+ 1;
if (part1->sort_order != part2->sort_order)
return part1->sort_order < part2->sort_order ? -1 : 1;
if (key_part_is_nullable(part1) != key_part_is_nullable(part2))
return key_part_is_nullable(part1) <
- key_part_is_nullable(part2) ? -1 : 1;
+ key_part_is_nullable(part2) ?
+ -1 :
+ 1;
int rc = json_path_cmp(part1->path, part1->path_len,
part2->path, part2->path_len,
TUPLE_INDEX_BASE);
@@ -429,7 +429,8 @@ key_def_update_optionality(struct key_def *def, uint32_t min_field_count)
struct key_part *part = &def->parts[i];
def->has_optional_parts |=
(min_field_count < part->fieldno + 1 ||
- part->path != NULL) && key_part_is_nullable(part);
+ part->path != NULL) &&
+ key_part_is_nullable(part);
/*
* One optional part is enough to switch to new
* comparators.
@@ -577,7 +578,7 @@ key_def_decode_parts_166(struct key_part_def *parts, uint32_t part_count,
return -1;
}
*part = key_part_def_default;
- part->fieldno = (uint32_t) mp_decode_uint(data);
+ part->fieldno = (uint32_t)mp_decode_uint(data);
if (mp_typeof(**data) != MP_STR) {
diag_set(ClientError, ER_WRONG_INDEX_PARTS,
"field type must be a string");
@@ -594,8 +595,8 @@ key_def_decode_parts_166(struct key_part_def *parts, uint32_t part_count,
return -1;
}
part->is_nullable = (part->fieldno < field_count ?
- fields[part->fieldno].is_nullable :
- key_part_def_default.is_nullable);
+ fields[part->fieldno].is_nullable :
+ key_part_def_default.is_nullable);
part->coll_id = COLL_NONE;
part->path = NULL;
}
@@ -608,8 +609,8 @@ key_def_decode_parts(struct key_part_def *parts, uint32_t part_count,
uint32_t field_count, struct region *region)
{
if (mp_typeof(**data) == MP_ARRAY) {
- return key_def_decode_parts_166(parts, part_count, data,
- fields, field_count);
+ return key_def_decode_parts_166(parts, part_count, data, fields,
+ field_count);
}
for (uint32_t i = 0; i < part_count; i++) {
struct key_part_def *part = &parts[i];
@@ -622,7 +623,7 @@ key_def_decode_parts(struct key_part_def *parts, uint32_t part_count,
int opts_count = mp_decode_map(data);
*part = key_part_def_default;
bool is_action_missing = true;
- uint32_t action_literal_len = strlen("nullable_action");
+ uint32_t action_literal_len = strlen("nullable_action");
for (int j = 0; j < opts_count; ++j) {
if (mp_typeof(**data) != MP_STR) {
diag_set(ClientError, ER_WRONG_INDEX_OPTIONS,
@@ -632,8 +633,8 @@ key_def_decode_parts(struct key_part_def *parts, uint32_t part_count,
}
uint32_t key_len;
const char *key = mp_decode_str(data, &key_len);
- if (opts_parse_key(part, part_def_reg, key, key_len, data,
- ER_WRONG_INDEX_OPTIONS,
+ if (opts_parse_key(part, part_def_reg, key, key_len,
+ data, ER_WRONG_INDEX_OPTIONS,
i + TUPLE_INDEX_BASE, region,
false) != 0)
return -1;
@@ -644,9 +645,9 @@ key_def_decode_parts(struct key_part_def *parts, uint32_t part_count,
is_action_missing = false;
}
if (is_action_missing) {
- part->nullable_action = part->is_nullable ?
- ON_CONFLICT_ACTION_NONE
- : ON_CONFLICT_ACTION_DEFAULT;
+ part->nullable_action =
+ part->is_nullable ? ON_CONFLICT_ACTION_NONE :
+ ON_CONFLICT_ACTION_DEFAULT;
}
if (part->type == field_type_MAX) {
diag_set(ClientError, ER_WRONG_INDEX_OPTIONS,
@@ -657,17 +658,15 @@ key_def_decode_parts(struct key_part_def *parts, uint32_t part_count,
if (part->coll_id != COLL_NONE &&
part->type != FIELD_TYPE_STRING &&
part->type != FIELD_TYPE_SCALAR) {
- diag_set(ClientError, ER_WRONG_INDEX_OPTIONS,
- i + 1,
+ diag_set(ClientError, ER_WRONG_INDEX_OPTIONS, i + 1,
"collation is reasonable only for "
"string and scalar parts");
return -1;
}
- if (!((part->is_nullable && part->nullable_action ==
- ON_CONFLICT_ACTION_NONE)
- || (!part->is_nullable
- && part->nullable_action !=
- ON_CONFLICT_ACTION_NONE))) {
+ if (!((part->is_nullable &&
+ part->nullable_action == ON_CONFLICT_ACTION_NONE) ||
+ (!part->is_nullable &&
+ part->nullable_action != ON_CONFLICT_ACTION_NONE))) {
diag_set(ClientError, ER_WRONG_INDEX_OPTIONS,
i + TUPLE_INDEX_BASE,
"index part: conflicting nullability and "
@@ -707,9 +706,8 @@ key_def_find(const struct key_def *key_def, const struct key_part *to_find)
const struct key_part *end = part + key_def->part_count;
for (; part != end; part++) {
if (part->fieldno == to_find->fieldno &&
- json_path_cmp(part->path, part->path_len,
- to_find->path, to_find->path_len,
- TUPLE_INDEX_BASE) == 0)
+ json_path_cmp(part->path, part->path_len, to_find->path,
+ to_find->path_len, TUPLE_INDEX_BASE) == 0)
return part;
}
return NULL;
@@ -840,17 +838,15 @@ key_def_merge(const struct key_def *first, const struct key_def *second)
struct key_def *
key_def_find_pk_in_cmp_def(const struct key_def *cmp_def,
- const struct key_def *pk_def,
- struct region *region)
+ const struct key_def *pk_def, struct region *region)
{
struct key_def *extracted_def = NULL;
size_t region_svp = region_used(region);
/* First, dump primary key parts as is. */
size_t size;
- struct key_part_def *parts =
- region_alloc_array(region, typeof(parts[0]), pk_def->part_count,
- &size);
+ struct key_part_def *parts = region_alloc_array(
+ region, typeof(parts[0]), pk_def->part_count, &size);
if (parts == NULL) {
diag_set(OutOfMemory, size, "region_alloc_array", "parts");
goto out;
@@ -862,8 +858,8 @@ key_def_find_pk_in_cmp_def(const struct key_def *cmp_def,
* parts in a secondary key.
*/
for (uint32_t i = 0; i < pk_def->part_count; i++) {
- const struct key_part *part = key_def_find(cmp_def,
- &pk_def->parts[i]);
+ const struct key_part *part =
+ key_def_find(cmp_def, &pk_def->parts[i]);
assert(part != NULL);
parts[i].fieldno = part - cmp_def->parts;
parts[i].path = NULL;
@@ -885,7 +881,7 @@ key_validate_parts(const struct key_def *key_def, const char *key,
const struct key_part *part = &key_def->parts[i];
if (key_part_validate(part->type, key, i,
key_part_is_nullable(part) &&
- allow_nullable))
+ allow_nullable))
return -1;
mp_next(&key);
}
diff --git a/src/box/key_def.h b/src/box/key_def.h
index f4d9e76..c3ea3db 100644
--- a/src/box/key_def.h
+++ b/src/box/key_def.h
@@ -130,38 +130,28 @@ key_part_is_nullable(const struct key_part *part)
}
/** @copydoc tuple_compare_with_key() */
-typedef int (*tuple_compare_with_key_t)(struct tuple *tuple,
- hint_t tuple_hint,
- const char *key,
- uint32_t part_count,
+typedef int (*tuple_compare_with_key_t)(struct tuple *tuple, hint_t tuple_hint,
+ const char *key, uint32_t part_count,
hint_t key_hint,
struct key_def *key_def);
/** @copydoc tuple_compare() */
-typedef int (*tuple_compare_t)(struct tuple *tuple_a,
- hint_t tuple_a_hint,
- struct tuple *tuple_b,
- hint_t tuple_b_hint,
+typedef int (*tuple_compare_t)(struct tuple *tuple_a, hint_t tuple_a_hint,
+ struct tuple *tuple_b, hint_t tuple_b_hint,
struct key_def *key_def);
/** @copydoc tuple_extract_key() */
typedef char *(*tuple_extract_key_t)(struct tuple *tuple,
- struct key_def *key_def,
- int multikey_idx,
+ struct key_def *key_def, int multikey_idx,
uint32_t *key_size);
/** @copydoc tuple_extract_key_raw() */
-typedef char *(*tuple_extract_key_raw_t)(const char *data,
- const char *data_end,
+typedef char *(*tuple_extract_key_raw_t)(const char *data, const char *data_end,
struct key_def *key_def,
- int multikey_idx,
- uint32_t *key_size);
+ int multikey_idx, uint32_t *key_size);
/** @copydoc tuple_hash() */
-typedef uint32_t (*tuple_hash_t)(struct tuple *tuple,
- struct key_def *key_def);
+typedef uint32_t (*tuple_hash_t)(struct tuple *tuple, struct key_def *key_def);
/** @copydoc key_hash() */
-typedef uint32_t (*key_hash_t)(const char *key,
- struct key_def *key_def);
+typedef uint32_t (*key_hash_t)(const char *key, struct key_def *key_def);
/** @copydoc tuple_hint() */
-typedef hint_t (*tuple_hint_t)(struct tuple *tuple,
- struct key_def *key_def);
+typedef hint_t (*tuple_hint_t)(struct tuple *tuple, struct key_def *key_def);
/** @copydoc key_hint() */
typedef hint_t (*key_hint_t)(const char *key, uint32_t part_count,
struct key_def *key_def);
@@ -459,8 +449,7 @@ key_def_merge(const struct key_def *first, const struct key_def *second);
*/
struct key_def *
key_def_find_pk_in_cmp_def(const struct key_def *cmp_def,
- const struct key_def *pk_def,
- struct region *region);
+ const struct key_def *pk_def, struct region *region);
/*
* Check that parts of the key match with the key definition.
@@ -525,10 +514,11 @@ key_def_has_collation(const struct key_def *key_def)
* @retval -1 mp_type is invalid.
*/
static inline int
-key_part_validate(enum field_type key_type, const char *key,
- uint32_t field_no, bool is_nullable)
+key_part_validate(enum field_type key_type, const char *key, uint32_t field_no,
+ bool is_nullable)
{
- if (unlikely(!field_mp_type_is_compatible(key_type, key, is_nullable))) {
+ if (unlikely(
+ !field_mp_type_is_compatible(key_type, key, is_nullable))) {
diag_set(ClientError, ER_KEY_PART_TYPE, field_no,
field_type_strs[key_type]);
return -1;
@@ -630,9 +620,8 @@ tuple_extract_key_raw(const char *data, const char *data_end,
* @retval >0 if key_a > key_b
*/
int
-key_compare(const char *key_a, hint_t key_a_hint,
- const char *key_b, hint_t key_b_hint,
- struct key_def *key_def);
+key_compare(const char *key_a, hint_t key_a_hint, const char *key_b,
+ hint_t key_b_hint, struct key_def *key_def);
/**
* Compare tuples using the key definition and comparison hints.
@@ -646,12 +635,11 @@ key_compare(const char *key_a, hint_t key_a_hint,
* @retval >0 if key_fields(tuple_a) > key_fields(tuple_b)
*/
static inline int
-tuple_compare(struct tuple *tuple_a, hint_t tuple_a_hint,
- struct tuple *tuple_b, hint_t tuple_b_hint,
- struct key_def *key_def)
+tuple_compare(struct tuple *tuple_a, hint_t tuple_a_hint, struct tuple *tuple_b,
+ hint_t tuple_b_hint, struct key_def *key_def)
{
- return key_def->tuple_compare(tuple_a, tuple_a_hint,
- tuple_b, tuple_b_hint, key_def);
+ return key_def->tuple_compare(tuple_a, tuple_a_hint, tuple_b,
+ tuple_b_hint, key_def);
}
/**
@@ -668,9 +656,9 @@ tuple_compare(struct tuple *tuple_a, hint_t tuple_a_hint,
* @retval >0 if key_fields(tuple) > parts(key)
*/
static inline int
-tuple_compare_with_key(struct tuple *tuple, hint_t tuple_hint,
- const char *key, uint32_t part_count,
- hint_t key_hint, struct key_def *key_def)
+tuple_compare_with_key(struct tuple *tuple, hint_t tuple_hint, const char *key,
+ uint32_t part_count, hint_t key_hint,
+ struct key_def *key_def)
{
return key_def->tuple_compare_with_key(tuple, tuple_hint, key,
part_count, key_hint, key_def);
@@ -730,7 +718,7 @@ key_hash(const char *key, struct key_def *key_def)
return key_def->key_hash(key, key_def);
}
- /*
+/*
* Get comparison hint for a tuple.
* @param tuple - tuple to compute the hint for
* @param key_def - key_def used for tuple comparison
diff --git a/src/box/key_list.c b/src/box/key_list.c
index 6143b84..a604a1d 100644
--- a/src/box/key_list.c
+++ b/src/box/key_list.c
@@ -97,7 +97,7 @@ key_list_iterator_create(struct key_list_iterator *it, struct tuple *tuple,
}
if (func->def->opts.is_multikey) {
if (mp_typeof(*key_data) != MP_ARRAY) {
- struct space * space = space_by_id(index_def->space_id);
+ struct space *space = space_by_id(index_def->space_id);
/*
* Multikey function must return an array
* of keys.
@@ -159,12 +159,12 @@ key_list_iterator_next(struct key_list_iterator *it, const char **value)
diag_set(ClientError, ER_FUNC_INDEX_FORMAT, it->index_def->name,
space ? space_name(space) : "",
tt_sprintf(tnt_errcode_desc(ER_EXACT_MATCH),
- key_def->part_count, part_count));
+ key_def->part_count, part_count));
return -1;
}
const char *key_end;
- if (key_validate_parts(key_def, rptr, part_count, true,
- &key_end) != 0) {
+ if (key_validate_parts(key_def, rptr, part_count, true, &key_end) !=
+ 0) {
struct space *space = space_by_id(it->index_def->space_id);
/*
* The key doesn't follow functional index key
diff --git a/src/box/key_list.h b/src/box/key_list.h
index ccc91e7..6b56eb3 100644
--- a/src/box/key_list.h
+++ b/src/box/key_list.h
@@ -49,8 +49,8 @@ struct tuple;
* key, since the key is only used to lookup the old tuple in the
* b+* tree, so we pass in a dummy allocator.
*/
-typedef const char *(*key_list_allocator_t)(struct tuple *tuple, const char *key,
- uint32_t key_sz);
+typedef const char *(*key_list_allocator_t)(struct tuple *tuple,
+ const char *key, uint32_t key_sz);
/**
* An iterator over key_data returned by a stored function function.
diff --git a/src/box/lua/call.c b/src/box/lua/call.c
index 0315e72..a78e15a 100644
--- a/src/box/lua/call.c
+++ b/src/box/lua/call.c
@@ -77,57 +77,55 @@ box_lua_find(lua_State *L, const char *name, const char *name_end)
int objstack = 0, top = lua_gettop(L);
const char *start = name, *end;
- while ((end = (const char *) memchr(start, '.', name_end - start))) {
+ while ((end = (const char *)memchr(start, '.', name_end - start))) {
lua_checkstack(L, 3);
lua_pushlstring(L, start, end - start);
lua_gettable(L, index);
- if (! lua_istable(L, -1)) {
- diag_set(ClientError, ER_NO_SUCH_PROC,
- name_end - name, name);
+ if (!lua_istable(L, -1)) {
+ diag_set(ClientError, ER_NO_SUCH_PROC, name_end - name,
+ name);
return -1;
}
- start = end + 1; /* next piece of a.b.c */
+ start = end + 1; /* next piece of a.b.c */
index = lua_gettop(L); /* top of the stack */
}
/* box.something:method */
- if ((end = (const char *) memchr(start, ':', name_end - start))) {
+ if ((end = (const char *)memchr(start, ':', name_end - start))) {
lua_checkstack(L, 3);
lua_pushlstring(L, start, end - start);
lua_gettable(L, index);
- if (! (lua_istable(L, -1) ||
- lua_islightuserdata(L, -1) || lua_isuserdata(L, -1) )) {
- diag_set(ClientError, ER_NO_SUCH_PROC,
- name_end - name, name);
- return -1;
+ if (!(lua_istable(L, -1) || lua_islightuserdata(L, -1) ||
+ lua_isuserdata(L, -1))) {
+ diag_set(ClientError, ER_NO_SUCH_PROC, name_end - name,
+ name);
+ return -1;
}
- start = end + 1; /* next piece of a.b.c */
+ start = end + 1; /* next piece of a.b.c */
index = lua_gettop(L); /* top of the stack */
objstack = index - top;
}
-
lua_pushlstring(L, start, name_end - start);
lua_gettable(L, index);
if (!lua_isfunction(L, -1) && !lua_istable(L, -1)) {
/* lua_call or lua_gettable would raise a type error
* for us, but our own message is more verbose. */
- diag_set(ClientError, ER_NO_SUCH_PROC,
- name_end - name, name);
+ diag_set(ClientError, ER_NO_SUCH_PROC, name_end - name, name);
return -1;
}
/* setting stack that it would contain only
* the function pointer. */
if (index != LUA_GLOBALSINDEX) {
- if (objstack == 0) { /* no object, only a function */
+ if (objstack == 0) { /* no object, only a function */
lua_replace(L, top + 1);
lua_pop(L, lua_gettop(L) - top - 1);
} else if (objstack == 1) { /* just two values, swap them */
lua_insert(L, -2);
lua_pop(L, lua_gettop(L) - top - 2);
- } else { /* long path */
+ } else { /* long path */
lua_insert(L, top + 1);
lua_insert(L, top + 2);
lua_pop(L, objstack - 1);
@@ -300,7 +298,7 @@ static const struct port_vtab port_lua_vtab;
void
port_lua_create(struct port *port, struct lua_State *L)
{
- struct port_lua *port_lua = (struct port_lua *) port;
+ struct port_lua *port_lua = (struct port_lua *)port;
memset(port_lua, 0, sizeof(*port_lua));
port_lua->vtab = &port_lua_vtab;
port_lua->L = L;
@@ -328,7 +326,7 @@ static int
execute_lua_call(lua_State *L)
{
struct execute_lua_ctx *ctx =
- (struct execute_lua_ctx *) lua_topointer(L, 1);
+ (struct execute_lua_ctx *)lua_topointer(L, 1);
lua_settop(L, 0); /* clear the stack to simplify the logic below */
const char *name = ctx->name;
@@ -356,7 +354,7 @@ static int
execute_lua_call_by_ref(lua_State *L)
{
struct execute_lua_ctx *ctx =
- (struct execute_lua_ctx *) lua_topointer(L, 1);
+ (struct execute_lua_ctx *)lua_topointer(L, 1);
lua_settop(L, 0); /* clear the stack to simplify the logic below */
lua_rawgeti(L, LUA_REGISTRYINDEX, ctx->lua_ref);
@@ -374,7 +372,7 @@ static int
execute_lua_eval(lua_State *L)
{
struct execute_lua_ctx *ctx =
- (struct execute_lua_ctx *) lua_topointer(L, 1);
+ (struct execute_lua_ctx *)lua_topointer(L, 1);
lua_settop(L, 0); /* clear the stack to simplify the logic below */
/* Compile expression */
@@ -404,7 +402,7 @@ static int
encode_lua_call(lua_State *L)
{
struct encode_lua_ctx *ctx =
- (struct encode_lua_ctx *) lua_topointer(L, 1);
+ (struct encode_lua_ctx *)lua_topointer(L, 1);
/*
* Add all elements from Lua stack to the buffer.
*
@@ -425,7 +423,7 @@ static int
encode_lua_call_16(lua_State *L)
{
struct encode_lua_ctx *ctx =
- (struct encode_lua_ctx *) lua_topointer(L, 1);
+ (struct encode_lua_ctx *)lua_topointer(L, 1);
/*
* Add all elements from Lua stack to the buffer.
*
@@ -441,7 +439,7 @@ static inline int
port_lua_do_dump(struct port *base, struct mpstream *stream,
lua_CFunction handler)
{
- struct port_lua *port = (struct port_lua *) base;
+ struct port_lua *port = (struct port_lua *)base;
assert(port->vtab == &port_lua_vtab);
/*
* Use the same global state, assuming the encoder doesn't
@@ -463,10 +461,10 @@ port_lua_do_dump(struct port *base, struct mpstream *stream,
static int
port_lua_dump(struct port *base, struct obuf *out)
{
- struct port_lua *port = (struct port_lua *) base;
+ struct port_lua *port = (struct port_lua *)base;
struct mpstream stream;
- mpstream_init(&stream, out, obuf_reserve_cb, obuf_alloc_cb,
- luamp_error, port->L);
+ mpstream_init(&stream, out, obuf_reserve_cb, obuf_alloc_cb, luamp_error,
+ port->L);
return port_lua_do_dump(base, &stream, encode_lua_call);
}
@@ -475,17 +473,17 @@ port_lua_dump_16(struct port *base, struct obuf *out)
{
struct port_lua *port = (struct port_lua *)base;
struct mpstream stream;
- mpstream_init(&stream, out, obuf_reserve_cb, obuf_alloc_cb,
- luamp_error, port->L);
+ mpstream_init(&stream, out, obuf_reserve_cb, obuf_alloc_cb, luamp_error,
+ port->L);
return port_lua_do_dump(base, &stream, encode_lua_call_16);
}
static void
port_lua_dump_lua(struct port *base, struct lua_State *L, bool is_flat)
{
- (void) is_flat;
+ (void)is_flat;
assert(is_flat == true);
- struct port_lua *port = (struct port_lua *) base;
+ struct port_lua *port = (struct port_lua *)base;
uint32_t size = lua_gettop(port->L);
lua_xmove(port->L, L, size);
port->size = size;
@@ -494,7 +492,7 @@ port_lua_dump_lua(struct port *base, struct lua_State *L, bool is_flat)
static const char *
port_lua_get_msgpack(struct port *base, uint32_t *size)
{
- struct port_lua *port = (struct port_lua *) base;
+ struct port_lua *port = (struct port_lua *)base;
struct region *region = &fiber()->gc;
uint32_t region_svp = region_used(region);
struct mpstream stream;
@@ -553,7 +551,7 @@ box_process_lua(enum handlers handler, struct execute_lua_ctx *ctx,
return -1;
int coro_ref = luaL_ref(tarantool_L, LUA_REGISTRYINDEX);
port_lua_create(ret, L);
- ((struct port_lua *) ret)->ref = coro_ref;
+ ((struct port_lua *)ret)->ref = coro_ref;
/*
* A code that need a temporary fiber-local Lua state may
@@ -593,8 +591,8 @@ box_process_lua(enum handlers handler, struct execute_lua_ctx *ctx,
}
int
-box_lua_call(const char *name, uint32_t name_len,
- struct port *args, struct port *ret)
+box_lua_call(const char *name, uint32_t name_len, struct port *args,
+ struct port *ret)
{
struct execute_lua_ctx ctx;
ctx.name = name;
@@ -604,8 +602,8 @@ box_lua_call(const char *name, uint32_t name_len,
}
int
-box_lua_eval(const char *expr, uint32_t expr_len,
- struct port *args, struct port *ret)
+box_lua_eval(const char *expr, uint32_t expr_len, struct port *args,
+ struct port *ret)
{
struct execute_lua_ctx ctx;
ctx.name = expr;
@@ -628,9 +626,9 @@ static struct func_vtab func_lua_vtab;
static struct func_vtab func_persistent_lua_vtab;
static const char *default_sandbox_exports[] = {
- "assert", "error", "ipairs", "math", "next", "pairs", "pcall", "print",
- "select", "string", "table", "tonumber", "tostring", "type", "unpack",
- "xpcall", "utf8",
+ "assert", "error", "ipairs", "math", "next", "pairs",
+ "pcall", "print", "select", "string", "table", "tonumber",
+ "tostring", "type", "unpack", "xpcall", "utf8",
};
/**
@@ -696,7 +694,8 @@ func_persistent_lua_load(struct func_lua *func)
diag_set(OutOfMemory, load_str_sz, "region", "load_str");
return -1;
}
- snprintf(load_str, load_str_sz, "%s%s", load_pref, func->base.def->body);
+ snprintf(load_str, load_str_sz, "%s%s", load_pref,
+ func->base.def->body);
/*
* Perform loading of the persistent Lua function
@@ -759,7 +758,7 @@ func_lua_new(struct func_def *def)
{
assert(def->language == FUNC_LANGUAGE_LUA);
struct func_lua *func =
- (struct func_lua *) malloc(sizeof(struct func_lua));
+ (struct func_lua *)malloc(sizeof(struct func_lua));
if (func == NULL) {
diag_set(OutOfMemory, sizeof(*func), "malloc", "func");
return NULL;
@@ -812,7 +811,7 @@ func_persistent_lua_destroy(struct func *base)
assert(base != NULL && base->def->language == FUNC_LANGUAGE_LUA &&
base->def->body != NULL);
assert(base->vtab == &func_persistent_lua_vtab);
- struct func_lua *func = (struct func_lua *) base;
+ struct func_lua *func = (struct func_lua *)base;
func_persistent_lua_unload(func);
free(func);
}
@@ -828,7 +827,6 @@ func_persistent_lua_call(struct func *base, struct port *args, struct port *ret)
ctx.lua_ref = func->lua_ref;
ctx.args = args;
return box_process_lua(HANDLER_CALL_BY_REF, &ctx, ret);
-
}
static struct func_vtab func_persistent_lua_vtab = {
@@ -872,7 +870,7 @@ lbox_func_call(struct lua_State *L)
lua_xmove(L, args_L, lua_gettop(L) - 1);
struct port args;
port_lua_create(&args, args_L);
- ((struct port_lua *) &args)->ref = coro_ref;
+ ((struct port_lua *)&args)->ref = coro_ref;
struct port ret;
if (func_call(func, &args, &ret) != 0) {
@@ -1011,7 +1009,7 @@ lbox_func_delete(struct lua_State *L, struct func *func)
static int
lbox_func_new_or_delete(struct trigger *trigger, void *event)
{
- struct lua_State *L = (struct lua_State *) trigger->data;
+ struct lua_State *L = (struct lua_State *)trigger->data;
struct func *func = (struct func *)event;
if (!func->def->exports.lua)
return 0;
@@ -1022,15 +1020,15 @@ lbox_func_new_or_delete(struct trigger *trigger, void *event)
return 0;
}
-static struct trigger on_alter_func_in_lua = {
- RLIST_LINK_INITIALIZER, lbox_func_new_or_delete, NULL, NULL
-};
+static struct trigger on_alter_func_in_lua = { RLIST_LINK_INITIALIZER,
+ lbox_func_new_or_delete, NULL,
+ NULL };
static const struct luaL_Reg boxlib_internal[] = {
- {"call_loadproc", lbox_call_loadproc},
- {"module_reload", lbox_module_reload},
- {"func_call", lbox_func_call},
- {NULL, NULL}
+ { "call_loadproc", lbox_call_loadproc },
+ { "module_reload", lbox_module_reload },
+ { "func_call", lbox_func_call },
+ { NULL, NULL }
};
void
diff --git a/src/box/lua/call.h b/src/box/lua/call.h
index 83aa439..8c0c5a1 100644
--- a/src/box/lua/call.h
+++ b/src/box/lua/call.h
@@ -51,12 +51,12 @@ struct func_def;
* (implementation of 'CALL' command code).
*/
int
-box_lua_call(const char *name, uint32_t name_len,
- struct port *args, struct port *ret);
+box_lua_call(const char *name, uint32_t name_len, struct port *args,
+ struct port *ret);
int
-box_lua_eval(const char *expr, uint32_t expr_len,
- struct port *args, struct port *ret);
+box_lua_eval(const char *expr, uint32_t expr_len, struct port *args,
+ struct port *ret);
/** Construct a Lua function object. */
struct func *
diff --git a/src/box/lua/cfg.cc b/src/box/lua/cfg.cc
index bbb92f0..c422870 100644
--- a/src/box/lua/cfg.cc
+++ b/src/box/lua/cfg.cc
@@ -40,7 +40,7 @@
#include "libeio/eio.h"
extern "C" {
- #include <lua.h>
+#include <lua.h>
} // extern "C"
static int
@@ -263,7 +263,7 @@ lbox_set_prepared_stmt_cache_size(struct lua_State *L)
static int
lbox_cfg_set_worker_pool_threads(struct lua_State *L)
{
- (void) L;
+ (void)L;
eio_set_min_parallel(cfg_geti("worker_pool_threads"));
eio_set_max_parallel(cfg_geti("worker_pool_threads"));
return 0;
@@ -378,7 +378,7 @@ lbox_cfg_set_replication_anon(struct lua_State *L)
static int
lbox_cfg_set_replication_skip_conflict(struct lua_State *L)
{
- (void) L;
+ (void)L;
box_set_replication_skip_conflict();
return 0;
}
@@ -387,40 +387,58 @@ void
box_lua_cfg_init(struct lua_State *L)
{
static const struct luaL_Reg cfglib_internal[] = {
- {"cfg_check", lbox_cfg_check},
- {"cfg_load", lbox_cfg_load},
- {"cfg_set_listen", lbox_cfg_set_listen},
- {"cfg_set_replication", lbox_cfg_set_replication},
- {"cfg_set_worker_pool_threads", lbox_cfg_set_worker_pool_threads},
- {"cfg_set_readahead", lbox_cfg_set_readahead},
- {"cfg_set_io_collect_interval", lbox_cfg_set_io_collect_interval},
- {"cfg_set_too_long_threshold", lbox_cfg_set_too_long_threshold},
- {"cfg_set_snap_io_rate_limit", lbox_cfg_set_snap_io_rate_limit},
- {"cfg_set_checkpoint_count", lbox_cfg_set_checkpoint_count},
- {"cfg_set_checkpoint_interval", lbox_cfg_set_checkpoint_interval},
- {"cfg_set_checkpoint_wal_threshold", lbox_cfg_set_checkpoint_wal_threshold},
- {"cfg_set_read_only", lbox_cfg_set_read_only},
- {"cfg_set_memtx_memory", lbox_cfg_set_memtx_memory},
- {"cfg_set_memtx_max_tuple_size", lbox_cfg_set_memtx_max_tuple_size},
- {"cfg_set_vinyl_memory", lbox_cfg_set_vinyl_memory},
- {"cfg_set_vinyl_max_tuple_size", lbox_cfg_set_vinyl_max_tuple_size},
- {"cfg_set_vinyl_cache", lbox_cfg_set_vinyl_cache},
- {"cfg_set_vinyl_timeout", lbox_cfg_set_vinyl_timeout},
- {"cfg_set_election_is_enabled", lbox_cfg_set_election_is_enabled},
- {"cfg_set_election_is_candidate", lbox_cfg_set_election_is_candidate},
- {"cfg_set_election_timeout", lbox_cfg_set_election_timeout},
- {"cfg_set_replication_timeout", lbox_cfg_set_replication_timeout},
- {"cfg_set_replication_connect_quorum", lbox_cfg_set_replication_connect_quorum},
- {"cfg_set_replication_connect_timeout", lbox_cfg_set_replication_connect_timeout},
- {"cfg_set_replication_sync_lag", lbox_cfg_set_replication_sync_lag},
- {"cfg_set_replication_synchro_quorum", lbox_cfg_set_replication_synchro_quorum},
- {"cfg_set_replication_synchro_timeout", lbox_cfg_set_replication_synchro_timeout},
- {"cfg_set_replication_sync_timeout", lbox_cfg_set_replication_sync_timeout},
- {"cfg_set_replication_skip_conflict", lbox_cfg_set_replication_skip_conflict},
- {"cfg_set_replication_anon", lbox_cfg_set_replication_anon},
- {"cfg_set_net_msg_max", lbox_cfg_set_net_msg_max},
- {"cfg_set_sql_cache_size", lbox_set_prepared_stmt_cache_size},
- {NULL, NULL}
+ { "cfg_check", lbox_cfg_check },
+ { "cfg_load", lbox_cfg_load },
+ { "cfg_set_listen", lbox_cfg_set_listen },
+ { "cfg_set_replication", lbox_cfg_set_replication },
+ { "cfg_set_worker_pool_threads",
+ lbox_cfg_set_worker_pool_threads },
+ { "cfg_set_readahead", lbox_cfg_set_readahead },
+ { "cfg_set_io_collect_interval",
+ lbox_cfg_set_io_collect_interval },
+ { "cfg_set_too_long_threshold",
+ lbox_cfg_set_too_long_threshold },
+ { "cfg_set_snap_io_rate_limit",
+ lbox_cfg_set_snap_io_rate_limit },
+ { "cfg_set_checkpoint_count", lbox_cfg_set_checkpoint_count },
+ { "cfg_set_checkpoint_interval",
+ lbox_cfg_set_checkpoint_interval },
+ { "cfg_set_checkpoint_wal_threshold",
+ lbox_cfg_set_checkpoint_wal_threshold },
+ { "cfg_set_read_only", lbox_cfg_set_read_only },
+ { "cfg_set_memtx_memory", lbox_cfg_set_memtx_memory },
+ { "cfg_set_memtx_max_tuple_size",
+ lbox_cfg_set_memtx_max_tuple_size },
+ { "cfg_set_vinyl_memory", lbox_cfg_set_vinyl_memory },
+ { "cfg_set_vinyl_max_tuple_size",
+ lbox_cfg_set_vinyl_max_tuple_size },
+ { "cfg_set_vinyl_cache", lbox_cfg_set_vinyl_cache },
+ { "cfg_set_vinyl_timeout", lbox_cfg_set_vinyl_timeout },
+ { "cfg_set_election_is_enabled",
+ lbox_cfg_set_election_is_enabled },
+ { "cfg_set_election_is_candidate",
+ lbox_cfg_set_election_is_candidate },
+ { "cfg_set_election_timeout", lbox_cfg_set_election_timeout },
+ { "cfg_set_replication_timeout",
+ lbox_cfg_set_replication_timeout },
+ { "cfg_set_replication_connect_quorum",
+ lbox_cfg_set_replication_connect_quorum },
+ { "cfg_set_replication_connect_timeout",
+ lbox_cfg_set_replication_connect_timeout },
+ { "cfg_set_replication_sync_lag",
+ lbox_cfg_set_replication_sync_lag },
+ { "cfg_set_replication_synchro_quorum",
+ lbox_cfg_set_replication_synchro_quorum },
+ { "cfg_set_replication_synchro_timeout",
+ lbox_cfg_set_replication_synchro_timeout },
+ { "cfg_set_replication_sync_timeout",
+ lbox_cfg_set_replication_sync_timeout },
+ { "cfg_set_replication_skip_conflict",
+ lbox_cfg_set_replication_skip_conflict },
+ { "cfg_set_replication_anon", lbox_cfg_set_replication_anon },
+ { "cfg_set_net_msg_max", lbox_cfg_set_net_msg_max },
+ { "cfg_set_sql_cache_size", lbox_set_prepared_stmt_cache_size },
+ { NULL, NULL }
};
luaL_register(L, "box.internal", cfglib_internal);
diff --git a/src/box/lua/console.c b/src/box/lua/console.c
index ea5385c..997244a 100644
--- a/src/box/lua/console.c
+++ b/src/box/lua/console.c
@@ -170,7 +170,6 @@ console_completion_handler(const char *text, int start, int end)
if (lua_pcall(readline_L, 3, 1, 0) != 0 ||
!lua_istable(readline_L, -1) ||
(n = lua_objlen(readline_L, -1)) == 0) {
-
lua_pop(readline_L, 1);
return NULL;
}
@@ -281,8 +280,8 @@ lbox_console_readline(struct lua_State *L)
rl_callback_handler_install(prompt, console_push_line);
top = lua_gettop(L);
while (top == lua_gettop(L)) {
- while (coio_wait(STDIN_FILENO, COIO_READ,
- TIMEOUT_INFINITY) == 0) {
+ while (coio_wait(STDIN_FILENO, COIO_READ, TIMEOUT_INFINITY) ==
+ 0) {
/*
* Make sure the user of interactive
* console has not hanged us, otherwise
@@ -336,8 +335,8 @@ lbox_console_completion_handler(struct lua_State *L)
lua_pushcfunction(L, console_completion_helper);
lua_pushlightuserdata(L, &res);
- res = lua_rl_complete(L, lua_tostring(L, 1),
- lua_tointeger(L, 2), lua_tointeger(L, 3));
+ res = lua_rl_complete(L, lua_tostring(L, 1), lua_tointeger(L, 2),
+ lua_tointeger(L, 3));
if (res == NULL) {
return 0;
@@ -387,7 +386,8 @@ lbox_console_add_history(struct lua_State *L)
const char *s = lua_tostring(L, 1);
if (*s) {
- HIST_ENTRY *hist_ent = history_get(history_length - 1 + history_base);
+ HIST_ENTRY *hist_ent =
+ history_get(history_length - 1 + history_base);
const char *prev_s = hist_ent ? hist_ent->line : "";
if (strcmp(prev_s, s) != 0)
add_history(s);
@@ -480,7 +480,7 @@ console_dump_plain(struct lua_State *L, uint32_t *size)
assert(lua_isstring(L, -1));
size_t len;
const char *result = lua_tolstring(L, -1, &len);
- *size = (uint32_t) len;
+ *size = (uint32_t)len;
return result;
}
@@ -526,14 +526,14 @@ port_msgpack_dump_plain_via_lua(struct lua_State *L)
port_msgpack_set_plain((struct port *)port, data, *size);
}
return 0;
- }
+}
/** Plain text converter for raw MessagePack. */
const char *
port_msgpack_dump_plain(struct port *base, uint32_t *size)
{
struct lua_State *L = tarantool_L;
- void *ctx[2] = {(void *)base, (void *)size};
+ void *ctx[2] = { (void *)base, (void *)size };
/*
* lua_cpcall() protects from errors thrown from Lua which
* may break a caller, not knowing about Lua and not
@@ -587,12 +587,12 @@ lua_serpent_init(struct lua_State *L)
lua_getfield(L, LUA_REGISTRYINDEX, "_LOADED");
modfile = lua_pushfstring(L, "@builtin/%s.lua", modname);
if (luaL_loadbuffer(L, serpent_lua, strlen(serpent_lua), modfile)) {
- panic("Error loading Lua module %s...: %s",
- modname, lua_tostring(L, -1));
+ panic("Error loading Lua module %s...: %s", modname,
+ lua_tostring(L, -1));
}
lua_call(L, 0, 1);
- lua_setfield(L, -3, modname); /* _LOADED[modname] = new table */
+ lua_setfield(L, -3, modname); /* _LOADED[modname] = new table */
lua_pop(L, 2);
}
@@ -600,13 +600,13 @@ void
tarantool_lua_console_init(struct lua_State *L)
{
static const struct luaL_Reg consolelib[] = {
- {"load_history", lbox_console_load_history},
- {"save_history", lbox_console_save_history},
- {"add_history", lbox_console_add_history},
- {"completion_handler", lbox_console_completion_handler},
- {"format_yaml", lbox_console_format_yaml},
- {"format_lua", lbox_console_format_lua},
- {NULL, NULL}
+ { "load_history", lbox_console_load_history },
+ { "save_history", lbox_console_save_history },
+ { "add_history", lbox_console_add_history },
+ { "completion_handler", lbox_console_completion_handler },
+ { "format_yaml", lbox_console_format_yaml },
+ { "format_lua", lbox_console_format_lua },
+ { NULL, NULL }
};
luaL_register_module(L, "console", consolelib);
@@ -641,11 +641,11 @@ tarantool_lua_console_init(struct lua_State *L)
};
serializer_lua = luaL_newserializer(L, NULL, lualib);
- serializer_lua->has_compact = 1;
- serializer_lua->encode_invalid_numbers = 1;
- serializer_lua->encode_load_metatables = 1;
- serializer_lua->encode_use_tostring = 1;
- serializer_lua->encode_invalid_as_nil = 1;
+ serializer_lua->has_compact = 1;
+ serializer_lua->encode_invalid_numbers = 1;
+ serializer_lua->encode_load_metatables = 1;
+ serializer_lua->encode_use_tostring = 1;
+ serializer_lua->encode_invalid_as_nil = 1;
/*
* Keep a reference to this module so it
@@ -657,9 +657,9 @@ tarantool_lua_console_init(struct lua_State *L)
lua_serializer_init(L);
struct session_vtab console_session_vtab = {
- .push = console_session_push,
- .fd = console_session_fd,
- .sync = generic_session_sync,
+ .push = console_session_push,
+ .fd = console_session_fd,
+ .sync = generic_session_sync,
};
session_vtab_registry[SESSION_TYPE_CONSOLE] = console_session_vtab;
session_vtab_registry[SESSION_TYPE_REPL] = console_session_vtab;
@@ -696,19 +696,21 @@ enum {
};
/* goto intentionally omited */
-static const char *
-const lua_rl_keywords[] = {
- "and", "break", "do", "else", "elseif", "end", "false",
- "for", "function", "if", "in", "local", "nil", "not", "or",
- "repeat", "return", "then", "true", "until", "while", NULL
+static const char *const lua_rl_keywords[] = {
+ "and", "break", "do", "else", "elseif", "end",
+ "false", "for", "function", "if", "in", "local",
+ "nil", "not", "or", "repeat", "return", "then",
+ "true", "until", "while", NULL
};
static int
valid_identifier(const char *s)
{
- if (!(isalpha(*s) || *s == '_')) return 0;
+ if (!(isalpha(*s) || *s == '_'))
+ return 0;
for (s++; *s; s++)
- if (!(isalpha(*s) || isdigit(*s) || *s == '_')) return 0;
+ if (!(isalpha(*s) || isdigit(*s) || *s == '_'))
+ return 0;
return 1;
}
@@ -746,10 +748,10 @@ lua_rl_dmadd(dmlist *ml, const char *p, size_t pn, const char *s, int suf)
{
char *t = NULL;
- if (ml->idx+1 >= ml->allocated) {
+ if (ml->idx + 1 >= ml->allocated) {
char **new_list;
- new_list = realloc(
- ml->list, sizeof(char *)*(ml->allocated += 32));
+ new_list = realloc(ml->list,
+ sizeof(char *) * (ml->allocated += 32));
if (!new_list)
return -1;
ml->list = new_list;
@@ -757,20 +759,23 @@ lua_rl_dmadd(dmlist *ml, const char *p, size_t pn, const char *s, int suf)
if (s) {
size_t n = strlen(s);
- if (!(t = (char *)malloc(sizeof(char)*(pn + n + 2))))
+ if (!(t = (char *)malloc(sizeof(char) * (pn + n + 2))))
return 1;
memcpy(t, p, pn);
memcpy(t + pn, s, n);
n += pn;
t[n] = suf;
- if (suf) t[++n] = '\0';
+ if (suf)
+ t[++n] = '\0';
if (ml->idx == 0) {
ml->matchlen = n;
} else {
size_t i;
for (i = 0; i < ml->matchlen && i < n &&
- ml->list[1][i] == t[i]; i++) ;
+ ml->list[1][i] == t[i];
+ i++)
+ ;
/* Set matchlen to common prefix. */
ml->matchlen = i;
}
@@ -797,7 +802,7 @@ lua_rl_getmetaindex(lua_State *L)
}
lua_replace(L, -2);
return 1;
-} /* 1: obj -- val, 0: obj -- */
+} /* 1: obj -- val, 0: obj -- */
/* Get field from object on top of stack. Avoid calling metamethods. */
static int
@@ -820,7 +825,7 @@ lua_rl_getfield(lua_State *L, const char *s, size_t n)
}
} while (lua_rl_getmetaindex(L));
return 0;
-} /* 1: obj -- val, 0: obj -- */
+} /* 1: obj -- val, 0: obj -- */
static char **
lua_rl_complete(lua_State *L, const char *text, int start, int end)
@@ -838,12 +843,12 @@ lua_rl_complete(lua_State *L, const char *text, int start, int end)
savetop = lua_gettop(L);
lua_pushglobaltable(L);
- for (n = (size_t)(end-start), i = dot = 0; i < n; i++) {
+ for (n = (size_t)(end - start), i = dot = 0; i < n; i++) {
if (text[i] == '.' || text[i] == ':') {
is_method_ref = (text[i] == ':');
- if (!lua_rl_getfield(L, text+dot, i-dot))
+ if (!lua_rl_getfield(L, text + dot, i - dot))
goto error; /* Invalid prefix. */
- dot = i+1;
+ dot = i + 1;
/* Points to first char after dot/colon. */
}
}
@@ -851,10 +856,8 @@ lua_rl_complete(lua_State *L, const char *text, int start, int end)
/* Add all matches against keywords if there is no dot/colon. */
if (dot == 0) {
for (i = 0; (s = lua_rl_keywords[i]) != NULL; i++) {
- if (n >= KEYWORD_MATCH_MIN &&
- !strncmp(s, text, n) &&
+ if (n >= KEYWORD_MATCH_MIN && !strncmp(s, text, n) &&
lua_rl_dmadd(&ml, NULL, 0, s, ' ')) {
-
goto error;
}
}
@@ -871,7 +874,6 @@ lua_rl_complete(lua_State *L, const char *text, int start, int end)
continue;
for (lua_pushnil(L); lua_next(L, -2); lua_pop(L, 1)) {
-
/* Beware huge tables */
if (++items_checked > ITEMS_CHECKED_MAX)
break;
@@ -884,9 +886,10 @@ lua_rl_complete(lua_State *L, const char *text, int start, int end)
* Only match names starting with '_'
* if explicitly requested.
*/
- if (strncmp(s, text+dot, n-dot) ||
+ if (strncmp(s, text + dot, n - dot) ||
!valid_identifier(s) ||
- (*s == '_' && text[dot] != '_')) continue;
+ (*s == '_' && text[dot] != '_'))
+ continue;
int suf = 0; /* Omit suffix by default. */
int type = lua_type(L, -1);
@@ -929,7 +932,7 @@ lua_rl_complete(lua_State *L, const char *text, int start, int end)
lua_pop(L, 1);
if (ml.idx == 0) {
-error:
+ error:
lua_rl_dmfree(&ml);
lua_settop(L, savetop);
return NULL;
@@ -937,13 +940,14 @@ error:
/* list[0] holds the common prefix of all matches (may
* be ""). If there is only one match, list[0] and
* list[1] will be the same. */
- ml.list[0] = malloc(sizeof(char)*(ml.matchlen+1));
+ ml.list[0] = malloc(sizeof(char) * (ml.matchlen + 1));
if (!ml.list[0])
goto error;
memcpy(ml.list[0], ml.list[1], ml.matchlen);
ml.list[0][ml.matchlen] = '\0';
/* Add the NULL list terminator. */
- if (lua_rl_dmadd(&ml, NULL, 0, NULL, 0)) goto error;
+ if (lua_rl_dmadd(&ml, NULL, 0, NULL, 0))
+ goto error;
}
lua_settop(L, savetop);
diff --git a/src/box/lua/ctl.c b/src/box/lua/ctl.c
index 2017ddc..72bfa81 100644
--- a/src/box/lua/ctl.c
+++ b/src/box/lua/ctl.c
@@ -81,18 +81,18 @@ lbox_ctl_on_schema_init(struct lua_State *L)
static int
lbox_ctl_clear_synchro_queue(struct lua_State *L)
{
- (void) L;
+ (void)L;
box_clear_synchro_queue();
return 0;
}
static const struct luaL_Reg lbox_ctl_lib[] = {
- {"wait_ro", lbox_ctl_wait_ro},
- {"wait_rw", lbox_ctl_wait_rw},
- {"on_shutdown", lbox_ctl_on_shutdown},
- {"on_schema_init", lbox_ctl_on_schema_init},
- {"clear_synchro_queue", lbox_ctl_clear_synchro_queue},
- {NULL, NULL}
+ { "wait_ro", lbox_ctl_wait_ro },
+ { "wait_rw", lbox_ctl_wait_rw },
+ { "on_shutdown", lbox_ctl_on_shutdown },
+ { "on_schema_init", lbox_ctl_on_schema_init },
+ { "clear_synchro_queue", lbox_ctl_clear_synchro_queue },
+ { NULL, NULL }
};
void
diff --git a/src/box/lua/error.cc b/src/box/lua/error.cc
index 54ec284..3a51d93 100644
--- a/src/box/lua/error.cc
+++ b/src/box/lua/error.cc
@@ -70,8 +70,8 @@ luaT_error_create(lua_State *L, int top_base)
lua_Debug info;
int top = lua_gettop(L);
int top_type = lua_type(L, top_base);
- if (top >= top_base && (top_type == LUA_TNUMBER ||
- top_type == LUA_TSTRING)) {
+ if (top >= top_base &&
+ (top_type == LUA_TNUMBER || top_type == LUA_TSTRING)) {
/* Shift of the "reason args". */
int shift = 1;
if (top_type == LUA_TNUMBER) {
@@ -182,8 +182,8 @@ luaT_error_new(lua_State *L)
{
struct error *e;
if (lua_gettop(L) == 0 || (e = luaT_error_create(L, 1)) == NULL) {
- return luaL_error(L, "Usage: box.error.new(code, args) or "\
- "box.error.new(type, args)");
+ return luaL_error(L, "Usage: box.error.new(code, args) or "
+ "box.error.new(type, args)");
}
lua_settop(L, 0);
luaT_pusherror(L, e);
@@ -213,12 +213,13 @@ luaT_error_set(struct lua_State *L)
static int
lbox_errinj_set(struct lua_State *L)
{
- char *name = (char*)luaL_checkstring(L, 1);
+ char *name = (char *)luaL_checkstring(L, 1);
struct errinj *errinj;
errinj = errinj_by_name(name);
if (errinj == NULL) {
say_error("%s", name);
- lua_pushfstring(L, "error: can't find error injection '%s'", name);
+ lua_pushfstring(L, "error: can't find error injection '%s'",
+ name);
return 1;
}
switch (errinj->type) {
@@ -262,7 +263,7 @@ lbox_errinj_push_value(struct lua_State *L, const struct errinj *e)
static int
lbox_errinj_get(struct lua_State *L)
{
- char *name = (char*)luaL_checkstring(L, 1);
+ char *name = (char *)luaL_checkstring(L, 1);
struct errinj *e = errinj_by_name(name);
if (e != NULL)
return lbox_errinj_push_value(L, e);
@@ -273,7 +274,7 @@ lbox_errinj_get(struct lua_State *L)
static inline int
lbox_errinj_cb(struct errinj *e, void *cb_ctx)
{
- struct lua_State *L = (struct lua_State*)cb_ctx;
+ struct lua_State *L = (struct lua_State *)cb_ctx;
lua_pushstring(L, e->name);
lua_newtable(L);
lua_pushstring(L, "state");
@@ -292,10 +293,9 @@ lbox_errinj_info(struct lua_State *L)
}
void
-box_lua_error_init(struct lua_State *L) {
- static const struct luaL_Reg errorlib[] = {
- {NULL, NULL}
- };
+box_lua_error_init(struct lua_State *L)
+{
+ static const struct luaL_Reg errorlib[] = { { NULL, NULL } };
luaL_register_module(L, "box.error", errorlib);
for (int i = 0; i < box_error_code_MAX; i++) {
const char *name = box_error_codes[i].errstr;
@@ -334,12 +334,11 @@ box_lua_error_init(struct lua_State *L) {
lua_pop(L, 1);
- static const struct luaL_Reg errinjlib[] = {
- {"info", lbox_errinj_info},
- {"set", lbox_errinj_set},
- {"get", lbox_errinj_get},
- {NULL, NULL}
- };
+ static const struct luaL_Reg errinjlib[] = { { "info",
+ lbox_errinj_info },
+ { "set", lbox_errinj_set },
+ { "get", lbox_errinj_get },
+ { NULL, NULL } };
/* box.error.injection is not set by register_module */
luaL_register_module(L, "box.error.injection", errinjlib);
lua_pop(L, 1);
diff --git a/src/box/lua/execute.c b/src/box/lua/execute.c
index 926a0a6..41bfee9 100644
--- a/src/box/lua/execute.c
+++ b/src/box/lua/execute.c
@@ -99,7 +99,7 @@ lbox_execute_prepared(struct lua_State *L)
{
int top = lua_gettop(L);
- if ((top != 1 && top != 2) || ! lua_istable(L, 1))
+ if ((top != 1 && top != 2) || !lua_istable(L, 1))
return luaL_error(L, "Usage: statement:execute([, params])");
lua_getfield(L, 1, "stmt_id");
if (!lua_isnumber(L, -1))
@@ -138,15 +138,15 @@ lbox_unprepare(struct lua_State *L)
{
int top = lua_gettop(L);
- if (top != 1 || (! lua_istable(L, 1) && ! lua_isnumber(L, 1))) {
- return luaL_error(L, "Usage: statement:unprepare() or "\
+ if (top != 1 || (!lua_istable(L, 1) && !lua_isnumber(L, 1))) {
+ return luaL_error(L, "Usage: statement:unprepare() or "
"box.unprepare(stmt_id)");
}
lua_Integer stmt_id;
if (lua_istable(L, 1)) {
lua_getfield(L, -1, "stmt_id");
- if (! lua_isnumber(L, -1)) {
- return luaL_error(L, "Statement id is expected "\
+ if (!lua_isnumber(L, -1)) {
+ return luaL_error(L, "Statement id is expected "
"to be numeric");
}
stmt_id = lua_tointeger(L, -1);
@@ -156,7 +156,7 @@ lbox_unprepare(struct lua_State *L)
}
if (stmt_id < 0)
return luaL_error(L, "Statement id can't be negative");
- if (sql_unprepare((uint32_t) stmt_id) != 0)
+ if (sql_unprepare((uint32_t)stmt_id) != 0)
return luaT_push_nil_and_error(L);
return 0;
}
@@ -164,7 +164,7 @@ lbox_unprepare(struct lua_State *L)
void
port_sql_dump_lua(struct port *port, struct lua_State *L, bool is_flat)
{
- (void) is_flat;
+ (void)is_flat;
assert(is_flat == false);
assert(port->vtab == &port_sql_vtab);
struct sql *db = sql_get();
@@ -180,9 +180,9 @@ port_sql_dump_lua(struct port *port, struct lua_State *L, bool is_flat)
break;
}
case DML_EXECUTE: {
- assert(((struct port_c *) port)->size == 0);
+ assert(((struct port_c *)port)->size == 0);
struct stailq *autoinc_id_list =
- vdbe_autoinc_id_list((struct Vdbe *) stmt);
+ vdbe_autoinc_id_list((struct Vdbe *)stmt);
lua_createtable(L, 0, stailq_empty(autoinc_id_list) ? 1 : 2);
luaL_pushuint64(L, db->nChange);
@@ -192,7 +192,8 @@ port_sql_dump_lua(struct port *port, struct lua_State *L, bool is_flat)
lua_newtable(L);
int i = 1;
struct autoinc_id_entry *id_entry;
- stailq_foreach_entry(id_entry, autoinc_id_list, link) {
+ stailq_foreach_entry(id_entry, autoinc_id_list, link)
+ {
if (id_entry->id >= 0)
luaL_pushuint64(L, id_entry->id);
else
@@ -236,8 +237,8 @@ port_sql_dump_lua(struct port *port, struct lua_State *L, bool is_flat)
lua_setfield(L, -2, "unprepare");
break;
}
- case DML_PREPARE : {
- assert(((struct port_c *) port)->size == 0);
+ case DML_PREPARE: {
+ assert(((struct port_c *)port)->size == 0);
/* Format is following:
* stmt_id,
* param_count,
@@ -264,7 +265,7 @@ port_sql_dump_lua(struct port *port, struct lua_State *L, bool is_flat)
lua_setfield(L, -2, "unprepare");
break;
}
- default:{
+ default: {
unreachable();
}
}
@@ -296,16 +297,18 @@ lua_sql_bind_decode(struct lua_State *L, struct sql_bind *bind, int idx, int i)
*/
lua_pushnil(L);
lua_next(L, -2);
- if (! lua_isstring(L, -2)) {
- diag_set(ClientError, ER_ILLEGAL_PARAMS, "name of the "\
+ if (!lua_isstring(L, -2)) {
+ diag_set(ClientError, ER_ILLEGAL_PARAMS,
+ "name of the "
"parameter should be a string.");
return -1;
}
/* Check that the table is one-row sized. */
lua_pushvalue(L, -2);
if (lua_next(L, -4) != 0) {
- diag_set(ClientError, ER_ILLEGAL_PARAMS, "SQL bind "\
- "named parameter should be a table with "\
+ diag_set(ClientError, ER_ILLEGAL_PARAMS,
+ "SQL bind "
+ "named parameter should be a table with "
"one key - {name = value}");
return -1;
}
@@ -399,7 +402,7 @@ lua_sql_bind_list_decode(struct lua_State *L, struct sql_bind **out_bind,
return 0;
if (bind_count > SQL_BIND_PARAMETER_MAX) {
diag_set(ClientError, ER_SQL_BIND_PARAMETER_MAX,
- (int) bind_count);
+ (int)bind_count);
return -1;
}
struct region *region = &fiber()->gc;
@@ -410,8 +413,8 @@ lua_sql_bind_list_decode(struct lua_State *L, struct sql_bind **out_bind,
* sql_stmt_finalize() or in txn_commit()/txn_rollback() if
* there is an active transaction.
*/
- struct sql_bind *bind = region_alloc_array(region, typeof(bind[0]),
- bind_count, &size);
+ struct sql_bind *bind =
+ region_alloc_array(region, typeof(bind[0]), bind_count, &size);
if (bind == NULL) {
diag_set(OutOfMemory, size, "region_alloc_array", "bind");
return -1;
@@ -435,12 +438,12 @@ lbox_execute(struct lua_State *L)
struct port port;
int top = lua_gettop(L);
- if ((top != 1 && top != 2) || ! lua_isstring(L, 1))
+ if ((top != 1 && top != 2) || !lua_isstring(L, 1))
return luaL_error(L, "Usage: box.execute(sqlstring[, params]) "
- "or box.execute(stmt_id[, params])");
+ "or box.execute(stmt_id[, params])");
if (top == 2) {
- if (! lua_istable(L, 2))
+ if (!lua_istable(L, 2))
return luaL_error(L, "Second argument must be a table");
bind_count = lua_sql_bind_list_decode(L, &bind, 2);
if (bind_count < 0)
@@ -452,8 +455,8 @@ lbox_execute(struct lua_State *L)
*/
if (lua_type(L, 1) == LUA_TSTRING) {
const char *sql = lua_tolstring(L, 1, &length);
- if (sql_prepare_and_execute(sql, length, bind, bind_count, &port,
- &fiber()->gc) != 0)
+ if (sql_prepare_and_execute(sql, length, bind, bind_count,
+ &port, &fiber()->gc) != 0)
return luaT_push_nil_and_error(L);
} else {
assert(lua_type(L, 1) == LUA_TNUMBER);
@@ -479,7 +482,7 @@ lbox_prepare(struct lua_State *L)
struct port port;
int top = lua_gettop(L);
- if ((top != 1 && top != 2) || ! lua_isstring(L, 1))
+ if ((top != 1 && top != 2) || !lua_isstring(L, 1))
return luaL_error(L, "Usage: box.prepare(sqlstring)");
const char *sql = lua_tolstring(L, 1, &length);
diff --git a/src/box/lua/index.c b/src/box/lua/index.c
index 4cf3c4d..5996979 100644
--- a/src/box/lua/index.c
+++ b/src/box/lua/index.c
@@ -90,8 +90,8 @@ lbox_index_update(lua_State *L)
const char *ops = lbox_encode_tuple_on_gc(L, 4, &ops_len);
struct tuple *result;
- if (box_update(space_id, index_id, key, key + key_len,
- ops, ops + ops_len, 1, &result) != 0)
+ if (box_update(space_id, index_id, key, key + key_len, ops,
+ ops + ops_len, 1, &result) != 0)
return luaT_error(L);
return luaT_pushtupleornil(L, result);
}
@@ -111,8 +111,8 @@ lbox_upsert(lua_State *L)
const char *ops = lbox_encode_tuple_on_gc(L, 3, &ops_len);
struct tuple *result;
- if (box_upsert(space_id, 0, tuple, tuple + tuple_len,
- ops, ops + ops_len, 1, &result) != 0)
+ if (box_upsert(space_id, 0, tuple, tuple + tuple_len, ops,
+ ops + ops_len, 1, &result) != 0)
return luaT_error(L);
return luaT_pushtupleornil(L, result);
}
@@ -140,7 +140,8 @@ lbox_index_random(lua_State *L)
{
if (lua_gettop(L) != 3 || !lua_isnumber(L, 1) || !lua_isnumber(L, 2) ||
!lua_isnumber(L, 3))
- return luaL_error(L, "Usage index.random(space_id, index_id, rnd)");
+ return luaL_error(
+ L, "Usage index.random(space_id, index_id, rnd)");
uint32_t space_id = lua_tonumber(L, 1);
uint32_t index_id = lua_tonumber(L, 2);
@@ -156,7 +157,8 @@ static int
lbox_index_get(lua_State *L)
{
if (lua_gettop(L) != 3 || !lua_isnumber(L, 1) || !lua_isnumber(L, 2))
- return luaL_error(L, "Usage index.get(space_id, index_id, key)");
+ return luaL_error(L,
+ "Usage index.get(space_id, index_id, key)");
uint32_t space_id = lua_tonumber(L, 1);
uint32_t index_id = lua_tonumber(L, 2);
@@ -173,7 +175,8 @@ static int
lbox_index_min(lua_State *L)
{
if (lua_gettop(L) != 3 || !lua_isnumber(L, 1) || !lua_isnumber(L, 2))
- return luaL_error(L, "usage index.min(space_id, index_id, key)");
+ return luaL_error(L,
+ "usage index.min(space_id, index_id, key)");
uint32_t space_id = lua_tonumber(L, 1);
uint32_t index_id = lua_tonumber(L, 2);
@@ -190,7 +193,8 @@ static int
lbox_index_max(lua_State *L)
{
if (lua_gettop(L) != 3 || !lua_isnumber(L, 1) || !lua_isnumber(L, 2))
- return luaL_error(L, "usage index.max(space_id, index_id, key)");
+ return luaL_error(L,
+ "usage index.max(space_id, index_id, key)");
uint32_t space_id = lua_tonumber(L, 1);
uint32_t index_id = lua_tonumber(L, 2);
@@ -209,7 +213,7 @@ lbox_index_count(lua_State *L)
if (lua_gettop(L) != 4 || !lua_isnumber(L, 1) || !lua_isnumber(L, 2) ||
!lua_isnumber(L, 3)) {
return luaL_error(L, "usage index.count(space_id, index_id, "
- "iterator, key)");
+ "iterator, key)");
}
uint32_t space_id = lua_tonumber(L, 1);
@@ -244,13 +248,16 @@ lbox_index_iterator(lua_State *L)
{
if (lua_gettop(L) != 4 || !lua_isnumber(L, 1) || !lua_isnumber(L, 2) ||
!lua_isnumber(L, 3))
- return luaL_error(L, "usage index.iterator(space_id, index_id, type, key)");
+ return luaL_error(
+ L,
+ "usage index.iterator(space_id, index_id, type, key)");
uint32_t space_id = lua_tonumber(L, 1);
uint32_t index_id = lua_tonumber(L, 2);
uint32_t iterator = lua_tonumber(L, 3);
size_t mpkey_len;
- const char *mpkey = lua_tolstring(L, 4, &mpkey_len); /* Key encoded by Lua */
+ const char *mpkey =
+ lua_tolstring(L, 4, &mpkey_len); /* Key encoded by Lua */
/* const char *key = lbox_encode_tuple_on_gc(L, 4, key_len); */
struct iterator *it = box_index_iterator(space_id, index_id, iterator,
mpkey, mpkey + mpkey_len);
@@ -258,8 +265,8 @@ lbox_index_iterator(lua_State *L)
return luaT_error(L);
assert(CTID_STRUCT_ITERATOR_REF != 0);
- struct iterator **ptr = (struct iterator **) luaL_pushcdata(L,
- CTID_STRUCT_ITERATOR_REF);
+ struct iterator **ptr =
+ (struct iterator **)luaL_pushcdata(L, CTID_STRUCT_ITERATOR_REF);
*ptr = it; /* NULL handled by Lua, gc also set by Lua */
return 1;
}
@@ -274,10 +281,10 @@ lbox_iterator_next(lua_State *L)
assert(CTID_STRUCT_ITERATOR_REF != 0);
uint32_t ctypeid;
void *data = luaL_checkcdata(L, 1, &ctypeid);
- if (ctypeid != (uint32_t) CTID_STRUCT_ITERATOR_REF)
+ if (ctypeid != (uint32_t)CTID_STRUCT_ITERATOR_REF)
return luaL_error(L, "usage: next(state)");
- struct iterator *itr = *(struct iterator **) data;
+ struct iterator *itr = *(struct iterator **)data;
struct tuple *tuple;
if (box_iterator_next(itr, &tuple) != 0)
return luaT_error(L);
@@ -336,13 +343,11 @@ box_lua_index_init(struct lua_State *L)
/* Get CTypeIDs */
int rc = luaL_cdef(L, "struct iterator;");
assert(rc == 0);
- (void) rc;
+ (void)rc;
CTID_STRUCT_ITERATOR_REF = luaL_ctypeid(L, "struct iterator&");
assert(CTID_STRUCT_ITERATOR_REF != 0);
- static const struct luaL_Reg indexlib [] = {
- {NULL, NULL}
- };
+ static const struct luaL_Reg indexlib[] = { { NULL, NULL } };
/* box.index */
luaL_register_module(L, "box.index", indexlib);
@@ -350,22 +355,22 @@ box_lua_index_init(struct lua_State *L)
lua_pop(L, 1);
static const struct luaL_Reg boxlib_internal[] = {
- {"insert", lbox_insert},
- {"replace", lbox_replace},
- {"update", lbox_index_update},
- {"upsert", lbox_upsert},
- {"delete", lbox_index_delete},
- {"random", lbox_index_random},
- {"get", lbox_index_get},
- {"min", lbox_index_min},
- {"max", lbox_index_max},
- {"count", lbox_index_count},
- {"iterator", lbox_index_iterator},
- {"iterator_next", lbox_iterator_next},
- {"truncate", lbox_truncate},
- {"stat", lbox_index_stat},
- {"compact", lbox_index_compact},
- {NULL, NULL}
+ { "insert", lbox_insert },
+ { "replace", lbox_replace },
+ { "update", lbox_index_update },
+ { "upsert", lbox_upsert },
+ { "delete", lbox_index_delete },
+ { "random", lbox_index_random },
+ { "get", lbox_index_get },
+ { "min", lbox_index_min },
+ { "max", lbox_index_max },
+ { "count", lbox_index_count },
+ { "iterator", lbox_index_iterator },
+ { "iterator_next", lbox_iterator_next },
+ { "truncate", lbox_truncate },
+ { "stat", lbox_index_stat },
+ { "compact", lbox_index_compact },
+ { NULL, NULL }
};
luaL_register(L, "box.internal", boxlib_internal);
diff --git a/src/box/lua/info.c b/src/box/lua/info.c
index cac3fd4..66d314c 100644
--- a/src/box/lua/info.c
+++ b/src/box/lua/info.c
@@ -60,7 +60,8 @@ lbox_pushvclock(struct lua_State *L, const struct vclock *vclock)
lua_createtable(L, 0, vclock_size(vclock));
struct vclock_iterator it;
vclock_iterator_init(&it, vclock);
- vclock_foreach(&it, replica) {
+ vclock_foreach(&it, replica)
+ {
lua_pushinteger(L, replica.id);
luaL_pushuint64(L, replica.lsn);
lua_settable(L, -3);
@@ -91,7 +92,8 @@ lbox_pushapplier(lua_State *L, struct applier *applier)
char *d = status;
const char *s = applier_state_strs[applier->state] + strlen("APPLIER_");
assert(strlen(s) < sizeof(status));
- while ((*(d++) = tolower(*(s++))));
+ while ((*(d++) = tolower(*(s++))))
+ ;
lua_pushstring(L, "status");
lua_pushstring(L, status);
@@ -104,11 +106,12 @@ lbox_pushapplier(lua_State *L, struct applier *applier)
lua_pushstring(L, "idle");
lua_pushnumber(L, ev_monotonic_now(loop()) -
- applier->last_row_time);
+ applier->last_row_time);
lua_settable(L, -3);
char name[APPLIER_SOURCE_MAXLEN];
- int total = uri_format(name, sizeof(name), &applier->uri, false);
+ int total =
+ uri_format(name, sizeof(name), &applier->uri, false);
/*
* total can be greater than sizeof(name) if
* name has insufficient length. Terminating
@@ -131,7 +134,7 @@ lbox_pushrelay(lua_State *L, struct relay *relay)
lua_newtable(L);
lua_pushstring(L, "status");
- switch(relay_get_state(relay)) {
+ switch (relay_get_state(relay)) {
case RELAY_FOLLOW:
lua_pushstring(L, "follow");
lua_settable(L, -3);
@@ -140,11 +143,10 @@ lbox_pushrelay(lua_State *L, struct relay *relay)
lua_settable(L, -3);
lua_pushstring(L, "idle");
lua_pushnumber(L, ev_monotonic_now(loop()) -
- relay_last_row_time(relay));
+ relay_last_row_time(relay));
lua_settable(L, -3);
break;
- case RELAY_STOPPED:
- {
+ case RELAY_STOPPED: {
lua_pushstring(L, "stopped");
lua_settable(L, -3);
@@ -153,7 +155,8 @@ lbox_pushrelay(lua_State *L, struct relay *relay)
lbox_push_replication_error_message(L, e, -1);
break;
}
- default: unreachable();
+ default:
+ unreachable();
}
}
@@ -202,7 +205,8 @@ lbox_info_replication(struct lua_State *L)
lua_setfield(L, -2, "__serialize");
lua_setmetatable(L, -2);
- replicaset_foreach(replica) {
+ replicaset_foreach(replica)
+ {
/* Applier hasn't received replica id yet */
if (replica->id == REPLICA_ID_NIL)
continue;
@@ -226,7 +230,8 @@ lbox_info_replication_anon_call(struct lua_State *L)
lua_setfield(L, -2, "__serialize");
lua_setmetatable(L, -2);
- replicaset_foreach(replica) {
+ replicaset_foreach(replica)
+ {
if (!replica->anon)
continue;
@@ -450,7 +455,8 @@ lbox_info_gc_call(struct lua_State *L)
count = 0;
struct gc_checkpoint *checkpoint;
- gc_foreach_checkpoint(checkpoint) {
+ gc_foreach_checkpoint(checkpoint)
+ {
lua_createtable(L, 0, 2);
lua_pushstring(L, "vclock");
@@ -465,7 +471,8 @@ lbox_info_gc_call(struct lua_State *L)
lua_newtable(L);
int ref_idx = 0;
struct gc_checkpoint_ref *ref;
- gc_foreach_checkpoint_ref(ref, checkpoint) {
+ gc_foreach_checkpoint_ref(ref, checkpoint)
+ {
lua_pushstring(L, ref->name);
lua_rawseti(L, -2, ++ref_idx);
}
@@ -594,38 +601,38 @@ lbox_info_election(struct lua_State *L)
}
static const struct luaL_Reg lbox_info_dynamic_meta[] = {
- {"id", lbox_info_id},
- {"uuid", lbox_info_uuid},
- {"lsn", lbox_info_lsn},
- {"signature", lbox_info_signature},
- {"vclock", lbox_info_vclock},
- {"ro", lbox_info_ro},
- {"replication", lbox_info_replication},
- {"replication_anon", lbox_info_replication_anon},
- {"status", lbox_info_status},
- {"uptime", lbox_info_uptime},
- {"pid", lbox_info_pid},
- {"cluster", lbox_info_cluster},
- {"memory", lbox_info_memory},
- {"gc", lbox_info_gc},
- {"vinyl", lbox_info_vinyl},
- {"sql", lbox_info_sql},
- {"listen", lbox_info_listen},
- {"election", lbox_info_election},
- {NULL, NULL}
+ { "id", lbox_info_id },
+ { "uuid", lbox_info_uuid },
+ { "lsn", lbox_info_lsn },
+ { "signature", lbox_info_signature },
+ { "vclock", lbox_info_vclock },
+ { "ro", lbox_info_ro },
+ { "replication", lbox_info_replication },
+ { "replication_anon", lbox_info_replication_anon },
+ { "status", lbox_info_status },
+ { "uptime", lbox_info_uptime },
+ { "pid", lbox_info_pid },
+ { "cluster", lbox_info_cluster },
+ { "memory", lbox_info_memory },
+ { "gc", lbox_info_gc },
+ { "vinyl", lbox_info_vinyl },
+ { "sql", lbox_info_sql },
+ { "listen", lbox_info_listen },
+ { "election", lbox_info_election },
+ { NULL, NULL }
};
static const struct luaL_Reg lbox_info_dynamic_meta_v16[] = {
- {"server", lbox_info_server},
- {NULL, NULL}
+ { "server", lbox_info_server },
+ { NULL, NULL }
};
/** Evaluate box.info.* function value and push it on the stack. */
static int
lbox_info_index(struct lua_State *L)
{
- lua_pushvalue(L, -1); /* dup key */
- lua_gettable(L, lua_upvalueindex(1)); /* table[key] */
+ lua_pushvalue(L, -1); /* dup key */
+ lua_gettable(L, lua_upvalueindex(1)); /* table[key] */
if (!lua_isfunction(L, -1)) {
/* No such key. Leave nil is on the stack. */
@@ -683,13 +690,11 @@ lbox_info_call(struct lua_State *L)
void
box_lua_info_init(struct lua_State *L)
{
- static const struct luaL_Reg infolib [] = {
- {NULL, NULL}
- };
+ static const struct luaL_Reg infolib[] = { { NULL, NULL } };
luaL_register_module(L, "box.info", infolib);
- lua_newtable(L); /* metatable for info */
+ lua_newtable(L); /* metatable for info */
lua_pushstring(L, "__index");
diff --git a/src/box/lua/init.c b/src/box/lua/init.c
index d0316ef..05c8b7e 100644
--- a/src/box/lua/init.c
+++ b/src/box/lua/init.c
@@ -70,41 +70,28 @@
static uint32_t CTID_STRUCT_TXN_SAVEPOINT_PTR = 0;
-extern char session_lua[],
- tuple_lua[],
- key_def_lua[],
- schema_lua[],
- load_cfg_lua[],
- xlog_lua[],
+extern char session_lua[], tuple_lua[], key_def_lua[], schema_lua[],
+ load_cfg_lua[], xlog_lua[],
#if ENABLE_FEEDBACK_DAEMON
feedback_daemon_lua[],
#endif
- net_box_lua[],
- upgrade_lua[],
- console_lua[],
- merger_lua[];
-
-static const char *lua_sources[] = {
- "box/session", session_lua,
- "box/tuple", tuple_lua,
- "box/schema", schema_lua,
+ net_box_lua[], upgrade_lua[], console_lua[], merger_lua[];
+
+static const char *lua_sources[] = { "box/session", session_lua, "box/tuple",
+ tuple_lua, "box/schema", schema_lua,
#if ENABLE_FEEDBACK_DAEMON
- /*
+ /*
* It is important to initialize the daemon before
* load_cfg, because the latter picks up some values
* from the feedback daemon.
*/
- "box/feedback_daemon", feedback_daemon_lua,
+ "box/feedback_daemon", feedback_daemon_lua,
#endif
- "box/upgrade", upgrade_lua,
- "box/net_box", net_box_lua,
- "box/console", console_lua,
- "box/load_cfg", load_cfg_lua,
- "box/xlog", xlog_lua,
- "box/key_def", key_def_lua,
- "box/merger", merger_lua,
- NULL
-};
+ "box/upgrade", upgrade_lua, "box/net_box",
+ net_box_lua, "box/console", console_lua,
+ "box/load_cfg", load_cfg_lua, "box/xlog",
+ xlog_lua, "box/key_def", key_def_lua,
+ "box/merger", merger_lua, NULL };
static int
lbox_commit(lua_State *L)
@@ -193,8 +180,8 @@ lbox_rollback_to_savepoint(struct lua_State *L)
if (lua_gettop(L) != 1 ||
(svp = luaT_check_txn_savepoint(L, 1, &svp_txn_id)) == NULL)
- return luaL_error(L,
- "Usage: box.rollback_to_savepoint(savepoint)");
+ return luaL_error(
+ L, "Usage: box.rollback_to_savepoint(savepoint)");
/*
* Verify that we're in a transaction and that it is the
@@ -242,7 +229,7 @@ lbox_txn_iterator_next(struct lua_State *L)
return luaT_error(L);
}
struct txn_stmt *stmt =
- (struct txn_stmt *) lua_topointer(L, lua_upvalueindex(2));
+ (struct txn_stmt *)lua_topointer(L, lua_upvalueindex(2));
if (stmt == NULL)
return 0;
while (stmt->row == NULL) {
@@ -302,7 +289,7 @@ lbox_txn_pairs(struct lua_State *L)
static int
lbox_push_txn(struct lua_State *L, void *event)
{
- struct txn *txn = (struct txn *) event;
+ struct txn *txn = (struct txn *)event;
luaL_pushint64(L, txn->id);
lua_pushcclosure(L, lbox_txn_pairs, 1);
return 1;
@@ -313,18 +300,20 @@ lbox_push_txn(struct lua_State *L, void *event)
* @sa lbox_trigger_reset.
*/
#define LBOX_TXN_TRIGGER(name) \
-static int \
-lbox_on_##name(struct lua_State *L) { \
- struct txn *txn = in_txn(); \
- int top = lua_gettop(L); \
- if (top > 2 || txn == NULL) { \
- return luaL_error(L, "Usage inside a transaction: " \
- "box.on_" #name "([function | nil, " \
- "[function | nil]])"); \
- } \
- txn_init_triggers(txn); \
- return lbox_trigger_reset(L, 2, &txn->on_##name, lbox_push_txn, NULL); \
-}
+ static int lbox_on_##name(struct lua_State *L) \
+ { \
+ struct txn *txn = in_txn(); \
+ int top = lua_gettop(L); \
+ if (top > 2 || txn == NULL) { \
+ return luaL_error(L, \
+ "Usage inside a transaction: " \
+ "box.on_" #name "([function | nil, " \
+ "[function | nil]])"); \
+ } \
+ txn_init_triggers(txn); \
+ return lbox_trigger_reset(L, 2, &txn->on_##name, \
+ lbox_push_txn, NULL); \
+ }
LBOX_TXN_TRIGGER(commit)
LBOX_TXN_TRIGGER(rollback)
@@ -384,21 +373,18 @@ lbox_backup_stop(struct lua_State *L)
return 0;
}
-static const struct luaL_Reg boxlib[] = {
- {"commit", lbox_commit},
- {"rollback", lbox_rollback},
- {"on_commit", lbox_on_commit},
- {"on_rollback", lbox_on_rollback},
- {"snapshot", lbox_snapshot},
- {"rollback_to_savepoint", lbox_rollback_to_savepoint},
- {NULL, NULL}
-};
+static const struct luaL_Reg boxlib[] = { { "commit", lbox_commit },
+ { "rollback", lbox_rollback },
+ { "on_commit", lbox_on_commit },
+ { "on_rollback", lbox_on_rollback },
+ { "snapshot", lbox_snapshot },
+ { "rollback_to_savepoint",
+ lbox_rollback_to_savepoint },
+ { NULL, NULL } };
-static const struct luaL_Reg boxlib_backup[] = {
- {"start", lbox_backup_start},
- {"stop", lbox_backup_stop},
- {NULL, NULL}
-};
+static const struct luaL_Reg boxlib_backup[] = { { "start", lbox_backup_start },
+ { "stop", lbox_backup_stop },
+ { NULL, NULL } };
/**
* A MsgPack extensions handler, for types defined in box.
@@ -452,8 +438,8 @@ void
box_lua_init(struct lua_State *L)
{
luaL_cdef(L, "struct txn_savepoint;");
- CTID_STRUCT_TXN_SAVEPOINT_PTR = luaL_ctypeid(L,
- "struct txn_savepoint*");
+ CTID_STRUCT_TXN_SAVEPOINT_PTR =
+ luaL_ctypeid(L, "struct txn_savepoint*");
/* Use luaL_register() to set _G.box */
luaL_register(L, "box", boxlib);
@@ -493,12 +479,12 @@ box_lua_init(struct lua_State *L)
for (const char **s = lua_sources; *s; s += 2) {
const char *modname = *s;
const char *modsrc = *(s + 1);
- const char *modfile = lua_pushfstring(L,
- "@builtin/%s.lua", modname);
+ const char *modfile =
+ lua_pushfstring(L, "@builtin/%s.lua", modname);
if (luaL_loadbuffer(L, modsrc, strlen(modsrc), modfile) != 0 ||
lua_pcall(L, 0, 0, 0) != 0)
- panic("Error loading Lua module %s...: %s",
- modname, lua_tostring(L, -1));
+ panic("Error loading Lua module %s...: %s", modname,
+ lua_tostring(L, -1));
lua_pop(L, 1); /* modfile */
}
diff --git a/src/box/lua/key_def.c b/src/box/lua/key_def.c
index 1a99fab..81f54ea 100644
--- a/src/box/lua/key_def.c
+++ b/src/box/lua/key_def.c
@@ -102,7 +102,7 @@ luaT_key_def_set_part(struct lua_State *L, struct key_part_def *part,
}
} else {
lua_getfield(L, -2, "field");
- if (! lua_isnil(L, -1)) {
+ if (!lua_isnil(L, -1)) {
diag_set(IllegalParams,
"Conflicting options: fieldno and field");
return -1;
@@ -171,14 +171,14 @@ luaT_key_def_set_part(struct lua_State *L, struct key_part_def *part,
/* Check for conflicting options. */
if (part->coll_id != COLL_NONE) {
diag_set(IllegalParams, "Conflicting options: "
- "collation_id and collation");
+ "collation_id and collation");
return -1;
}
size_t coll_name_len;
const char *coll_name = lua_tolstring(L, -1, &coll_name_len);
- struct coll_id *coll_id = coll_by_name(coll_name,
- coll_name_len);
+ struct coll_id *coll_id =
+ coll_by_name(coll_name, coll_name_len);
if (coll_id == NULL) {
diag_set(IllegalParams, "Unknown collation: \"%s\"",
coll_name);
@@ -198,8 +198,8 @@ luaT_key_def_set_part(struct lua_State *L, struct key_part_def *part,
diag_set(IllegalParams, "invalid path");
return -1;
}
- if ((size_t)json_path_multikey_offset(path, path_len,
- TUPLE_INDEX_BASE) != path_len) {
+ if ((size_t)json_path_multikey_offset(
+ path, path_len, TUPLE_INDEX_BASE) != path_len) {
diag_set(IllegalParams, "multikey path is unsupported");
return -1;
}
@@ -358,15 +358,14 @@ lbox_key_def_compare_with_key(struct lua_State *L)
size_t key_len;
const char *key_end, *key = lbox_encode_tuple_on_gc(L, 3, &key_len);
uint32_t part_count = mp_decode_array(&key);
- if (key_validate_parts(key_def, key, part_count, true,
- &key_end) != 0) {
+ if (key_validate_parts(key_def, key, part_count, true, &key_end) != 0) {
region_truncate(region, region_svp);
tuple_unref(tuple);
return luaT_error(L);
}
- int rc = tuple_compare_with_key(tuple, HINT_NONE, key,
- part_count, HINT_NONE, key_def);
+ int rc = tuple_compare_with_key(tuple, HINT_NONE, key, part_count,
+ HINT_NONE, key_def);
region_truncate(region, region_svp);
tuple_unref(tuple);
lua_pushinteger(L, rc);
@@ -395,14 +394,13 @@ lbox_key_def_merge(struct lua_State *L)
if (new_key_def == NULL)
return luaT_error(L);
- *(struct key_def **) luaL_pushcdata(L,
- CTID_STRUCT_KEY_DEF_REF) = new_key_def;
+ *(struct key_def **)luaL_pushcdata(L, CTID_STRUCT_KEY_DEF_REF) =
+ new_key_def;
lua_pushcfunction(L, lbox_key_def_gc);
luaL_setcdatagc(L, -2);
return 1;
}
-
/**
* Push a new table representing a key_def to a Lua stack.
*/
@@ -431,11 +429,11 @@ lbox_key_def_new(struct lua_State *L)
{
if (lua_gettop(L) != 1 || lua_istable(L, 1) != 1)
return luaL_error(L, "Bad params, use: key_def.new({"
- "{fieldno = fieldno, type = type"
- "[, is_nullable = <boolean>]"
- "[, path = <string>]"
- "[, collation_id = <number>]"
- "[, collation = <string>]}, ...}");
+ "{fieldno = fieldno, type = type"
+ "[, is_nullable = <boolean>]"
+ "[, path = <string>]"
+ "[, collation_id = <number>]"
+ "[, collation = <string>]}, ...}");
uint32_t part_count = lua_objlen(L, 1);
@@ -478,8 +476,8 @@ lbox_key_def_new(struct lua_State *L)
*/
key_def_update_optionality(key_def, 0);
- *(struct key_def **) luaL_pushcdata(L,
- CTID_STRUCT_KEY_DEF_REF) = key_def;
+ *(struct key_def **)luaL_pushcdata(L, CTID_STRUCT_KEY_DEF_REF) =
+ key_def;
lua_pushcfunction(L, lbox_key_def_gc);
luaL_setcdatagc(L, -2);
@@ -494,13 +492,13 @@ luaopen_key_def(struct lua_State *L)
/* Export C functions to Lua. */
static const struct luaL_Reg meta[] = {
- {"new", lbox_key_def_new},
- {"extract_key", lbox_key_def_extract_key},
- {"compare", lbox_key_def_compare},
- {"compare_with_key", lbox_key_def_compare_with_key},
- {"merge", lbox_key_def_merge},
- {"totable", lbox_key_def_to_table},
- {NULL, NULL}
+ { "new", lbox_key_def_new },
+ { "extract_key", lbox_key_def_extract_key },
+ { "compare", lbox_key_def_compare },
+ { "compare_with_key", lbox_key_def_compare_with_key },
+ { "merge", lbox_key_def_merge },
+ { "totable", lbox_key_def_to_table },
+ { NULL, NULL }
};
luaL_register_module(L, "key_def", meta);
return 1;
diff --git a/src/box/lua/merger.c b/src/box/lua/merger.c
index 583946c..1dbca5b 100644
--- a/src/box/lua/merger.c
+++ b/src/box/lua/merger.c
@@ -37,26 +37,26 @@
#include <stdlib.h>
#include <string.h>
-#include <lua.h> /* lua_*() */
-#include <lauxlib.h> /* luaL_*() */
+#include <lua.h> /* lua_*() */
+#include <lauxlib.h> /* luaL_*() */
-#include "fiber.h" /* fiber() */
-#include "diag.h" /* diag_set() */
+#include "fiber.h" /* fiber() */
+#include "diag.h" /* diag_set() */
-#include "box/tuple.h" /* tuple_format_runtime,
+#include "box/tuple.h" /* tuple_format_runtime,
tuple_*(), ... */
-#include "lua/error.h" /* luaT_error() */
-#include "lua/utils.h" /* luaL_pushcdata(),
+#include "lua/error.h" /* luaT_error() */
+#include "lua/utils.h" /* luaL_pushcdata(),
luaL_iterator_*() */
#include "box/lua/key_def.h" /* luaT_check_key_def() */
#include "box/lua/tuple.h" /* luaT_tuple_new() */
-#include "small/ibuf.h" /* struct ibuf */
-#include "msgpuck.h" /* mp_*() */
+#include "small/ibuf.h" /* struct ibuf */
+#include "msgpuck.h" /* mp_*() */
-#include "box/merger.h" /* merge_source_*, merger_*() */
+#include "box/merger.h" /* merge_source_*, merger_*() */
static uint32_t CTID_STRUCT_MERGE_SOURCE_REF = 0;
@@ -105,7 +105,7 @@ decode_header(struct ibuf *buf, size_t *len_p)
if (ok)
ok = mp_check_array(buf->rpos, buf->wpos) <= 0;
if (ok)
- *len_p = mp_decode_array((const char **) &buf->rpos);
+ *len_p = mp_decode_array((const char **)&buf->rpos);
return ok ? 0 : -1;
}
@@ -270,8 +270,8 @@ lbox_merge_source_new(struct lua_State *L, const char *func_name,
merge_source_unref(source);
return luaT_error(L);
}
- *(struct merge_source **)
- luaL_pushcdata(L, CTID_STRUCT_MERGE_SOURCE_REF) = source;
+ *(struct merge_source **)luaL_pushcdata(
+ L, CTID_STRUCT_MERGE_SOURCE_REF) = source;
lua_pushcfunction(L, lbox_merge_source_gc);
luaL_setcdatagc(L, -2);
@@ -310,8 +310,8 @@ luaT_merger_new_parse_sources(struct lua_State *L, int idx,
{
/* Allocate sources array. */
uint32_t source_count = lua_objlen(L, idx);
- const size_t sources_size = sizeof(struct merge_source *) *
- source_count;
+ const size_t sources_size =
+ sizeof(struct merge_source *) * source_count;
struct merge_source **sources = malloc(sources_size);
if (sources == NULL) {
diag_set(OutOfMemory, sources_size, "malloc", "sources");
@@ -352,12 +352,12 @@ lbox_merger_new(struct lua_State *L)
struct key_def *key_def;
int top = lua_gettop(L);
bool ok = (top == 2 || top == 3) &&
- /* key_def. */
- (key_def = luaT_check_key_def(L, 1)) != NULL &&
- /* Sources. */
- lua_istable(L, 2) == 1 &&
- /* Opts. */
- (lua_isnoneornil(L, 3) == 1 || lua_istable(L, 3) == 1);
+ /* key_def. */
+ (key_def = luaT_check_key_def(L, 1)) != NULL &&
+ /* Sources. */
+ lua_istable(L, 2) == 1 &&
+ /* Opts. */
+ (lua_isnoneornil(L, 3) == 1 || lua_istable(L, 3) == 1);
if (!ok)
return lbox_merger_new_usage(L, NULL);
@@ -379,21 +379,21 @@ lbox_merger_new(struct lua_State *L)
}
uint32_t source_count = 0;
- struct merge_source **sources = luaT_merger_new_parse_sources(L, 2,
- &source_count);
+ struct merge_source **sources =
+ luaT_merger_new_parse_sources(L, 2, &source_count);
if (sources == NULL)
return luaT_error(L);
- struct merge_source *merger = merger_new(key_def, sources, source_count,
- reverse);
+ struct merge_source *merger =
+ merger_new(key_def, sources, source_count, reverse);
free(sources);
if (merger == NULL) {
merge_source_unref(merger);
return luaT_error(L);
}
- *(struct merge_source **)
- luaL_pushcdata(L, CTID_STRUCT_MERGE_SOURCE_REF) = merger;
+ *(struct merge_source **)luaL_pushcdata(
+ L, CTID_STRUCT_MERGE_SOURCE_REF) = merger;
lua_pushcfunction(L, lbox_merge_source_gc);
luaL_setcdatagc(L, -2);
@@ -435,8 +435,7 @@ static void
luaL_merge_source_buffer_destroy(struct merge_source *base);
static int
luaL_merge_source_buffer_next(struct merge_source *base,
- struct tuple_format *format,
- struct tuple **out);
+ struct tuple_format *format, struct tuple **out);
/* Non-virtual methods */
@@ -455,8 +454,8 @@ luaL_merge_source_buffer_new(struct lua_State *L)
.next = luaL_merge_source_buffer_next,
};
- struct merge_source_buffer *source = malloc(
- sizeof(struct merge_source_buffer));
+ struct merge_source_buffer *source =
+ malloc(sizeof(struct merge_source_buffer));
if (source == NULL) {
diag_set(OutOfMemory, sizeof(struct merge_source_buffer),
"malloc", "merge_source_buffer");
@@ -492,8 +491,10 @@ luaL_merge_source_buffer_fetch_impl(struct merge_source_buffer *source,
/* Handle incorrect results count. */
if (nresult != 2) {
- diag_set(IllegalParams, "Expected <state>, <buffer>, got %d "
- "return values", nresult);
+ diag_set(IllegalParams,
+ "Expected <state>, <buffer>, got %d "
+ "return values",
+ nresult);
return -1;
}
@@ -550,8 +551,8 @@ luaL_merge_source_buffer_fetch(struct merge_source_buffer *source)
static void
luaL_merge_source_buffer_destroy(struct merge_source *base)
{
- struct merge_source_buffer *source = container_of(base,
- struct merge_source_buffer, base);
+ struct merge_source_buffer *source =
+ container_of(base, struct merge_source_buffer, base);
assert(source->fetch_it != NULL);
luaL_iterator_delete(source->fetch_it);
@@ -568,11 +569,10 @@ luaL_merge_source_buffer_destroy(struct merge_source *base)
*/
static int
luaL_merge_source_buffer_next(struct merge_source *base,
- struct tuple_format *format,
- struct tuple **out)
+ struct tuple_format *format, struct tuple **out)
{
- struct merge_source_buffer *source = container_of(base,
- struct merge_source_buffer, base);
+ struct merge_source_buffer *source =
+ container_of(base, struct merge_source_buffer, base);
/*
* Handle the case when all data were processed: ask a
@@ -599,7 +599,7 @@ luaL_merge_source_buffer_next(struct merge_source *base,
return -1;
}
--source->remaining_tuple_count;
- source->buf->rpos = (char *) tuple_end;
+ source->buf->rpos = (char *)tuple_end;
if (format == NULL)
format = tuple_format_runtime;
struct tuple *tuple = tuple_new(format, tuple_beg, tuple_end);
@@ -648,8 +648,7 @@ static void
luaL_merge_source_table_destroy(struct merge_source *base);
static int
luaL_merge_source_table_next(struct merge_source *base,
- struct tuple_format *format,
- struct tuple **out);
+ struct tuple_format *format, struct tuple **out);
/* Non-virtual methods */
@@ -666,8 +665,8 @@ luaL_merge_source_table_new(struct lua_State *L)
.next = luaL_merge_source_table_next,
};
- struct merge_source_table *source = malloc(
- sizeof(struct merge_source_table));
+ struct merge_source_table *source =
+ malloc(sizeof(struct merge_source_table));
if (source == NULL) {
diag_set(OutOfMemory, sizeof(struct merge_source_table),
"malloc", "merge_source_table");
@@ -705,8 +704,10 @@ luaL_merge_source_table_fetch(struct merge_source_table *source,
/* Handle incorrect results count. */
if (nresult != 2) {
- diag_set(IllegalParams, "Expected <state>, <table>, got %d "
- "return values", nresult);
+ diag_set(IllegalParams,
+ "Expected <state>, <table>, got %d "
+ "return values",
+ nresult);
return -1;
}
@@ -737,8 +738,8 @@ luaL_merge_source_table_fetch(struct merge_source_table *source,
static void
luaL_merge_source_table_destroy(struct merge_source *base)
{
- struct merge_source_table *source = container_of(base,
- struct merge_source_table, base);
+ struct merge_source_table *source =
+ container_of(base, struct merge_source_table, base);
assert(source->fetch_it != NULL);
luaL_iterator_delete(source->fetch_it);
@@ -754,11 +755,10 @@ luaL_merge_source_table_destroy(struct merge_source *base)
static int
luaL_merge_source_table_next_impl(struct merge_source *base,
struct tuple_format *format,
- struct tuple **out,
- struct lua_State *L)
+ struct tuple **out, struct lua_State *L)
{
- struct merge_source_table *source = container_of(base,
- struct merge_source_table, base);
+ struct merge_source_table *source =
+ container_of(base, struct merge_source_table, base);
if (source->ref > 0) {
lua_rawgeti(L, LUA_REGISTRYINDEX, source->ref);
@@ -806,8 +806,7 @@ luaL_merge_source_table_next_impl(struct merge_source *base,
*/
static int
luaL_merge_source_table_next(struct merge_source *base,
- struct tuple_format *format,
- struct tuple **out)
+ struct tuple_format *format, struct tuple **out)
{
int coro_ref = LUA_NOREF;
int top = -1;
@@ -850,8 +849,7 @@ static void
luaL_merge_source_tuple_destroy(struct merge_source *base);
static int
luaL_merge_source_tuple_next(struct merge_source *base,
- struct tuple_format *format,
- struct tuple **out);
+ struct tuple_format *format, struct tuple **out);
/* Non-virtual methods */
@@ -868,8 +866,8 @@ luaL_merge_source_tuple_new(struct lua_State *L)
.next = luaL_merge_source_tuple_next,
};
- struct merge_source_tuple *source = malloc(
- sizeof(struct merge_source_tuple));
+ struct merge_source_tuple *source =
+ malloc(sizeof(struct merge_source_tuple));
if (source == NULL) {
diag_set(OutOfMemory, sizeof(struct merge_source_tuple),
"malloc", "merge_source_tuple");
@@ -896,7 +894,7 @@ luaL_merge_source_tuple_new(struct lua_State *L)
*/
static int
luaL_merge_source_tuple_fetch(struct merge_source_tuple *source,
- struct lua_State *L)
+ struct lua_State *L)
{
int nresult = luaL_iterator_next(L, source->fetch_it);
@@ -910,14 +908,16 @@ luaL_merge_source_tuple_fetch(struct merge_source_tuple *source,
/* Handle incorrect results count. */
if (nresult != 2) {
- diag_set(IllegalParams, "Expected <state>, <tuple>, got %d "
- "return values", nresult);
+ diag_set(IllegalParams,
+ "Expected <state>, <tuple>, got %d "
+ "return values",
+ nresult);
return -1;
}
/* Set a new tuple as the current chunk. */
lua_insert(L, -2); /* Swap state and tuple. */
- lua_pop(L, 1); /* Pop state. */
+ lua_pop(L, 1); /* Pop state. */
return 1;
}
@@ -932,8 +932,8 @@ luaL_merge_source_tuple_fetch(struct merge_source_tuple *source,
static void
luaL_merge_source_tuple_destroy(struct merge_source *base)
{
- struct merge_source_tuple *source = container_of(base,
- struct merge_source_tuple, base);
+ struct merge_source_tuple *source =
+ container_of(base, struct merge_source_tuple, base);
assert(source->fetch_it != NULL);
luaL_iterator_delete(source->fetch_it);
@@ -947,11 +947,10 @@ luaL_merge_source_tuple_destroy(struct merge_source *base)
static int
luaL_merge_source_tuple_next_impl(struct merge_source *base,
struct tuple_format *format,
- struct tuple **out,
- struct lua_State *L)
+ struct tuple **out, struct lua_State *L)
{
- struct merge_source_tuple *source = container_of(base,
- struct merge_source_tuple, base);
+ struct merge_source_tuple *source =
+ container_of(base, struct merge_source_tuple, base);
int rc = luaL_merge_source_tuple_fetch(source, L);
if (rc < 0)
@@ -981,8 +980,7 @@ luaL_merge_source_tuple_next_impl(struct merge_source *base,
*/
static int
luaL_merge_source_tuple_next(struct merge_source *base,
- struct tuple_format *format,
- struct tuple **out)
+ struct tuple_format *format, struct tuple **out)
{
int coro_ref = LUA_NOREF;
int top = -1;
@@ -1024,10 +1022,10 @@ lbox_merge_source_gen(struct lua_State *L)
{
struct merge_source *source;
bool ok = lua_gettop(L) == 2 && lua_isnil(L, 1) &&
- (source = luaT_check_merge_source(L, 2)) != NULL;
+ (source = luaT_check_merge_source(L, 2)) != NULL;
if (!ok)
return luaL_error(L, "Bad params, use: lbox_merge_source_gen("
- "nil, merge_source)");
+ "nil, merge_source)");
struct tuple *tuple;
if (merge_source_next(source, NULL, &tuple) != 0)
@@ -1039,8 +1037,8 @@ lbox_merge_source_gen(struct lua_State *L)
}
/* Push merge_source, tuple. */
- *(struct merge_source **)
- luaL_pushcdata(L, CTID_STRUCT_MERGE_SOURCE_REF) = source;
+ *(struct merge_source **)luaL_pushcdata(
+ L, CTID_STRUCT_MERGE_SOURCE_REF) = source;
luaT_pushtuple(L, tuple);
/*
@@ -1066,7 +1064,7 @@ lbox_merge_source_ipairs(struct lua_State *L)
{
struct merge_source *source;
bool ok = lua_gettop(L) == 1 &&
- (source = luaT_check_merge_source(L, 1)) != NULL;
+ (source = luaT_check_merge_source(L, 1)) != NULL;
if (!ok)
return luaL_error(L, "Usage: merge_source:ipairs()");
/* Stack: merge_source. */
@@ -1116,8 +1114,8 @@ encode_result_buffer(struct lua_State *L, struct merge_source *source,
/* Fetch, merge and copy tuples to the buffer. */
struct tuple *tuple;
int rc = 0;
- while (result_len < limit && (rc =
- merge_source_next(source, NULL, &tuple)) == 0 &&
+ while (result_len < limit &&
+ (rc = merge_source_next(source, NULL, &tuple)) == 0 &&
tuple != NULL) {
uint32_t bsize = tuple->bsize;
ibuf_reserve(output_buffer, bsize);
@@ -1156,8 +1154,8 @@ create_result_table(struct lua_State *L, struct merge_source *source,
/* Fetch, merge and save tuples to the table. */
struct tuple *tuple;
int rc = 0;
- while (cur - 1 < limit && (rc =
- merge_source_next(source, NULL, &tuple)) == 0 &&
+ while (cur - 1 < limit &&
+ (rc = merge_source_next(source, NULL, &tuple)) == 0 &&
tuple != NULL) {
luaT_pushtuple(L, tuple);
lua_rawseti(L, -2, cur);
@@ -1209,10 +1207,10 @@ lbox_merge_source_select(struct lua_State *L)
struct merge_source *source;
int top = lua_gettop(L);
bool ok = (top == 1 || top == 2) &&
- /* Merge source. */
- (source = luaT_check_merge_source(L, 1)) != NULL &&
- /* Opts. */
- (lua_isnoneornil(L, 2) == 1 || lua_istable(L, 2) == 1);
+ /* Merge source. */
+ (source = luaT_check_merge_source(L, 1)) != NULL &&
+ /* Opts. */
+ (lua_isnoneornil(L, 2) == 1 || lua_istable(L, 2) == 1);
if (!ok)
return lbox_merge_source_select_usage(L, NULL);
@@ -1227,7 +1225,7 @@ lbox_merge_source_select(struct lua_State *L)
if (!lua_isnil(L, -1)) {
if ((output_buffer = luaL_checkibuf(L, -1)) == NULL)
return lbox_merge_source_select_usage(L,
- "buffer");
+ "buffer");
}
lua_pop(L, 1);
@@ -1239,7 +1237,7 @@ lbox_merge_source_select(struct lua_State *L)
limit = lua_tointeger(L, -1);
else
return lbox_merge_source_select_usage(L,
- "limit");
+ "limit");
}
lua_pop(L, 1);
}
@@ -1263,11 +1261,11 @@ luaopen_merger(struct lua_State *L)
/* Export C functions to Lua. */
static const struct luaL_Reg meta[] = {
- {"new_buffer_source", lbox_merger_new_buffer_source},
- {"new_table_source", lbox_merger_new_table_source},
- {"new_tuple_source", lbox_merger_new_tuple_source},
- {"new", lbox_merger_new},
- {NULL, NULL}
+ { "new_buffer_source", lbox_merger_new_buffer_source },
+ { "new_table_source", lbox_merger_new_table_source },
+ { "new_tuple_source", lbox_merger_new_tuple_source },
+ { "new", lbox_merger_new },
+ { NULL, NULL }
};
luaL_register_module(L, "merger", meta);
diff --git a/src/box/lua/misc.cc b/src/box/lua/misc.cc
index e356f2d..fd378b5 100644
--- a/src/box/lua/misc.cc
+++ b/src/box/lua/misc.cc
@@ -54,11 +54,11 @@ lbox_encode_tuple_on_gc(lua_State *L, int idx, size_t *p_len)
size_t used = region_used(gc);
struct mpstream stream;
mpstream_init(&stream, gc, region_reserve_cb, region_alloc_cb,
- luamp_error, L);
+ luamp_error, L);
luamp_encode_tuple(L, luaL_msgpack_default, &stream, idx);
mpstream_flush(&stream);
*p_len = region_used(gc) - used;
- return (char *) region_join_xc(gc, *p_len);
+ return (char *)region_join_xc(gc, *p_len);
}
extern "C" void
@@ -84,9 +84,9 @@ port_c_dump_lua(struct port *base, struct lua_State *L, bool is_flat)
extern "C" void
port_msgpack_dump_lua(struct port *base, struct lua_State *L, bool is_flat)
{
- (void) is_flat;
+ (void)is_flat;
assert(is_flat == true);
- struct port_msgpack *port = (struct port_msgpack *) base;
+ struct port_msgpack *port = (struct port_msgpack *)base;
const char *args = port->data;
uint32_t arg_count = mp_decode_array(&args);
@@ -102,9 +102,9 @@ static int
lbox_select(lua_State *L)
{
if (lua_gettop(L) != 6 || !lua_isnumber(L, 1) || !lua_isnumber(L, 2) ||
- !lua_isnumber(L, 3) || !lua_isnumber(L, 4) || !lua_isnumber(L, 5)) {
+ !lua_isnumber(L, 3) || !lua_isnumber(L, 4) || !lua_isnumber(L, 5)) {
return luaL_error(L, "Usage index:select(iterator, offset, "
- "limit, key)");
+ "limit, key)");
}
uint32_t space_id = lua_tonumber(L, 1);
@@ -117,8 +117,8 @@ lbox_select(lua_State *L)
const char *key = lbox_encode_tuple_on_gc(L, 6, &key_len);
struct port port;
- if (box_select(space_id, index_id, iterator, offset, limit,
- key, key + key_len, &port) != 0) {
+ if (box_select(space_id, index_id, iterator, offset, limit, key,
+ key + key_len, &port) != 0) {
return luaT_error(L);
}
@@ -147,7 +147,8 @@ lbox_check_tuple_format(struct lua_State *L, int narg)
struct tuple_format *format =
*(struct tuple_format **)luaL_checkcdata(L, narg, &ctypeid);
if (ctypeid != CTID_STRUCT_TUPLE_FORMAT_PTR) {
- luaL_error(L, "Invalid argument: 'struct tuple_format *' "
+ luaL_error(L,
+ "Invalid argument: 'struct tuple_format *' "
"expected, got %s)",
lua_typename(L, lua_type(L, narg)));
}
@@ -157,7 +158,7 @@ lbox_check_tuple_format(struct lua_State *L, int narg)
static int
lbox_tuple_format_gc(struct lua_State *L)
{
- struct tuple_format *format = lbox_check_tuple_format(L, 1);
+ struct tuple_format *format = lbox_check_tuple_format(L, 1);
tuple_format_unref(format);
return 0;
}
@@ -165,8 +166,8 @@ lbox_tuple_format_gc(struct lua_State *L)
static int
lbox_push_tuple_format(struct lua_State *L, struct tuple_format *format)
{
- struct tuple_format **ptr = (struct tuple_format **)
- luaL_pushcdata(L, CTID_STRUCT_TUPLE_FORMAT_PTR);
+ struct tuple_format **ptr = (struct tuple_format **)luaL_pushcdata(
+ L, CTID_STRUCT_TUPLE_FORMAT_PTR);
*ptr = format;
tuple_format_ref(format);
lua_pushcfunction(L, lbox_tuple_format_gc);
@@ -188,8 +189,8 @@ lbox_tuple_format_new(struct lua_State *L)
size_t size;
struct region *region = &fiber()->gc;
size_t region_svp = region_used(region);
- struct field_def *fields = region_alloc_array(region, typeof(fields[0]),
- count, &size);
+ struct field_def *fields =
+ region_alloc_array(region, typeof(fields[0]), count, &size);
if (fields == NULL) {
diag_set(OutOfMemory, size, "region_alloc_array", "fields");
return luaT_error(L);
@@ -204,7 +205,7 @@ lbox_tuple_format_new(struct lua_State *L)
lua_pushstring(L, "type");
lua_gettable(L, -2);
- if (! lua_isnil(L, -1)) {
+ if (!lua_isnil(L, -1)) {
const char *type_name = lua_tolstring(L, -1, &len);
fields[i].type = field_type_by_name(type_name, len);
assert(fields[i].type != field_type_MAX);
@@ -213,7 +214,7 @@ lbox_tuple_format_new(struct lua_State *L)
lua_pushstring(L, "name");
lua_gettable(L, -2);
- assert(! lua_isnil(L, -1));
+ assert(!lua_isnil(L, -1));
const char *name = lua_tolstring(L, -1, &len);
fields[i].name = (char *)region_alloc(region, len + 1);
if (fields == NULL) {
@@ -251,9 +252,9 @@ void
box_lua_misc_init(struct lua_State *L)
{
static const struct luaL_Reg boxlib_internal[] = {
- {"select", lbox_select},
- {"new_tuple_format", lbox_tuple_format_new},
- {NULL, NULL}
+ { "select", lbox_select },
+ { "new_tuple_format", lbox_tuple_format_new },
+ { NULL, NULL }
};
luaL_register(L, "box.internal", boxlib_internal);
@@ -261,7 +262,7 @@ box_lua_misc_init(struct lua_State *L)
int rc = luaL_cdef(L, "struct tuple_format;");
assert(rc == 0);
- (void) rc;
+ (void)rc;
CTID_STRUCT_TUPLE_FORMAT_PTR = luaL_ctypeid(L, "struct tuple_format *");
assert(CTID_STRUCT_TUPLE_FORMAT_PTR != 0);
}
diff --git a/src/box/lua/net_box.c b/src/box/lua/net_box.c
index 0b6c362..d2aa3ef 100644
--- a/src/box/lua/net_box.c
+++ b/src/box/lua/net_box.c
@@ -55,11 +55,11 @@
static inline size_t
netbox_prepare_request(lua_State *L, struct mpstream *stream, uint32_t r_type)
{
- struct ibuf *ibuf = (struct ibuf *) lua_topointer(L, 1);
+ struct ibuf *ibuf = (struct ibuf *)lua_topointer(L, 1);
uint64_t sync = luaL_touint64(L, 2);
- mpstream_init(stream, ibuf, ibuf_reserve_cb, ibuf_alloc_cb,
- luamp_error, L);
+ mpstream_init(stream, ibuf, ibuf_reserve_cb, ibuf_alloc_cb, luamp_error,
+ L);
/* Remember initial size of ibuf (see netbox_encode_request()) */
size_t used = ibuf_used(ibuf);
@@ -87,7 +87,7 @@ netbox_encode_request(struct mpstream *stream, size_t initial_size)
{
mpstream_flush(stream);
- struct ibuf *ibuf = (struct ibuf *) stream->ctx;
+ struct ibuf *ibuf = (struct ibuf *)stream->ctx;
/*
* Calculation the start position in ibuf by getting current size
@@ -428,8 +428,7 @@ netbox_decode_greeting(lua_State *L)
buf = lua_tolstring(L, 1, &len);
if (buf == NULL || len != IPROTO_GREETING_SIZE ||
- greeting_decode(buf, &greeting) != 0) {
-
+ greeting_decode(buf, &greeting) != 0) {
lua_pushboolean(L, 0);
lua_pushstring(L, "Invalid greeting");
return 2;
@@ -469,8 +468,8 @@ netbox_communicate(lua_State *L)
{
uint32_t fd = lua_tonumber(L, 1);
const int NETBOX_READAHEAD = 16320;
- struct ibuf *send_buf = (struct ibuf *) lua_topointer(L, 2);
- struct ibuf *recv_buf = (struct ibuf *) lua_topointer(L, 3);
+ struct ibuf *send_buf = (struct ibuf *)lua_topointer(L, 2);
+ struct ibuf *recv_buf = (struct ibuf *)lua_topointer(L, 3);
/* limit or boundary */
size_t limit = SIZE_MAX;
@@ -494,20 +493,18 @@ netbox_communicate(lua_State *L)
int revents = COIO_READ;
while (true) {
/* reader serviced first */
-check_limit:
+ check_limit:
if (ibuf_used(recv_buf) >= limit) {
lua_pushnil(L);
lua_pushinteger(L, (lua_Integer)limit);
return 2;
}
const char *p;
- if (boundary != NULL && (p = memmem(
- recv_buf->rpos,
- ibuf_used(recv_buf),
- boundary, boundary_len)) != NULL) {
+ if (boundary != NULL &&
+ (p = memmem(recv_buf->rpos, ibuf_used(recv_buf), boundary,
+ boundary_len)) != NULL) {
lua_pushnil(L);
- lua_pushinteger(L, (lua_Integer)(
- p - recv_buf->rpos));
+ lua_pushinteger(L, (lua_Integer)(p - recv_buf->rpos));
return 2;
}
@@ -515,13 +512,14 @@ check_limit:
void *p = ibuf_reserve(recv_buf, NETBOX_READAHEAD);
if (p == NULL)
luaL_error(L, "out of memory");
- ssize_t rc = recv(
- fd, recv_buf->wpos, ibuf_unused(recv_buf), 0);
+ ssize_t rc = recv(fd, recv_buf->wpos,
+ ibuf_unused(recv_buf), 0);
if (rc == 0) {
lua_pushinteger(L, ER_NO_CONNECTION);
lua_pushstring(L, "Peer closed");
return 2;
- } if (rc > 0) {
+ }
+ if (rc > 0) {
recv_buf->wpos += rc;
goto check_limit;
} else if (errno == EAGAIN || errno == EWOULDBLOCK)
@@ -531,8 +529,8 @@ check_limit:
}
while ((revents & COIO_WRITE) && ibuf_used(send_buf) != 0) {
- ssize_t rc = send(
- fd, send_buf->rpos, ibuf_used(send_buf), 0);
+ ssize_t rc = send(fd, send_buf->rpos,
+ ibuf_used(send_buf), 0);
if (rc >= 0)
send_buf->rpos += rc;
else if (errno == EAGAIN || errno == EWOULDBLOCK)
@@ -542,8 +540,9 @@ check_limit:
}
ev_tstamp deadline = ev_monotonic_now(loop()) + timeout;
- revents = coio_wait(fd, EV_READ | (ibuf_used(send_buf) != 0 ?
- EV_WRITE : 0), timeout);
+ revents = coio_wait(
+ fd, EV_READ | (ibuf_used(send_buf) != 0 ? EV_WRITE : 0),
+ timeout);
luaL_testcancel(L);
timeout = deadline - ev_monotonic_now(loop());
timeout = MAX(0.0, timeout);
@@ -563,8 +562,8 @@ static int
netbox_encode_execute(lua_State *L)
{
if (lua_gettop(L) < 5)
- return luaL_error(L, "Usage: netbox.encode_execute(ibuf, "\
- "sync, query, parameters, options)");
+ return luaL_error(L, "Usage: netbox.encode_execute(ibuf, "
+ "sync, query, parameters, options)");
struct mpstream stream;
size_t svp = netbox_prepare_request(L, &stream, IPROTO_EXECUTE);
@@ -595,7 +594,7 @@ static int
netbox_encode_prepare(lua_State *L)
{
if (lua_gettop(L) < 3)
- return luaL_error(L, "Usage: netbox.encode_prepare(ibuf, "\
+ return luaL_error(L, "Usage: netbox.encode_prepare(ibuf, "
"sync, query)");
struct mpstream stream;
size_t svp = netbox_prepare_request(L, &stream, IPROTO_PREPARE);
@@ -631,8 +630,7 @@ netbox_decode_data(struct lua_State *L, const char **data,
for (uint32_t j = 0; j < count; ++j) {
const char *begin = *data;
mp_next(data);
- struct tuple *tuple =
- box_tuple_new(format, begin, *data);
+ struct tuple *tuple = box_tuple_new(format, begin, *data);
if (tuple == NULL)
luaT_error(L);
luaT_pushtuple(L, tuple);
@@ -661,10 +659,10 @@ netbox_decode_select(struct lua_State *L)
uint32_t map_size = mp_decode_map(&data);
/* Until 2.0 body has no keys except DATA. */
assert(map_size == 1);
- (void) map_size;
+ (void)map_size;
uint32_t key = mp_decode_uint(&data);
assert(key == IPROTO_DATA);
- (void) key;
+ (void)key;
netbox_decode_data(L, &data, format);
*(const char **)luaL_pushcdata(L, ctypeid) = data;
return 2;
@@ -729,7 +727,7 @@ netbox_decode_metadata(struct lua_State *L, const char **data)
assert(map_size >= 2 && map_size <= 6);
uint32_t key = mp_decode_uint(data);
assert(key == IPROTO_FIELD_NAME);
- (void) key;
+ (void)key;
lua_createtable(L, 0, map_size);
uint32_t name_len, type_len;
const char *str = mp_decode_str(data, &name_len);
@@ -796,7 +794,7 @@ netbox_decode_execute(struct lua_State *L)
int rows_index = 0, meta_index = 0, info_index = 0;
for (uint32_t i = 0; i < map_size; ++i) {
uint32_t key = mp_decode_uint(&data);
- switch(key) {
+ switch (key) {
case IPROTO_DATA:
netbox_decode_data(L, &data, tuple_format_runtime);
rows_index = i - map_size;
@@ -840,7 +838,7 @@ netbox_decode_prepare(struct lua_State *L)
uint32_t stmt_id = 0;
for (uint32_t i = 0; i < map_size; ++i) {
uint32_t key = mp_decode_uint(&data);
- switch(key) {
+ switch (key) {
case IPROTO_STMT_ID: {
stmt_id = mp_decode_uint(&data);
luaL_pushuint64(L, stmt_id);
@@ -863,7 +861,8 @@ netbox_decode_prepare(struct lua_State *L)
luaL_pushuint64(L, bind_count);
bind_count_idx = i - map_size;
break;
- }}
+ }
+ }
}
/* These fields must be present in response. */
assert(stmt_id_idx * bind_meta_idx * bind_count_idx != 0);
@@ -888,25 +887,25 @@ int
luaopen_net_box(struct lua_State *L)
{
static const luaL_Reg net_box_lib[] = {
- { "encode_ping", netbox_encode_ping },
+ { "encode_ping", netbox_encode_ping },
{ "encode_call_16", netbox_encode_call_16 },
- { "encode_call", netbox_encode_call },
- { "encode_eval", netbox_encode_eval },
- { "encode_select", netbox_encode_select },
- { "encode_insert", netbox_encode_insert },
+ { "encode_call", netbox_encode_call },
+ { "encode_eval", netbox_encode_eval },
+ { "encode_select", netbox_encode_select },
+ { "encode_insert", netbox_encode_insert },
{ "encode_replace", netbox_encode_replace },
- { "encode_delete", netbox_encode_delete },
- { "encode_update", netbox_encode_update },
- { "encode_upsert", netbox_encode_upsert },
- { "encode_execute", netbox_encode_execute},
- { "encode_prepare", netbox_encode_prepare},
- { "encode_auth", netbox_encode_auth },
- { "decode_greeting",netbox_decode_greeting },
- { "communicate", netbox_communicate },
- { "decode_select", netbox_decode_select },
+ { "encode_delete", netbox_encode_delete },
+ { "encode_update", netbox_encode_update },
+ { "encode_upsert", netbox_encode_upsert },
+ { "encode_execute", netbox_encode_execute },
+ { "encode_prepare", netbox_encode_prepare },
+ { "encode_auth", netbox_encode_auth },
+ { "decode_greeting", netbox_decode_greeting },
+ { "communicate", netbox_communicate },
+ { "decode_select", netbox_decode_select },
{ "decode_execute", netbox_decode_execute },
{ "decode_prepare", netbox_decode_prepare },
- { NULL, NULL}
+ { NULL, NULL }
};
/* luaL_register_module polutes _G */
lua_newtable(L);
diff --git a/src/box/lua/sequence.c b/src/box/lua/sequence.c
index bf0714c..e33904f 100644
--- a/src/box/lua/sequence.c
+++ b/src/box/lua/sequence.c
@@ -173,16 +173,16 @@ void
box_lua_sequence_init(struct lua_State *L)
{
static const struct luaL_Reg sequence_internal_lib[] = {
- {"next", lbox_sequence_next},
- {"set", lbox_sequence_set},
- {"reset", lbox_sequence_reset},
- {NULL, NULL}
+ { "next", lbox_sequence_next },
+ { "set", lbox_sequence_set },
+ { "reset", lbox_sequence_reset },
+ { NULL, NULL }
};
luaL_register(L, "box.internal.sequence", sequence_internal_lib);
lua_pop(L, 1);
static struct trigger on_alter_sequence_in_lua;
- trigger_create(&on_alter_sequence_in_lua,
- lbox_sequence_new_or_delete, L, NULL);
+ trigger_create(&on_alter_sequence_in_lua, lbox_sequence_new_or_delete,
+ L, NULL);
trigger_add(&on_alter_sequence, &on_alter_sequence_in_lua);
}
diff --git a/src/box/lua/serialize_lua.c b/src/box/lua/serialize_lua.c
index caa08a6..32dcf47 100644
--- a/src/box/lua/serialize_lua.c
+++ b/src/box/lua/serialize_lua.c
@@ -45,36 +45,36 @@
#include "serialize_lua.h"
#if 0
-# define SERIALIZER_TRACE
+#define SERIALIZER_TRACE
#endif
/* Serializer for Lua output mode */
static struct luaL_serializer *serializer_lua;
enum {
- NODE_NONE_BIT = 0,
- NODE_ROOT_BIT = 1,
- NODE_RAW_BIT = 2,
- NODE_LVALUE_BIT = 3,
- NODE_RVALUE_BIT = 4,
- NODE_MAP_KEY_BIT = 5,
- NODE_MAP_VALUE_BIT = 6,
- NODE_EMBRACE_BIT = 7,
- NODE_QUOTE_BIT = 8,
+ NODE_NONE_BIT = 0,
+ NODE_ROOT_BIT = 1,
+ NODE_RAW_BIT = 2,
+ NODE_LVALUE_BIT = 3,
+ NODE_RVALUE_BIT = 4,
+ NODE_MAP_KEY_BIT = 5,
+ NODE_MAP_VALUE_BIT = 6,
+ NODE_EMBRACE_BIT = 7,
+ NODE_QUOTE_BIT = 8,
NODE_MAX
};
enum {
- NODE_NONE = (1u << NODE_NONE_BIT),
- NODE_ROOT = (1u << NODE_ROOT_BIT),
- NODE_RAW = (1u << NODE_RAW_BIT),
- NODE_LVALUE = (1u << NODE_LVALUE_BIT),
- NODE_RVALUE = (1u << NODE_RVALUE_BIT),
- NODE_MAP_KEY = (1u << NODE_MAP_KEY_BIT),
- NODE_MAP_VALUE = (1u << NODE_MAP_VALUE_BIT),
- NODE_EMBRACE = (1u << NODE_EMBRACE_BIT),
- NODE_QUOTE = (1u << NODE_QUOTE_BIT),
+ NODE_NONE = (1u << NODE_NONE_BIT),
+ NODE_ROOT = (1u << NODE_ROOT_BIT),
+ NODE_RAW = (1u << NODE_RAW_BIT),
+ NODE_LVALUE = (1u << NODE_LVALUE_BIT),
+ NODE_RVALUE = (1u << NODE_RVALUE_BIT),
+ NODE_MAP_KEY = (1u << NODE_MAP_KEY_BIT),
+ NODE_MAP_VALUE = (1u << NODE_MAP_VALUE_BIT),
+ NODE_EMBRACE = (1u << NODE_EMBRACE_BIT),
+ NODE_QUOTE = (1u << NODE_QUOTE_BIT),
};
struct node {
@@ -136,18 +136,13 @@ struct lua_dumper {
#ifdef SERIALIZER_TRACE
-#define __gen_mp_name(__v) [__v] = # __v
+#define __gen_mp_name(__v) [__v] = #__v
static const char *mp_type_names[] = {
- __gen_mp_name(MP_NIL),
- __gen_mp_name(MP_UINT),
- __gen_mp_name(MP_INT),
- __gen_mp_name(MP_STR),
- __gen_mp_name(MP_BIN),
- __gen_mp_name(MP_ARRAY),
- __gen_mp_name(MP_MAP),
- __gen_mp_name(MP_BOOL),
- __gen_mp_name(MP_FLOAT),
- __gen_mp_name(MP_DOUBLE),
+ __gen_mp_name(MP_NIL), __gen_mp_name(MP_UINT),
+ __gen_mp_name(MP_INT), __gen_mp_name(MP_STR),
+ __gen_mp_name(MP_BIN), __gen_mp_name(MP_ARRAY),
+ __gen_mp_name(MP_MAP), __gen_mp_name(MP_BOOL),
+ __gen_mp_name(MP_FLOAT), __gen_mp_name(MP_DOUBLE),
__gen_mp_name(MP_EXT),
};
@@ -158,16 +153,12 @@ static const char *mp_ext_type_names[] = {
};
#undef __gen_mp_name
-#define __gen_nd_name(__v) [__v ##_BIT] = # __v
+#define __gen_nd_name(__v) [__v##_BIT] = #__v
static const char *nd_type_names[] = {
- __gen_nd_name(NODE_NONE),
- __gen_nd_name(NODE_ROOT),
- __gen_nd_name(NODE_RAW),
- __gen_nd_name(NODE_LVALUE),
- __gen_nd_name(NODE_RVALUE),
- __gen_nd_name(NODE_MAP_KEY),
- __gen_nd_name(NODE_MAP_VALUE),
- __gen_nd_name(NODE_EMBRACE),
+ __gen_nd_name(NODE_NONE), __gen_nd_name(NODE_ROOT),
+ __gen_nd_name(NODE_RAW), __gen_nd_name(NODE_LVALUE),
+ __gen_nd_name(NODE_RVALUE), __gen_nd_name(NODE_MAP_KEY),
+ __gen_nd_name(NODE_MAP_VALUE), __gen_nd_name(NODE_EMBRACE),
__gen_nd_name(NODE_QUOTE),
};
#undef __gen_nd_name
@@ -204,8 +195,8 @@ static void
trace_node(struct lua_dumper *d)
{
int ltype = lua_type(d->L, -1);
- say_info("serializer-trace: node : lua type %d -> %s",
- ltype, lua_typename(d->L, ltype));
+ say_info("serializer-trace: node : lua type %d -> %s", ltype,
+ lua_typename(d->L, ltype));
if (d->err != 0)
return;
@@ -223,8 +214,8 @@ trace_node(struct lua_dumper *d)
snprintf(mp_type, sizeof(mp_type), "%s/%s",
mp_type_names[field.type],
field.ext_type < max_ext ?
- mp_ext_type_names[field.ext_type] :
- "UNKNOWN");
+ mp_ext_type_names[field.ext_type] :
+ "UNKNOWN");
} else {
type_str = (char *)mp_type_names[field.type];
}
@@ -235,8 +226,8 @@ trace_node(struct lua_dumper *d)
memset(&field, 0, sizeof(field));
luaL_checkfield(d->L, d->cfg, top, &field);
- say_info("serializer-trace: node :\tfield type %s (%d)",
- type_str, field.type);
+ say_info("serializer-trace: node :\tfield type %s (%d)", type_str,
+ field.type);
}
static char *
@@ -245,8 +236,8 @@ trace_string(const char *src, size_t len)
static char buf[128];
size_t pos = 0;
- if (len > sizeof(buf)-1)
- len = sizeof(buf)-1;
+ if (len > sizeof(buf) - 1)
+ len = sizeof(buf) - 1;
while (pos < len) {
if (src[pos] == '\n') {
@@ -262,31 +253,27 @@ trace_string(const char *src, size_t len)
}
static void
-trace_emit(struct lua_dumper *d, int nd_mask, int indent,
- const char *str, size_t len)
+trace_emit(struct lua_dumper *d, int nd_mask, int indent, const char *str,
+ size_t len)
{
if (d->suffix_len) {
say_info("serializer-trace: emit-sfx: \"%s\"",
- trace_string(d->suffix_buf,
- d->suffix_len));
+ trace_string(d->suffix_buf, d->suffix_len));
}
- static_assert(NODE_MAX < sizeof(int) * 8,
- "NODE_MAX is too big");
+ static_assert(NODE_MAX < sizeof(int) * 8, "NODE_MAX is too big");
char *names = trace_nd_mask_str(nd_mask);
say_info("serializer-trace: emit : type %s (0x%x) "
"indent %d val \"%s\" len %zu",
- names, nd_mask, indent,
- trace_string(str, len), len);
+ names, nd_mask, indent, trace_string(str, len), len);
}
static void
trace_anchor(const char *s, bool alias)
{
- say_info("serializer-trace: anchor : alias %d name %s",
- alias, s);
+ say_info("serializer-trace: anchor : alias %d name %s", alias, s);
}
#else /* SERIALIZER_TRACE */
@@ -298,8 +285,8 @@ trace_node(struct lua_dumper *d)
}
static void
-trace_emit(struct lua_dumper *d, int nd_mask, int indent,
- const char *str, size_t len)
+trace_emit(struct lua_dumper *d, int nd_mask, int indent, const char *str,
+ size_t len)
{
(void)d;
(void)nd_mask;
@@ -318,20 +305,18 @@ trace_anchor(const char *s, bool alias)
#endif /* SERIALIZER_TRACE */
static const char *lua_keywords[] = {
- "and", "break", "do", "else",
- "elseif", "end", "false", "for",
- "function", "if", "in", "local",
- "nil", "not", "or", "repeat",
- "return", "then", "true", "until",
- "while", "and",
+ "and", "break", "do", "else", "elseif", "end",
+ "false", "for", "function", "if", "in", "local",
+ "nil", "not", "or", "repeat", "return", "then",
+ "true", "until", "while", "and",
};
static int
dump_node(struct lua_dumper *d, struct node *nd, int indent);
static int
-emit_node(struct lua_dumper *d, struct node *nd, int indent,
- const char *str, size_t len);
+emit_node(struct lua_dumper *d, struct node *nd, int indent, const char *str,
+ size_t len);
/**
* Generate anchor numbers for self references.
@@ -405,12 +390,11 @@ suffix_flush(struct lua_dumper *d)
static int
gen_indent(struct lua_dumper *d, int indent)
{
- static_assert(sizeof(d->indent_buf) > 0,
- "indent buffer is too small");
+ static_assert(sizeof(d->indent_buf) > 0, "indent buffer is too small");
if (indent > 0 && d->opts->block_mode && !d->noindent) {
- snprintf(d->indent_buf, sizeof(d->indent_buf),
- "%*s", indent, "");
+ snprintf(d->indent_buf, sizeof(d->indent_buf), "%*s", indent,
+ "");
size_t len = strlen(d->indent_buf);
d->indent_buf[len] = '\0';
return len;
@@ -425,12 +409,12 @@ emit_hex_char(struct lua_dumper *d, unsigned char c)
luaL_addchar(&d->luabuf, '\\');
luaL_addchar(&d->luabuf, 'x');
-#define __emit_hex(v) \
- do { \
- if (v <= 9) \
- luaL_addchar(&d->luabuf, '0' + v); \
- else \
- luaL_addchar(&d->luabuf, v - 10 + 'a'); \
+#define __emit_hex(v) \
+ do { \
+ if (v <= 9) \
+ luaL_addchar(&d->luabuf, '0' + v); \
+ else \
+ luaL_addchar(&d->luabuf, v - 10 + 'a'); \
} while (0)
__emit_hex((c >> 4));
@@ -477,9 +461,8 @@ emit_string(struct lua_dumper *d, const char *str, size_t len)
luaL_addchar(&d->luabuf, '\\');
luaL_addchar(&d->luabuf, 't');
} else if (str[i] == '\xef') {
- if (i < len-1 && i < len-2 &&
- str[i+1] == '\xbb' &&
- str[i+2] == '\xbf') {
+ if (i < len - 1 && i < len - 2 &&
+ str[i + 1] == '\xbb' && str[i + 2] == '\xbf') {
emit_hex_char(d, 0xef);
emit_hex_char(d, 0xbb);
emit_hex_char(d, 0xbf);
@@ -498,8 +481,8 @@ emit_string(struct lua_dumper *d, const char *str, size_t len)
* Emit value into output buffer.
*/
static void
-emit_value(struct lua_dumper *d, struct node *nd,
- int indent, const char *str, size_t len)
+emit_value(struct lua_dumper *d, struct node *nd, int indent, const char *str,
+ size_t len)
{
trace_emit(d, nd->mask, indent, str, len);
@@ -511,8 +494,7 @@ emit_value(struct lua_dumper *d, struct node *nd,
*/
suffix_flush(d);
- luaL_addlstring(&d->luabuf, d->indent_buf,
- gen_indent(d, indent));
+ luaL_addlstring(&d->luabuf, d->indent_buf, gen_indent(d, indent));
if (nd->mask & NODE_EMBRACE)
luaL_addlstring(&d->luabuf, "[", 1);
@@ -535,8 +517,7 @@ emit_value(struct lua_dumper *d, struct node *nd,
* Emit a raw string into output.
*/
static void
-emit_raw_value(struct lua_dumper *d, int indent,
- const char *str, size_t len)
+emit_raw_value(struct lua_dumper *d, int indent, const char *str, size_t len)
{
struct node node = {
.mask = NODE_RAW,
@@ -650,16 +631,16 @@ dump_table(struct lua_dumper *d, struct node *nd, int indent)
while (lua_next(d->L, -2)) {
lua_pushvalue(d->L, -2);
struct node node_key = {
- .prev = nd,
- .mask = NODE_LVALUE | NODE_MAP_KEY,
- .index = index++,
+ .prev = nd,
+ .mask = NODE_LVALUE | NODE_MAP_KEY,
+ .index = index++,
};
dump_node(d, &node_key, indent);
lua_pop(d->L, 1);
struct node node_val = {
- .key = &node_key,
- .mask = NODE_RVALUE | NODE_MAP_VALUE,
+ .key = &node_key,
+ .mask = NODE_RVALUE | NODE_MAP_VALUE,
};
dump_node(d, &node_val, indent);
lua_pop(d->L, 1);
@@ -708,8 +689,8 @@ decorate_key(struct node *nd, const char *str, size_t len)
}
static int
-emit_node(struct lua_dumper *d, struct node *nd, int indent,
- const char *str, size_t len)
+emit_node(struct lua_dumper *d, struct node *nd, int indent, const char *str,
+ size_t len)
{
struct luaL_field *field = &nd->field;
@@ -724,8 +705,7 @@ emit_node(struct lua_dumper *d, struct node *nd, int indent,
* the current position in the table we
* can simply skip it and print value only.
*/
- if (nd->field.type == MP_INT ||
- nd->field.type == MP_UINT) {
+ if (nd->field.type == MP_INT || nd->field.type == MP_UINT) {
if (nd->index == (int)field->ival) {
d->noindent = false;
return 0;
@@ -837,14 +817,12 @@ dump_node(struct lua_dumper *d, struct node *nd, int indent)
}
break;
case MP_FLOAT:
- fpconv_g_fmt(buf, field->fval,
- d->cfg->encode_number_precision);
+ fpconv_g_fmt(buf, field->fval, d->cfg->encode_number_precision);
len = strlen(buf);
str = buf;
break;
case MP_DOUBLE:
- fpconv_g_fmt(buf, field->dval,
- d->cfg->encode_number_precision);
+ fpconv_g_fmt(buf, field->dval, d->cfg->encode_number_precision);
len = strlen(buf);
str = buf;
break;
@@ -872,8 +850,7 @@ dump_node(struct lua_dumper *d, struct node *nd, int indent)
default:
d->err = EINVAL;
snprintf(d->err_msg, sizeof(d->err_msg),
- "serializer: Unknown field %d type",
- field->type);
+ "serializer: Unknown field %d type", field->type);
len = strlen(d->err_msg);
return -1;
}
@@ -966,10 +943,10 @@ lua_encode(lua_State *L, struct luaL_serializer *serializer,
lua_dumper_opts_t *opts)
{
struct lua_dumper dumper = {
- .L = L,
- .cfg = serializer,
- .outputL= luaT_newthread(L),
- .opts = opts,
+ .L = L,
+ .cfg = serializer,
+ .outputL = luaT_newthread(L),
+ .opts = opts,
};
if (!dumper.outputL)
@@ -1045,11 +1022,11 @@ lua_serializer_init(struct lua_State *L)
};
serializer_lua = luaL_newserializer(L, NULL, lualib);
- serializer_lua->has_compact = 1;
- serializer_lua->encode_invalid_numbers = 1;
- serializer_lua->encode_load_metatables = 1;
- serializer_lua->encode_use_tostring = 1;
- serializer_lua->encode_invalid_as_nil = 1;
+ serializer_lua->has_compact = 1;
+ serializer_lua->encode_invalid_numbers = 1;
+ serializer_lua->encode_load_metatables = 1;
+ serializer_lua->encode_use_tostring = 1;
+ serializer_lua->encode_invalid_as_nil = 1;
/*
* Keep a reference to this module so it
diff --git a/src/box/lua/session.c b/src/box/lua/session.c
index 0a20aaa..2cc23ee 100644
--- a/src/box/lua/session.c
+++ b/src/box/lua/session.c
@@ -247,7 +247,6 @@ lbox_session_fd(struct lua_State *L)
return 1;
}
-
/**
* Pretty print peer name.
*/
@@ -286,15 +285,15 @@ lbox_session_peer(struct lua_State *L)
static int
lbox_push_on_connect_event(struct lua_State *L, void *event)
{
- (void) L;
- (void) event;
+ (void)L;
+ (void)event;
return 0;
}
static int
lbox_push_on_auth_event(struct lua_State *L, void *event)
{
- struct on_auth_trigger_ctx *ctx = (struct on_auth_trigger_ctx *) event;
+ struct on_auth_trigger_ctx *ctx = (struct on_auth_trigger_ctx *)event;
lua_pushstring(L, ctx->username);
lua_pushboolean(L, ctx->is_authenticated);
return 2;
@@ -328,7 +327,7 @@ lbox_session_run_on_disconnect(struct lua_State *L)
{
struct session *session = current_session();
session_run_on_disconnect_triggers(session);
- (void) L;
+ (void)L;
return 0;
}
@@ -360,7 +359,7 @@ lbox_session_run_on_auth(struct lua_State *L)
static int
lbox_push_on_access_denied_event(struct lua_State *L, void *event)
{
- struct on_access_denied_ctx *ctx = (struct on_access_denied_ctx *) event;
+ struct on_access_denied_ctx *ctx = (struct on_access_denied_ctx *)event;
lua_pushstring(L, ctx->access_type);
lua_pushstring(L, ctx->object_type);
lua_pushstring(L, ctx->object_name);
@@ -427,8 +426,10 @@ lbox_session_setting_get(struct lua_State *L)
const char *setting_name = lua_tostring(L, -1);
int sid = session_setting_find(setting_name);
if (sid < 0) {
- diag_set(ClientError, ER_PROC_LUA, tt_sprintf("Session "\
- "setting %s doesn't exist", setting_name));
+ diag_set(ClientError, ER_PROC_LUA,
+ tt_sprintf("Session "
+ "setting %s doesn't exist",
+ setting_name));
return luaT_error(L);
}
return lbox_session_setting_get_by_id(L, sid);
@@ -450,7 +451,7 @@ lbox_session_setting_set(struct lua_State *L)
case LUA_TBOOLEAN: {
bool value = lua_toboolean(L, -1);
size_t size = mp_sizeof_bool(value);
- char *mp_value = (char *) static_alloc(size);
+ char *mp_value = (char *)static_alloc(size);
mp_encode_bool(mp_value, value);
if (setting->set(sid, mp_value) != 0)
return luaT_error(L);
@@ -460,10 +461,9 @@ lbox_session_setting_set(struct lua_State *L)
const char *str = lua_tostring(L, -1);
size_t len = strlen(str);
uint32_t size = mp_sizeof_str(len);
- char *mp_value = (char *) static_alloc(size);
+ char *mp_value = (char *)static_alloc(size);
if (mp_value == NULL) {
- diag_set(OutOfMemory, size, "static_alloc",
- "mp_value");
+ diag_set(OutOfMemory, size, "static_alloc", "mp_value");
return luaT_error(L);
}
mp_encode_str(mp_value, str, len);
@@ -544,33 +544,33 @@ void
box_lua_session_init(struct lua_State *L)
{
static const struct luaL_Reg session_internal_lib[] = {
- {"create", lbox_session_create},
- {"run_on_connect", lbox_session_run_on_connect},
- {"run_on_disconnect", lbox_session_run_on_disconnect},
- {"run_on_auth", lbox_session_run_on_auth},
- {NULL, NULL}
+ { "create", lbox_session_create },
+ { "run_on_connect", lbox_session_run_on_connect },
+ { "run_on_disconnect", lbox_session_run_on_disconnect },
+ { "run_on_auth", lbox_session_run_on_auth },
+ { NULL, NULL }
};
luaL_register(L, "box.internal.session", session_internal_lib);
lua_pop(L, 1);
static const struct luaL_Reg sessionlib[] = {
- {"id", lbox_session_id},
- {"type", lbox_session_type},
- {"sync", lbox_session_sync},
- {"uid", lbox_session_uid},
- {"euid", lbox_session_euid},
- {"user", lbox_session_user},
- {"effective_user", lbox_session_effective_user},
- {"su", lbox_session_su},
- {"fd", lbox_session_fd},
- {"exists", lbox_session_exists},
- {"peer", lbox_session_peer},
- {"on_connect", lbox_session_on_connect},
- {"on_disconnect", lbox_session_on_disconnect},
- {"on_auth", lbox_session_on_auth},
- {"on_access_denied", lbox_session_on_access_denied},
- {"push", lbox_session_push},
- {NULL, NULL}
+ { "id", lbox_session_id },
+ { "type", lbox_session_type },
+ { "sync", lbox_session_sync },
+ { "uid", lbox_session_uid },
+ { "euid", lbox_session_euid },
+ { "user", lbox_session_user },
+ { "effective_user", lbox_session_effective_user },
+ { "su", lbox_session_su },
+ { "fd", lbox_session_fd },
+ { "exists", lbox_session_exists },
+ { "peer", lbox_session_peer },
+ { "on_connect", lbox_session_on_connect },
+ { "on_disconnect", lbox_session_on_disconnect },
+ { "on_auth", lbox_session_on_auth },
+ { "on_access_denied", lbox_session_on_access_denied },
+ { "push", lbox_session_push },
+ { NULL, NULL }
};
luaL_register_module(L, sessionlib_name, sessionlib);
lbox_session_settings_init(L);
diff --git a/src/box/lua/slab.c b/src/box/lua/slab.c
index 9f5e7e9..e6e0202 100644
--- a/src/box/lua/slab.c
+++ b/src/box/lua/slab.c
@@ -47,8 +47,8 @@
static int
small_stats_noop_cb(const struct mempool_stats *stats, void *cb_ctx)
{
- (void) stats;
- (void) cb_ctx;
+ (void)stats;
+ (void)cb_ctx;
return 0;
}
@@ -59,7 +59,7 @@ small_stats_lua_cb(const struct mempool_stats *stats, void *cb_ctx)
if (stats->slabcount == 0)
return 0;
- struct lua_State *L = (struct lua_State *) cb_ctx;
+ struct lua_State *L = (struct lua_State *)cb_ctx;
/*
* Create a Lua table for every slab class. A class is
@@ -142,8 +142,7 @@ lbox_slab_info(struct lua_State *L)
double ratio;
char ratio_buf[32];
- ratio = 100 * ((double) totals.used
- / ((double) totals.total + 0.0001));
+ ratio = 100 * ((double)totals.used / ((double)totals.total + 0.0001));
snprintf(ratio_buf, sizeof(ratio_buf), "%0.2lf%%", ratio);
/** How much address space has been already touched */
@@ -190,8 +189,8 @@ lbox_slab_info(struct lua_State *L)
luaL_pushuint64(L, totals.used + index_stats.totals.used);
lua_settable(L, -3);
- ratio = 100 * ((double) (totals.used + index_stats.totals.used)
- / (double) arena_size);
+ ratio = 100 * ((double)(totals.used + index_stats.totals.used) /
+ (double)arena_size);
snprintf(ratio_buf, sizeof(ratio_buf), "%0.1lf%%", ratio);
lua_pushstring(L, "arena_used_ratio");
@@ -220,8 +219,8 @@ lbox_slab_info(struct lua_State *L)
* factor, it's the quota that give you OOM error in the
* end of the day.
*/
- ratio = 100 * ((double) quota_used(&memtx->quota) /
- ((double) quota_total(&memtx->quota) + 0.0001));
+ ratio = 100 * ((double)quota_used(&memtx->quota) /
+ ((double)quota_total(&memtx->quota) + 0.0001));
snprintf(ratio_buf, sizeof(ratio_buf), "%0.2lf%%", ratio);
lua_pushstring(L, "quota_used_ratio");
diff --git a/src/box/lua/slab.h b/src/box/lua/slab.h
index fd4ef88..9d73f8a 100644
--- a/src/box/lua/slab.h
+++ b/src/box/lua/slab.h
@@ -35,7 +35,8 @@ extern "C" {
#endif /* defined(__cplusplus) */
struct lua_State;
-void box_lua_slab_init(struct lua_State *L);
+void
+box_lua_slab_init(struct lua_State *L);
#if defined(__cplusplus)
} /* extern "C" */
diff --git a/src/box/lua/space.cc b/src/box/lua/space.cc
index 177c588..d074bc9 100644
--- a/src/box/lua/space.cc
+++ b/src/box/lua/space.cc
@@ -37,9 +37,9 @@
#include "lua/trigger.h"
extern "C" {
- #include <lua.h>
- #include <lauxlib.h>
- #include <lualib.h>
+#include <lua.h>
+#include <lauxlib.h>
+#include <lualib.h>
} /* extern "C" */
#include "box/func.h"
@@ -61,7 +61,7 @@ extern "C" {
static int
lbox_push_txn_stmt(struct lua_State *L, void *event)
{
- struct txn_stmt *stmt = txn_current_stmt((struct txn *) event);
+ struct txn_stmt *stmt = txn_current_stmt((struct txn *)event);
if (stmt->old_tuple) {
luaT_pushtuple(L, stmt->old_tuple);
@@ -84,7 +84,7 @@ lbox_push_txn_stmt(struct lua_State *L, void *event)
static int
lbox_pop_txn_stmt(struct lua_State *L, int nret, void *event)
{
- struct txn_stmt *stmt = txn_current_stmt((struct txn *) event);
+ struct txn_stmt *stmt = txn_current_stmt((struct txn *)event);
if (nret < 1) {
/* No return value - nothing to do. */
@@ -117,16 +117,17 @@ lbox_space_on_replace(struct lua_State *L)
int top = lua_gettop(L);
if (top < 1 || !lua_istable(L, 1)) {
- luaL_error(L,
- "usage: space:on_replace(function | nil, [function | nil])");
+ luaL_error(
+ L,
+ "usage: space:on_replace(function | nil, [function | nil])");
}
lua_getfield(L, 1, "id"); /* Get space id. */
uint32_t id = lua_tonumber(L, lua_gettop(L));
struct space *space = space_cache_find_xc(id);
lua_pop(L, 1);
- return lbox_trigger_reset(L, 3, &space->on_replace,
- lbox_push_txn_stmt, NULL);
+ return lbox_trigger_reset(L, 3, &space->on_replace, lbox_push_txn_stmt,
+ NULL);
}
/**
@@ -138,8 +139,9 @@ lbox_space_before_replace(struct lua_State *L)
int top = lua_gettop(L);
if (top < 1 || !lua_istable(L, 1)) {
- luaL_error(L,
- "usage: space:before_replace(function | nil, [function | nil])");
+ luaL_error(
+ L,
+ "usage: space:before_replace(function | nil, [function | nil])");
}
lua_getfield(L, 1, "id"); /* Get space id. */
uint32_t id = lua_tonumber(L, lua_gettop(L));
@@ -175,8 +177,8 @@ lbox_push_ck_constraint(struct lua_State *L, struct space *space, int i)
* Remove ck_constraint only if it was
* deleted.
*/
- if (space_ck_constraint_by_name(space, name,
- (uint32_t)name_len) == NULL) {
+ if (space_ck_constraint_by_name(
+ space, name, (uint32_t)name_len) == NULL) {
lua_pushlstring(L, name, name_len);
lua_pushnil(L);
lua_settable(L, -5);
@@ -262,16 +264,15 @@ lbox_fillspace(struct lua_State *L, struct space *space, int i)
lua_pushboolean(L, space_index(space, 0) != 0);
lua_settable(L, i);
+ /* space:on_replace */
+ lua_pushstring(L, "on_replace");
+ lua_pushcfunction(L, lbox_space_on_replace);
+ lua_settable(L, i);
- /* space:on_replace */
- lua_pushstring(L, "on_replace");
- lua_pushcfunction(L, lbox_space_on_replace);
- lua_settable(L, i);
-
- /* space:before_replace */
- lua_pushstring(L, "before_replace");
- lua_pushcfunction(L, lbox_space_before_replace);
- lua_settable(L, i);
+ /* space:before_replace */
+ lua_pushstring(L, "before_replace");
+ lua_pushcfunction(L, lbox_space_before_replace);
+ lua_settable(L, i);
lua_getfield(L, i, "index");
if (lua_isnil(L, -1)) {
@@ -279,13 +280,13 @@ lbox_fillspace(struct lua_State *L, struct space *space, int i)
/* space.index */
lua_pushstring(L, "index");
lua_newtable(L);
- lua_settable(L, i); /* push space.index */
+ lua_settable(L, i); /* push space.index */
lua_getfield(L, i, "index");
} else {
lua_pushnil(L);
while (lua_next(L, -2) != 0) {
if (lua_isnumber(L, -2)) {
- uint32_t iid = (uint32_t) lua_tonumber(L, -2);
+ uint32_t iid = (uint32_t)lua_tonumber(L, -2);
/*
* Remove index only if it was deleted.
* If an existing index was
@@ -334,7 +335,7 @@ lbox_fillspace(struct lua_State *L, struct space *space, int i)
lua_newtable(L);
lua_settable(L, -3);
lua_rawgeti(L, -1, index_def->iid);
- assert(! lua_isnil(L, -1));
+ assert(!lua_isnil(L, -1));
}
if (index_def->type == HASH || index_def->type == TREE) {
@@ -399,7 +400,7 @@ lbox_fillspace(struct lua_State *L, struct space *space, int i)
lua_pushstring(L, "sequence_fieldno");
if (k == 0 && space->sequence != NULL)
lua_pushnumber(L, space->sequence_fieldno +
- TUPLE_INDEX_BASE);
+ TUPLE_INDEX_BASE);
else
lua_pushnil(L);
lua_rawset(L, -3);
@@ -449,9 +450,9 @@ lbox_fillspace(struct lua_State *L, struct space *space, int i)
lua_pushstring(L, "bless");
lua_gettable(L, -2);
- lua_pushvalue(L, i); /* space */
+ lua_pushvalue(L, i); /* space */
lua_call(L, 1, 0);
- lua_pop(L, 3); /* cleanup stack - box, schema, space */
+ lua_pop(L, 3); /* cleanup stack - box, schema, space */
}
/** Export a space to Lua */
@@ -511,8 +512,8 @@ box_lua_space_delete(struct lua_State *L, uint32_t id)
static int
box_lua_space_new_or_delete(struct trigger *trigger, void *event)
{
- struct lua_State *L = (struct lua_State *) trigger->data;
- struct space *space = (struct space *) event;
+ struct lua_State *L = (struct lua_State *)trigger->data;
+ struct space *space = (struct space *)event;
if (space_by_id(space->def->id) != NULL) {
box_lua_space_new(L, space);
@@ -522,9 +523,9 @@ box_lua_space_new_or_delete(struct trigger *trigger, void *event)
return 0;
}
-static struct trigger on_alter_space_in_lua = {
- RLIST_LINK_INITIALIZER, box_lua_space_new_or_delete, NULL, NULL
-};
+static struct trigger on_alter_space_in_lua = { RLIST_LINK_INITIALIZER,
+ box_lua_space_new_or_delete,
+ NULL, NULL };
/**
* Make a tuple or a table Lua object by map.
@@ -561,8 +562,9 @@ lbox_space_frommap(struct lua_State *L)
space = space_by_id(id);
if (space == NULL) {
lua_pushnil(L);
- lua_pushstring(L, tt_sprintf("Space with id '%d' "\
- "doesn't exist", id));
+ lua_pushstring(L, tt_sprintf("Space with id '%d' "
+ "doesn't exist",
+ id));
return 2;
}
assert(space->format != NULL);
@@ -579,11 +581,11 @@ lbox_space_frommap(struct lua_State *L)
if (tuple_fieldno_by_name(dict, key, key_len, key_hash,
&fieldno)) {
lua_pushnil(L);
- lua_pushstring(L, tt_sprintf("Unknown field '%s'",
- key));
+ lua_pushstring(L,
+ tt_sprintf("Unknown field '%s'", key));
return 2;
}
- lua_rawseti(L, -3, fieldno+1);
+ lua_rawseti(L, -3, fieldno + 1);
}
lua_replace(L, 1);
@@ -701,8 +703,7 @@ box_lua_space_init(struct lua_State *L)
lua_pop(L, 2); /* box, schema */
static const struct luaL_Reg space_internal_lib[] = {
- {"frommap", lbox_space_frommap},
- {NULL, NULL}
+ { "frommap", lbox_space_frommap }, { NULL, NULL }
};
luaL_register(L, "box.internal.space", space_internal_lib);
lua_pop(L, 1);
diff --git a/src/box/lua/stat.c b/src/box/lua/stat.c
index 29ec38b..81dacdd 100644
--- a/src/box/lua/stat.c
+++ b/src/box/lua/stat.c
@@ -68,7 +68,7 @@ fill_stat_item(struct lua_State *L, int rps, int64_t total)
static int
set_stat_item(const char *name, int rps, int64_t total, void *cb_ctx)
{
- struct lua_State *L = (struct lua_State *) cb_ctx;
+ struct lua_State *L = (struct lua_State *)cb_ctx;
lua_pushstring(L, name);
lua_newtable(L);
@@ -87,7 +87,7 @@ set_stat_item(const char *name, int rps, int64_t total, void *cb_ctx)
static int
seek_stat_item(const char *name, int rps, int64_t total, void *cb_ctx)
{
- struct lua_State *L = (struct lua_State *) cb_ctx;
+ struct lua_State *L = (struct lua_State *)cb_ctx;
if (strcmp(name, lua_tostring(L, -1)) != 0)
return 0;
@@ -211,28 +211,25 @@ lbox_stat_sql(struct lua_State *L)
return 1;
}
-static const struct luaL_Reg lbox_stat_meta [] = {
- {"__index", lbox_stat_index},
- {"__call", lbox_stat_call},
- {NULL, NULL}
-};
+static const struct luaL_Reg lbox_stat_meta[] = { { "__index",
+ lbox_stat_index },
+ { "__call", lbox_stat_call },
+ { NULL, NULL } };
-static const struct luaL_Reg lbox_stat_net_meta [] = {
- {"__index", lbox_stat_net_index},
- {"__call", lbox_stat_net_call},
- {NULL, NULL}
+static const struct luaL_Reg lbox_stat_net_meta[] = {
+ { "__index", lbox_stat_net_index },
+ { "__call", lbox_stat_net_call },
+ { NULL, NULL }
};
/** Initialize box.stat package. */
void
box_lua_stat_init(struct lua_State *L)
{
- static const struct luaL_Reg statlib [] = {
- {"vinyl", lbox_stat_vinyl},
- {"reset", lbox_stat_reset},
- {"sql", lbox_stat_sql},
- {NULL, NULL}
- };
+ static const struct luaL_Reg statlib[] = { { "vinyl", lbox_stat_vinyl },
+ { "reset", lbox_stat_reset },
+ { "sql", lbox_stat_sql },
+ { NULL, NULL } };
luaL_register_module(L, "box.stat", statlib);
@@ -241,9 +238,7 @@ box_lua_stat_init(struct lua_State *L)
lua_setmetatable(L, -2);
lua_pop(L, 1); /* stat module */
- static const struct luaL_Reg netstatlib [] = {
- {NULL, NULL}
- };
+ static const struct luaL_Reg netstatlib[] = { { NULL, NULL } };
luaL_register_module(L, "box.stat.net", netstatlib);
@@ -252,4 +247,3 @@ box_lua_stat_init(struct lua_State *L)
lua_setmetatable(L, -2);
lua_pop(L, 1); /* stat net module */
}
-
diff --git a/src/box/lua/stat.h b/src/box/lua/stat.h
index bd22383..c5d46c0 100644
--- a/src/box/lua/stat.h
+++ b/src/box/lua/stat.h
@@ -35,7 +35,8 @@ extern "C" {
#endif /* defined(__cplusplus) */
struct lua_State;
-void box_lua_stat_init(struct lua_State *L);
+void
+box_lua_stat_init(struct lua_State *L);
#if defined(__cplusplus)
} /* extern "C" */
diff --git a/src/box/lua/tuple.c b/src/box/lua/tuple.c
index 03b4b8a..e800fec 100644
--- a/src/box/lua/tuple.c
+++ b/src/box/lua/tuple.c
@@ -31,9 +31,9 @@
#include "box/lua/tuple.h"
#include "box/xrow_update.h"
-#include "lua/utils.h" /* luaT_error() */
+#include "lua/utils.h" /* luaT_error() */
#include "lua/msgpack.h" /* luamp_encode_XXX() */
-#include "diag.h" /* diag_set() */
+#include "diag.h" /* diag_set() */
#include <small/ibuf.h>
#include <small/region.h>
#include <fiber.h>
@@ -73,9 +73,10 @@ box_tuple_t *
luaT_checktuple(struct lua_State *L, int idx)
{
struct tuple *tuple = luaT_istuple(L, idx);
- if (tuple == NULL) {
- luaL_error(L, "Invalid argument #%d (box.tuple expected, got %s)",
- idx, lua_typename(L, lua_type(L, idx)));
+ if (tuple == NULL) {
+ luaL_error(L,
+ "Invalid argument #%d (box.tuple expected, got %s)",
+ idx, lua_typename(L, lua_type(L, idx)));
}
return tuple;
@@ -95,7 +96,7 @@ luaT_istuple(struct lua_State *L, int narg)
if (ctypeid != CTID_STRUCT_TUPLE_REF)
return NULL;
- return *(struct tuple **) data;
+ return *(struct tuple **)data;
}
struct tuple *
@@ -110,8 +111,8 @@ luaT_tuple_new(struct lua_State *L, int idx, box_tuple_format_t *format)
struct ibuf *buf = tarantool_lua_ibuf;
ibuf_reset(buf);
struct mpstream stream;
- mpstream_init(&stream, buf, ibuf_reserve_cb, ibuf_alloc_cb,
- luamp_error, L);
+ mpstream_init(&stream, buf, ibuf_reserve_cb, ibuf_alloc_cb, luamp_error,
+ L);
if (idx == 0) {
/*
* Create the tuple from lua stack
@@ -127,8 +128,8 @@ luaT_tuple_new(struct lua_State *L, int idx, box_tuple_format_t *format)
luamp_encode_tuple(L, &tuple_serializer, &stream, idx);
}
mpstream_flush(&stream);
- struct tuple *tuple = box_tuple_new(format, buf->buf,
- buf->buf + ibuf_used(buf));
+ struct tuple *tuple =
+ box_tuple_new(format, buf->buf, buf->buf + ibuf_used(buf));
if (tuple == NULL)
return NULL;
ibuf_reinit(tarantool_lua_ibuf);
@@ -148,8 +149,7 @@ lbox_tuple_new(lua_State *L)
* box.tuple.new(1, 2, 3) (idx == 0), or the new one:
* box.tuple.new({1, 2, 3}) (idx == 1).
*/
- int idx = argc == 1 && (lua_istable(L, 1) ||
- luaT_istuple(L, 1));
+ int idx = argc == 1 && (lua_istable(L, 1) || luaT_istuple(L, 1));
box_tuple_format_t *fmt = box_tuple_format_default();
struct tuple *tuple = luaT_tuple_new(L, idx, fmt);
if (tuple == NULL)
@@ -170,8 +170,7 @@ lbox_tuple_gc(struct lua_State *L)
static int
lbox_tuple_slice_wrapper(struct lua_State *L)
{
- box_tuple_iterator_t *it = (box_tuple_iterator_t *)
- lua_topointer(L, 1);
+ box_tuple_iterator_t *it = (box_tuple_iterator_t *)lua_topointer(L, 1);
uint32_t start = lua_tonumber(L, 2);
uint32_t end = lua_tonumber(L, 3);
assert(end >= start);
@@ -221,13 +220,15 @@ lbox_tuple_slice(struct lua_State *L)
} else if (offset < 0 && -offset < field_count) {
end = offset + field_count;
} else {
- return luaL_error(L, "tuple.slice(): end > field count");
+ return luaL_error(L,
+ "tuple.slice(): end > field count");
}
} else {
end = field_count;
}
if (end <= start)
- return luaL_error(L, "tuple.slice(): start must be less than end");
+ return luaL_error(L,
+ "tuple.slice(): start must be less than end");
box_tuple_iterator_t *it = box_tuple_iterator(tuple);
lua_pushcfunction(L, lbox_tuple_slice_wrapper);
@@ -365,7 +366,8 @@ lbox_tuple_transform(struct lua_State *L)
int argc = lua_gettop(L);
if (argc < 3)
luaL_error(L, "tuple.transform(): bad arguments");
- lua_Integer offset = lua_tointeger(L, 2); /* Can be negative and can be > INT_MAX */
+ lua_Integer offset =
+ lua_tointeger(L, 2); /* Can be negative and can be > INT_MAX */
lua_Integer len = lua_tointeger(L, 3);
lua_Integer field_count = box_tuple_field_count(tuple);
@@ -374,7 +376,8 @@ lbox_tuple_transform(struct lua_State *L)
luaL_error(L, "tuple.transform(): offset is out of bound");
} else if (offset < 0) {
if (-offset > field_count)
- luaL_error(L, "tuple.transform(): offset is out of bound");
+ luaL_error(L,
+ "tuple.transform(): offset is out of bound");
offset += field_count + 1;
} else if (offset > field_count) {
offset = field_count + 1;
@@ -403,8 +406,8 @@ lbox_tuple_transform(struct lua_State *L)
struct ibuf *buf = tarantool_lua_ibuf;
ibuf_reset(buf);
struct mpstream stream;
- mpstream_init(&stream, buf, ibuf_reserve_cb, ibuf_alloc_cb,
- luamp_error, L);
+ mpstream_init(&stream, buf, ibuf_reserve_cb, ibuf_alloc_cb, luamp_error,
+ L);
/*
* Prepare UPDATE expression
@@ -417,7 +420,7 @@ lbox_tuple_transform(struct lua_State *L)
mpstream_encode_uint(&stream, len);
}
- for (int i = argc ; i > 3; i--) {
+ for (int i = argc; i > 3; i--) {
mpstream_encode_array(&stream, 3);
mpstream_encode_str(&stream, "!");
mpstream_encode_uint(&stream, offset);
@@ -438,13 +441,13 @@ lbox_tuple_transform(struct lua_State *L)
* to use the default one with no restrictions on field
* count or types.
*/
- const char *new_data =
- xrow_update_execute(buf->buf, buf->buf + ibuf_used(buf),
- old_data, old_data + bsize, format,
- &new_size, 1, NULL);
+ const char *new_data = xrow_update_execute(buf->buf,
+ buf->buf + ibuf_used(buf),
+ old_data, old_data + bsize,
+ format, &new_size, 1, NULL);
if (new_data != NULL)
- new_tuple = tuple_new(box_tuple_format_default(),
- new_data, new_data + new_size);
+ new_tuple = tuple_new(box_tuple_format_default(), new_data,
+ new_data + new_size);
region_truncate(region, used);
if (new_tuple == NULL)
@@ -478,11 +481,9 @@ lbox_tuple_field_by_path(struct lua_State *L)
const char *field = NULL, *path = lua_tolstring(L, 2, &len);
if (len == 0)
return 0;
- field = tuple_field_raw_by_full_path(tuple_format(tuple),
- tuple_data(tuple),
- tuple_field_map(tuple),
- path, (uint32_t)len,
- lua_hashstring(L, 2));
+ field = tuple_field_raw_by_full_path(
+ tuple_format(tuple), tuple_data(tuple), tuple_field_map(tuple),
+ path, (uint32_t)len, lua_hashstring(L, 2));
if (field == NULL)
return 0;
luamp_decode(L, luaL_msgpack_default, &field);
@@ -508,8 +509,8 @@ void
luaT_pushtuple(struct lua_State *L, box_tuple_t *tuple)
{
assert(CTID_STRUCT_TUPLE_REF != 0);
- struct tuple **ptr = (struct tuple **)
- luaL_pushcdata(L, CTID_STRUCT_TUPLE_REF);
+ struct tuple **ptr =
+ (struct tuple **)luaL_pushcdata(L, CTID_STRUCT_TUPLE_REF);
*ptr = tuple;
/* The order is important - first reference tuple, next set gc */
box_tuple_ref(tuple);
@@ -518,23 +519,19 @@ luaT_pushtuple(struct lua_State *L, box_tuple_t *tuple)
}
static const struct luaL_Reg lbox_tuple_meta[] = {
- {"__gc", lbox_tuple_gc},
- {"tostring", lbox_tuple_to_string},
- {"slice", lbox_tuple_slice},
- {"transform", lbox_tuple_transform},
- {"tuple_to_map", lbox_tuple_to_map},
- {"tuple_field_by_path", lbox_tuple_field_by_path},
- {NULL, NULL}
+ { "__gc", lbox_tuple_gc },
+ { "tostring", lbox_tuple_to_string },
+ { "slice", lbox_tuple_slice },
+ { "transform", lbox_tuple_transform },
+ { "tuple_to_map", lbox_tuple_to_map },
+ { "tuple_field_by_path", lbox_tuple_field_by_path },
+ { NULL, NULL }
};
-static const struct luaL_Reg lbox_tuplelib[] = {
- {"new", lbox_tuple_new},
- {NULL, NULL}
-};
+static const struct luaL_Reg lbox_tuplelib[] = { { "new", lbox_tuple_new },
+ { NULL, NULL } };
-static const struct luaL_Reg lbox_tuple_iterator_meta[] = {
- {NULL, NULL}
-};
+static const struct luaL_Reg lbox_tuple_iterator_meta[] = { { NULL, NULL } };
/* }}} */
@@ -548,8 +545,8 @@ tuple_serializer_update_options(void)
static int
on_msgpack_serializer_update(struct trigger *trigger, void *event)
{
- (void) trigger;
- (void) event;
+ (void)trigger;
+ (void)event;
tuple_serializer_update_options();
return 0;
}
@@ -563,8 +560,7 @@ box_lua_tuple_init(struct lua_State *L)
luaL_register(L, NULL, lbox_tuple_meta);
lua_setfield(L, -2, "tuple");
lua_pop(L, 1); /* box.internal */
- luaL_register_type(L, tuple_iteratorlib_name,
- lbox_tuple_iterator_meta);
+ luaL_register_type(L, tuple_iteratorlib_name, lbox_tuple_iterator_meta);
luaL_register_module(L, tuplelib_name, lbox_tuplelib);
lua_pop(L, 1);
@@ -577,7 +573,7 @@ box_lua_tuple_init(struct lua_State *L)
/* Get CTypeID for `struct tuple' */
int rc = luaL_cdef(L, "struct tuple;");
assert(rc == 0);
- (void) rc;
+ (void)rc;
CTID_STRUCT_TUPLE_REF = luaL_ctypeid(L, "struct tuple &");
assert(CTID_STRUCT_TUPLE_REF != 0);
}
diff --git a/src/box/lua/xlog.c b/src/box/lua/xlog.c
index 971a26a..1074d3b 100644
--- a/src/box/lua/xlog.c
+++ b/src/box/lua/xlog.c
@@ -53,8 +53,8 @@ static int
lbox_pushcursor(struct lua_State *L, struct xlog_cursor *cur)
{
struct xlog_cursor **pcur = NULL;
- pcur = (struct xlog_cursor **)luaL_pushcdata(L,
- CTID_STRUCT_XLOG_CURSOR_REF);
+ pcur = (struct xlog_cursor **)luaL_pushcdata(
+ L, CTID_STRUCT_XLOG_CURSOR_REF);
*pcur = cur;
return 1;
}
@@ -66,7 +66,7 @@ lbox_checkcursor(struct lua_State *L, int narg, const char *src)
void *data = NULL;
data = (struct xlog_cursor *)luaL_checkcdata(L, narg, &ctypeid);
assert(ctypeid == CTID_STRUCT_XLOG_CURSOR_REF);
- if (ctypeid != (uint32_t )CTID_STRUCT_XLOG_CURSOR_REF)
+ if (ctypeid != (uint32_t)CTID_STRUCT_XLOG_CURSOR_REF)
luaL_error(L, "%s: expecting xlog_cursor object", src);
return *(struct xlog_cursor **)data;
}
@@ -90,7 +90,8 @@ lbox_xlog_pushkey(lua_State *L, const char *key)
}
static void
-lbox_xlog_parse_body_kv(struct lua_State *L, int type, const char **beg, const char *end)
+lbox_xlog_parse_body_kv(struct lua_State *L, int type, const char **beg,
+ const char *end)
{
if (mp_typeof(**beg) != MP_UINT)
luaL_error(L, "Broken type of body key");
@@ -146,7 +147,8 @@ lbox_xlog_parse_body(struct lua_State *L, int type, const char *ptr, size_t len)
lbox_xlog_parse_body_kv(L, type, beg, end);
if (i != size)
say_warn("warning: decoded %u values from"
- " MP_MAP, %u expected", i, size);
+ " MP_MAP, %u expected",
+ i, size);
return 0;
}
@@ -244,7 +246,7 @@ lbox_xlog_parser_iterate(struct lua_State *L)
lua_newtable(L);
lbox_xlog_parse_body(L, row.type, row.body[0].iov_base,
row.body[0].iov_len);
- lua_settable(L, -3); /* BODY */
+ lua_settable(L, -3); /* BODY */
}
return 2;
}
@@ -252,7 +254,8 @@ lbox_xlog_parser_iterate(struct lua_State *L)
/* }}} */
static void
-lbox_xlog_parser_close(struct xlog_cursor *cur) {
+lbox_xlog_parser_close(struct xlog_cursor *cur)
+{
if (cur == NULL)
return;
xlog_cursor_close(cur, false);
@@ -277,11 +280,11 @@ lbox_xlog_parser_open_pairs(struct lua_State *L)
const char *filename = luaL_checkstring(L, 1);
/* Construct xlog cursor */
- struct xlog_cursor *cur = (struct xlog_cursor *)calloc(1,
- sizeof(struct xlog_cursor));
+ struct xlog_cursor *cur =
+ (struct xlog_cursor *)calloc(1, sizeof(struct xlog_cursor));
if (cur == NULL) {
- diag_set(OutOfMemory, sizeof(struct xlog_cursor),
- "malloc", "struct xlog_cursor");
+ diag_set(OutOfMemory, sizeof(struct xlog_cursor), "malloc",
+ "struct xlog_cursor");
return luaT_error(L);
}
/* Construct xlog object */
@@ -296,8 +299,7 @@ lbox_xlog_parser_open_pairs(struct lua_State *L)
strncmp(cur->meta.filetype, "VYLOG", 4) != 0) {
char buf[1024];
snprintf(buf, sizeof(buf), "'%.*s' file type",
- (int) strlen(cur->meta.filetype),
- cur->meta.filetype);
+ (int)strlen(cur->meta.filetype), cur->meta.filetype);
diag_set(ClientError, ER_UNSUPPORTED, "xlog reader", buf);
xlog_cursor_close(cur, false);
free(cur);
@@ -314,9 +316,9 @@ lbox_xlog_parser_open_pairs(struct lua_State *L)
return 3;
}
-static const struct luaL_Reg lbox_xlog_parser_lib [] = {
- { "pairs", lbox_xlog_parser_open_pairs },
- { NULL, NULL }
+static const struct luaL_Reg lbox_xlog_parser_lib[] = {
+ { "pairs", lbox_xlog_parser_open_pairs },
+ { NULL, NULL }
};
void
@@ -324,7 +326,9 @@ box_lua_xlog_init(struct lua_State *L)
{
int rc = 0;
/* Get CTypeIDs */
- rc = luaL_cdef(L, "struct xlog_cursor;"); assert(rc == 0); (void) rc;
+ rc = luaL_cdef(L, "struct xlog_cursor;");
+ assert(rc == 0);
+ (void)rc;
CTID_STRUCT_XLOG_CURSOR_REF = luaL_ctypeid(L, "struct xlog_cursor&");
assert(CTID_STRUCT_XLOG_CURSOR_REF != 0);
diff --git a/src/box/memtx_bitset.c b/src/box/memtx_bitset.c
index 2283a47..9006edd 100644
--- a/src/box/memtx_bitset.c
+++ b/src/box/memtx_bitset.c
@@ -69,7 +69,9 @@ struct bitset_hash_entry {
#if UINTPTR_MAX == 0xffffffff
#define mh_hash_key(a, arg) ((uintptr_t)(a))
#else
-#define mh_hash_key(a, arg) ((uint32_t)(((uintptr_t)(a)) >> 33 ^ ((uintptr_t)(a)) ^ ((uintptr_t)(a)) << 11))
+#define mh_hash_key(a, arg) \
+ ((uint32_t)(((uintptr_t)(a)) >> 33 ^ ((uintptr_t)(a)) ^ \
+ ((uintptr_t)(a)) << 11))
#endif
#define mh_hash(a, arg) mh_hash_key((a)->tuple, arg)
#define mh_cmp(a, b, arg) ((a)->tuple != (b)->tuple)
@@ -81,9 +83,7 @@ struct bitset_hash_entry {
#define MH_SOURCE 1
#include <salad/mhash.h>
-enum {
- SPARE_ID_END = 0xFFFFFFFF
-};
+enum { SPARE_ID_END = 0xFFFFFFFF };
static int
memtx_bitset_index_register_tuple(struct memtx_bitset_index *index,
@@ -108,7 +108,7 @@ memtx_bitset_index_register_tuple(struct memtx_bitset_index *index,
if (pos == mh_end(index->tuple_to_id)) {
*(uint32_t *)tuple = index->spare_id;
index->spare_id = id;
- diag_set(OutOfMemory, (ssize_t) pos, "hash", "key");
+ diag_set(OutOfMemory, (ssize_t)pos, "hash", "key");
return -1;
}
return 0;
@@ -118,9 +118,9 @@ static void
memtx_bitset_index_unregister_tuple(struct memtx_bitset_index *index,
struct tuple *tuple)
{
-
uint32_t k = mh_bitset_index_find(index->tuple_to_id, tuple, 0);
- struct bitset_hash_entry *e = mh_bitset_index_node(index->tuple_to_id, k);
+ struct bitset_hash_entry *e =
+ mh_bitset_index_node(index->tuple_to_id, k);
void *mem = matras_get(index->id_to_tuple, e->id);
*(uint32_t *)mem = index->spare_id;
index->spare_id = e->id;
@@ -132,7 +132,8 @@ memtx_bitset_index_tuple_to_value(struct memtx_bitset_index *index,
struct tuple *tuple)
{
uint32_t k = mh_bitset_index_find(index->tuple_to_id, tuple, 0);
- struct bitset_hash_entry *e = mh_bitset_index_node(index->tuple_to_id, k);
+ struct bitset_hash_entry *e =
+ mh_bitset_index_node(index->tuple_to_id, k);
return e->id;
}
@@ -155,7 +156,7 @@ tuple_to_value(struct tuple *tuple)
* https://github.com/tarantool/tarantool/issues/49
*/
/* size_t value = small_ptr_compress(tuple); */
- size_t value = (intptr_t) tuple >> 2;
+ size_t value = (intptr_t)tuple >> 2;
assert(value_to_tuple(value) == tuple);
return value;
}
@@ -164,7 +165,7 @@ static inline struct tuple *
value_to_tuple(size_t value)
{
/* return (struct tuple *) salloc_ptr_from_index(value); */
- return (struct tuple *) (value << 2);
+ return (struct tuple *)(value << 2);
}
#endif /* #ifndef OLD_GOOD_BITSET */
@@ -182,7 +183,7 @@ static_assert(sizeof(struct bitset_index_iterator) <= MEMTX_ITERATOR_SIZE,
static struct bitset_index_iterator *
bitset_index_iterator(struct iterator *it)
{
- return (struct bitset_index_iterator *) it;
+ return (struct bitset_index_iterator *)it;
}
static void
@@ -266,7 +267,7 @@ make_key(const char *field, uint32_t *key_len)
case MP_UINT:
u64key = mp_decode_uint(&field);
*key_len = sizeof(uint64_t);
- return (const char *) &u64key;
+ return (const char *)&u64key;
break;
case MP_STR:
return mp_decode_str(&field, key_len);
@@ -288,18 +289,19 @@ memtx_bitset_index_replace(struct index *base, struct tuple *old_tuple,
assert(!base->def->opts.is_unique);
assert(!base->def->key_def->is_multikey);
assert(old_tuple != NULL || new_tuple != NULL);
- (void) mode;
+ (void)mode;
*result = NULL;
if (old_tuple != NULL) {
#ifndef OLD_GOOD_BITSET
- uint32_t value = memtx_bitset_index_tuple_to_value(index, old_tuple);
+ uint32_t value =
+ memtx_bitset_index_tuple_to_value(index, old_tuple);
#else /* #ifndef OLD_GOOD_BITSET */
size_t value = tuple_to_value(old_tuple);
#endif /* #ifndef OLD_GOOD_BITSET */
if (tt_bitset_index_contains_value(&index->index,
- (size_t) value)) {
+ (size_t)value)) {
*result = old_tuple;
assert(old_tuple != new_tuple);
@@ -311,23 +313,25 @@ memtx_bitset_index_replace(struct index *base, struct tuple *old_tuple,
}
if (new_tuple != NULL) {
- const char *field = tuple_field_by_part(new_tuple,
- base->def->key_def->parts, MULTIKEY_NONE);
+ const char *field = tuple_field_by_part(
+ new_tuple, base->def->key_def->parts, MULTIKEY_NONE);
uint32_t key_len;
const void *key = make_key(field, &key_len);
#ifndef OLD_GOOD_BITSET
if (memtx_bitset_index_register_tuple(index, new_tuple) != 0)
return -1;
- uint32_t value = memtx_bitset_index_tuple_to_value(index, new_tuple);
+ uint32_t value =
+ memtx_bitset_index_tuple_to_value(index, new_tuple);
#else /* #ifndef OLD_GOOD_BITSET */
uint32_t value = tuple_to_value(new_tuple);
#endif /* #ifndef OLD_GOOD_BITSET */
- if (tt_bitset_index_insert(&index->index, key, key_len,
- value) < 0) {
+ if (tt_bitset_index_insert(&index->index, key, key_len, value) <
+ 0) {
#ifndef OLD_GOOD_BITSET
memtx_bitset_index_unregister_tuple(index, new_tuple);
#endif /* #ifndef OLD_GOOD_BITSET */
- diag_set(OutOfMemory, 0, "memtx_bitset_index", "insert");
+ diag_set(OutOfMemory, 0, "memtx_bitset_index",
+ "insert");
return -1;
}
}
@@ -342,13 +346,13 @@ memtx_bitset_index_create_iterator(struct index *base, enum iterator_type type,
struct memtx_engine *memtx = (struct memtx_engine *)base->engine;
assert(part_count == 0 || key != NULL);
- (void) part_count;
+ (void)part_count;
struct bitset_index_iterator *it;
it = mempool_alloc(&memtx->iterator_pool);
if (!it) {
- diag_set(OutOfMemory, sizeof(*it),
- "memtx_bitset_index", "iterator");
+ diag_set(OutOfMemory, sizeof(*it), "memtx_bitset_index",
+ "iterator");
return NULL;
}
@@ -474,7 +478,7 @@ memtx_bitset_index_count(struct index *base, enum iterator_type type,
*/
if (bit_iterator_next(&bit_it) == SIZE_MAX)
return tt_bitset_index_size(&index->index) -
- tt_bitset_index_count(&index->index, bit);
+ tt_bitset_index_count(&index->index, bit);
}
/* Call generic method */
@@ -490,7 +494,7 @@ static const struct index_vtab memtx_bitset_index_vtab = {
/* .update_def = */ generic_index_update_def,
/* .depends_on_pk = */ generic_index_depends_on_pk,
/* .def_change_requires_rebuild = */
- memtx_index_def_change_requires_rebuild,
+ memtx_index_def_change_requires_rebuild,
/* .size = */ memtx_bitset_index_size,
/* .bsize = */ memtx_bitset_index_bsize,
/* .min = */ generic_index_min,
@@ -501,7 +505,7 @@ static const struct index_vtab memtx_bitset_index_vtab = {
/* .replace = */ memtx_bitset_index_replace,
/* .create_iterator = */ memtx_bitset_index_create_iterator,
/* .create_snapshot_iterator = */
- generic_index_create_snapshot_iterator,
+ generic_index_create_snapshot_iterator,
/* .stat = */ generic_index_stat,
/* .compact = */ generic_index_compact,
/* .reset_stat = */ generic_index_reset_stat,
@@ -520,8 +524,8 @@ memtx_bitset_index_new(struct memtx_engine *memtx, struct index_def *def)
struct memtx_bitset_index *index =
(struct memtx_bitset_index *)calloc(1, sizeof(*index));
if (index == NULL) {
- diag_set(OutOfMemory, sizeof(*index),
- "malloc", "struct memtx_bitset_index");
+ diag_set(OutOfMemory, sizeof(*index), "malloc",
+ "struct memtx_bitset_index");
return NULL;
}
if (index_create(&index->base, (struct engine *)memtx,
@@ -532,11 +536,13 @@ memtx_bitset_index_new(struct memtx_engine *memtx, struct index_def *def)
#ifndef OLD_GOOD_BITSET
index->spare_id = SPARE_ID_END;
- index->id_to_tuple = (struct matras *)malloc(sizeof(*index->id_to_tuple));
+ index->id_to_tuple =
+ (struct matras *)malloc(sizeof(*index->id_to_tuple));
if (index->id_to_tuple == NULL)
panic("failed to allocate memtx bitset index");
- matras_create(index->id_to_tuple, MEMTX_EXTENT_SIZE, sizeof(struct tuple *),
- memtx_index_extent_alloc, memtx_index_extent_free, memtx);
+ matras_create(index->id_to_tuple, MEMTX_EXTENT_SIZE,
+ sizeof(struct tuple *), memtx_index_extent_alloc,
+ memtx_index_extent_free, memtx);
index->tuple_to_id = mh_bitset_index_new();
if (index->tuple_to_id == NULL)
diff --git a/src/box/memtx_engine.c b/src/box/memtx_engine.c
index 8147557..cc8670e 100644
--- a/src/box/memtx_engine.c
+++ b/src/box/memtx_engine.c
@@ -52,7 +52,7 @@
#include "raft.h"
/* sync snapshot every 16MB */
-#define SNAP_SYNC_INTERVAL (1 << 24)
+#define SNAP_SYNC_INTERVAL (1 << 24)
static void
checkpoint_cancel(struct checkpoint *ckpt);
@@ -159,8 +159,8 @@ memtx_engine_recover_snapshot(struct memtx_engine *memtx,
/* Process existing snapshot */
say_info("recovery start");
int64_t signature = vclock_sum(vclock);
- const char *filename = xdir_format_filename(&memtx->snap_dir,
- signature, NONE);
+ const char *filename =
+ xdir_format_filename(&memtx->snap_dir, signature, NONE);
say_info("recovering from `%s'", filename);
struct xlog_cursor cursor;
@@ -170,8 +170,8 @@ memtx_engine_recover_snapshot(struct memtx_engine *memtx,
int rc;
struct xrow_header row;
uint64_t row_count = 0;
- while ((rc = xlog_cursor_next(&cursor, &row,
- memtx->force_recovery)) == 0) {
+ while ((rc = xlog_cursor_next(&cursor, &row, memtx->force_recovery)) ==
+ 0) {
row.lsn = signature;
rc = memtx_engine_recover_snapshot_row(memtx, &row);
if (rc < 0) {
@@ -182,8 +182,7 @@ memtx_engine_recover_snapshot(struct memtx_engine *memtx,
}
++row_count;
if (row_count % 100000 == 0) {
- say_info("%.1fM rows processed",
- row_count / 1000000.);
+ say_info("%.1fM rows processed", row_count / 1000000.);
fiber_yield_timeout(0);
}
}
@@ -223,7 +222,7 @@ memtx_engine_recover_snapshot_row(struct memtx_engine *memtx,
if (row->type == IPROTO_RAFT)
return memtx_engine_recover_raft(row);
diag_set(ClientError, ER_UNKNOWN_REQUEST_TYPE,
- (uint32_t) row->type);
+ (uint32_t)row->type);
return -1;
}
int rc;
@@ -287,8 +286,8 @@ memtx_engine_begin_initial_recovery(struct engine *engine,
* recovery mode. Enable all keys on start, to detect and
* discard duplicates in the snapshot.
*/
- memtx->state = (memtx->force_recovery ?
- MEMTX_OK : MEMTX_INITIAL_RECOVERY);
+ memtx->state =
+ (memtx->force_recovery ? MEMTX_OK : MEMTX_INITIAL_RECOVERY);
return 0;
}
@@ -365,7 +364,8 @@ memtx_engine_prepare(struct engine *engine, struct txn *txn)
{
(void)engine;
struct txn_stmt *stmt;
- stailq_foreach_entry(stmt, &txn->stmts, next) {
+ stailq_foreach_entry(stmt, &txn->stmts, next)
+ {
if (stmt->add_story != NULL || stmt->del_story != NULL)
memtx_tx_history_prepare_stmt(stmt);
}
@@ -377,7 +377,8 @@ memtx_engine_commit(struct engine *engine, struct txn *txn)
{
(void)engine;
struct txn_stmt *stmt;
- stailq_foreach_entry(stmt, &txn->stmts, next) {
+ stailq_foreach_entry(stmt, &txn->stmts, next)
+ {
if (stmt->add_story != NULL || stmt->del_story != NULL) {
ssize_t bsize = memtx_tx_history_commit_stmt(stmt);
assert(stmt->space->engine == engine);
@@ -486,9 +487,9 @@ checkpoint_write_row(struct xlog *l, struct xrow_header *row)
return -1;
if ((l->rows + l->tx_rows) % 100000 == 0)
- say_crit("%.1fM rows written", (l->rows + l->tx_rows) / 1000000.0);
+ say_crit("%.1fM rows written",
+ (l->rows + l->tx_rows) / 1000000.0);
return 0;
-
}
static int
@@ -611,8 +612,8 @@ checkpoint_add_space(struct space *sp, void *data)
struct checkpoint *ckpt = (struct checkpoint *)data;
struct checkpoint_entry *entry = malloc(sizeof(*entry));
if (entry == NULL) {
- diag_set(OutOfMemory, sizeof(*entry),
- "malloc", "struct checkpoint_entry");
+ diag_set(OutOfMemory, sizeof(*entry), "malloc",
+ "struct checkpoint_entry");
return -1;
}
rlist_add_tail_entry(&ckpt->entries, entry, link);
@@ -672,7 +673,8 @@ checkpoint_f(va_list ap)
struct snapshot_iterator *it = entry->iterator;
while ((rc = it->next(it, &data, &size)) == 0 && data != NULL) {
if (checkpoint_write_tuple(&snap, entry->space_id,
- entry->group_id, data, size) != 0)
+ entry->group_id, data,
+ size) != 0)
goto fail;
}
if (rc != 0)
@@ -694,7 +696,7 @@ fail:
static int
memtx_engine_begin_checkpoint(struct engine *engine, bool is_scheduled)
{
- (void) is_scheduled;
+ (void)is_scheduled;
struct memtx_engine *memtx = (struct memtx_engine *)engine;
assert(memtx->checkpoint == NULL);
@@ -712,8 +714,7 @@ memtx_engine_begin_checkpoint(struct engine *engine, bool is_scheduled)
}
static int
-memtx_engine_wait_checkpoint(struct engine *engine,
- const struct vclock *vclock)
+memtx_engine_wait_checkpoint(struct engine *engine, const struct vclock *vclock)
{
struct memtx_engine *memtx = (struct memtx_engine *)engine;
@@ -728,8 +729,8 @@ memtx_engine_wait_checkpoint(struct engine *engine,
}
vclock_copy(&memtx->checkpoint->vclock, vclock);
- if (cord_costart(&memtx->checkpoint->cord, "snapshot",
- checkpoint_f, memtx->checkpoint)) {
+ if (cord_costart(&memtx->checkpoint->cord, "snapshot", checkpoint_f,
+ memtx->checkpoint)) {
return -1;
}
memtx->checkpoint->waiting_for_snap_thread = true;
@@ -747,7 +748,7 @@ static void
memtx_engine_commit_checkpoint(struct engine *engine,
const struct vclock *vclock)
{
- (void) vclock;
+ (void)vclock;
struct memtx_engine *memtx = (struct memtx_engine *)engine;
/* beginCheckpoint() must have been done */
@@ -796,11 +797,10 @@ memtx_engine_abort_checkpoint(struct engine *engine)
}
/** Remove garbage .inprogress file. */
- const char *filename =
- xdir_format_filename(&memtx->checkpoint->dir,
- vclock_sum(&memtx->checkpoint->vclock),
- INPROGRESS);
- (void) coio_unlink(filename);
+ const char *filename = xdir_format_filename(
+ &memtx->checkpoint->dir, vclock_sum(&memtx->checkpoint->vclock),
+ INPROGRESS);
+ (void)coio_unlink(filename);
checkpoint_delete(memtx->checkpoint);
memtx->checkpoint = NULL;
@@ -851,8 +851,8 @@ memtx_join_add_space(struct space *space, void *arg)
return 0;
struct memtx_join_entry *entry = malloc(sizeof(*entry));
if (entry == NULL) {
- diag_set(OutOfMemory, sizeof(*entry),
- "malloc", "struct memtx_join_entry");
+ diag_set(OutOfMemory, sizeof(*entry), "malloc",
+ "struct memtx_join_entry");
return -1;
}
entry->space_id = space_id(space);
@@ -871,8 +871,8 @@ memtx_engine_prepare_join(struct engine *engine, void **arg)
(void)engine;
struct memtx_join_ctx *ctx = malloc(sizeof(*ctx));
if (ctx == NULL) {
- diag_set(OutOfMemory, sizeof(*ctx),
- "malloc", "struct memtx_join_ctx");
+ diag_set(OutOfMemory, sizeof(*ctx), "malloc",
+ "struct memtx_join_ctx");
return -1;
}
rlist_create(&ctx->entries);
@@ -1018,8 +1018,8 @@ memtx_engine_run_gc(struct memtx_engine *memtx, bool *stop)
if (*stop)
return;
- struct memtx_gc_task *task = stailq_first_entry(&memtx->gc_queue,
- struct memtx_gc_task, link);
+ struct memtx_gc_task *task = stailq_first_entry(
+ &memtx->gc_queue, struct memtx_gc_task, link);
bool task_done;
task->vtab->run(task, &task_done);
if (task_done) {
@@ -1056,8 +1056,8 @@ memtx_engine_new(const char *snap_dirname, bool force_recovery,
{
struct memtx_engine *memtx = calloc(1, sizeof(*memtx));
if (memtx == NULL) {
- diag_set(OutOfMemory, sizeof(*memtx),
- "malloc", "struct memtx_engine");
+ diag_set(OutOfMemory, sizeof(*memtx), "malloc",
+ "struct memtx_engine");
return NULL;
}
@@ -1081,8 +1081,8 @@ memtx_engine_new(const char *snap_dirname, bool force_recovery,
int64_t snap_signature = xdir_last_vclock(&memtx->snap_dir, NULL);
if (snap_signature >= 0) {
struct xlog_cursor cursor;
- if (xdir_open_cursor(&memtx->snap_dir,
- snap_signature, &cursor) != 0)
+ if (xdir_open_cursor(&memtx->snap_dir, snap_signature,
+ &cursor) != 0)
goto fail;
INSTANCE_UUID = cursor.meta.instance_uuid;
xlog_cursor_close(&cursor, false);
@@ -1109,8 +1109,8 @@ memtx_engine_new(const char *snap_dirname, bool force_recovery,
tuple_arena_create(&memtx->arena, &memtx->quota, tuple_arena_max_size,
SLAB_SIZE, dontdump, "memtx");
slab_cache_create(&memtx->slab_cache, &memtx->arena);
- small_alloc_create(&memtx->alloc, &memtx->slab_cache,
- objsize_min, alloc_factor);
+ small_alloc_create(&memtx->alloc, &memtx->slab_cache, objsize_min,
+ alloc_factor);
/* Initialize index extent allocator. */
slab_cache_create(&memtx->index_slab_cache, &memtx->arena);
@@ -1139,8 +1139,7 @@ fail:
}
void
-memtx_engine_schedule_gc(struct memtx_engine *memtx,
- struct memtx_gc_task *task)
+memtx_engine_schedule_gc(struct memtx_engine *memtx, struct memtx_gc_task *task)
{
stailq_add_tail_entry(&memtx->gc_queue, task, link);
fiber_wakeup(memtx->gc_fiber);
@@ -1175,7 +1174,8 @@ memtx_enter_delayed_free_mode(struct memtx_engine *memtx)
{
memtx->snapshot_version++;
if (memtx->delayed_free_mode++ == 0)
- small_alloc_setopt(&memtx->alloc, SMALL_DELAYED_FREE_MODE, true);
+ small_alloc_setopt(&memtx->alloc, SMALL_DELAYED_FREE_MODE,
+ true);
}
void
@@ -1183,7 +1183,8 @@ memtx_leave_delayed_free_mode(struct memtx_engine *memtx)
{
assert(memtx->delayed_free_mode > 0);
if (--memtx->delayed_free_mode == 0)
- small_alloc_setopt(&memtx->alloc, SMALL_DELAYED_FREE_MODE, false);
+ small_alloc_setopt(&memtx->alloc, SMALL_DELAYED_FREE_MODE,
+ false);
}
struct tuple *
@@ -1244,7 +1245,7 @@ memtx_tuple_new(struct tuple_format *format, const char *data, const char *end)
tuple_format_ref(format);
tuple->data_offset = data_offset;
tuple->is_dirty = false;
- char *raw = (char *) tuple + tuple->data_offset;
+ char *raw = (char *)tuple + tuple->data_offset;
field_map_build(&builder, raw - field_map_size);
memcpy(raw, data, tuple_len);
say_debug("%s(%zu) = %p", __func__, tuple_len, memtx_tuple);
@@ -1276,8 +1277,7 @@ metmx_tuple_chunk_delete(struct tuple_format *format, const char *data)
{
struct memtx_engine *memtx = (struct memtx_engine *)format->engine;
struct tuple_chunk *tuple_chunk =
- container_of((const char (*)[0])data,
- struct tuple_chunk, data);
+ container_of((const char(*)[0])data, struct tuple_chunk, data);
uint32_t sz = tuple_chunk_sz(tuple_chunk->data_sz);
smfree(&memtx->alloc, tuple_chunk, sz);
}
@@ -1289,7 +1289,7 @@ memtx_tuple_chunk_new(struct tuple_format *format, struct tuple *tuple,
struct memtx_engine *memtx = (struct memtx_engine *)format->engine;
uint32_t sz = tuple_chunk_sz(data_sz);
struct tuple_chunk *tuple_chunk =
- (struct tuple_chunk *) smalloc(&memtx->alloc, sz);
+ (struct tuple_chunk *)smalloc(&memtx->alloc, sz);
if (tuple == NULL) {
diag_set(OutOfMemory, sz, "smalloc", "tuple");
return NULL;
@@ -1322,8 +1322,7 @@ memtx_index_extent_alloc(void *ctx)
}
ERROR_INJECT(ERRINJ_INDEX_ALLOC, {
/* same error as in mempool_alloc */
- diag_set(OutOfMemory, MEMTX_EXTENT_SIZE,
- "mempool", "new slab");
+ diag_set(OutOfMemory, MEMTX_EXTENT_SIZE, "mempool", "new slab");
return NULL;
});
void *ret;
@@ -1334,8 +1333,7 @@ memtx_index_extent_alloc(void *ctx)
break;
}
if (ret == NULL)
- diag_set(OutOfMemory, MEMTX_EXTENT_SIZE,
- "mempool", "new slab");
+ diag_set(OutOfMemory, MEMTX_EXTENT_SIZE, "mempool", "new slab");
return ret;
}
@@ -1358,8 +1356,7 @@ memtx_index_extent_reserve(struct memtx_engine *memtx, int num)
{
ERROR_INJECT(ERRINJ_INDEX_ALLOC, {
/* same error as in mempool_alloc */
- diag_set(OutOfMemory, MEMTX_EXTENT_SIZE,
- "mempool", "new slab");
+ diag_set(OutOfMemory, MEMTX_EXTENT_SIZE, "mempool", "new slab");
return -1;
});
struct mempool *pool = &memtx->index_extent_pool;
@@ -1372,8 +1369,8 @@ memtx_index_extent_reserve(struct memtx_engine *memtx, int num)
break;
}
if (ext == NULL) {
- diag_set(OutOfMemory, MEMTX_EXTENT_SIZE,
- "mempool", "new slab");
+ diag_set(OutOfMemory, MEMTX_EXTENT_SIZE, "mempool",
+ "new slab");
return -1;
}
*(void **)ext = memtx->reserved_extents;
diff --git a/src/box/memtx_engine.h b/src/box/memtx_engine.h
index 8b380bf..a033055 100644
--- a/src/box/memtx_engine.h
+++ b/src/box/memtx_engine.h
@@ -211,9 +211,8 @@ memtx_engine_schedule_gc(struct memtx_engine *memtx,
struct memtx_engine *
memtx_engine_new(const char *snap_dirname, bool force_recovery,
- uint64_t tuple_arena_max_size,
- uint32_t objsize_min, bool dontdump,
- float alloc_factor);
+ uint64_t tuple_arena_max_size, uint32_t objsize_min,
+ bool dontdump, float alloc_factor);
int
memtx_engine_recover_snapshot(struct memtx_engine *memtx,
@@ -256,10 +255,7 @@ memtx_tuple_delete(struct tuple_format *format, struct tuple *tuple);
/** Tuple format vtab for memtx engine. */
extern struct tuple_format_vtab memtx_tuple_format_vtab;
-enum {
- MEMTX_EXTENT_SIZE = 16 * 1024,
- MEMTX_SLAB_SIZE = 4 * 1024 * 1024
-};
+enum { MEMTX_EXTENT_SIZE = 16 * 1024, MEMTX_SLAB_SIZE = 4 * 1024 * 1024 };
/**
* Allocate a block of size MEMTX_EXTENT_SIZE for memtx index
@@ -297,14 +293,12 @@ memtx_index_def_change_requires_rebuild(struct index *index,
static inline struct memtx_engine *
memtx_engine_new_xc(const char *snap_dirname, bool force_recovery,
- uint64_t tuple_arena_max_size,
- uint32_t objsize_min, bool dontdump,
- float alloc_factor)
+ uint64_t tuple_arena_max_size, uint32_t objsize_min,
+ bool dontdump, float alloc_factor)
{
struct memtx_engine *memtx;
memtx = memtx_engine_new(snap_dirname, force_recovery,
- tuple_arena_max_size,
- objsize_min, dontdump,
+ tuple_arena_max_size, objsize_min, dontdump,
alloc_factor);
if (memtx == NULL)
diag_raise();
diff --git a/src/box/memtx_hash.c b/src/box/memtx_hash.c
index ed4dba9..a8b6fb2 100644
--- a/src/box/memtx_hash.c
+++ b/src/box/memtx_hash.c
@@ -46,16 +46,17 @@ static inline bool
memtx_hash_equal(struct tuple *tuple_a, struct tuple *tuple_b,
struct key_def *key_def)
{
- return tuple_compare(tuple_a, HINT_NONE,
- tuple_b, HINT_NONE, key_def) == 0;
+ return tuple_compare(tuple_a, HINT_NONE, tuple_b, HINT_NONE, key_def) ==
+ 0;
}
static inline bool
memtx_hash_equal_key(struct tuple *tuple, const char *key,
struct key_def *key_def)
{
- return tuple_compare_with_key(tuple, HINT_NONE, key, key_def->part_count,
- HINT_NONE, key_def) == 0;
+ return tuple_compare_with_key(tuple, HINT_NONE, key,
+ key_def->part_count, HINT_NONE,
+ key_def) == 0;
}
#define LIGHT_NAME _index
@@ -98,7 +99,7 @@ static void
hash_iterator_free(struct iterator *iterator)
{
assert(iterator->free == hash_iterator_free);
- struct hash_iterator *it = (struct hash_iterator *) iterator;
+ struct hash_iterator *it = (struct hash_iterator *)iterator;
mempool_free(it->pool, it);
}
@@ -106,10 +107,10 @@ static int
hash_iterator_ge_base(struct iterator *ptr, struct tuple **ret)
{
assert(ptr->free == hash_iterator_free);
- struct hash_iterator *it = (struct hash_iterator *) ptr;
+ struct hash_iterator *it = (struct hash_iterator *)ptr;
struct memtx_hash_index *index = (struct memtx_hash_index *)ptr->index;
- struct tuple **res = light_index_iterator_get_and_next(&index->hash_table,
- &it->iterator);
+ struct tuple **res = light_index_iterator_get_and_next(
+ &index->hash_table, &it->iterator);
*ret = res != NULL ? *res : NULL;
return 0;
}
@@ -119,10 +120,10 @@ hash_iterator_gt_base(struct iterator *ptr, struct tuple **ret)
{
assert(ptr->free == hash_iterator_free);
ptr->next = hash_iterator_ge_base;
- struct hash_iterator *it = (struct hash_iterator *) ptr;
+ struct hash_iterator *it = (struct hash_iterator *)ptr;
struct memtx_hash_index *index = (struct memtx_hash_index *)ptr->index;
- struct tuple **res = light_index_iterator_get_and_next(&index->hash_table,
- &it->iterator);
+ struct tuple **res = light_index_iterator_get_and_next(
+ &index->hash_table, &it->iterator);
if (res != NULL)
res = light_index_iterator_get_and_next(&index->hash_table,
&it->iterator);
@@ -130,26 +131,27 @@ hash_iterator_gt_base(struct iterator *ptr, struct tuple **ret)
return 0;
}
-#define WRAP_ITERATOR_METHOD(name) \
-static int \
-name(struct iterator *iterator, struct tuple **ret) \
-{ \
- struct txn *txn = in_txn(); \
- struct space *space = space_by_id(iterator->space_id); \
- bool is_rw = txn != NULL; \
- uint32_t iid = iterator->index->def->iid; \
- bool is_first = true; \
- do { \
- int rc = is_first ? name##_base(iterator, ret) \
- : hash_iterator_ge_base(iterator, ret); \
- if (rc != 0 || *ret == NULL) \
- return rc; \
- is_first = false; \
- *ret = memtx_tx_tuple_clarify(txn, space, *ret, iid, 0, is_rw); \
- } while (*ret == NULL); \
- return 0; \
-} \
-struct forgot_to_add_semicolon
+#define WRAP_ITERATOR_METHOD(name) \
+ static int name(struct iterator *iterator, struct tuple **ret) \
+ { \
+ struct txn *txn = in_txn(); \
+ struct space *space = space_by_id(iterator->space_id); \
+ bool is_rw = txn != NULL; \
+ uint32_t iid = iterator->index->def->iid; \
+ bool is_first = true; \
+ do { \
+ int rc = is_first ? \
+ name##_base(iterator, ret) : \
+ hash_iterator_ge_base(iterator, ret); \
+ if (rc != 0 || *ret == NULL) \
+ return rc; \
+ is_first = false; \
+ *ret = memtx_tx_tuple_clarify(txn, space, *ret, iid, \
+ 0, is_rw); \
+ } while (*ret == NULL); \
+ return 0; \
+ } \
+ struct forgot_to_add_semicolon
WRAP_ITERATOR_METHOD(hash_iterator_ge);
WRAP_ITERATOR_METHOD(hash_iterator_gt);
@@ -173,8 +175,8 @@ hash_iterator_eq(struct iterator *it, struct tuple **ret)
struct txn *txn = in_txn();
struct space *sp = space_by_id(it->space_id);
bool is_rw = txn != NULL;
- *ret = memtx_tx_tuple_clarify(txn, sp, *ret, it->index->def->iid,
- 0, is_rw);
+ *ret = memtx_tx_tuple_clarify(txn, sp, *ret, it->index->def->iid, 0,
+ is_rw);
return 0;
}
@@ -202,8 +204,8 @@ memtx_hash_index_gc_run(struct memtx_gc_task *task, bool *done)
enum { YIELD_LOOPS = 10 };
#endif
- struct memtx_hash_index *index = container_of(task,
- struct memtx_hash_index, gc_task);
+ struct memtx_hash_index *index =
+ container_of(task, struct memtx_hash_index, gc_task);
struct light_index_core *hash = &index->hash_table;
struct light_index_iterator *itr = &index->gc_iterator;
@@ -222,8 +224,8 @@ memtx_hash_index_gc_run(struct memtx_gc_task *task, bool *done)
static void
memtx_hash_index_gc_free(struct memtx_gc_task *task)
{
- struct memtx_hash_index *index = container_of(task,
- struct memtx_hash_index, gc_task);
+ struct memtx_hash_index *index =
+ container_of(task, struct memtx_hash_index, gc_task);
memtx_hash_index_free(index);
}
@@ -275,7 +277,7 @@ memtx_hash_index_bsize(struct index *base)
{
struct memtx_hash_index *index = (struct memtx_hash_index *)base;
return matras_extent_count(&index->hash_table.mtable) *
- MEMTX_EXTENT_SIZE;
+ MEMTX_EXTENT_SIZE;
}
static int
@@ -306,14 +308,14 @@ memtx_hash_index_count(struct index *base, enum iterator_type type,
}
static int
-memtx_hash_index_get(struct index *base, const char *key,
- uint32_t part_count, struct tuple **result)
+memtx_hash_index_get(struct index *base, const char *key, uint32_t part_count,
+ struct tuple **result)
{
struct memtx_hash_index *index = (struct memtx_hash_index *)base;
assert(base->def->opts.is_unique &&
part_count == base->def->key_def->part_count);
- (void) part_count;
+ (void)part_count;
struct space *space = space_by_id(base->def->space_id);
*result = NULL;
@@ -324,8 +326,8 @@ memtx_hash_index_get(struct index *base, const char *key,
uint32_t iid = base->def->iid;
struct txn *txn = in_txn();
bool is_rw = txn != NULL;
- *result = memtx_tx_tuple_clarify(txn, space, tuple, iid,
- 0, is_rw);
+ *result = memtx_tx_tuple_clarify(txn, space, tuple, iid, 0,
+ is_rw);
}
return 0;
}
@@ -346,8 +348,7 @@ memtx_hash_index_replace(struct index *base, struct tuple *old_tuple,
if (pos == light_index_end)
pos = light_index_insert(hash_table, h, new_tuple);
- ERROR_INJECT(ERRINJ_INDEX_ALLOC,
- {
+ ERROR_INJECT(ERRINJ_INDEX_ALLOC, {
light_index_delete(hash_table, pos);
pos = light_index_end;
});
@@ -357,18 +358,20 @@ memtx_hash_index_replace(struct index *base, struct tuple *old_tuple,
"hash_table", "key");
return -1;
}
- uint32_t errcode = replace_check_dup(old_tuple,
- dup_tuple, mode);
+ uint32_t errcode =
+ replace_check_dup(old_tuple, dup_tuple, mode);
if (errcode) {
light_index_delete(hash_table, pos);
if (dup_tuple) {
- uint32_t pos = light_index_insert(hash_table, h, dup_tuple);
+ uint32_t pos = light_index_insert(hash_table, h,
+ dup_tuple);
if (pos == light_index_end) {
panic("Failed to allocate memory in "
"recover of int hash_table");
}
}
- struct space *sp = space_cache_find(base->def->space_id);
+ struct space *sp =
+ space_cache_find(base->def->space_id);
if (sp != NULL)
diag_set(ClientError, errcode, base->def->name,
space_name(sp));
@@ -384,7 +387,8 @@ memtx_hash_index_replace(struct index *base, struct tuple *old_tuple,
if (old_tuple) {
uint32_t h = tuple_hash(old_tuple, base->def->key_def);
int res = light_index_delete_value(hash_table, h, old_tuple);
- assert(res == 0); (void) res;
+ assert(res == 0);
+ (void)res;
}
*result = old_tuple;
return 0;
@@ -413,11 +417,13 @@ memtx_hash_index_create_iterator(struct index *base, enum iterator_type type,
switch (type) {
case ITER_GT:
if (part_count != 0) {
- light_index_iterator_key(&index->hash_table, &it->iterator,
- key_hash(key, base->def->key_def), key);
+ light_index_iterator_key(
+ &index->hash_table, &it->iterator,
+ key_hash(key, base->def->key_def), key);
it->base.next = hash_iterator_gt;
} else {
- light_index_iterator_begin(&index->hash_table, &it->iterator);
+ light_index_iterator_begin(&index->hash_table,
+ &it->iterator);
it->base.next = hash_iterator_ge;
}
break;
@@ -428,7 +434,8 @@ memtx_hash_index_create_iterator(struct index *base, enum iterator_type type,
case ITER_EQ:
assert(part_count > 0);
light_index_iterator_key(&index->hash_table, &it->iterator,
- key_hash(key, base->def->key_def), key);
+ key_hash(key, base->def->key_def),
+ key);
it->base.next = hash_iterator_eq;
break;
default:
@@ -457,9 +464,9 @@ hash_snapshot_iterator_free(struct snapshot_iterator *iterator)
{
assert(iterator->free == hash_snapshot_iterator_free);
struct hash_snapshot_iterator *it =
- (struct hash_snapshot_iterator *) iterator;
- memtx_leave_delayed_free_mode((struct memtx_engine *)
- it->index->base.engine);
+ (struct hash_snapshot_iterator *)iterator;
+ memtx_leave_delayed_free_mode(
+ (struct memtx_engine *)it->index->base.engine);
light_index_iterator_destroy(&it->index->hash_table, &it->iterator);
index_unref(&it->index->base);
memtx_tx_snapshot_cleaner_destroy(&it->cleaner);
@@ -477,13 +484,12 @@ hash_snapshot_iterator_next(struct snapshot_iterator *iterator,
{
assert(iterator->free == hash_snapshot_iterator_free);
struct hash_snapshot_iterator *it =
- (struct hash_snapshot_iterator *) iterator;
+ (struct hash_snapshot_iterator *)iterator;
struct light_index_core *hash_table = &it->index->hash_table;
while (true) {
- struct tuple **res =
- light_index_iterator_get_and_next(hash_table,
- &it->iterator);
+ struct tuple **res = light_index_iterator_get_and_next(
+ hash_table, &it->iterator);
if (res == NULL) {
*data = NULL;
return 0;
@@ -509,8 +515,8 @@ static struct snapshot_iterator *
memtx_hash_index_create_snapshot_iterator(struct index *base)
{
struct memtx_hash_index *index = (struct memtx_hash_index *)base;
- struct hash_snapshot_iterator *it = (struct hash_snapshot_iterator *)
- calloc(1, sizeof(*it));
+ struct hash_snapshot_iterator *it =
+ (struct hash_snapshot_iterator *)calloc(1, sizeof(*it));
if (it == NULL) {
diag_set(OutOfMemory, sizeof(struct hash_snapshot_iterator),
"memtx_hash_index", "iterator");
@@ -524,7 +530,7 @@ memtx_hash_index_create_snapshot_iterator(struct index *base)
light_index_iterator_begin(&index->hash_table, &it->iterator);
light_index_iterator_freeze(&index->hash_table, &it->iterator);
memtx_enter_delayed_free_mode((struct memtx_engine *)base->engine);
- return (struct snapshot_iterator *) it;
+ return (struct snapshot_iterator *)it;
}
static const struct index_vtab memtx_hash_index_vtab = {
@@ -536,7 +542,7 @@ static const struct index_vtab memtx_hash_index_vtab = {
/* .update_def = */ memtx_hash_index_update_def,
/* .depends_on_pk = */ generic_index_depends_on_pk,
/* .def_change_requires_rebuild = */
- memtx_index_def_change_requires_rebuild,
+ memtx_index_def_change_requires_rebuild,
/* .size = */ memtx_hash_index_size,
/* .bsize = */ memtx_hash_index_bsize,
/* .min = */ generic_index_min,
@@ -547,7 +553,7 @@ static const struct index_vtab memtx_hash_index_vtab = {
/* .replace = */ memtx_hash_index_replace,
/* .create_iterator = */ memtx_hash_index_create_iterator,
/* .create_snapshot_iterator = */
- memtx_hash_index_create_snapshot_iterator,
+ memtx_hash_index_create_snapshot_iterator,
/* .stat = */ generic_index_stat,
/* .compact = */ generic_index_compact,
/* .reset_stat = */ generic_index_reset_stat,
@@ -563,8 +569,8 @@ memtx_hash_index_new(struct memtx_engine *memtx, struct index_def *def)
struct memtx_hash_index *index =
(struct memtx_hash_index *)calloc(1, sizeof(*index));
if (index == NULL) {
- diag_set(OutOfMemory, sizeof(*index),
- "malloc", "struct memtx_hash_index");
+ diag_set(OutOfMemory, sizeof(*index), "malloc",
+ "struct memtx_hash_index");
return NULL;
}
if (index_create(&index->base, (struct engine *)memtx,
diff --git a/src/box/memtx_rtree.c b/src/box/memtx_rtree.c
index b734daa..911da37 100644
--- a/src/box/memtx_rtree.c
+++ b/src/box/memtx_rtree.c
@@ -71,8 +71,8 @@ mp_decode_num(const char **data, uint32_t fieldno, double *ret)
* There must be <count> or <count * 2> numbers in that string.
*/
static inline int
-mp_decode_rect(struct rtree_rect *rect, unsigned dimension,
- const char *mp, unsigned count, const char *what)
+mp_decode_rect(struct rtree_rect *rect, unsigned dimension, const char *mp,
+ unsigned count, const char *what)
{
coord_t c = 0;
if (count == dimension) { /* point */
@@ -94,8 +94,8 @@ mp_decode_rect(struct rtree_rect *rect, unsigned dimension,
rect->coords[i * 2 + 1] = c;
}
} else {
- diag_set(ClientError, ER_RTREE_RECT,
- what, dimension, dimension * 2);
+ diag_set(ClientError, ER_RTREE_RECT, what, dimension,
+ dimension * 2);
return -1;
}
rtree_rect_normalize(rect, dimension);
@@ -124,8 +124,8 @@ extract_rectangle(struct rtree_rect *rect, struct tuple *tuple,
{
assert(index_def->key_def->part_count == 1);
assert(!index_def->key_def->is_multikey);
- const char *elems = tuple_field_by_part(tuple,
- index_def->key_def->parts, MULTIKEY_NONE);
+ const char *elems = tuple_field_by_part(
+ tuple, index_def->key_def->parts, MULTIKEY_NONE);
unsigned dimension = index_def->opts.dimension;
uint32_t count = mp_decode_array(&elems);
return mp_decode_rect(rect, dimension, elems, count, "Field");
@@ -133,8 +133,8 @@ extract_rectangle(struct rtree_rect *rect, struct tuple *tuple,
/* {{{ MemtxRTree Iterators ****************************************/
struct index_rtree_iterator {
- struct iterator base;
- struct rtree_iterator impl;
+ struct iterator base;
+ struct rtree_iterator impl;
/** Memory pool the iterator was allocated from. */
struct mempool *pool;
};
@@ -152,7 +152,7 @@ index_rtree_iterator_next(struct iterator *i, struct tuple **ret)
{
struct index_rtree_iterator *itr = (struct index_rtree_iterator *)i;
do {
- *ret = (struct tuple *) rtree_iterator_next(&itr->impl);
+ *ret = (struct tuple *)rtree_iterator_next(&itr->impl);
if (*ret == NULL)
break;
uint32_t iid = i->index->def->iid;
@@ -186,7 +186,6 @@ memtx_rtree_index_def_change_requires_rebuild(struct index *index,
index->def->opts.dimension != new_def->opts.dimension)
return true;
return false;
-
}
static ssize_t
@@ -213,8 +212,8 @@ memtx_rtree_index_count(struct index *base, enum iterator_type type,
}
static int
-memtx_rtree_index_get(struct index *base, const char *key,
- uint32_t part_count, struct tuple **result)
+memtx_rtree_index_get(struct index *base, const char *key, uint32_t part_count,
+ struct tuple **result)
{
struct memtx_rtree_index *index = (struct memtx_rtree_index *)base;
struct rtree_iterator iterator;
@@ -230,16 +229,16 @@ memtx_rtree_index_get(struct index *base, const char *key,
return 0;
}
do {
- struct tuple *tuple = (struct tuple *)
- rtree_iterator_next(&iterator);
+ struct tuple *tuple =
+ (struct tuple *)rtree_iterator_next(&iterator);
if (tuple == NULL)
break;
uint32_t iid = base->def->iid;
struct txn *txn = in_txn();
struct space *space = space_by_id(base->def->space_id);
bool is_rw = txn != NULL;
- *result = memtx_tx_tuple_clarify(txn, space, tuple, iid,
- 0, is_rw);
+ *result = memtx_tx_tuple_clarify(txn, space, tuple, iid, 0,
+ is_rw);
} while (*result == NULL);
rtree_iterator_destroy(&iterator);
return 0;
@@ -283,11 +282,12 @@ memtx_rtree_index_reserve(struct index *base, uint32_t size_hint)
return -1;
});
struct memtx_engine *memtx = (struct memtx_engine *)base->engine;
- return memtx_index_extent_reserve(memtx, RESERVE_EXTENTS_BEFORE_REPLACE);
+ return memtx_index_extent_reserve(memtx,
+ RESERVE_EXTENTS_BEFORE_REPLACE);
}
static struct iterator *
-memtx_rtree_index_create_iterator(struct index *base, enum iterator_type type,
+memtx_rtree_index_create_iterator(struct index *base, enum iterator_type type,
const char *key, uint32_t part_count)
{
struct memtx_rtree_index *index = (struct memtx_rtree_index *)base;
@@ -300,8 +300,8 @@ memtx_rtree_index_create_iterator(struct index *base, enum iterator_type type,
"empty keys for requested iterator type");
return NULL;
}
- } else if (mp_decode_rect_from_key(&rect, index->dimension,
- key, part_count)) {
+ } else if (mp_decode_rect_from_key(&rect, index->dimension, key,
+ part_count)) {
return NULL;
}
@@ -337,7 +337,8 @@ memtx_rtree_index_create_iterator(struct index *base, enum iterator_type type,
return NULL;
}
- struct index_rtree_iterator *it = mempool_alloc(&memtx->rtree_iterator_pool);
+ struct index_rtree_iterator *it =
+ mempool_alloc(&memtx->rtree_iterator_pool);
if (it == NULL) {
diag_set(OutOfMemory, sizeof(struct index_rtree_iterator),
"memtx_rtree_index", "iterator");
@@ -368,7 +369,7 @@ static const struct index_vtab memtx_rtree_index_vtab = {
/* .update_def = */ generic_index_update_def,
/* .depends_on_pk = */ generic_index_depends_on_pk,
/* .def_change_requires_rebuild = */
- memtx_rtree_index_def_change_requires_rebuild,
+ memtx_rtree_index_def_change_requires_rebuild,
/* .size = */ memtx_rtree_index_size,
/* .bsize = */ memtx_rtree_index_bsize,
/* .min = */ generic_index_min,
@@ -379,7 +380,7 @@ static const struct index_vtab memtx_rtree_index_vtab = {
/* .replace = */ memtx_rtree_index_replace,
/* .create_iterator = */ memtx_rtree_index_create_iterator,
/* .create_snapshot_iterator = */
- generic_index_create_snapshot_iterator,
+ generic_index_create_snapshot_iterator,
/* .stat = */ generic_index_stat,
/* .compact = */ generic_index_compact,
/* .reset_stat = */ generic_index_reset_stat,
@@ -401,13 +402,15 @@ memtx_rtree_index_new(struct memtx_engine *memtx, struct index_def *def)
def->opts.dimension > RTREE_MAX_DIMENSION) {
diag_set(UnsupportedIndexFeature, def,
tt_sprintf("dimension (%lld): must belong to "
- "range [%u, %u]", def->opts.dimension,
- 1, RTREE_MAX_DIMENSION));
+ "range [%u, %u]",
+ def->opts.dimension, 1,
+ RTREE_MAX_DIMENSION));
return NULL;
}
assert((int)RTREE_EUCLID == (int)RTREE_INDEX_DISTANCE_TYPE_EUCLID);
- assert((int)RTREE_MANHATTAN == (int)RTREE_INDEX_DISTANCE_TYPE_MANHATTAN);
+ assert((int)RTREE_MANHATTAN ==
+ (int)RTREE_INDEX_DISTANCE_TYPE_MANHATTAN);
enum rtree_distance_type distance_type =
(enum rtree_distance_type)def->opts.distance;
@@ -419,8 +422,8 @@ memtx_rtree_index_new(struct memtx_engine *memtx, struct index_def *def)
struct memtx_rtree_index *index =
(struct memtx_rtree_index *)calloc(1, sizeof(*index));
if (index == NULL) {
- diag_set(OutOfMemory, sizeof(*index),
- "malloc", "struct memtx_rtree_index");
+ diag_set(OutOfMemory, sizeof(*index), "malloc",
+ "struct memtx_rtree_index");
return NULL;
}
if (index_create(&index->base, (struct engine *)memtx,
diff --git a/src/box/memtx_space.c b/src/box/memtx_space.c
index d4b18d9..9060ca3 100644
--- a/src/box/memtx_space.c
+++ b/src/box/memtx_space.c
@@ -91,8 +91,8 @@ memtx_space_update_bsize(struct space *space, struct tuple *old_tuple,
*/
int
memtx_space_replace_no_keys(struct space *space, struct tuple *old_tuple,
- struct tuple *new_tuple,
- enum dup_replace_mode mode, struct tuple **result)
+ struct tuple *new_tuple, enum dup_replace_mode mode,
+ struct tuple **result)
{
(void)old_tuple;
(void)new_tuple;
@@ -100,7 +100,7 @@ memtx_space_replace_no_keys(struct space *space, struct tuple *old_tuple,
(void)result;
struct index *index = index_find(space, 0);
assert(index == NULL); /* not reached. */
- (void) index;
+ (void)index;
return -1;
}
@@ -144,8 +144,8 @@ memtx_space_replace_primary_key(struct space *space, struct tuple *old_tuple,
enum dup_replace_mode mode,
struct tuple **result)
{
- if (index_replace(space->index[0], old_tuple,
- new_tuple, mode, &old_tuple) != 0)
+ if (index_replace(space->index[0], old_tuple, new_tuple, mode,
+ &old_tuple) != 0)
return -1;
memtx_space_update_bsize(space, old_tuple, new_tuple);
if (new_tuple != NULL)
@@ -241,17 +241,17 @@ memtx_space_replace_primary_key(struct space *space, struct tuple *old_tuple,
int
memtx_space_replace_all_keys(struct space *space, struct tuple *old_tuple,
struct tuple *new_tuple,
- enum dup_replace_mode mode,
- struct tuple **result)
+ enum dup_replace_mode mode, struct tuple **result)
{
struct memtx_engine *memtx = (struct memtx_engine *)space->engine;
/*
* Ensure we have enough slack memory to guarantee
* successful statement-level rollback.
*/
- if (memtx_index_extent_reserve(memtx, new_tuple != NULL ?
- RESERVE_EXTENTS_BEFORE_REPLACE :
- RESERVE_EXTENTS_BEFORE_DELETE) != 0)
+ if (memtx_index_extent_reserve(
+ memtx, new_tuple != NULL ?
+ RESERVE_EXTENTS_BEFORE_REPLACE :
+ RESERVE_EXTENTS_BEFORE_DELETE) != 0)
return -1;
uint32_t i = 0;
@@ -264,11 +264,11 @@ memtx_space_replace_all_keys(struct space *space, struct tuple *old_tuple,
if (memtx_tx_manager_use_mvcc_engine) {
struct txn *txn = in_txn();
- struct txn_stmt *stmt =
- txn == NULL ? NULL : txn_current_stmt(txn);
+ struct txn_stmt *stmt = txn == NULL ? NULL :
+ txn_current_stmt(txn);
if (stmt != NULL) {
- return memtx_tx_history_add_stmt(stmt, old_tuple, new_tuple,
- mode, result);
+ return memtx_tx_history_add_stmt(
+ stmt, old_tuple, new_tuple, mode, result);
} else {
/** Ephemeral space */
assert(space->def->id == 0);
@@ -287,8 +287,8 @@ memtx_space_replace_all_keys(struct space *space, struct tuple *old_tuple,
for (i++; i < space->index_count; i++) {
struct tuple *unused;
struct index *index = space->index[i];
- if (index_replace(index, old_tuple, new_tuple,
- DUP_INSERT, &unused) != 0)
+ if (index_replace(index, old_tuple, new_tuple, DUP_INSERT,
+ &unused) != 0)
goto rollback;
}
@@ -303,8 +303,8 @@ rollback:
struct tuple *unused;
struct index *index = space->index[i - 1];
/* Rollback must not fail. */
- if (index_replace(index, new_tuple, old_tuple,
- DUP_INSERT, &unused) != 0) {
+ if (index_replace(index, new_tuple, old_tuple, DUP_INSERT,
+ &unused) != 0) {
diag_log();
unreachable();
panic("failed to rollback change");
@@ -335,8 +335,8 @@ memtx_space_execute_replace(struct space *space, struct txn *txn,
if (mode == DUP_INSERT)
stmt->does_require_old_tuple = true;
- if (memtx_space->replace(space, NULL, stmt->new_tuple,
- mode, &stmt->old_tuple) != 0)
+ if (memtx_space->replace(space, NULL, stmt->new_tuple, mode,
+ &stmt->old_tuple) != 0)
return -1;
stmt->engine_savepoint = stmt;
/** The new tuple is referenced by the primary key. */
@@ -369,8 +369,8 @@ memtx_space_execute_delete(struct space *space, struct txn *txn,
stmt->does_require_old_tuple = true;
if (old_tuple != NULL &&
- memtx_space->replace(space, old_tuple, NULL,
- DUP_REPLACE_OR_INSERT, &stmt->old_tuple) != 0)
+ memtx_space->replace(space, old_tuple, NULL, DUP_REPLACE_OR_INSERT,
+ &stmt->old_tuple) != 0)
return -1;
stmt->engine_savepoint = stmt;
*result = stmt->old_tuple;
@@ -404,23 +404,22 @@ memtx_space_execute_update(struct space *space, struct txn *txn,
uint32_t new_size = 0, bsize;
struct tuple_format *format = space->format;
const char *old_data = tuple_data_range(old_tuple, &bsize);
- const char *new_data =
- xrow_update_execute(request->tuple, request->tuple_end,
- old_data, old_data + bsize, format,
- &new_size, request->index_base, NULL);
+ const char *new_data = xrow_update_execute(
+ request->tuple, request->tuple_end, old_data, old_data + bsize,
+ format, &new_size, request->index_base, NULL);
if (new_data == NULL)
return -1;
- stmt->new_tuple = memtx_tuple_new(format, new_data,
- new_data + new_size);
+ stmt->new_tuple =
+ memtx_tuple_new(format, new_data, new_data + new_size);
if (stmt->new_tuple == NULL)
return -1;
tuple_ref(stmt->new_tuple);
stmt->does_require_old_tuple = true;
- if (memtx_space->replace(space, old_tuple, stmt->new_tuple,
- DUP_REPLACE, &stmt->old_tuple) != 0)
+ if (memtx_space->replace(space, old_tuple, stmt->new_tuple, DUP_REPLACE,
+ &stmt->old_tuple) != 0)
return -1;
stmt->engine_savepoint = stmt;
*result = stmt->new_tuple;
@@ -446,10 +445,9 @@ memtx_space_execute_upsert(struct space *space, struct txn *txn,
uint32_t part_count = index->def->key_def->part_count;
/* Extract the primary key from tuple. */
- const char *key = tuple_extract_key_raw(request->tuple,
- request->tuple_end,
- index->def->key_def,
- MULTIKEY_NONE, NULL);
+ const char *key =
+ tuple_extract_key_raw(request->tuple, request->tuple_end,
+ index->def->key_def, MULTIKEY_NONE, NULL);
if (key == NULL)
return -1;
/* Cut array header */
@@ -497,17 +495,15 @@ memtx_space_execute_upsert(struct space *space, struct txn *txn,
* for the tuple.
*/
uint64_t column_mask = COLUMN_MASK_FULL;
- const char *new_data =
- xrow_upsert_execute(request->ops, request->ops_end,
- old_data, old_data + bsize,
- format, &new_size,
- request->index_base, false,
- &column_mask);
+ const char *new_data = xrow_upsert_execute(
+ request->ops, request->ops_end, old_data,
+ old_data + bsize, format, &new_size,
+ request->index_base, false, &column_mask);
if (new_data == NULL)
return -1;
- stmt->new_tuple = memtx_tuple_new(format, new_data,
- new_data + new_size);
+ stmt->new_tuple =
+ memtx_tuple_new(format, new_data, new_data + new_size);
if (stmt->new_tuple == NULL)
return -1;
tuple_ref(stmt->new_tuple);
@@ -555,16 +551,16 @@ memtx_space_execute_upsert(struct space *space, struct txn *txn,
*/
static int
memtx_space_ephemeral_replace(struct space *space, const char *tuple,
- const char *tuple_end)
+ const char *tuple_end)
{
struct memtx_space *memtx_space = (struct memtx_space *)space;
- struct tuple *new_tuple = memtx_tuple_new(space->format, tuple,
- tuple_end);
+ struct tuple *new_tuple =
+ memtx_tuple_new(space->format, tuple, tuple_end);
if (new_tuple == NULL)
return -1;
struct tuple *old_tuple;
- if (memtx_space->replace(space, NULL, new_tuple,
- DUP_REPLACE_OR_INSERT, &old_tuple) != 0) {
+ if (memtx_space->replace(space, NULL, new_tuple, DUP_REPLACE_OR_INSERT,
+ &old_tuple) != 0) {
memtx_tuple_delete(space->format, new_tuple);
return -1;
}
@@ -598,8 +594,8 @@ memtx_space_ephemeral_delete(struct space *space, const char *key)
if (index_get(primary_index, key, part_count, &old_tuple) != 0)
return -1;
if (old_tuple != NULL &&
- memtx_space->replace(space, old_tuple, NULL,
- DUP_REPLACE, &old_tuple) != 0)
+ memtx_space->replace(space, old_tuple, NULL, DUP_REPLACE,
+ &old_tuple) != 0)
return -1;
tuple_unref(old_tuple);
return 0;
@@ -638,21 +634,21 @@ memtx_space_check_index_def(struct space *space, struct index_def *index_def)
}
switch (index_def->type) {
case HASH:
- if (! index_def->opts.is_unique) {
- diag_set(ClientError, ER_MODIFY_INDEX,
- index_def->name, space_name(space),
+ if (!index_def->opts.is_unique) {
+ diag_set(ClientError, ER_MODIFY_INDEX, index_def->name,
+ space_name(space),
"HASH index must be unique");
return -1;
}
if (key_def->is_multikey) {
- diag_set(ClientError, ER_MODIFY_INDEX,
- index_def->name, space_name(space),
+ diag_set(ClientError, ER_MODIFY_INDEX, index_def->name,
+ space_name(space),
"HASH index cannot be multikey");
return -1;
}
if (key_def->for_func_index) {
- diag_set(ClientError, ER_MODIFY_INDEX,
- index_def->name, space_name(space),
+ diag_set(ClientError, ER_MODIFY_INDEX, index_def->name,
+ space_name(space),
"HASH index can not use a function");
return -1;
}
@@ -662,32 +658,32 @@ memtx_space_check_index_def(struct space *space, struct index_def *index_def)
break;
case RTREE:
if (key_def->part_count != 1) {
- diag_set(ClientError, ER_MODIFY_INDEX,
- index_def->name, space_name(space),
+ diag_set(ClientError, ER_MODIFY_INDEX, index_def->name,
+ space_name(space),
"RTREE index key can not be multipart");
return -1;
}
if (index_def->opts.is_unique) {
- diag_set(ClientError, ER_MODIFY_INDEX,
- index_def->name, space_name(space),
+ diag_set(ClientError, ER_MODIFY_INDEX, index_def->name,
+ space_name(space),
"RTREE index can not be unique");
return -1;
}
if (key_def->parts[0].type != FIELD_TYPE_ARRAY) {
- diag_set(ClientError, ER_MODIFY_INDEX,
- index_def->name, space_name(space),
+ diag_set(ClientError, ER_MODIFY_INDEX, index_def->name,
+ space_name(space),
"RTREE index field type must be ARRAY");
return -1;
}
if (key_def->is_multikey) {
- diag_set(ClientError, ER_MODIFY_INDEX,
- index_def->name, space_name(space),
+ diag_set(ClientError, ER_MODIFY_INDEX, index_def->name,
+ space_name(space),
"RTREE index cannot be multikey");
return -1;
}
if (key_def->for_func_index) {
- diag_set(ClientError, ER_MODIFY_INDEX,
- index_def->name, space_name(space),
+ diag_set(ClientError, ER_MODIFY_INDEX, index_def->name,
+ space_name(space),
"RTREE index can not use a function");
return -1;
}
@@ -695,41 +691,40 @@ memtx_space_check_index_def(struct space *space, struct index_def *index_def)
return 0;
case BITSET:
if (key_def->part_count != 1) {
- diag_set(ClientError, ER_MODIFY_INDEX,
- index_def->name, space_name(space),
+ diag_set(ClientError, ER_MODIFY_INDEX, index_def->name,
+ space_name(space),
"BITSET index key can not be multipart");
return -1;
}
if (index_def->opts.is_unique) {
- diag_set(ClientError, ER_MODIFY_INDEX,
- index_def->name, space_name(space),
- "BITSET can not be unique");
+ diag_set(ClientError, ER_MODIFY_INDEX, index_def->name,
+ space_name(space), "BITSET can not be unique");
return -1;
}
if (key_def->parts[0].type != FIELD_TYPE_UNSIGNED &&
key_def->parts[0].type != FIELD_TYPE_STRING) {
- diag_set(ClientError, ER_MODIFY_INDEX,
- index_def->name, space_name(space),
+ diag_set(ClientError, ER_MODIFY_INDEX, index_def->name,
+ space_name(space),
"BITSET index field type must be NUM or STR");
return -1;
}
if (key_def->is_multikey) {
- diag_set(ClientError, ER_MODIFY_INDEX,
- index_def->name, space_name(space),
+ diag_set(ClientError, ER_MODIFY_INDEX, index_def->name,
+ space_name(space),
"BITSET index cannot be multikey");
return -1;
}
if (key_def->for_func_index) {
- diag_set(ClientError, ER_MODIFY_INDEX,
- index_def->name, space_name(space),
+ diag_set(ClientError, ER_MODIFY_INDEX, index_def->name,
+ space_name(space),
"BITSET index can not use a function");
return -1;
}
/* no furter checks of parts needed */
return 0;
default:
- diag_set(ClientError, ER_INDEX_TYPE,
- index_def->name, space_name(space));
+ diag_set(ClientError, ER_INDEX_TYPE, index_def->name,
+ space_name(space));
return -1;
}
/* Only HASH and TREE indexes checks parts there */
@@ -738,8 +733,8 @@ memtx_space_check_index_def(struct space *space, struct index_def *index_def)
struct key_part *part = &key_def->parts[i];
if (part->type <= FIELD_TYPE_ANY ||
part->type >= FIELD_TYPE_ARRAY) {
- diag_set(ClientError, ER_MODIFY_INDEX,
- index_def->name, space_name(space),
+ diag_set(ClientError, ER_MODIFY_INDEX, index_def->name,
+ space_name(space),
tt_sprintf("field type '%s' is not supported",
field_type_strs[part->type]));
return -1;
@@ -987,7 +982,7 @@ memtx_build_on_replace(struct trigger *trigger, void *event)
struct txn_stmt *stmt = txn_current_stmt(txn);
struct tuple *cmp_tuple = stmt->new_tuple != NULL ? stmt->new_tuple :
- stmt->old_tuple;
+ stmt->old_tuple;
/*
* Only update the already built part of an index. All the other
* tuples will be inserted when build continues.
@@ -1004,9 +999,9 @@ memtx_build_on_replace(struct trigger *trigger, void *event)
}
struct tuple *delete = NULL;
- enum dup_replace_mode mode =
- state->index->def->opts.is_unique ? DUP_INSERT :
- DUP_REPLACE_OR_INSERT;
+ enum dup_replace_mode mode = state->index->def->opts.is_unique ?
+ DUP_INSERT :
+ DUP_REPLACE_OR_INSERT;
state->rc = index_replace(state->index, stmt->old_tuple,
stmt->new_tuple, mode, &delete);
if (state->rc != 0) {
@@ -1098,12 +1093,12 @@ memtx_space_build_index(struct space *src_space, struct index *new_index,
* @todo: better message if there is a duplicate.
*/
struct tuple *old_tuple;
- rc = index_replace(new_index, NULL, tuple,
- DUP_INSERT, &old_tuple);
+ rc = index_replace(new_index, NULL, tuple, DUP_INSERT,
+ &old_tuple);
if (rc != 0)
break;
assert(old_tuple == NULL); /* Guaranteed by DUP_INSERT. */
- (void) old_tuple;
+ (void)old_tuple;
/*
* All tuples stored in a memtx space must be
* referenced by the primary index.
@@ -1187,13 +1182,13 @@ static const struct space_vtab memtx_space_vtab = {
};
struct space *
-memtx_space_new(struct memtx_engine *memtx,
- struct space_def *def, struct rlist *key_list)
+memtx_space_new(struct memtx_engine *memtx, struct space_def *def,
+ struct rlist *key_list)
{
struct memtx_space *memtx_space = malloc(sizeof(*memtx_space));
if (memtx_space == NULL) {
- diag_set(OutOfMemory, sizeof(*memtx_space),
- "malloc", "struct memtx_space");
+ diag_set(OutOfMemory, sizeof(*memtx_space), "malloc",
+ "struct memtx_space");
return NULL;
}
@@ -1204,11 +1199,10 @@ memtx_space_new(struct memtx_engine *memtx,
free(memtx_space);
return NULL;
}
- struct tuple_format *format =
- tuple_format_new(&memtx_tuple_format_vtab, memtx, keys, key_count,
- def->fields, def->field_count,
- def->exact_field_count, def->dict,
- def->opts.is_temporary, def->opts.is_ephemeral);
+ struct tuple_format *format = tuple_format_new(
+ &memtx_tuple_format_vtab, memtx, keys, key_count, def->fields,
+ def->field_count, def->exact_field_count, def->dict,
+ def->opts.is_temporary, def->opts.is_ephemeral);
if (format == NULL) {
free(memtx_space);
return NULL;
diff --git a/src/box/memtx_space.h b/src/box/memtx_space.h
index a14065f..b3082f8 100644
--- a/src/box/memtx_space.h
+++ b/src/box/memtx_space.h
@@ -84,8 +84,8 @@ memtx_space_replace_all_keys(struct space *, struct tuple *, struct tuple *,
enum dup_replace_mode, struct tuple **);
struct space *
-memtx_space_new(struct memtx_engine *memtx,
- struct space_def *def, struct rlist *key_list);
+memtx_space_new(struct memtx_engine *memtx, struct space_def *def,
+ struct rlist *key_list);
#if defined(__cplusplus)
} /* extern "C" */
diff --git a/src/box/memtx_tree.c b/src/box/memtx_tree.c
index 5af482f..43fa8e2 100644
--- a/src/box/memtx_tree.c
+++ b/src/box/memtx_tree.c
@@ -82,10 +82,10 @@ memtx_tree_data_is_equal(const struct memtx_tree_data *a,
#define BPS_TREE_NAME memtx_tree
#define BPS_TREE_BLOCK_SIZE (512)
#define BPS_TREE_EXTENT_SIZE MEMTX_EXTENT_SIZE
-#define BPS_TREE_COMPARE(a, b, arg)\
+#define BPS_TREE_COMPARE(a, b, arg) \
tuple_compare((&a)->tuple, (&a)->hint, (&b)->tuple, (&b)->hint, arg)
-#define BPS_TREE_COMPARE_KEY(a, b, arg)\
- tuple_compare_with_key((&a)->tuple, (&a)->hint, (b)->key,\
+#define BPS_TREE_COMPARE_KEY(a, b, arg) \
+ tuple_compare_with_key((&a)->tuple, (&a)->hint, (b)->key, \
(b)->part_count, (b)->hint, arg)
#define BPS_TREE_IS_IDENTICAL(a, b) memtx_tree_data_is_equal(&a, &b)
#define BPS_TREE_NO_DEBUG 1
@@ -124,7 +124,7 @@ memtx_tree_cmp_def(struct memtx_tree *tree)
}
static int
-memtx_tree_qcompare(const void* a, const void *b, void *c)
+memtx_tree_qcompare(const void *a, const void *b, void *c)
{
const struct memtx_tree_data *data_a = a;
const struct memtx_tree_data *data_b = b;
@@ -155,7 +155,7 @@ static inline struct tree_iterator *
tree_iterator(struct iterator *it)
{
assert(it->free == tree_iterator_free);
- return (struct tree_iterator *) it;
+ return (struct tree_iterator *)it;
}
static void
@@ -186,8 +186,8 @@ tree_iterator_next_base(struct iterator *iterator, struct tuple **ret)
struct memtx_tree_data *check =
memtx_tree_iterator_get_elem(&index->tree, &it->tree_iterator);
if (check == NULL || !memtx_tree_data_is_equal(check, &it->current)) {
- it->tree_iterator = memtx_tree_upper_bound_elem(&index->tree,
- it->current, NULL);
+ it->tree_iterator = memtx_tree_upper_bound_elem(
+ &index->tree, it->current, NULL);
} else {
memtx_tree_iterator_next(&index->tree, &it->tree_iterator);
}
@@ -216,8 +216,8 @@ tree_iterator_prev_base(struct iterator *iterator, struct tuple **ret)
struct memtx_tree_data *check =
memtx_tree_iterator_get_elem(&index->tree, &it->tree_iterator);
if (check == NULL || !memtx_tree_data_is_equal(check, &it->current)) {
- it->tree_iterator = memtx_tree_lower_bound_elem(&index->tree,
- it->current, NULL);
+ it->tree_iterator = memtx_tree_lower_bound_elem(
+ &index->tree, it->current, NULL);
}
memtx_tree_iterator_prev(&index->tree, &it->tree_iterator);
tuple_unref(it->current.tuple);
@@ -245,8 +245,8 @@ tree_iterator_next_equal_base(struct iterator *iterator, struct tuple **ret)
struct memtx_tree_data *check =
memtx_tree_iterator_get_elem(&index->tree, &it->tree_iterator);
if (check == NULL || !memtx_tree_data_is_equal(check, &it->current)) {
- it->tree_iterator = memtx_tree_upper_bound_elem(&index->tree,
- it->current, NULL);
+ it->tree_iterator = memtx_tree_upper_bound_elem(
+ &index->tree, it->current, NULL);
} else {
memtx_tree_iterator_next(&index->tree, &it->tree_iterator);
}
@@ -255,10 +255,8 @@ tree_iterator_next_equal_base(struct iterator *iterator, struct tuple **ret)
memtx_tree_iterator_get_elem(&index->tree, &it->tree_iterator);
/* Use user key def to save a few loops. */
if (res == NULL ||
- tuple_compare_with_key(res->tuple, res->hint,
- it->key_data.key,
- it->key_data.part_count,
- it->key_data.hint,
+ tuple_compare_with_key(res->tuple, res->hint, it->key_data.key,
+ it->key_data.part_count, it->key_data.hint,
index->base.def->key_def) != 0) {
iterator->next = tree_iterator_dummie;
it->current.tuple = NULL;
@@ -281,8 +279,8 @@ tree_iterator_prev_equal_base(struct iterator *iterator, struct tuple **ret)
struct memtx_tree_data *check =
memtx_tree_iterator_get_elem(&index->tree, &it->tree_iterator);
if (check == NULL || !memtx_tree_data_is_equal(check, &it->current)) {
- it->tree_iterator = memtx_tree_lower_bound_elem(&index->tree,
- it->current, NULL);
+ it->tree_iterator = memtx_tree_lower_bound_elem(
+ &index->tree, it->current, NULL);
}
memtx_tree_iterator_prev(&index->tree, &it->tree_iterator);
tuple_unref(it->current.tuple);
@@ -290,10 +288,8 @@ tree_iterator_prev_equal_base(struct iterator *iterator, struct tuple **ret)
memtx_tree_iterator_get_elem(&index->tree, &it->tree_iterator);
/* Use user key def to save a few loops. */
if (res == NULL ||
- tuple_compare_with_key(res->tuple, res->hint,
- it->key_data.key,
- it->key_data.part_count,
- it->key_data.hint,
+ tuple_compare_with_key(res->tuple, res->hint, it->key_data.key,
+ it->key_data.part_count, it->key_data.hint,
index->base.def->key_def) != 0) {
iterator->next = tree_iterator_dummie;
it->current.tuple = NULL;
@@ -306,39 +302,39 @@ tree_iterator_prev_equal_base(struct iterator *iterator, struct tuple **ret)
return 0;
}
-#define WRAP_ITERATOR_METHOD(name) \
-static int \
-name(struct iterator *iterator, struct tuple **ret) \
-{ \
- struct memtx_tree *tree = \
- &((struct memtx_tree_index *)iterator->index)->tree; \
- struct tree_iterator *it = tree_iterator(iterator); \
- struct memtx_tree_iterator *ti = &it->tree_iterator; \
- uint32_t iid = iterator->index->def->iid; \
- bool is_multikey = iterator->index->def->key_def->is_multikey; \
- struct txn *txn = in_txn(); \
- struct space *space = space_by_id(iterator->space_id); \
- bool is_rw = txn != NULL; \
- do { \
- int rc = name##_base(iterator, ret); \
- if (rc != 0 || *ret == NULL) \
- return rc; \
- uint32_t mk_index = 0; \
- if (is_multikey) { \
- struct memtx_tree_data *check = \
- memtx_tree_iterator_get_elem(tree, ti); \
- assert(check != NULL); \
- mk_index = check->hint; \
- } \
- *ret = memtx_tx_tuple_clarify(txn, space, *ret, \
- iid, mk_index, is_rw); \
- } while (*ret == NULL); \
- tuple_unref(it->current.tuple); \
- it->current.tuple = *ret; \
- tuple_ref(it->current.tuple); \
- return 0; \
-} \
-struct forgot_to_add_semicolon
+#define WRAP_ITERATOR_METHOD(name) \
+ static int name(struct iterator *iterator, struct tuple **ret) \
+ { \
+ struct memtx_tree *tree = \
+ &((struct memtx_tree_index *)iterator->index)->tree; \
+ struct tree_iterator *it = tree_iterator(iterator); \
+ struct memtx_tree_iterator *ti = &it->tree_iterator; \
+ uint32_t iid = iterator->index->def->iid; \
+ bool is_multikey = iterator->index->def->key_def->is_multikey; \
+ struct txn *txn = in_txn(); \
+ struct space *space = space_by_id(iterator->space_id); \
+ bool is_rw = txn != NULL; \
+ do { \
+ int rc = name##_base(iterator, ret); \
+ if (rc != 0 || *ret == NULL) \
+ return rc; \
+ uint32_t mk_index = 0; \
+ if (is_multikey) { \
+ struct memtx_tree_data *check = \
+ memtx_tree_iterator_get_elem(tree, \
+ ti); \
+ assert(check != NULL); \
+ mk_index = check->hint; \
+ } \
+ *ret = memtx_tx_tuple_clarify(txn, space, *ret, iid, \
+ mk_index, is_rw); \
+ } while (*ret == NULL); \
+ tuple_unref(it->current.tuple); \
+ it->current.tuple = *ret; \
+ tuple_ref(it->current.tuple); \
+ return 0; \
+ } \
+ struct forgot_to_add_semicolon
WRAP_ITERATOR_METHOD(tree_iterator_next);
WRAP_ITERATOR_METHOD(tree_iterator_prev);
@@ -393,17 +389,15 @@ tree_iterator_start(struct iterator *iterator, struct tuple **ret)
else
it->tree_iterator = memtx_tree_iterator_first(tree);
} else {
- if (type == ITER_ALL || type == ITER_EQ ||
- type == ITER_GE || type == ITER_LT) {
- it->tree_iterator =
- memtx_tree_lower_bound(tree, &it->key_data,
- &exact);
+ if (type == ITER_ALL || type == ITER_EQ || type == ITER_GE ||
+ type == ITER_LT) {
+ it->tree_iterator = memtx_tree_lower_bound(
+ tree, &it->key_data, &exact);
if (type == ITER_EQ && !exact)
return 0;
} else { // ITER_GT, ITER_REQ, ITER_LE
- it->tree_iterator =
- memtx_tree_upper_bound(tree, &it->key_data,
- &exact);
+ it->tree_iterator = memtx_tree_upper_bound(
+ tree, &it->key_data, &exact);
if (type == ITER_REQ && !exact)
return 0;
}
@@ -423,8 +417,8 @@ tree_iterator_start(struct iterator *iterator, struct tuple **ret)
}
}
- struct memtx_tree_data *res = memtx_tree_iterator_get_elem(tree,
- &it->tree_iterator);
+ struct memtx_tree_data *res =
+ memtx_tree_iterator_get_elem(tree, &it->tree_iterator);
if (!res)
return 0;
*ret = res->tuple;
@@ -475,8 +469,8 @@ memtx_tree_index_gc_run(struct memtx_gc_task *task, bool *done)
enum { YIELD_LOOPS = 10 };
#endif
- struct memtx_tree_index *index = container_of(task,
- struct memtx_tree_index, gc_task);
+ struct memtx_tree_index *index =
+ container_of(task, struct memtx_tree_index, gc_task);
struct memtx_tree *tree = &index->tree;
struct memtx_tree_iterator *itr = &index->gc_iterator;
@@ -497,8 +491,8 @@ memtx_tree_index_gc_run(struct memtx_gc_task *task, bool *done)
static void
memtx_tree_index_gc_free(struct memtx_gc_task *task)
{
- struct memtx_tree_index *index = container_of(task,
- struct memtx_tree_index, gc_task);
+ struct memtx_tree_index *index =
+ container_of(task, struct memtx_tree_index, gc_task);
memtx_tree_index_free(index);
}
@@ -542,7 +536,8 @@ memtx_tree_index_update_def(struct index *base)
* def must be used. For details @sa tuple_compare.cc.
*/
index->tree.arg = def->opts.is_unique && !def->key_def->is_nullable ?
- def->key_def : def->cmp_def;
+ def->key_def :
+ def->cmp_def;
}
static bool
@@ -586,8 +581,8 @@ memtx_tree_index_count(struct index *base, enum iterator_type type,
}
static int
-memtx_tree_index_get(struct index *base, const char *key,
- uint32_t part_count, struct tuple **result)
+memtx_tree_index_get(struct index *base, const char *key, uint32_t part_count,
+ struct tuple **result)
{
assert(base->def->opts.is_unique &&
part_count == base->def->key_def->part_count);
@@ -626,21 +621,22 @@ memtx_tree_index_replace(struct index *base, struct tuple *old_tuple,
dup_data.tuple = NULL;
/* Try to optimistically replace the new_tuple. */
- int tree_res = memtx_tree_insert(&index->tree, new_data,
- &dup_data);
+ int tree_res =
+ memtx_tree_insert(&index->tree, new_data, &dup_data);
if (tree_res) {
diag_set(OutOfMemory, MEMTX_EXTENT_SIZE,
"memtx_tree_index", "replace");
return -1;
}
- uint32_t errcode = replace_check_dup(old_tuple,
- dup_data.tuple, mode);
+ uint32_t errcode =
+ replace_check_dup(old_tuple, dup_data.tuple, mode);
if (errcode) {
memtx_tree_delete(&index->tree, new_data);
if (dup_data.tuple != NULL)
memtx_tree_insert(&index->tree, dup_data, NULL);
- struct space *sp = space_cache_find(base->def->space_id);
+ struct space *sp =
+ space_cache_find(base->def->space_id);
if (sp != NULL)
diag_set(ClientError, errcode, base->def->name,
space_name(sp));
@@ -668,10 +664,11 @@ memtx_tree_index_replace(struct index *base, struct tuple *old_tuple,
*/
static int
memtx_tree_index_replace_multikey_one(struct memtx_tree_index *index,
- struct tuple *old_tuple, struct tuple *new_tuple,
- enum dup_replace_mode mode, hint_t hint,
- struct memtx_tree_data *replaced_data,
- bool *is_multikey_conflict)
+ struct tuple *old_tuple,
+ struct tuple *new_tuple,
+ enum dup_replace_mode mode, hint_t hint,
+ struct memtx_tree_data *replaced_data,
+ bool *is_multikey_conflict)
{
struct memtx_tree_data new_data, dup_data;
new_data.tuple = new_tuple;
@@ -692,7 +689,7 @@ memtx_tree_index_replace_multikey_one(struct memtx_tree_index *index,
*/
*is_multikey_conflict = true;
} else if ((errcode = replace_check_dup(old_tuple, dup_data.tuple,
- mode)) != 0) {
+ mode)) != 0) {
/* Rollback replace. */
memtx_tree_delete(&index->tree, new_data);
if (dup_data.tuple != NULL)
@@ -721,8 +718,9 @@ memtx_tree_index_replace_multikey_one(struct memtx_tree_index *index,
*/
static void
memtx_tree_index_replace_multikey_rollback(struct memtx_tree_index *index,
- struct tuple *new_tuple, struct tuple *replaced_tuple,
- int err_multikey_idx)
+ struct tuple *new_tuple,
+ struct tuple *replaced_tuple,
+ int err_multikey_idx)
{
struct memtx_tree_data data;
if (replaced_tuple != NULL) {
@@ -731,7 +729,7 @@ memtx_tree_index_replace_multikey_rollback(struct memtx_tree_index *index,
data.tuple = replaced_tuple;
uint32_t multikey_count =
tuple_multikey_count(replaced_tuple, cmp_def);
- for (int i = 0; (uint32_t) i < multikey_count; i++) {
+ for (int i = 0; (uint32_t)i < multikey_count; i++) {
data.hint = i;
memtx_tree_insert(&index->tree, data, NULL);
}
@@ -795,8 +793,9 @@ memtx_tree_index_replace_multikey_rollback(struct memtx_tree_index *index,
*/
static int
memtx_tree_index_replace_multikey(struct index *base, struct tuple *old_tuple,
- struct tuple *new_tuple, enum dup_replace_mode mode,
- struct tuple **result)
+ struct tuple *new_tuple,
+ enum dup_replace_mode mode,
+ struct tuple **result)
{
struct memtx_tree_index *index = (struct memtx_tree_index *)base;
struct key_def *cmp_def = memtx_tree_cmp_def(&index->tree);
@@ -805,14 +804,13 @@ memtx_tree_index_replace_multikey(struct index *base, struct tuple *old_tuple,
int multikey_idx = 0, err = 0;
uint32_t multikey_count =
tuple_multikey_count(new_tuple, cmp_def);
- for (; (uint32_t) multikey_idx < multikey_count;
+ for (; (uint32_t)multikey_idx < multikey_count;
multikey_idx++) {
bool is_multikey_conflict;
struct memtx_tree_data replaced_data;
- err = memtx_tree_index_replace_multikey_one(index,
- old_tuple, new_tuple, mode,
- multikey_idx, &replaced_data,
- &is_multikey_conflict);
+ err = memtx_tree_index_replace_multikey_one(
+ index, old_tuple, new_tuple, mode, multikey_idx,
+ &replaced_data, &is_multikey_conflict);
if (err != 0)
break;
if (replaced_data.tuple != NULL &&
@@ -823,8 +821,8 @@ memtx_tree_index_replace_multikey(struct index *base, struct tuple *old_tuple,
}
}
if (err != 0) {
- memtx_tree_index_replace_multikey_rollback(index,
- new_tuple, *result, multikey_idx);
+ memtx_tree_index_replace_multikey_rollback(
+ index, new_tuple, *result, multikey_idx);
return -1;
}
if (*result != NULL) {
@@ -837,7 +835,7 @@ memtx_tree_index_replace_multikey(struct index *base, struct tuple *old_tuple,
data.tuple = old_tuple;
uint32_t multikey_count =
tuple_multikey_count(old_tuple, cmp_def);
- for (int i = 0; (uint32_t) i < multikey_count; i++) {
+ for (int i = 0; (uint32_t)i < multikey_count; i++) {
data.hint = i;
memtx_tree_delete_value(&index->tree, data, NULL);
}
@@ -850,9 +848,9 @@ static const char *
func_index_key_dummy_alloc(struct tuple *tuple, const char *key,
uint32_t key_sz)
{
- (void) tuple;
- (void) key_sz;
- return (void*) key;
+ (void)tuple;
+ (void)key_sz;
+ return (void *)key;
}
/**
@@ -873,8 +871,8 @@ struct func_key_undo *
func_key_undo_new(struct region *region)
{
size_t size;
- struct func_key_undo *undo = region_alloc_object(region, typeof(*undo),
- &size);
+ struct func_key_undo *undo =
+ region_alloc_object(region, typeof(*undo), &size);
if (undo == NULL) {
diag_set(OutOfMemory, size, "region_alloc_object", "undo");
return NULL;
@@ -916,8 +914,8 @@ memtx_tree_func_index_replace_rollback(struct memtx_tree_index *index,
*/
static int
memtx_tree_func_index_replace(struct index *base, struct tuple *old_tuple,
- struct tuple *new_tuple, enum dup_replace_mode mode,
- struct tuple **result)
+ struct tuple *new_tuple,
+ enum dup_replace_mode mode, struct tuple **result)
{
struct memtx_tree_index *index = (struct memtx_tree_index *)base;
struct index_def *index_def = index->base.def;
@@ -940,7 +938,7 @@ memtx_tree_func_index_replace(struct index *base, struct tuple *old_tuple,
const char *key;
struct func_key_undo *undo;
while ((err = key_list_iterator_next(&it, &key)) == 0 &&
- key != NULL) {
+ key != NULL) {
/* Perform insertion, log it in list. */
undo = func_key_undo_new(region);
if (undo == NULL) {
@@ -954,10 +952,9 @@ memtx_tree_func_index_replace(struct index *base, struct tuple *old_tuple,
bool is_multikey_conflict;
struct memtx_tree_data old_data;
old_data.tuple = NULL;
- err = memtx_tree_index_replace_multikey_one(index,
- old_tuple, new_tuple,
- mode, (hint_t)key, &old_data,
- &is_multikey_conflict);
+ err = memtx_tree_index_replace_multikey_one(
+ index, old_tuple, new_tuple, mode, (hint_t)key,
+ &old_data, &is_multikey_conflict);
if (err != 0)
break;
if (old_data.tuple != NULL && !is_multikey_conflict) {
@@ -984,7 +981,7 @@ memtx_tree_func_index_replace(struct index *base, struct tuple *old_tuple,
* from undo list.
*/
tuple_chunk_delete(new_tuple,
- (const char *)old_data.hint);
+ (const char *)old_data.hint);
rlist_foreach_entry(undo, &new_keys, link) {
if (undo->key.hint == old_data.hint) {
rlist_del(&undo->link);
@@ -994,8 +991,8 @@ memtx_tree_func_index_replace(struct index *base, struct tuple *old_tuple,
}
}
if (key != NULL || err != 0) {
- memtx_tree_func_index_replace_rollback(index,
- &old_keys, &new_keys);
+ memtx_tree_func_index_replace_rollback(index, &old_keys,
+ &new_keys);
goto end;
}
if (*result != NULL) {
@@ -1019,7 +1016,7 @@ memtx_tree_func_index_replace(struct index *base, struct tuple *old_tuple,
data.tuple = old_tuple;
const char *key;
while (key_list_iterator_next(&it, &key) == 0 && key != NULL) {
- data.hint = (hint_t) key;
+ data.hint = (hint_t)key;
deleted_data.tuple = NULL;
memtx_tree_delete_value(&index->tree, data,
&deleted_data);
@@ -1028,7 +1025,8 @@ memtx_tree_func_index_replace(struct index *base, struct tuple *old_tuple,
* Release related hint on
* successful node deletion.
*/
- tuple_chunk_delete(deleted_data.tuple,
+ tuple_chunk_delete(
+ deleted_data.tuple,
(const char *)deleted_data.hint);
}
}
@@ -1126,14 +1124,16 @@ memtx_tree_index_build_array_append(struct memtx_tree_index *index,
}
assert(index->build_array_size <= index->build_array_alloc_size);
if (index->build_array_size == index->build_array_alloc_size) {
- index->build_array_alloc_size = index->build_array_alloc_size +
- DIV_ROUND_UP(index->build_array_alloc_size, 2);
+ index->build_array_alloc_size =
+ index->build_array_alloc_size +
+ DIV_ROUND_UP(index->build_array_alloc_size, 2);
struct memtx_tree_data *tmp =
realloc(index->build_array,
index->build_array_alloc_size * sizeof(*tmp));
if (tmp == NULL) {
- diag_set(OutOfMemory, index->build_array_alloc_size *
- sizeof(*tmp), "memtx_tree_index", "build_next");
+ diag_set(OutOfMemory,
+ index->build_array_alloc_size * sizeof(*tmp),
+ "memtx_tree_index", "build_next");
return -1;
}
index->build_array = tmp;
@@ -1210,7 +1210,8 @@ error:
*/
static void
memtx_tree_index_build_array_deduplicate(struct memtx_tree_index *index,
- void (*destroy)(struct tuple *tuple, const char *hint))
+ void (*destroy)(struct tuple *tuple,
+ const char *hint))
{
if (index->build_array_size == 0)
return;
@@ -1218,7 +1219,7 @@ memtx_tree_index_build_array_deduplicate(struct memtx_tree_index *index,
size_t w_idx = 0, r_idx = 1;
while (r_idx < index->build_array_size) {
if (index->build_array[w_idx].tuple !=
- index->build_array[r_idx].tuple ||
+ index->build_array[r_idx].tuple ||
tuple_compare(index->build_array[w_idx].tuple,
index->build_array[w_idx].hint,
index->build_array[r_idx].tuple,
@@ -1234,8 +1235,8 @@ memtx_tree_index_build_array_deduplicate(struct memtx_tree_index *index,
}
if (destroy != NULL) {
/* Destroy deduplicated entries. */
- for (r_idx = w_idx + 1;
- r_idx < index->build_array_size; r_idx++) {
+ for (r_idx = w_idx + 1; r_idx < index->build_array_size;
+ r_idx++) {
destroy(index->build_array[r_idx].tuple,
(const char *)index->build_array[r_idx].hint);
}
@@ -1285,8 +1286,8 @@ tree_snapshot_iterator_free(struct snapshot_iterator *iterator)
assert(iterator->free == tree_snapshot_iterator_free);
struct tree_snapshot_iterator *it =
(struct tree_snapshot_iterator *)iterator;
- memtx_leave_delayed_free_mode((struct memtx_engine *)
- it->index->base.engine);
+ memtx_leave_delayed_free_mode(
+ (struct memtx_engine *)it->index->base.engine);
memtx_tree_iterator_destroy(&it->index->tree, &it->tree_iterator);
index_unref(&it->index->base);
memtx_tx_snapshot_cleaner_destroy(&it->cleaner);
@@ -1335,7 +1336,7 @@ memtx_tree_index_create_snapshot_iterator(struct index *base)
{
struct memtx_tree_index *index = (struct memtx_tree_index *)base;
struct tree_snapshot_iterator *it =
- (struct tree_snapshot_iterator *) calloc(1, sizeof(*it));
+ (struct tree_snapshot_iterator *)calloc(1, sizeof(*it));
if (it == NULL) {
diag_set(OutOfMemory, sizeof(struct tree_snapshot_iterator),
"memtx_tree_index", "create_snapshot_iterator");
@@ -1356,7 +1357,7 @@ memtx_tree_index_create_snapshot_iterator(struct index *base)
it->tree_iterator = memtx_tree_iterator_first(&index->tree);
memtx_tree_iterator_freeze(&index->tree, &it->tree_iterator);
memtx_enter_delayed_free_mode((struct memtx_engine *)base->engine);
- return (struct snapshot_iterator *) it;
+ return (struct snapshot_iterator *)it;
}
static const struct index_vtab memtx_tree_index_vtab = {
@@ -1368,7 +1369,7 @@ static const struct index_vtab memtx_tree_index_vtab = {
/* .update_def = */ memtx_tree_index_update_def,
/* .depends_on_pk = */ memtx_tree_index_depends_on_pk,
/* .def_change_requires_rebuild = */
- memtx_index_def_change_requires_rebuild,
+ memtx_index_def_change_requires_rebuild,
/* .size = */ memtx_tree_index_size,
/* .bsize = */ memtx_tree_index_bsize,
/* .min = */ generic_index_min,
@@ -1379,7 +1380,7 @@ static const struct index_vtab memtx_tree_index_vtab = {
/* .replace = */ memtx_tree_index_replace,
/* .create_iterator = */ memtx_tree_index_create_iterator,
/* .create_snapshot_iterator = */
- memtx_tree_index_create_snapshot_iterator,
+ memtx_tree_index_create_snapshot_iterator,
/* .stat = */ generic_index_stat,
/* .compact = */ generic_index_compact,
/* .reset_stat = */ generic_index_reset_stat,
@@ -1398,7 +1399,7 @@ static const struct index_vtab memtx_tree_index_multikey_vtab = {
/* .update_def = */ memtx_tree_index_update_def,
/* .depends_on_pk = */ memtx_tree_index_depends_on_pk,
/* .def_change_requires_rebuild = */
- memtx_index_def_change_requires_rebuild,
+ memtx_index_def_change_requires_rebuild,
/* .size = */ memtx_tree_index_size,
/* .bsize = */ memtx_tree_index_bsize,
/* .min = */ generic_index_min,
@@ -1409,7 +1410,7 @@ static const struct index_vtab memtx_tree_index_multikey_vtab = {
/* .replace = */ memtx_tree_index_replace_multikey,
/* .create_iterator = */ memtx_tree_index_create_iterator,
/* .create_snapshot_iterator = */
- memtx_tree_index_create_snapshot_iterator,
+ memtx_tree_index_create_snapshot_iterator,
/* .stat = */ generic_index_stat,
/* .compact = */ generic_index_compact,
/* .reset_stat = */ generic_index_reset_stat,
@@ -1428,7 +1429,7 @@ static const struct index_vtab memtx_tree_func_index_vtab = {
/* .update_def = */ memtx_tree_index_update_def,
/* .depends_on_pk = */ memtx_tree_index_depends_on_pk,
/* .def_change_requires_rebuild = */
- memtx_index_def_change_requires_rebuild,
+ memtx_index_def_change_requires_rebuild,
/* .size = */ memtx_tree_index_size,
/* .bsize = */ memtx_tree_index_bsize,
/* .min = */ generic_index_min,
@@ -1439,7 +1440,7 @@ static const struct index_vtab memtx_tree_func_index_vtab = {
/* .replace = */ memtx_tree_func_index_replace,
/* .create_iterator = */ memtx_tree_index_create_iterator,
/* .create_snapshot_iterator = */
- memtx_tree_index_create_snapshot_iterator,
+ memtx_tree_index_create_snapshot_iterator,
/* .stat = */ generic_index_stat,
/* .compact = */ generic_index_compact,
/* .reset_stat = */ generic_index_reset_stat,
@@ -1464,7 +1465,7 @@ static const struct index_vtab memtx_tree_disabled_index_vtab = {
/* .update_def = */ generic_index_update_def,
/* .depends_on_pk = */ generic_index_depends_on_pk,
/* .def_change_requires_rebuild = */
- generic_index_def_change_requires_rebuild,
+ generic_index_def_change_requires_rebuild,
/* .size = */ generic_index_size,
/* .bsize = */ generic_index_bsize,
/* .min = */ generic_index_min,
@@ -1475,7 +1476,7 @@ static const struct index_vtab memtx_tree_disabled_index_vtab = {
/* .replace = */ disabled_index_replace,
/* .create_iterator = */ generic_index_create_iterator,
/* .create_snapshot_iterator = */
- generic_index_create_snapshot_iterator,
+ generic_index_create_snapshot_iterator,
/* .stat = */ generic_index_stat,
/* .compact = */ generic_index_compact,
/* .reset_stat = */ generic_index_reset_stat,
@@ -1491,8 +1492,8 @@ memtx_tree_index_new(struct memtx_engine *memtx, struct index_def *def)
struct memtx_tree_index *index =
(struct memtx_tree_index *)calloc(1, sizeof(*index));
if (index == NULL) {
- diag_set(OutOfMemory, sizeof(*index),
- "malloc", "struct memtx_tree_index");
+ diag_set(OutOfMemory, sizeof(*index), "malloc",
+ "struct memtx_tree_index");
return NULL;
}
const struct index_vtab *vtab;
@@ -1506,8 +1507,8 @@ memtx_tree_index_new(struct memtx_engine *memtx, struct index_def *def)
} else {
vtab = &memtx_tree_index_vtab;
}
- if (index_create(&index->base, (struct engine *)memtx,
- vtab, def) != 0) {
+ if (index_create(&index->base, (struct engine *)memtx, vtab, def) !=
+ 0) {
free(index);
return NULL;
}
@@ -1515,7 +1516,8 @@ memtx_tree_index_new(struct memtx_engine *memtx, struct index_def *def)
/* See comment to memtx_tree_index_update_def(). */
struct key_def *cmp_def;
cmp_def = def->opts.is_unique && !def->key_def->is_nullable ?
- index->base.def->key_def : index->base.def->cmp_def;
+ index->base.def->key_def :
+ index->base.def->cmp_def;
memtx_tree_create(&index->tree, cmp_def, memtx_index_extent_alloc,
memtx_index_extent_free, memtx);
diff --git a/src/box/memtx_tx.c b/src/box/memtx_tx.c
index 55748ad..c1d8c10 100644
--- a/src/box/memtx_tx.c
+++ b/src/box/memtx_tx.c
@@ -59,8 +59,7 @@ memtx_tx_story_key_hash(const struct tuple *a)
#define MH_SOURCE
#include "salad/mhash.h"
-struct tx_manager
-{
+struct tx_manager {
/**
* List of all transactions that are in a read view.
* New transactions are added to the tail of this list,
@@ -83,7 +82,7 @@ enum {
* searching and deleting no more used memtx_tx_stories per creation of
* a new story.
*/
- TX_MANAGER_GC_STEPS_SIZE = 2,
+ TX_MANAGER_GC_STEPS_SIZE = 2,
};
/** That's a definition, see declaration for description. */
@@ -99,8 +98,8 @@ memtx_tx_manager_init()
for (size_t i = 0; i < BOX_INDEX_MAX; i++) {
size_t item_size = sizeof(struct memtx_story) +
i * sizeof(struct memtx_story_link);
- mempool_create(&txm.memtx_tx_story_pool[i],
- cord_slab_cache(), item_size);
+ mempool_create(&txm.memtx_tx_story_pool[i], cord_slab_cache(),
+ item_size);
}
txm.history = mh_history_new();
rlist_create(&txm.all_stories);
@@ -109,8 +108,7 @@ memtx_tx_manager_init()
void
memtx_tx_manager_free()
-{
-}
+{}
int
memtx_tx_cause_conflict(struct txn *breaker, struct txn *victim)
@@ -121,12 +119,12 @@ memtx_tx_cause_conflict(struct txn *breaker, struct txn *victim)
while (r1 != &breaker->conflict_list &&
r2 != &victim->conflicted_by_list) {
tracker = rlist_entry(r1, struct tx_conflict_tracker,
- in_conflict_list);
+ in_conflict_list);
assert(tracker->breaker == breaker);
if (tracker->victim == victim)
break;
tracker = rlist_entry(r2, struct tx_conflict_tracker,
- in_conflicted_by_list);
+ in_conflicted_by_list);
assert(tracker->victim == victim);
if (tracker->breaker == breaker)
break;
@@ -143,9 +141,8 @@ memtx_tx_cause_conflict(struct txn *breaker, struct txn *victim)
rlist_del(&tracker->in_conflicted_by_list);
} else {
size_t size;
- tracker = region_alloc_object(&victim->region,
- struct tx_conflict_tracker,
- &size);
+ tracker = region_alloc_object(
+ &victim->region, struct tx_conflict_tracker, &size);
if (tracker == NULL) {
diag_set(OutOfMemory, size, "tx region",
"conflict_tracker");
@@ -196,18 +193,18 @@ memtx_tx_story_new(struct space *space, struct tuple *tuple)
uint32_t index_count = space->index_count;
assert(index_count < BOX_INDEX_MAX);
struct mempool *pool = &txm.memtx_tx_story_pool[index_count];
- struct memtx_story *story = (struct memtx_story *) mempool_alloc(pool);
+ struct memtx_story *story = (struct memtx_story *)mempool_alloc(pool);
if (story == NULL) {
- size_t item_size = sizeof(struct memtx_story) +
- index_count *
- sizeof(struct memtx_story_link);
+ size_t item_size =
+ sizeof(struct memtx_story) +
+ index_count * sizeof(struct memtx_story_link);
diag_set(OutOfMemory, item_size, "mempool_alloc", "story");
return NULL;
}
story->tuple = tuple;
const struct memtx_story **put_story =
- (const struct memtx_story **) &story;
+ (const struct memtx_story **)&story;
struct memtx_story **empty = NULL;
mh_int_t pos = mh_history_put(txm.history, put_story, &empty, 0);
if (pos == mh_end(txm.history)) {
@@ -289,7 +286,6 @@ memtx_tx_story_delete_del_stmt(struct memtx_story *story)
memtx_tx_story_delete(story);
}
-
/**
* Find a story of a @a tuple. The story expected to be present (assert).
*/
@@ -309,8 +305,8 @@ memtx_tx_story_get(struct tuple *tuple)
static struct tuple *
memtx_tx_story_older_tuple(struct memtx_story_link *link)
{
- return link->older.is_story ? link->older.story->tuple
- : link->older.tuple;
+ return link->older.is_story ? link->older.story->tuple :
+ link->older.tuple;
}
/**
@@ -318,8 +314,7 @@ memtx_tx_story_older_tuple(struct memtx_story_link *link)
*/
static void
memtx_tx_story_link_story(struct memtx_story *story,
- struct memtx_story *older_story,
- uint32_t index)
+ struct memtx_story *older_story, uint32_t index)
{
assert(older_story != NULL);
struct memtx_story_link *link = &story->link[index];
@@ -336,8 +331,7 @@ memtx_tx_story_link_story(struct memtx_story *story,
* dirty -find and link with the corresponding story.
*/
static void
-memtx_tx_story_link_tuple(struct memtx_story *story,
- struct tuple *older_tuple,
+memtx_tx_story_link_tuple(struct memtx_story *story, struct tuple *older_tuple,
uint32_t index)
{
struct memtx_story_link *link = &story->link[index];
@@ -347,9 +341,8 @@ memtx_tx_story_link_tuple(struct memtx_story *story,
if (older_tuple == NULL)
return;
if (older_tuple->is_dirty) {
- memtx_tx_story_link_story(story,
- memtx_tx_story_get(older_tuple),
- index);
+ memtx_tx_story_link_story(
+ story, memtx_tx_story_get(older_tuple), index);
return;
}
link->older.tuple = older_tuple;
@@ -389,16 +382,14 @@ memtx_tx_story_gc_step()
/* Lowest read view PSN */
int64_t lowest_rv_psm = txn_last_psn;
if (!rlist_empty(&txm.read_view_txs)) {
- struct txn *txn =
- rlist_first_entry(&txm.read_view_txs, struct txn,
- in_read_view_txs);
+ struct txn *txn = rlist_first_entry(
+ &txm.read_view_txs, struct txn, in_read_view_txs);
assert(txn->rv_psn != 0);
lowest_rv_psm = txn->rv_psn;
}
- struct memtx_story *story =
- rlist_entry(txm.traverse_all_stories, struct memtx_story,
- in_all_stories);
+ struct memtx_story *story = rlist_entry(
+ txm.traverse_all_stories, struct memtx_story, in_all_stories);
txm.traverse_all_stories = txm.traverse_all_stories->next;
if (story->add_stmt != NULL || story->del_stmt != NULL ||
@@ -511,8 +502,7 @@ memtx_tx_story_is_visible(struct memtx_story *story, struct txn *txn,
/**
* Temporary (allocated on region) struct that stores a conflicting TX.
*/
-struct memtx_tx_conflict
-{
+struct memtx_tx_conflict {
/* The transaction that will conflict us upon commit. */
struct txn *breaker;
/* Link in single-linked list. */
@@ -552,12 +542,10 @@ memtx_tx_save_conflict(struct txn *breaker,
* @return 0 on success, -1 on memory error.
*/
static int
-memtx_tx_story_find_visible_tuple(struct memtx_story *story,
- struct txn_stmt *stmt,
- uint32_t index,
- struct tuple **visible_replaced,
- struct memtx_tx_conflict **collected_conflicts,
- struct region *region)
+memtx_tx_story_find_visible_tuple(
+ struct memtx_story *story, struct txn_stmt *stmt, uint32_t index,
+ struct tuple **visible_replaced,
+ struct memtx_tx_conflict **collected_conflicts, struct region *region)
{
while (true) {
if (!story->link[index].older.is_story) {
@@ -603,7 +591,6 @@ memtx_tx_story_find_visible_tuple(struct memtx_story *story,
collected_conflicts,
region) != 0)
return -1;
-
}
}
return 0;
@@ -647,18 +634,16 @@ memtx_tx_history_add_stmt(struct txn_stmt *stmt, struct tuple *old_tuple,
add_story_linked++;
struct tuple *visible_replaced = NULL;
- if (memtx_tx_story_find_visible_tuple(add_story, stmt, i,
- &visible_replaced,
- &collected_conflicts,
- region) != 0)
+ if (memtx_tx_story_find_visible_tuple(
+ add_story, stmt, i, &visible_replaced,
+ &collected_conflicts, region) != 0)
goto fail;
uint32_t errcode;
errcode = replace_check_dup(old_tuple, visible_replaced,
i == 0 ? mode : DUP_INSERT);
if (errcode != 0) {
- diag_set(ClientError, errcode,
- index->def->name,
+ diag_set(ClientError, errcode, index->def->name,
space_name(space));
goto fail;
}
@@ -685,8 +670,8 @@ memtx_tx_history_add_stmt(struct txn_stmt *stmt, struct tuple *old_tuple,
if (del_tuple->is_dirty) {
del_story = memtx_tx_story_get(del_tuple);
} else {
- del_story = memtx_tx_story_new_del_stmt(del_tuple,
- stmt);
+ del_story =
+ memtx_tx_story_new_del_stmt(del_tuple, stmt);
if (del_story == NULL)
goto fail;
del_story_created = true;
@@ -729,7 +714,7 @@ memtx_tx_history_add_stmt(struct txn_stmt *stmt, struct tuple *old_tuple,
tuple_ref(*result);
return 0;
- fail:
+fail:
if (add_story != NULL) {
while (add_story_linked > 0) {
--add_story_linked;
@@ -739,15 +724,14 @@ memtx_tx_history_add_stmt(struct txn_stmt *stmt, struct tuple *old_tuple,
struct memtx_story_link *link = &add_story->link[i];
struct tuple *was = memtx_tx_story_older_tuple(link);
struct tuple *unused;
- if (index_replace(index, new_tuple, was,
- DUP_INSERT, &unused) != 0) {
+ if (index_replace(index, new_tuple, was, DUP_INSERT,
+ &unused) != 0) {
diag_log();
unreachable();
panic("failed to rollback change");
}
memtx_tx_story_unlink(stmt->add_story, i);
-
}
memtx_tx_story_delete_add_stmt(stmt->add_story);
}
@@ -778,7 +762,8 @@ memtx_tx_history_rollback_stmt(struct txn_stmt *stmt)
if (link->newer_story == NULL) {
struct tuple *unused;
struct index *index = stmt->space->index[i];
- struct tuple *was = memtx_tx_story_older_tuple(link);
+ struct tuple *was =
+ memtx_tx_story_older_tuple(link);
if (index_replace(index, story->tuple, was,
DUP_INSERT, &unused) != 0) {
diag_log();
@@ -791,7 +776,8 @@ memtx_tx_history_rollback_stmt(struct txn_stmt *stmt)
assert(newer->link[i].older.story == story);
memtx_tx_story_unlink(newer, i);
if (link->older.is_story) {
- struct memtx_story *to = link->older.story;
+ struct memtx_story *to =
+ link->older.story;
memtx_tx_story_link_story(newer, to, i);
} else {
struct tuple *to = link->older.tuple;
@@ -834,7 +820,7 @@ memtx_tx_history_prepare_stmt(struct txn_stmt *stmt)
* Note that if stmt->add_story == NULL, the index_count is set to 0,
* and we will not enter the loop.
*/
- for (uint32_t i = 0; i < index_count; ) {
+ for (uint32_t i = 0; i < index_count;) {
if (!story->link[i].older.is_story) {
/* tuple is old. */
i++;
@@ -894,13 +880,11 @@ memtx_tx_history_prepare_stmt(struct txn_stmt *stmt)
memtx_tx_story_unlink(story, i);
if (old_story->link[i].older.is_story) {
- struct memtx_story *to =
- old_story->link[i].older.story;
+ struct memtx_story *to = old_story->link[i].older.story;
memtx_tx_story_unlink(old_story, i);
memtx_tx_story_link_story(story, to, i);
} else {
- struct tuple *to =
- old_story->link[i].older.tuple;
+ struct tuple *to = old_story->link[i].older.tuple;
memtx_tx_story_unlink(old_story, i);
memtx_tx_story_link_tuple(story, to, i);
}
@@ -1019,10 +1003,9 @@ memtx_tx_on_space_delete(struct space *space)
{
/* Just clear pointer to space, it will be handled in GC. */
while (!rlist_empty(&space->memtx_stories)) {
- struct memtx_story *story
- = rlist_first_entry(&space->memtx_stories,
- struct memtx_story,
- in_space_stories);
+ struct memtx_story *story =
+ rlist_first_entry(&space->memtx_stories,
+ struct memtx_story, in_space_stories);
story->space = NULL;
rlist_del(&story->in_space_stories);
}
@@ -1095,13 +1078,12 @@ memtx_tx_track_read(struct txn *txn, struct space *space, struct tuple *tuple)
struct rlist *r1 = story->reader_list.next;
struct rlist *r2 = txn->read_set.next;
while (r1 != &story->reader_list && r2 != &txn->read_set) {
- tracker = rlist_entry(r1, struct tx_read_tracker,
- in_reader_list);
+ tracker =
+ rlist_entry(r1, struct tx_read_tracker, in_reader_list);
assert(tracker->story == story);
if (tracker->reader == txn)
break;
- tracker = rlist_entry(r2, struct tx_read_tracker,
- in_read_set);
+ tracker = rlist_entry(r2, struct tx_read_tracker, in_read_set);
assert(tracker->reader == txn);
if (tracker->story == story)
break;
@@ -1139,8 +1121,7 @@ memtx_tx_snapshot_cleaner_hash(const struct tuple *a)
return u ^ (u >> 32);
}
-struct memtx_tx_snapshot_cleaner_entry
-{
+struct memtx_tx_snapshot_cleaner_entry {
struct tuple *from;
struct tuple *to;
};
@@ -1165,8 +1146,8 @@ memtx_tx_snapshot_cleaner_create(struct memtx_tx_snapshot_cleaner *cleaner,
return 0;
struct mh_snapshot_cleaner_t *ht = mh_snapshot_cleaner_new();
if (ht == NULL) {
- diag_set(OutOfMemory, sizeof(*ht),
- index_name, "snapshot cleaner");
+ diag_set(OutOfMemory, sizeof(*ht), index_name,
+ "snapshot cleaner");
free(ht);
return -1;
}
@@ -1174,19 +1155,18 @@ memtx_tx_snapshot_cleaner_create(struct memtx_tx_snapshot_cleaner *cleaner,
struct memtx_story *story;
rlist_foreach_entry(story, &space->memtx_stories, in_space_stories) {
struct tuple *tuple = story->tuple;
- struct tuple *clean =
- memtx_tx_tuple_clarify_slow(NULL, space, tuple, 0, 0,
- true);
+ struct tuple *clean = memtx_tx_tuple_clarify_slow(
+ NULL, space, tuple, 0, 0, true);
if (clean == tuple)
continue;
struct memtx_tx_snapshot_cleaner_entry entry;
entry.from = tuple;
entry.to = clean;
- mh_int_t res = mh_snapshot_cleaner_put(ht, &entry, NULL, 0);
+ mh_int_t res = mh_snapshot_cleaner_put(ht, &entry, NULL, 0);
if (res == mh_end(ht)) {
- diag_set(OutOfMemory, sizeof(entry),
- index_name, "snapshot rollback entry");
+ diag_set(OutOfMemory, sizeof(entry), index_name,
+ "snapshot rollback entry");
mh_snapshot_cleaner_delete(ht);
return -1;
}
@@ -1204,7 +1184,7 @@ memtx_tx_snapshot_clarify_slow(struct memtx_tx_snapshot_cleaner *cleaner,
struct mh_snapshot_cleaner_t *ht = cleaner->ht;
while (true) {
- mh_int_t pos = mh_snapshot_cleaner_find(ht, tuple, 0);
+ mh_int_t pos = mh_snapshot_cleaner_find(ht, tuple, 0);
if (pos == mh_end(ht))
break;
struct memtx_tx_snapshot_cleaner_entry *entry =
@@ -1216,7 +1196,6 @@ memtx_tx_snapshot_clarify_slow(struct memtx_tx_snapshot_cleaner *cleaner,
return tuple;
}
-
void
memtx_tx_snapshot_cleaner_destroy(struct memtx_tx_snapshot_cleaner *cleaner)
{
diff --git a/src/box/memtx_tx.h b/src/box/memtx_tx.h
index 25a2038..22a5872 100644
--- a/src/box/memtx_tx.h
+++ b/src/box/memtx_tx.h
@@ -297,8 +297,8 @@ memtx_tx_track_read(struct txn *txn, struct space *space, struct tuple *tuple);
*/
static inline struct tuple *
memtx_tx_tuple_clarify(struct txn *txn, struct space *space,
- struct tuple *tuple, uint32_t index,
- uint32_t mk_index, bool is_prepared_ok)
+ struct tuple *tuple, uint32_t index, uint32_t mk_index,
+ bool is_prepared_ok)
{
if (!memtx_tx_manager_use_mvcc_engine)
return tuple;
diff --git a/src/box/merger.c b/src/box/merger.c
index fff12f9..8d33ba4 100644
--- a/src/box/merger.c
+++ b/src/box/merger.c
@@ -39,8 +39,8 @@
#define HEAP_FORWARD_DECLARATION
#include "salad/heap.h"
-#include "diag.h" /* diag_set() */
-#include "box/tuple.h" /* tuple_ref(), tuple_unref(),
+#include "diag.h" /* diag_set() */
+#include "box/tuple.h" /* tuple_ref(), tuple_unref(),
tuple_validate() */
#include "box/tuple_format.h" /* box_tuple_format_new(),
tuple_format_*() */
@@ -210,8 +210,8 @@ static int
merger_set_sources(struct merger *merger, struct merge_source **sources,
uint32_t source_count)
{
- const size_t nodes_size = sizeof(struct merger_heap_node) *
- source_count;
+ const size_t nodes_size =
+ sizeof(struct merger_heap_node) * source_count;
struct merger_heap_node *nodes = malloc(nodes_size);
if (nodes == NULL) {
diag_set(OutOfMemory, nodes_size, "malloc",
@@ -227,7 +227,6 @@ merger_set_sources(struct merger *merger, struct merge_source **sources,
return 0;
}
-
struct merge_source *
merger_new(struct key_def *key_def, struct merge_source **sources,
uint32_t source_count, bool reverse)
diff --git a/src/box/mp_error.cc b/src/box/mp_error.cc
index 36fbcef..4f5fa53 100644
--- a/src/box/mp_error.cc
+++ b/src/box/mp_error.cc
@@ -69,9 +69,7 @@
/**
* MP_ERROR keys
*/
-enum {
- MP_ERROR_STACK = 0x00
-};
+enum { MP_ERROR_STACK = 0x00 };
/**
* Keys of individual error in the stack.
@@ -98,13 +96,8 @@ enum {
};
static const char *const mp_error_field_to_json_key[MP_ERROR_MAX] = {
- "\"type\": ",
- "\"file\": ",
- "\"line\": ",
- "\"message\": ",
- "\"errno\": ",
- "\"code\": ",
- "\"fields\": ",
+ "\"type\": ", "\"file\": ", "\"line\": ", "\"message\": ",
+ "\"errno\": ", "\"code\": ", "\"fields\": ",
};
/**
@@ -253,7 +246,7 @@ error_build_xc(struct mp_error *mp_error)
struct error *err = NULL;
if (mp_error->type == NULL || mp_error->message == NULL ||
mp_error->file == NULL) {
-missing_fields:
+ missing_fields:
diag_set(ClientError, ER_INVALID_MSGPACK,
"Missing mandatory error fields");
return NULL;
@@ -286,14 +279,14 @@ missing_fields:
err = new XlogGapError(mp_error->file, mp_error->line,
mp_error->message);
} else if (strcmp(mp_error->type, "SystemError") == 0) {
- err = new SystemError(mp_error->file, mp_error->line,
- "%s", mp_error->message);
+ err = new SystemError(mp_error->file, mp_error->line, "%s",
+ mp_error->message);
} else if (strcmp(mp_error->type, "SocketError") == 0) {
err = new SocketError(mp_error->file, mp_error->line, "", "");
error_format_msg(err, "%s", mp_error->message);
} else if (strcmp(mp_error->type, "OutOfMemory") == 0) {
- err = new OutOfMemory(mp_error->file, mp_error->line,
- 0, "", "");
+ err = new OutOfMemory(mp_error->file, mp_error->line, 0, "",
+ "");
} else if (strcmp(mp_error->type, "TimedOut") == 0) {
err = new TimedOut(mp_error->file, mp_error->line);
} else if (strcmp(mp_error->type, "ChannelIsClosed") == 0) {
@@ -304,17 +297,17 @@ missing_fields:
err = new LuajitError(mp_error->file, mp_error->line,
mp_error->message);
} else if (strcmp(mp_error->type, "IllegalParams") == 0) {
- err = new IllegalParams(mp_error->file, mp_error->line,
- "%s", mp_error->message);
+ err = new IllegalParams(mp_error->file, mp_error->line, "%s",
+ mp_error->message);
} else if (strcmp(mp_error->type, "CollationError") == 0) {
- err = new CollationError(mp_error->file, mp_error->line,
- "%s", mp_error->message);
+ err = new CollationError(mp_error->file, mp_error->line, "%s",
+ mp_error->message);
} else if (strcmp(mp_error->type, "SwimError") == 0) {
- err = new SwimError(mp_error->file, mp_error->line,
- "%s", mp_error->message);
+ err = new SwimError(mp_error->file, mp_error->line, "%s",
+ mp_error->message);
} else if (strcmp(mp_error->type, "CryptoError") == 0) {
- err = new CryptoError(mp_error->file, mp_error->line,
- "%s", mp_error->message);
+ err = new CryptoError(mp_error->file, mp_error->line, "%s",
+ mp_error->message);
} else {
err = new ClientError(mp_error->file, mp_error->line,
ER_UNKNOWN);
@@ -347,7 +340,8 @@ mp_decode_and_copy_str(const char **data, struct region *region)
}
uint32_t str_len;
const char *str = mp_decode_str(data, &str_len);
- return region_strdup(region, str, str_len);;
+ return region_strdup(region, str, str_len);
+ ;
}
static inline bool
@@ -415,7 +409,7 @@ mp_decode_error_one(const char **data)
goto error;
uint64_t key = mp_decode_uint(data);
- switch(key) {
+ switch (key) {
case MP_ERROR_TYPE:
mp_err.type = mp_decode_and_copy_str(data, region);
if (mp_err.type == NULL)
@@ -540,7 +534,7 @@ error_unpack_unsafe(const char **data)
return NULL;
}
uint64_t key = mp_decode_uint(data);
- switch(key) {
+ switch (key) {
case MP_ERROR_STACK: {
if (mp_typeof(**data) != MP_ARRAY) {
diag_set(ClientError, ER_INVALID_MSGPACK,
@@ -579,7 +573,7 @@ error_unpack_unsafe(const char **data)
#define MP_ERROR_PRINT_DEFINITION
#define MP_PRINT_FUNC snprintf
#define MP_PRINT_SUFFIX snprint
-#define MP_PRINT_2(total, func, ...) \
+#define MP_PRINT_2(total, func, ...) \
SNPRINT(total, func, buf, size, __VA_ARGS__)
#define MP_PRINT_ARGS_DECL char *buf, int size
#include __FILE__
@@ -587,12 +581,13 @@ error_unpack_unsafe(const char **data)
#define MP_ERROR_PRINT_DEFINITION
#define MP_PRINT_FUNC fprintf
#define MP_PRINT_SUFFIX fprint
-#define MP_PRINT_2(total, func, ...) do { \
- int bytes = func(file, __VA_ARGS__); \
- if (bytes < 0) \
- return -1; \
- total += bytes; \
-} while (0)
+#define MP_PRINT_2(total, func, ...) \
+ do { \
+ int bytes = func(file, __VA_ARGS__); \
+ if (bytes < 0) \
+ return -1; \
+ total += bytes; \
+ } while (0)
#define MP_PRINT_ARGS_DECL FILE *file
#include __FILE__
@@ -624,16 +619,15 @@ error_unpack_unsafe(const char **data)
* turn it into a template.
*/
-#define MP_CONCAT4_R(a, b, c, d) a##b##c##d
-#define MP_CONCAT4(a, b, c, d) MP_CONCAT4_R(a, b, c, d)
-#define MP_PRINT(total, ...) MP_PRINT_2(total, MP_PRINT_FUNC, \
- __VA_ARGS__)
+#define MP_CONCAT4_R(a, b, c, d) a##b##c##d
+#define MP_CONCAT4(a, b, c, d) MP_CONCAT4_R(a, b, c, d)
+#define MP_PRINT(total, ...) MP_PRINT_2(total, MP_PRINT_FUNC, __VA_ARGS__)
-#define mp_func_name(name) MP_CONCAT4(mp_, MP_PRINT_SUFFIX, _, name)
-#define mp_print_error_one mp_func_name(error_one)
-#define mp_print_error_stack mp_func_name(error_stack)
-#define mp_print_error mp_func_name(error)
-#define mp_print_common mp_func_name(recursion)
+#define mp_func_name(name) MP_CONCAT4(mp_, MP_PRINT_SUFFIX, _, name)
+#define mp_print_error_one mp_func_name(error_one)
+#define mp_print_error_stack mp_func_name(error_stack)
+#define mp_print_error mp_func_name(error)
+#define mp_print_common mp_func_name(recursion)
static int
mp_print_error_one(MP_PRINT_ARGS_DECL, const char **data, int depth)
@@ -710,7 +704,7 @@ mp_print_error(MP_PRINT_ARGS_DECL, const char **data, int depth)
if (mp_typeof(**data) != MP_UINT)
return -1;
uint64_t key = mp_decode_uint(data);
- switch(key) {
+ switch (key) {
case MP_ERROR_STACK: {
MP_PRINT(total, "\"stack\": ");
MP_PRINT_2(total, mp_print_error_stack, data, depth);
diff --git a/src/box/msgpack.c b/src/box/msgpack.c
index 1723dea..7a4aad3 100644
--- a/src/box/msgpack.c
+++ b/src/box/msgpack.c
@@ -42,7 +42,7 @@ msgpack_fprint_ext(FILE *file, const char **data, int depth)
const char **orig = data;
int8_t type;
uint32_t len = mp_decode_extl(data, &type);
- switch(type) {
+ switch (type) {
case MP_DECIMAL:
return mp_fprint_decimal(file, data, len);
case MP_UUID:
@@ -60,7 +60,7 @@ msgpack_snprint_ext(char *buf, int size, const char **data, int depth)
const char **orig = data;
int8_t type;
uint32_t len = mp_decode_extl(data, &type);
- switch(type) {
+ switch (type) {
case MP_DECIMAL:
return mp_snprint_decimal(buf, size, data, len);
case MP_UUID:
diff --git a/src/box/opt_def.c b/src/box/opt_def.c
index e282085..76ed42f 100644
--- a/src/box/opt_def.c
+++ b/src/box/opt_def.c
@@ -60,7 +60,7 @@ opt_set(void *opts, const struct opt_def *def, const char **val,
uint32_t str_len;
const char *str;
char *ptr;
- char *opt = ((char *) opts) + def->offset;
+ char *opt = ((char *)opts) + def->offset;
switch (def->type) {
case OPT_BOOL:
if (mp_typeof(**val) != MP_BOOL)
@@ -98,7 +98,7 @@ opt_set(void *opts, const struct opt_def *def, const char **val,
goto type_mismatch_err;
str = mp_decode_str(val, &str_len);
if (str_len > 0) {
- ptr = (char *) region_alloc(region, str_len + 1);
+ ptr = (char *)region_alloc(region, str_len + 1);
if (ptr == NULL) {
diag_set(OutOfMemory, str_len + 1, "region",
"opt string");
@@ -106,7 +106,7 @@ opt_set(void *opts, const struct opt_def *def, const char **val,
}
memcpy(ptr, str, str_len);
ptr[str_len] = '\0';
- assert (strlen(ptr) == str_len);
+ assert(strlen(ptr) == str_len);
} else {
ptr = NULL;
}
@@ -122,7 +122,7 @@ opt_set(void *opts, const struct opt_def *def, const char **val,
} else {
ival = def->to_enum(str, str_len);
}
- switch(def->enum_size) {
+ switch (def->enum_size) {
case sizeof(uint8_t):
store_u8(opt, (uint8_t)ival);
break;
@@ -175,7 +175,7 @@ opts_parse_key(void *opts, const struct opt_def *reg, const char *key,
return opt_set(opts, def, data, region, errcode, field_no);
}
- if (! skip_unknown_options) {
+ if (!skip_unknown_options) {
char *errmsg = tt_static_buf();
snprintf(errmsg, TT_STATIC_BUF_LEN, "unexpected option '%.*s'",
key_len, key);
diff --git a/src/box/opt_def.h b/src/box/opt_def.h
index 2154441..6640678 100644
--- a/src/box/opt_def.h
+++ b/src/box/opt_def.h
@@ -40,15 +40,15 @@ extern "C" {
#endif /* defined(__cplusplus) */
enum opt_type {
- OPT_BOOL, /* bool */
- OPT_UINT32, /* uint32_t */
- OPT_INT64, /* int64_t */
- OPT_FLOAT, /* double */
- OPT_STR, /* char[] */
- OPT_STRPTR, /* char* */
- OPT_ENUM, /* enum */
- OPT_ARRAY, /* array */
- OPT_LEGACY, /* any type, skipped */
+ OPT_BOOL, /* bool */
+ OPT_UINT32, /* uint32_t */
+ OPT_INT64, /* int64_t */
+ OPT_FLOAT, /* double */
+ OPT_STR, /* char[] */
+ OPT_STRPTR, /* char* */
+ OPT_ENUM, /* enum */
+ OPT_ARRAY, /* array */
+ OPT_LEGACY, /* any type, skipped */
opt_type_MAX,
};
@@ -94,23 +94,43 @@ struct opt_def {
};
};
-#define OPT_DEF(key, type, opts, field) \
- { key, type, offsetof(opts, field), sizeof(((opts *)0)->field), \
- NULL, 0, NULL, 0, {NULL} }
-
-#define OPT_DEF_ENUM(key, enum_name, opts, field, to_enum) \
- { key, OPT_ENUM, offsetof(opts, field), sizeof(int), #enum_name, \
- sizeof(enum enum_name), enum_name##_strs, enum_name##_MAX, \
- {(void *)to_enum} }
-
-#define OPT_DEF_ARRAY(key, opts, field, to_array) \
- { key, OPT_ARRAY, offsetof(opts, field), sizeof(((opts *)0)->field), \
- NULL, 0, NULL, 0, {(void *)to_array} }
-
-#define OPT_DEF_LEGACY(key) \
- { key, OPT_LEGACY, 0, 0, NULL, 0, NULL, 0, {NULL} }
-
-#define OPT_END {NULL, opt_type_MAX, 0, 0, NULL, 0, NULL, 0, {NULL}}
+#define OPT_DEF(key, type, opts, field) \
+ { \
+ key, type, offsetof(opts, field), sizeof(((opts *)0)->field), \
+ NULL, 0, NULL, 0, \
+ { \
+ NULL \
+ } \
+ }
+
+#define OPT_DEF_ENUM(key, enum_name, opts, field, to_enum) \
+ { \
+ key, OPT_ENUM, offsetof(opts, field), sizeof(int), #enum_name, \
+ sizeof(enum enum_name), enum_name##_strs, \
+ enum_name##_MAX, \
+ { \
+ (void *)to_enum \
+ } \
+ }
+
+#define OPT_DEF_ARRAY(key, opts, field, to_array) \
+ { \
+ key, OPT_ARRAY, offsetof(opts, field), \
+ sizeof(((opts *)0)->field), NULL, 0, NULL, 0, \
+ { \
+ (void *)to_array \
+ } \
+ }
+
+#define OPT_DEF_LEGACY(key) \
+ { \
+ key, OPT_LEGACY, 0, 0, NULL, 0, NULL, 0, { NULL } \
+ }
+
+#define OPT_END \
+ { \
+ NULL, opt_type_MAX, 0, 0, NULL, 0, NULL, 0, { NULL } \
+ }
struct region;
diff --git a/src/box/port.h b/src/box/port.h
index 43d0f9d..8714422 100644
--- a/src/box/port.h
+++ b/src/box/port.h
@@ -93,9 +93,9 @@ struct sql_value;
/** Port implementation used with vdbe memory variables. */
struct port_vdbemem {
- const struct port_vtab *vtab;
- struct sql_value *mem;
- uint32_t mem_count;
+ const struct port_vtab *vtab;
+ struct sql_value *mem;
+ uint32_t mem_count;
};
static_assert(sizeof(struct port_vdbemem) <= sizeof(struct port),
diff --git a/src/box/raft.c b/src/box/raft.c
index 0b6c373..b752c97 100644
--- a/src/box/raft.c
+++ b/src/box/raft.c
@@ -338,20 +338,23 @@ raft_process_msg(const struct raft_request *req, uint32_t source)
assert(source > 0);
assert(source != instance_id);
if (req->term == 0 || req->state == 0) {
- diag_set(ClientError, ER_PROTOCOL, "Raft term and state can't "
+ diag_set(ClientError, ER_PROTOCOL,
+ "Raft term and state can't "
"be zero");
return -1;
}
if (req->state == RAFT_STATE_CANDIDATE &&
(req->vote != source || req->vclock == NULL)) {
- diag_set(ClientError, ER_PROTOCOL, "Candidate should always "
+ diag_set(ClientError, ER_PROTOCOL,
+ "Candidate should always "
"vote for self and provide its vclock");
return -1;
}
/* Outdated request. */
if (req->term < raft.volatile_term) {
say_info("RAFT: the message is ignored due to outdated term - "
- "current term is %u", raft.volatile_term);
+ "current term is %u",
+ raft.volatile_term);
return 0;
}
@@ -428,8 +431,8 @@ raft_process_msg(const struct raft_request *req, uint32_t source)
raft.vote_count += !was_set;
if (raft.vote_count < quorum) {
say_info("RAFT: accepted vote for self, vote "
- "count is %d/%d", raft.vote_count,
- quorum);
+ "count is %d/%d",
+ raft.vote_count, quorum);
break;
}
raft_sm_become_leader();
@@ -441,7 +444,8 @@ raft_process_msg(const struct raft_request *req, uint32_t source)
if (req->state != RAFT_STATE_LEADER) {
if (source == raft.leader) {
say_info("RAFT: the node %u has resigned from the "
- "leader role", raft.leader);
+ "leader role",
+ raft.leader);
raft_sm_schedule_new_election();
}
return 0;
@@ -457,7 +461,8 @@ raft_process_msg(const struct raft_request *req, uint32_t source)
*/
if (raft.leader != 0) {
say_warn("RAFT: conflicting leader detected in one term - "
- "known is %u, received %u", raft.leader, source);
+ "known is %u, received %u",
+ raft.leader, source);
return 0;
}
@@ -531,8 +536,7 @@ raft_write_request(const struct raft_request *req)
struct region *region = &fiber()->gc;
uint32_t svp = region_used(region);
struct xrow_header row;
- char buf[sizeof(struct journal_entry) +
- sizeof(struct xrow_header *)];
+ char buf[sizeof(struct journal_entry) + sizeof(struct xrow_header *)];
struct journal_entry *entry = (struct journal_entry *)buf;
entry->rows[0] = &row;
@@ -567,7 +571,7 @@ raft_worker_handle_io(void)
struct raft_request req;
if (raft_is_fully_on_disk()) {
-end_dump:
+ end_dump:
raft.is_write_in_progress = false;
/*
* The state machine is stable. Can see now, to what state to
@@ -634,8 +638,7 @@ raft_worker_handle_broadcast(void)
assert(raft.vote == instance_id);
req.vclock = &replicaset.vclock;
}
- replicaset_foreach(replica)
- relay_push_raft(replica->relay, &req);
+ replicaset_foreach(replica) relay_push_raft(replica->relay, &req);
raft.is_broadcast_scheduled = false;
}
@@ -820,8 +823,8 @@ raft_sm_wait_election_end(void)
(raft.state == RAFT_STATE_CANDIDATE &&
raft.volatile_vote == instance_id));
assert(raft.leader == 0);
- double election_timeout = raft.election_timeout +
- raft_new_random_election_shift();
+ double election_timeout =
+ raft.election_timeout + raft_new_random_election_shift();
ev_timer_set(&raft.timer, election_timeout, election_timeout);
ev_timer_start(loop(), &raft.timer);
}
diff --git a/src/box/recovery.cc b/src/box/recovery.cc
index cd33e76..35693b9 100644
--- a/src/box/recovery.cc
+++ b/src/box/recovery.cc
@@ -83,17 +83,13 @@ struct recovery *
recovery_new(const char *wal_dirname, bool force_recovery,
const struct vclock *vclock)
{
- struct recovery *r = (struct recovery *)
- calloc(1, sizeof(*r));
+ struct recovery *r = (struct recovery *)calloc(1, sizeof(*r));
if (r == NULL) {
- tnt_raise(OutOfMemory, sizeof(*r), "malloc",
- "struct recovery");
+ tnt_raise(OutOfMemory, sizeof(*r), "malloc", "struct recovery");
}
- auto guard = make_scoped_guard([=]{
- free(r);
- });
+ auto guard = make_scoped_guard([=] { free(r); });
xdir_create(&r->wal_dir, wal_dirname, XLOG, &INSTANCE_UUID,
&xlog_opts_default);
@@ -152,8 +148,7 @@ recovery_close_log(struct recovery *r)
if (xlog_cursor_is_eof(&r->cursor)) {
say_info("done `%s'", r->cursor.name);
} else {
- say_warn("file `%s` wasn't correctly closed",
- r->cursor.name);
+ say_warn("file `%s` wasn't correctly closed", r->cursor.name);
}
xlog_cursor_close(&r->cursor, false);
trigger_run_xc(&r->on_close_log, NULL);
@@ -325,8 +320,7 @@ recover_remaining_wals(struct recovery *r, struct xstream *stream,
}
for (clock = vclockset_match(&r->wal_dir.index, &r->vclock);
- clock != NULL;
- clock = vclockset_next(&r->wal_dir.index, clock)) {
+ clock != NULL; clock = vclockset_next(&r->wal_dir.index, clock)) {
if (stop_vclock != NULL &&
clock->signature >= stop_vclock->signature) {
break;
@@ -345,7 +339,7 @@ recover_remaining_wals(struct recovery *r, struct xstream *stream,
say_info("recover from `%s'", r->cursor.name);
-recover_current_wal:
+ recover_current_wal:
recover_xlog(r, stream, stop_vclock);
}
@@ -364,7 +358,6 @@ recovery_finalize(struct recovery *r)
recovery_close_log(r);
}
-
/* }}} */
/* {{{ Local recovery: support of hot standby and replication relay */
@@ -405,9 +398,8 @@ public:
{
f = fiber();
events = 0;
- if ((size_t)snprintf(dir_path, sizeof(dir_path), "%s", wal_dir) >=
- sizeof(dir_path)) {
-
+ if ((size_t)snprintf(dir_path, sizeof(dir_path), "%s",
+ wal_dir) >= sizeof(dir_path)) {
panic("path too long: %s", wal_dir);
}
@@ -433,8 +425,7 @@ public:
* Note: .file_path valid iff file_stat is active.
*/
if (path && ev_is_active(&file_stat) &&
- strcmp(file_path, path) == 0) {
-
+ strcmp(file_path, path) == 0) {
return;
}
@@ -443,9 +434,8 @@ public:
if (path == NULL)
return;
- if ((size_t)snprintf(file_path, sizeof(file_path), "%s", path) >=
- sizeof(file_path)) {
-
+ if ((size_t)snprintf(file_path, sizeof(file_path), "%s",
+ path) >= sizeof(file_path)) {
panic("path too long: %s", path);
}
ev_stat_set(&file_stat, file_path, 0.0);
@@ -465,8 +455,7 @@ hot_standby_f(va_list ap)
WalSubscription subscription(r->wal_dir.dirname);
- while (! fiber_is_cancelled()) {
-
+ while (!fiber_is_cancelled()) {
/*
* Recover until there is no new stuff which appeared in
* the log dir while recovery was running.
@@ -491,7 +480,8 @@ hot_standby_f(va_list ap)
} while (end > start && !xlog_cursor_is_open(&r->cursor));
subscription.set_log_path(xlog_cursor_is_open(&r->cursor) ?
- r->cursor.name : NULL);
+ r->cursor.name :
+ NULL);
bool timed_out = false;
if (subscription.events == 0) {
@@ -505,7 +495,7 @@ hot_standby_f(va_list ap)
}
scan_dir = timed_out ||
- (subscription.events & WAL_EVENT_ROTATE) != 0;
+ (subscription.events & WAL_EVENT_ROTATE) != 0;
subscription.events = 0;
}
diff --git a/src/box/recovery.h b/src/box/recovery.h
index b8d8395..774e76e 100644
--- a/src/box/recovery.h
+++ b/src/box/recovery.h
@@ -73,7 +73,7 @@ recovery_delete(struct recovery *r);
* WAL directory.
*/
void
-recovery_scan(struct recovery *r, struct vclock *end_vclock,
+recovery_scan(struct recovery *r, struct vclock *end_vclock,
struct vclock *gc_vclock);
void
diff --git a/src/box/relay.cc b/src/box/relay.cc
index 096f455..941c60d 100644
--- a/src/box/relay.cc
+++ b/src/box/relay.cc
@@ -144,8 +144,8 @@ struct relay {
struct {
/* Align to prevent false-sharing with tx thread */
alignas(CACHELINE_SIZE)
- /** Known relay vclock. */
- struct vclock vclock;
+ /** Known relay vclock. */
+ struct vclock vclock;
/**
* True if the relay needs Raft updates. It can live fine
* without sending Raft updates, if it is a relay to an
@@ -155,7 +155,7 @@ struct relay {
} tx;
};
-struct diag*
+struct diag *
relay_get_diag(struct relay *relay)
{
return &relay->diag;
@@ -189,10 +189,10 @@ relay_send_row(struct xstream *stream, struct xrow_header *row);
struct relay *
relay_new(struct replica *replica)
{
- struct relay *relay = (struct relay *) calloc(1, sizeof(struct relay));
+ struct relay *relay = (struct relay *)calloc(1, sizeof(struct relay));
if (relay == NULL) {
diag_set(OutOfMemory, sizeof(struct relay), "malloc",
- "struct relay");
+ "struct relay");
return NULL;
}
relay->replica = replica;
@@ -206,7 +206,7 @@ relay_new(struct replica *replica)
static void
relay_start(struct relay *relay, int fd, uint64_t sync,
- void (*stream_write)(struct xstream *, struct xrow_header *))
+ void (*stream_write)(struct xstream *, struct xrow_header *))
{
xstream_create(&relay->stream, stream_write);
/*
@@ -256,8 +256,9 @@ static void
relay_stop(struct relay *relay)
{
struct relay_gc_msg *gc_msg, *next_gc_msg;
- stailq_foreach_entry_safe(gc_msg, next_gc_msg,
- &relay->pending_gc, in_pending) {
+ stailq_foreach_entry_safe(gc_msg, next_gc_msg, &relay->pending_gc,
+ in_pending)
+ {
free(gc_msg);
}
stailq_create(&relay->pending_gc);
@@ -290,7 +291,7 @@ relay_set_cord_name(int fd)
char name[FIBER_NAME_MAX];
struct sockaddr_storage peer;
socklen_t addrlen = sizeof(peer);
- if (getpeername(fd, ((struct sockaddr*)&peer), &addrlen) == 0) {
+ if (getpeername(fd, ((struct sockaddr *)&peer), &addrlen) == 0) {
snprintf(name, sizeof(name), "relay/%s",
sio_strfaddr((struct sockaddr *)&peer, addrlen));
} else {
@@ -315,9 +316,8 @@ relay_initial_join(int fd, uint64_t sync, struct vclock *vclock)
/* Freeze a read view in engines. */
struct engine_join_ctx ctx;
engine_prepare_join_xc(&ctx);
- auto join_guard = make_scoped_guard([&] {
- engine_complete_join(&ctx);
- });
+ auto join_guard =
+ make_scoped_guard([&] { engine_complete_join(&ctx); });
/*
* Sync WAL to make sure that all changes visible from
@@ -355,8 +355,8 @@ relay_final_join_f(va_list ap)
/* Send all WALs until stop_vclock */
assert(relay->stream.write != NULL);
- recover_remaining_wals(relay->r, &relay->stream,
- &relay->stop_vclock, true);
+ recover_remaining_wals(relay->r, &relay->stream, &relay->stop_vclock,
+ true);
assert(vclock_compare(&relay->r->vclock, &relay->stop_vclock) == 0);
return 0;
}
@@ -378,8 +378,8 @@ relay_final_join(int fd, uint64_t sync, struct vclock *start_vclock,
relay->r = recovery_new(wal_dir(), false, start_vclock);
vclock_copy(&relay->stop_vclock, stop_vclock);
- int rc = cord_costart(&relay->cord, "final_join",
- relay_final_join_f, relay);
+ int rc = cord_costart(&relay->cord, "final_join", relay_final_join_f,
+ relay);
if (rc == 0)
rc = cord_cojoin(&relay->cord);
if (rc != 0)
@@ -423,9 +423,8 @@ tx_status_update(struct cmsg *msg)
txn_limbo_ack(&txn_limbo, status->relay->replica->id,
vclock_get(&status->vclock, instance_id));
}
- static const struct cmsg_hop route[] = {
- {relay_status_update, NULL}
- };
+ static const struct cmsg_hop route[] = { { relay_status_update,
+ NULL } };
cmsg_init(msg, route);
cpipe_push(&status->relay->relay_pipe, msg);
}
@@ -444,9 +443,7 @@ tx_gc_advance(struct cmsg *msg)
static int
relay_on_close_log_f(struct trigger *trigger, void * /* event */)
{
- static const struct cmsg_hop route[] = {
- {tx_gc_advance, NULL}
- };
+ static const struct cmsg_hop route[] = { { tx_gc_advance, NULL } };
struct relay *relay = (struct relay *)trigger->data;
struct relay_gc_msg *m = (struct relay_gc_msg *)malloc(sizeof(*m));
if (m == NULL) {
@@ -477,7 +474,8 @@ static inline void
relay_schedule_pending_gc(struct relay *relay, const struct vclock *vclock)
{
struct relay_gc_msg *curr, *next, *gc_msg = NULL;
- stailq_foreach_entry_safe(curr, next, &relay->pending_gc, in_pending) {
+ stailq_foreach_entry_safe(curr, next, &relay->pending_gc, in_pending)
+ {
/*
* We may delete a WAL file only if its vclock is
* less than or equal to the vclock acknowledged by
@@ -548,8 +546,9 @@ relay_reader_f(va_list ap)
try {
while (!fiber_is_cancelled()) {
struct xrow_header xrow;
- coio_read_xrow_timeout_xc(&io, &ibuf, &xrow,
- replication_disconnect_timeout());
+ coio_read_xrow_timeout_xc(
+ &io, &ibuf, &xrow,
+ replication_disconnect_timeout());
/* vclock is followed while decoding, zeroing it. */
vclock_create(&relay->recv_vclock);
xrow_decode_vclock_xc(&xrow, &relay->recv_vclock);
@@ -706,8 +705,8 @@ relay_subscribe_f(va_list ap)
*/
while (!fiber_is_cancelled()) {
double timeout = replication_timeout;
- struct errinj *inj = errinj(ERRINJ_RELAY_REPORT_INTERVAL,
- ERRINJ_DOUBLE);
+ struct errinj *inj =
+ errinj(ERRINJ_RELAY_REPORT_INTERVAL, ERRINJ_DOUBLE);
if (inj != NULL && inj->dparam != 0)
timeout = inj->dparam;
@@ -741,9 +740,8 @@ relay_subscribe_f(va_list ap)
if (vclock_sum(&relay->status_msg.vclock) ==
vclock_sum(send_vclock))
continue;
- static const struct cmsg_hop route[] = {
- {tx_status_update, NULL}
- };
+ static const struct cmsg_hop route[] = { { tx_status_update,
+ NULL } };
cmsg_init(&relay->status_msg.msg, route);
vclock_copy(&relay->status_msg.vclock, send_vclock);
relay->status_msg.relay = relay;
@@ -775,8 +773,8 @@ relay_subscribe_f(va_list ap)
fiber_join(reader);
/* Destroy cpipe to tx. */
- cbus_unpair(&relay->tx_pipe, &relay->relay_pipe,
- NULL, NULL, cbus_process);
+ cbus_unpair(&relay->tx_pipe, &relay->relay_pipe, NULL, NULL,
+ cbus_process);
cbus_endpoint_destroy(&relay->endpoint, cbus_process);
relay_exit(relay);
@@ -817,8 +815,8 @@ relay_subscribe(struct replica *replica, int fd, uint64_t sync,
relay->id_filter = replica_id_filter;
- int rc = cord_costart(&relay->cord, "subscribe",
- relay_subscribe_f, relay);
+ int rc = cord_costart(&relay->cord, "subscribe", relay_subscribe_f,
+ relay);
if (rc == 0)
rc = cord_cojoin(&relay->cord);
if (rc != 0)
@@ -982,12 +980,11 @@ relay_send_row(struct xstream *stream, struct xrow_header *packet)
packet->replica_id != relay->replica->id ||
packet->lsn <= vclock_get(&relay->local_vclock_at_subscribe,
packet->replica_id)) {
- struct errinj *inj = errinj(ERRINJ_RELAY_BREAK_LSN,
- ERRINJ_INT);
+ struct errinj *inj = errinj(ERRINJ_RELAY_BREAK_LSN, ERRINJ_INT);
if (inj != NULL && packet->lsn == inj->iparam) {
packet->lsn = inj->iparam - 1;
say_warn("injected broken lsn: %lld",
- (long long) packet->lsn);
+ (long long)packet->lsn);
}
relay_send(relay, packet);
}
diff --git a/src/box/relay.h b/src/box/relay.h
index b32e2ea..6707bba 100644
--- a/src/box/relay.h
+++ b/src/box/relay.h
@@ -70,7 +70,7 @@ void
relay_delete(struct relay *relay);
/** Get last relay's diagnostic error */
-struct diag*
+struct diag *
relay_get_diag(struct relay *relay);
/** Return the current state of relay. */
diff --git a/src/box/replication.cc b/src/box/replication.cc
index c19f8c6..a4d0dae 100644
--- a/src/box/replication.cc
+++ b/src/box/replication.cc
@@ -47,13 +47,13 @@ uint32_t instance_id = REPLICA_ID_NIL;
struct tt_uuid INSTANCE_UUID;
struct tt_uuid REPLICASET_UUID;
-double replication_timeout = 1.0; /* seconds */
+double replication_timeout = 1.0; /* seconds */
double replication_connect_timeout = 30.0; /* seconds */
int replication_connect_quorum = REPLICATION_CONNECT_QUORUM_ALL;
double replication_sync_lag = 10.0; /* seconds */
int replication_synchro_quorum = 1;
double replication_synchro_timeout = 5.0; /* seconds */
-double replication_sync_timeout = 300.0; /* seconds */
+double replication_sync_timeout = 300.0; /* seconds */
bool replication_skip_conflict = false;
bool replication_anon = false;
@@ -65,11 +65,11 @@ replica_compare_by_uuid(const struct replica *a, const struct replica *b)
return tt_uuid_compare(&a->uuid, &b->uuid);
}
-rb_gen(MAYBE_UNUSED static, replica_hash_, replica_hash_t,
- struct replica, in_hash, replica_compare_by_uuid);
+rb_gen(MAYBE_UNUSED static, replica_hash_, replica_hash_t, struct replica,
+ in_hash, replica_compare_by_uuid);
-#define replica_hash_foreach_safe(hash, item, next) \
- for (item = replica_hash_first(hash); \
+#define replica_hash_foreach_safe(hash, item, next) \
+ for (item = replica_hash_first(hash); \
item != NULL && ((next = replica_hash_next(hash, item)) || 1); \
item = next)
@@ -109,8 +109,7 @@ replication_free(void)
* cbus upon shutdown, which could lead to segfaults.
* So cancel them.
*/
- replicaset_foreach(replica)
- relay_cancel(replica->relay);
+ replicaset_foreach(replica) relay_cancel(replica->relay);
diag_destroy(&replicaset.applier.diag);
}
@@ -120,12 +119,11 @@ replica_check_id(uint32_t replica_id)
{
if (replica_id == REPLICA_ID_NIL) {
diag_set(ClientError, ER_REPLICA_ID_IS_RESERVED,
- (unsigned) replica_id);
+ (unsigned)replica_id);
return -1;
}
if (replica_id >= VCLOCK_MAX) {
- diag_set(ClientError, ER_REPLICA_MAX,
- (unsigned) replica_id);
+ diag_set(ClientError, ER_REPLICA_MAX, (unsigned)replica_id);
return -1;
}
/*
@@ -140,7 +138,7 @@ replica_check_id(uint32_t replica_id)
*/
if (!replicaset.is_joining && replica_id == instance_id) {
diag_set(ClientError, ER_LOCAL_INSTANCE_ID_IS_READ_ONLY,
- (unsigned) replica_id);
+ (unsigned)replica_id);
return -1;
}
return 0;
@@ -161,8 +159,8 @@ replica_on_applier_state_f(struct trigger *trigger, void *event);
static struct replica *
replica_new(void)
{
- struct replica *replica = (struct replica *)
- malloc(sizeof(struct replica));
+ struct replica *replica =
+ (struct replica *)malloc(sizeof(struct replica));
if (replica == NULL) {
tnt_raise(OutOfMemory, sizeof(*replica), "malloc",
"struct replica");
@@ -178,8 +176,8 @@ replica_new(void)
replica->applier = NULL;
replica->gc = NULL;
rlist_create(&replica->in_anon);
- trigger_create(&replica->on_applier_state,
- replica_on_applier_state_f, NULL, NULL);
+ trigger_create(&replica->on_applier_state, replica_on_applier_state_f,
+ NULL, NULL);
replica->applier_sync_state = APPLIER_DISCONNECTED;
latch_create(&replica->order_latch);
return replica;
@@ -248,8 +246,8 @@ replica_set_id(struct replica *replica, uint32_t replica_id)
}
replicaset.replica_by_id[replica_id] = replica;
++replicaset.registered_count;
- say_info("assigned id %d to replica %s",
- replica->id, tt_uuid_str(&replica->uuid));
+ say_info("assigned id %d to replica %s", replica->id,
+ tt_uuid_str(&replica->uuid));
replica->anon = false;
}
@@ -306,8 +304,7 @@ replica_set_applier(struct replica *replica, struct applier *applier)
{
assert(replica->applier == NULL);
replica->applier = applier;
- trigger_add(&replica->applier->on_state,
- &replica->on_applier_state);
+ trigger_add(&replica->applier->on_state, &replica->on_applier_state);
}
void
@@ -449,8 +446,8 @@ static int
replica_on_applier_state_f(struct trigger *trigger, void *event)
{
(void)event;
- struct replica *replica = container_of(trigger,
- struct replica, on_applier_state);
+ struct replica *replica =
+ container_of(trigger, struct replica, on_applier_state);
switch (replica->applier->state) {
case APPLIER_INITIAL_JOIN:
replicaset.is_joining = true;
@@ -508,8 +505,9 @@ replicaset_update(struct applier **appliers, int count)
struct replica *replica, *next;
struct applier *applier;
- auto uniq_guard = make_scoped_guard([&]{
- replica_hash_foreach_safe(&uniq, replica, next) {
+ auto uniq_guard = make_scoped_guard([&] {
+ replica_hash_foreach_safe(&uniq, replica, next)
+ {
replica_hash_remove(&uniq, replica);
replica_clear_applier(replica);
replica_delete(replica);
@@ -555,8 +553,8 @@ replicaset_update(struct applier **appliers, int count)
/* Prune old appliers */
while (!rlist_empty(&replicaset.anon)) {
- replica = rlist_first_entry(&replicaset.anon,
- typeof(*replica), in_anon);
+ replica = rlist_first_entry(&replicaset.anon, typeof(*replica),
+ in_anon);
applier = replica->applier;
replica_clear_applier(replica);
rlist_del_entry(replica, in_anon);
@@ -564,7 +562,8 @@ replicaset_update(struct applier **appliers, int count)
applier_stop(applier);
applier_delete(applier);
}
- replicaset_foreach(replica) {
+ replicaset_foreach(replica)
+ {
if (replica->applier == NULL)
continue;
applier = replica->applier;
@@ -580,11 +579,12 @@ replicaset_update(struct applier **appliers, int count)
replicaset.applier.loading = 0;
replicaset.applier.synced = 0;
- replica_hash_foreach_safe(&uniq, replica, next) {
+ replica_hash_foreach_safe(&uniq, replica, next)
+ {
replica_hash_remove(&uniq, replica);
- struct replica *orig = replica_hash_search(&replicaset.hash,
- replica);
+ struct replica *orig =
+ replica_hash_search(&replicaset.hash, replica);
if (orig != NULL) {
/* Use existing struct replica */
replica_set_applier(orig, replica->applier);
@@ -603,7 +603,8 @@ replicaset_update(struct applier **appliers, int count)
rlist_swap(&replicaset.anon, &anon_replicas);
assert(replica_hash_first(&uniq) == NULL);
- replica_hash_foreach_safe(&replicaset.hash, replica, next) {
+ replica_hash_foreach_safe(&replicaset.hash, replica, next)
+ {
if (replica_is_orphan(replica)) {
replica_hash_remove(&replicaset.hash, replica);
replicaset.anon_count -= replica->anon;
@@ -633,8 +634,8 @@ struct applier_on_connect {
static int
applier_on_connect_f(struct trigger *trigger, void *event)
{
- struct applier_on_connect *on_connect = container_of(trigger,
- struct applier_on_connect, base);
+ struct applier_on_connect *on_connect =
+ container_of(trigger, struct applier_on_connect, base);
struct replicaset_connect_state *state = on_connect->state;
struct applier *applier = (struct applier *)event;
@@ -655,8 +656,7 @@ applier_on_connect_f(struct trigger *trigger, void *event)
}
void
-replicaset_connect(struct applier **appliers, int count,
- bool connect_quorum)
+replicaset_connect(struct applier **appliers, int count, bool connect_quorum)
{
if (count == 0) {
/* Cleanup the replica set. */
@@ -707,7 +707,8 @@ replicaset_connect(struct applier **appliers, int count,
struct applier *applier = appliers[i];
struct applier_on_connect *trigger = &triggers[i];
/* Register a trigger to wake us up when peer is connected */
- trigger_create(&trigger->base, applier_on_connect_f, NULL, NULL);
+ trigger_create(&trigger->base, applier_on_connect_f, NULL,
+ NULL);
trigger->state = &state;
trigger_add(&applier->on_state, &trigger->base);
/* Start background connection */
@@ -782,19 +783,20 @@ bool
replicaset_needs_rejoin(struct replica **master)
{
struct replica *leader = NULL;
- replicaset_foreach(replica) {
+ replicaset_foreach(replica)
+ {
struct applier *applier = replica->applier;
/*
* Skip the local instance, we shouldn't perform a
* check against our own gc vclock.
*/
- if (applier == NULL || tt_uuid_is_equal(&replica->uuid,
- &INSTANCE_UUID))
+ if (applier == NULL ||
+ tt_uuid_is_equal(&replica->uuid, &INSTANCE_UUID))
continue;
const struct ballot *ballot = &applier->ballot;
- if (vclock_compare(&ballot->gc_vclock,
- &replicaset.vclock) <= 0) {
+ if (vclock_compare(&ballot->gc_vclock, &replicaset.vclock) <=
+ 0) {
/*
* There's at least one master that still stores
* WALs needed by this instance. Proceed to local
@@ -804,11 +806,14 @@ replicaset_needs_rejoin(struct replica **master)
}
const char *uuid_str = tt_uuid_str(&replica->uuid);
- const char *addr_str = sio_strfaddr(&applier->addr,
- applier->addr_len);
- const char *local_vclock_str = vclock_to_string(&replicaset.vclock);
- const char *remote_vclock_str = vclock_to_string(&ballot->vclock);
- const char *gc_vclock_str = vclock_to_string(&ballot->gc_vclock);
+ const char *addr_str =
+ sio_strfaddr(&applier->addr, applier->addr_len);
+ const char *local_vclock_str =
+ vclock_to_string(&replicaset.vclock);
+ const char *remote_vclock_str =
+ vclock_to_string(&ballot->vclock);
+ const char *gc_vclock_str =
+ vclock_to_string(&ballot->gc_vclock);
say_info("can't follow %s at %s: required %s available %s",
uuid_str, addr_str, local_vclock_str, gc_vclock_str);
@@ -829,7 +834,7 @@ replicaset_needs_rejoin(struct replica **master)
/* Prefer a master with the max vclock. */
if (leader == NULL ||
vclock_sum(&ballot->vclock) >
- vclock_sum(&leader->applier->ballot.vclock))
+ vclock_sum(&leader->applier->ballot.vclock))
leader = replica;
}
if (leader == NULL)
@@ -843,7 +848,8 @@ void
replicaset_follow(void)
{
struct replica *replica;
- replicaset_foreach(replica) {
+ replicaset_foreach(replica)
+ {
/* Resume connected appliers. */
if (replica->applier != NULL)
applier_resume(replica->applier);
@@ -877,8 +883,8 @@ replicaset_sync(void)
*/
double deadline = ev_monotonic_now(loop()) + replication_sync_timeout;
while (replicaset.applier.synced < quorum &&
- replicaset.applier.connected +
- replicaset.applier.loading >= quorum) {
+ replicaset.applier.connected + replicaset.applier.loading >=
+ quorum) {
if (fiber_cond_wait_deadline(&replicaset.applier.cond,
deadline) != 0)
break;
@@ -957,7 +963,8 @@ static struct replica *
replicaset_round(bool skip_ro)
{
struct replica *leader = NULL;
- replicaset_foreach(replica) {
+ replicaset_foreach(replica)
+ {
struct applier *applier = replica->applier;
if (applier == NULL)
continue;
@@ -978,7 +985,8 @@ replicaset_round(bool skip_ro)
* Try to find a replica which has already left
* orphan mode.
*/
- if (applier->ballot.is_loading && ! leader->applier->ballot.is_loading)
+ if (applier->ballot.is_loading &&
+ !leader->applier->ballot.is_loading)
continue;
/*
* Choose the replica with the most advanced
@@ -986,12 +994,13 @@ replicaset_round(bool skip_ro)
* with the same vclock, prefer the one with
* the lowest uuid.
*/
- int cmp = vclock_compare_ignore0(&applier->ballot.vclock,
- &leader->applier->ballot.vclock);
+ int cmp =
+ vclock_compare_ignore0(&applier->ballot.vclock,
+ &leader->applier->ballot.vclock);
if (cmp < 0)
continue;
- if (cmp == 0 && tt_uuid_compare(&replica->uuid,
- &leader->uuid) > 0)
+ if (cmp == 0 &&
+ tt_uuid_compare(&replica->uuid, &leader->uuid) > 0)
continue;
leader = replica;
}
diff --git a/src/box/replication.h b/src/box/replication.h
index 3e46c59..3282b76 100644
--- a/src/box/replication.h
+++ b/src/box/replication.h
@@ -357,9 +357,9 @@ replicaset_first(void);
struct replica *
replicaset_next(struct replica *replica);
-#define replicaset_foreach(var) \
- for (struct replica *var = replicaset_first(); \
- var != NULL; var = replicaset_next(var))
+#define replicaset_foreach(var) \
+ for (struct replica *var = replicaset_first(); var != NULL; \
+ var = replicaset_next(var))
/**
* Set numeric replica-set-local id of remote replica.
@@ -380,7 +380,7 @@ void
replica_clear_applier(struct replica *replica);
void
-replica_set_applier(struct replica * replica, struct applier * applier);
+replica_set_applier(struct replica *replica, struct applier *applier);
/**
* Unregister \a relay from the \a replica.
@@ -421,8 +421,7 @@ replicaset_add_anon(const struct tt_uuid *replica_uuid);
* appliers have successfully connected.
*/
void
-replicaset_connect(struct applier **appliers, int count,
- bool connect_quorum);
+replicaset_connect(struct applier **appliers, int count, bool connect_quorum);
/**
* Check if the current instance fell too much behind its
diff --git a/src/box/request.c b/src/box/request.c
index 994f2da..f47cd1b 100644
--- a/src/box/request.c
+++ b/src/box/request.c
@@ -91,9 +91,9 @@ request_create_from_tuple(struct request *request, struct space *space,
if (new_tuple == NULL) {
uint32_t size, key_size;
const char *data = tuple_data_range(old_tuple, &size);
- request->key = tuple_extract_key_raw(data, data + size,
- space->index[0]->def->key_def, MULTIKEY_NONE,
- &key_size);
+ request->key = tuple_extract_key_raw(
+ data, data + size, space->index[0]->def->key_def,
+ MULTIKEY_NONE, &key_size);
if (request->key == NULL)
return -1;
request->key_end = request->key + key_size;
@@ -151,8 +151,7 @@ request_handle_sequence(struct request *request, struct space *space)
* An automatically generated sequence inherits
* privileges of the space it is used with.
*/
- if (!seq->is_generated &&
- access_check_sequence(seq) != 0)
+ if (!seq->is_generated && access_check_sequence(seq) != 0)
return -1;
struct index *pk = space_index(space, 0);
@@ -199,10 +198,11 @@ request_handle_sequence(struct request *request, struct space *space)
mp_decode_nil(&key_end);
size_t buf_size = (request->tuple_end - request->tuple) +
- mp_sizeof_uint(UINT64_MAX);
+ mp_sizeof_uint(UINT64_MAX);
char *tuple = region_alloc(&fiber()->gc, buf_size);
if (tuple == NULL) {
- diag_set(OutOfMemory, buf_size, "region_alloc", "tuple");
+ diag_set(OutOfMemory, buf_size, "region_alloc",
+ "tuple");
return -1;
}
char *tuple_end = mp_encode_array(tuple, len);
diff --git a/src/box/schema.cc b/src/box/schema.cc
index 60e4a7f..a8d5425 100644
--- a/src/box/schema.cc
+++ b/src/box/schema.cc
@@ -81,7 +81,7 @@ bool
space_is_system(struct space *space)
{
return space->def->id > BOX_SYSTEM_ID_MIN &&
- space->def->id < BOX_SYSTEM_ID_MAX;
+ space->def->id < BOX_SYSTEM_ID_MAX;
}
/** Return space by its number */
@@ -91,15 +91,15 @@ space_by_id(uint32_t id)
mh_int_t space = mh_i32ptr_find(spaces, id, NULL);
if (space == mh_end(spaces))
return NULL;
- return (struct space *) mh_i32ptr_node(spaces, space)->val;
+ return (struct space *)mh_i32ptr_node(spaces, space)->val;
}
/** Return space by its name */
struct space *
space_by_name(const char *name)
{
- mh_int_t space = mh_strnptr_find_inp(spaces_by_name, name,
- strlen(name));
+ mh_int_t space =
+ mh_strnptr_find_inp(spaces_by_name, name, strlen(name));
if (space == mh_end(spaces_by_name))
return NULL;
return (struct space *)mh_strnptr_node(spaces_by_name, space)->val;
@@ -133,20 +133,21 @@ space_foreach(int (*func)(struct space *sp, void *udata), void *udata)
space = space_by_id(BOX_SPACE_ID);
struct index *pk = space ? space_index(space, 0) : NULL;
if (pk) {
- struct iterator *it = index_create_iterator(pk, ITER_GE,
- key, 1);
+ struct iterator *it =
+ index_create_iterator(pk, ITER_GE, key, 1);
if (it == NULL)
return -1;
int rc;
struct tuple *tuple;
while ((rc = iterator_next(it, &tuple)) == 0 && tuple != NULL) {
uint32_t id;
- if (tuple_field_u32(tuple, BOX_SPACE_FIELD_ID, &id) != 0)
+ if (tuple_field_u32(tuple, BOX_SPACE_FIELD_ID, &id) !=
+ 0)
continue;
space = space_cache_find(id);
if (space == NULL)
continue;
- if (! space_is_system(space))
+ if (!space_is_system(space))
break;
rc = func(space, udata);
if (rc != 0)
@@ -157,8 +158,9 @@ space_foreach(int (*func)(struct space *sp, void *udata), void *udata)
return -1;
}
- mh_foreach(spaces, i) {
- space = (struct space *) mh_i32ptr_node(spaces, i)->val;
+ mh_foreach(spaces, i)
+ {
+ space = (struct space *)mh_i32ptr_node(spaces, i)->val;
if (space_is_system(space))
continue;
if (func(space, udata) != 0)
@@ -180,14 +182,15 @@ space_cache_replace(struct space *old_space, struct space *new_space)
* don't need to do so for @spaces cache.
*/
struct space *old_space_by_name = NULL;
- if (old_space != NULL && strcmp(space_name(old_space),
- space_name(new_space)) != 0) {
+ if (old_space != NULL &&
+ strcmp(space_name(old_space), space_name(new_space)) != 0) {
const char *name = space_name(old_space);
mh_int_t k = mh_strnptr_find_inp(spaces_by_name, name,
strlen(name));
assert(k != mh_end(spaces_by_name));
- old_space_by_name = (struct space *)
- mh_strnptr_node(spaces_by_name, k)->val;
+ old_space_by_name = (struct space *)mh_strnptr_node(
+ spaces_by_name, k)
+ ->val;
mh_strnptr_del(spaces_by_name, k, NULL);
}
/*
@@ -202,8 +205,8 @@ space_cache_replace(struct space *old_space, struct space *new_space)
panic_syserror("Out of memory for the data "
"dictionary cache.");
}
- struct space *old_space_by_id = p_old != NULL ?
- (struct space *)p_old->val : NULL;
+ struct space *old_space_by_id =
+ p_old != NULL ? (struct space *)p_old->val : NULL;
assert(old_space_by_id == old_space);
(void)old_space_by_id;
/*
@@ -213,7 +216,8 @@ space_cache_replace(struct space *old_space, struct space *new_space)
uint32_t name_len = strlen(name);
uint32_t name_hash = mh_strn_hash(name, name_len);
const struct mh_strnptr_node_t node_s = { name, name_len,
- name_hash, new_space };
+ name_hash,
+ new_space };
struct mh_strnptr_node_t old_s, *p_old_s = &old_s;
k = mh_strnptr_put(spaces_by_name, &node_s, &p_old_s, NULL);
if (k == mh_end(spaces_by_name)) {
@@ -249,8 +253,8 @@ space_cache_replace(struct space *old_space, struct space *new_space)
}
space_cache_version++;
- if (trigger_run(&on_alter_space, new_space != NULL ?
- new_space : old_space) != 0) {
+ if (trigger_run(&on_alter_space,
+ new_space != NULL ? new_space : old_space) != 0) {
diag_log();
panic("Can't update space cache");
}
@@ -261,23 +265,19 @@ space_cache_replace(struct space *old_space, struct space *new_space)
/** A wrapper around space_new() for data dictionary spaces. */
static void
-sc_space_new(uint32_t id, const char *name,
- struct key_part_def *key_parts,
- uint32_t key_part_count,
- struct trigger *replace_trigger)
+sc_space_new(uint32_t id, const char *name, struct key_part_def *key_parts,
+ uint32_t key_part_count, struct trigger *replace_trigger)
{
struct key_def *key_def = key_def_new(key_parts, key_part_count, false);
if (key_def == NULL)
diag_raise();
auto key_def_guard =
make_scoped_guard([=] { key_def_delete(key_def); });
- struct index_def *index_def = index_def_new(id, /* space id */
- 0 /* index id */,
- "primary", /* name */
- strlen("primary"),
- TREE /* index type */,
- &index_opts_default,
- key_def, NULL);
+ struct index_def *index_def =
+ index_def_new(id, /* space id */
+ 0 /* index id */, "primary", /* name */
+ strlen("primary"), TREE /* index type */,
+ &index_opts_default, key_def, NULL);
if (index_def == NULL)
diag_raise();
auto index_def_guard =
@@ -307,8 +307,8 @@ sc_space_new(uint32_t id, const char *name,
}
int
-schema_find_id(uint32_t system_space_id, uint32_t index_id,
- const char *name, uint32_t len, uint32_t *object_id)
+schema_find_id(uint32_t system_space_id, uint32_t index_id, const char *name,
+ uint32_t len, uint32_t *object_id)
{
if (len > BOX_NAME_MAX) {
*object_id = BOX_ID_NIL;
@@ -318,8 +318,8 @@ schema_find_id(uint32_t system_space_id, uint32_t index_id,
if (space == NULL)
return -1;
if (!space_is_memtx(space)) {
- diag_set(ClientError, ER_UNSUPPORTED,
- space->engine->name, "system data");
+ diag_set(ClientError, ER_UNSUPPORTED, space->engine->name,
+ "system data");
return -1;
}
struct index *index = index_find(space, index_id);
@@ -492,9 +492,8 @@ schema_init(void)
def = space_def_new_xc(BOX_VINYL_DEFERRED_DELETE_ID, ADMIN, 0,
name, strlen(name), engine,
strlen(engine), &opts, NULL, 0);
- auto def_guard = make_scoped_guard([=] {
- space_def_delete(def);
- });
+ auto def_guard =
+ make_scoped_guard([=] { space_def_delete(def); });
RLIST_HEAD(key_list);
struct space *space = space_new_xc(def, &key_list);
space_cache_replace(NULL, space);
@@ -516,8 +515,8 @@ schema_free(void)
while (mh_size(spaces) > 0) {
mh_int_t i = mh_first(spaces);
- struct space *space = (struct space *)
- mh_i32ptr_node(spaces, i)->val;
+ struct space *space =
+ (struct space *)mh_i32ptr_node(spaces, i)->val;
space_cache_replace(space, NULL);
space_delete(space);
}
@@ -526,8 +525,8 @@ schema_free(void)
while (mh_size(funcs) > 0) {
mh_int_t i = mh_first(funcs);
- struct func *func = ((struct func *)
- mh_i32ptr_node(funcs, i)->val);
+ struct func *func =
+ ((struct func *)mh_i32ptr_node(funcs, i)->val);
func_cache_delete(func->def->fid);
func_delete(func);
}
@@ -535,8 +534,8 @@ schema_free(void)
while (mh_size(sequences) > 0) {
mh_int_t i = mh_first(sequences);
- struct sequence *seq = ((struct sequence *)
- mh_i32ptr_node(sequences, i)->val);
+ struct sequence *seq =
+ ((struct sequence *)mh_i32ptr_node(sequences, i)->val);
sequence_cache_delete(seq->def->id);
}
mh_i32ptr_delete(sequences);
@@ -550,14 +549,15 @@ func_cache_insert(struct func *func)
const struct mh_i32ptr_node_t node = { func->def->fid, func };
mh_int_t k1 = mh_i32ptr_put(funcs, &node, NULL, NULL);
if (k1 == mh_end(funcs)) {
-error:
+ error:
panic_syserror("Out of memory for the data "
"dictionary cache (stored function).");
}
size_t def_name_len = strlen(func->def->name);
uint32_t name_hash = mh_strn_hash(func->def->name, def_name_len);
- const struct mh_strnptr_node_t strnode = {
- func->def->name, def_name_len, name_hash, func };
+ const struct mh_strnptr_node_t strnode = { func->def->name,
+ def_name_len, name_hash,
+ func };
mh_int_t k2 = mh_strnptr_put(funcs_by_name, &strnode, NULL, NULL);
if (k2 == mh_end(funcs_by_name)) {
mh_i32ptr_del(funcs, k1, NULL);
@@ -571,8 +571,7 @@ func_cache_delete(uint32_t fid)
mh_int_t k = mh_i32ptr_find(funcs, fid, NULL);
if (k == mh_end(funcs))
return;
- struct func *func = (struct func *)
- mh_i32ptr_node(funcs, k)->val;
+ struct func *func = (struct func *)mh_i32ptr_node(funcs, k)->val;
mh_i32ptr_del(funcs, k, NULL);
k = mh_strnptr_find_inp(funcs_by_name, func->def->name,
strlen(func->def->name));
@@ -586,7 +585,7 @@ func_by_id(uint32_t fid)
mh_int_t func = mh_i32ptr_find(funcs, fid, NULL);
if (func == mh_end(funcs))
return NULL;
- return (struct func *) mh_i32ptr_node(funcs, func)->val;
+ return (struct func *)mh_i32ptr_node(funcs, func)->val;
}
struct func *
@@ -595,7 +594,7 @@ func_by_name(const char *name, uint32_t name_len)
mh_int_t func = mh_strnptr_find_inp(funcs_by_name, name, name_len);
if (func == mh_end(funcs_by_name))
return NULL;
- return (struct func *) mh_strnptr_node(funcs_by_name, func)->val;
+ return (struct func *)mh_strnptr_node(funcs_by_name, func)->val;
}
int
@@ -607,8 +606,8 @@ schema_find_grants(const char *type, uint32_t id, bool *out)
/** "object" index */
if (!space_is_memtx(priv)) {
- diag_set(ClientError, ER_UNSUPPORTED,
- priv->engine->name, "system data");
+ diag_set(ClientError, ER_UNSUPPORTED, priv->engine->name,
+ "system data");
return -1;
}
struct index *index = index_find(priv, 2);
@@ -638,7 +637,7 @@ sequence_by_id(uint32_t id)
mh_int_t k = mh_i32ptr_find(sequences, id, NULL);
if (k == mh_end(sequences))
return NULL;
- return (struct sequence *) mh_i32ptr_node(sequences, k)->val;
+ return (struct sequence *)mh_i32ptr_node(sequences, k)->val;
}
struct sequence *
@@ -682,39 +681,34 @@ schema_find_name(enum schema_object_type type, uint32_t object_id)
case SC_ENTITY_ROLE:
case SC_ENTITY_USER:
return "";
- case SC_SPACE:
- {
- struct space *space = space_by_id(object_id);
- if (space == NULL)
- break;
- return space->def->name;
- }
- case SC_FUNCTION:
- {
- struct func *func = func_by_id(object_id);
- if (func == NULL)
- break;
- return func->def->name;
- }
- case SC_SEQUENCE:
- {
- struct sequence *seq = sequence_by_id(object_id);
- if (seq == NULL)
- break;
- return seq->def->name;
- }
+ case SC_SPACE: {
+ struct space *space = space_by_id(object_id);
+ if (space == NULL)
+ break;
+ return space->def->name;
+ }
+ case SC_FUNCTION: {
+ struct func *func = func_by_id(object_id);
+ if (func == NULL)
+ break;
+ return func->def->name;
+ }
+ case SC_SEQUENCE: {
+ struct sequence *seq = sequence_by_id(object_id);
+ if (seq == NULL)
+ break;
+ return seq->def->name;
+ }
case SC_ROLE:
- case SC_USER:
- {
- struct user *role = user_by_id(object_id);
- if (role == NULL)
- break;
- return role->def->name;
- }
+ case SC_USER: {
+ struct user *role = user_by_id(object_id);
+ if (role == NULL)
+ break;
+ return role->def->name;
+ }
default:
break;
}
assert(false);
return "(nil)";
}
-
diff --git a/src/box/schema.h b/src/box/schema.h
index 25ac6f1..6b22039 100644
--- a/src/box/schema.h
+++ b/src/box/schema.h
@@ -169,8 +169,8 @@ schema_init(void);
void
schema_free(void);
-struct space *schema_space(uint32_t id);
-
+struct space *
+schema_space(uint32_t id);
/**
* Check whether or not an object has grants on it (restrict
@@ -241,18 +241,17 @@ struct on_access_denied_ctx {
/** Global grants to classes of objects. */
struct entity_access {
- struct access space[BOX_USER_MAX];
- struct access function[BOX_USER_MAX];
- struct access user[BOX_USER_MAX];
- struct access role[BOX_USER_MAX];
- struct access sequence[BOX_USER_MAX];
+ struct access space[BOX_USER_MAX];
+ struct access function[BOX_USER_MAX];
+ struct access user[BOX_USER_MAX];
+ struct access role[BOX_USER_MAX];
+ struct access sequence[BOX_USER_MAX];
};
/** A single instance of the global entities. */
extern struct entity_access entity_access;
-static inline
-struct access *
+static inline struct access *
entity_access_get(enum schema_object_type type)
{
switch (type) {
diff --git a/src/box/schema_def.c b/src/box/schema_def.c
index b974703..4f315c6 100644
--- a/src/box/schema_def.c
+++ b/src/box/schema_def.c
@@ -75,14 +75,14 @@ schema_object_type(const char *name)
* name, and they are case-sensitive, so be case-sensitive
* here too.
*/
- int n_strs = sizeof(object_type_strs)/sizeof(*object_type_strs);
+ int n_strs = sizeof(object_type_strs) / sizeof(*object_type_strs);
int index = strindex(object_type_strs, name, n_strs);
- return (enum schema_object_type) (index == n_strs ? 0 : index);
+ return (enum schema_object_type)(index == n_strs ? 0 : index);
}
const char *
schema_object_name(enum schema_object_type type)
{
- assert((int) type < (int) schema_object_type_MAX);
+ assert((int)type < (int)schema_object_type_MAX);
return object_type_strs[type];
}
diff --git a/src/box/schema_def.h b/src/box/schema_def.h
index f86cd42..9dedf73 100644
--- a/src/box/schema_def.h
+++ b/src/box/schema_def.h
@@ -316,9 +316,9 @@ enum schema_object_type {
/** SQL Storage engine. */
enum sql_storage_engine {
- SQL_STORAGE_ENGINE_MEMTX = 0,
- SQL_STORAGE_ENGINE_VINYL = 1,
- sql_storage_engine_MAX = 2
+ SQL_STORAGE_ENGINE_MEMTX = 0,
+ SQL_STORAGE_ENGINE_VINYL = 1,
+ sql_storage_engine_MAX = 2
};
extern const char *sql_storage_engine_strs[];
diff --git a/src/box/sequence.c b/src/box/sequence.c
index 4afbc26..4bdc090 100644
--- a/src/box/sequence.c
+++ b/src/box/sequence.c
@@ -92,8 +92,8 @@ sequence_data_extent_alloc(void *ctx)
(void)ctx;
void *ret = mempool_alloc(&sequence_data_extent_pool);
if (ret == NULL)
- diag_set(OutOfMemory, SEQUENCE_DATA_EXTENT_SIZE,
- "mempool", "sequence_data_extent");
+ diag_set(OutOfMemory, SEQUENCE_DATA_EXTENT_SIZE, "mempool",
+ "sequence_data_extent");
return ret;
}
@@ -166,11 +166,11 @@ sequence_set(struct sequence *seq, int64_t value)
struct sequence_data new_data, old_data;
new_data.id = key;
new_data.value = value;
- if (light_sequence_replace(&sequence_data_index, hash,
- new_data, &old_data) != light_sequence_end)
+ if (light_sequence_replace(&sequence_data_index, hash, new_data,
+ &old_data) != light_sequence_end)
return 0;
- if (light_sequence_insert(&sequence_data_index, hash,
- new_data) != light_sequence_end)
+ if (light_sequence_insert(&sequence_data_index, hash, new_data) !=
+ light_sequence_end)
return 0;
return -1;
}
@@ -189,7 +189,8 @@ sequence_update(struct sequence *seq, int64_t value)
if ((seq->def->step > 0 && value > data.value) ||
(seq->def->step < 0 && value < data.value)) {
if (light_sequence_replace(&sequence_data_index, hash,
- new_data, &data) == light_sequence_end)
+ new_data,
+ &data) == light_sequence_end)
unreachable();
}
} else {
@@ -246,8 +247,8 @@ done:
assert(value >= def->min && value <= def->max);
new_data.id = key;
new_data.value = value;
- if (light_sequence_replace(&sequence_data_index, hash,
- new_data, &old_data) == light_sequence_end)
+ if (light_sequence_replace(&sequence_data_index, hash, new_data,
+ &old_data) == light_sequence_end)
unreachable();
*result = value;
return 0;
@@ -272,25 +273,23 @@ access_check_sequence(struct sequence *seq)
user_access_t access = PRIV_U | PRIV_W;
user_access_t sequence_access = access & ~cr->universal_access;
- sequence_access &= ~entity_access_get(SC_SEQUENCE)[cr->auth_token].effective;
+ sequence_access &=
+ ~entity_access_get(SC_SEQUENCE)[cr->auth_token].effective;
if (sequence_access &&
/* Check for missing Usage access, ignore owner rights. */
(sequence_access & PRIV_U ||
/* Check for missing specific access, respect owner rights. */
(seq->def->uid != cr->uid &&
sequence_access & ~seq->access[cr->auth_token].effective))) {
-
/* Access violation, report error. */
struct user *user = user_find(cr->uid);
if (user != NULL) {
if (!(cr->universal_access & PRIV_U)) {
- diag_set(AccessDeniedError,
- priv_name(PRIV_U),
+ diag_set(AccessDeniedError, priv_name(PRIV_U),
schema_object_name(SC_UNIVERSE), "",
user->def->name);
} else {
- diag_set(AccessDeniedError,
- priv_name(access),
+ diag_set(AccessDeniedError, priv_name(access),
schema_object_name(SC_SEQUENCE),
seq->def->name, user->def->name);
}
@@ -308,19 +307,18 @@ struct sequence_data_iterator {
char tuple[0];
};
-#define SEQUENCE_TUPLE_BUF_SIZE (mp_sizeof_array(2) + \
- 2 * mp_sizeof_uint(UINT64_MAX))
+#define SEQUENCE_TUPLE_BUF_SIZE \
+ (mp_sizeof_array(2) + 2 * mp_sizeof_uint(UINT64_MAX))
static int
-sequence_data_iterator_next(struct snapshot_iterator *base,
- const char **data, uint32_t *size)
+sequence_data_iterator_next(struct snapshot_iterator *base, const char **data,
+ uint32_t *size)
{
struct sequence_data_iterator *iter =
(struct sequence_data_iterator *)base;
- struct sequence_data *sd =
- light_sequence_iterator_get_and_next(&sequence_data_index,
- &iter->iter);
+ struct sequence_data *sd = light_sequence_iterator_get_and_next(
+ &sequence_data_index, &iter->iter);
if (sd == NULL) {
*data = NULL;
return 0;
@@ -329,9 +327,8 @@ sequence_data_iterator_next(struct snapshot_iterator *base,
char *buf_end = iter->tuple;
buf_end = mp_encode_array(buf_end, 2);
buf_end = mp_encode_uint(buf_end, sd->id);
- buf_end = (sd->value >= 0 ?
- mp_encode_uint(buf_end, sd->value) :
- mp_encode_int(buf_end, sd->value));
+ buf_end = (sd->value >= 0 ? mp_encode_uint(buf_end, sd->value) :
+ mp_encode_int(buf_end, sd->value));
assert(buf_end <= iter->tuple + SEQUENCE_TUPLE_BUF_SIZE);
*data = iter->tuple;
*size = buf_end - iter->tuple;
@@ -351,8 +348,8 @@ sequence_data_iterator_free(struct snapshot_iterator *base)
struct snapshot_iterator *
sequence_data_iterator_create(void)
{
- struct sequence_data_iterator *iter = calloc(1, sizeof(*iter) +
- SEQUENCE_TUPLE_BUF_SIZE);
+ struct sequence_data_iterator *iter =
+ calloc(1, sizeof(*iter) + SEQUENCE_TUPLE_BUF_SIZE);
if (iter == NULL) {
diag_set(OutOfMemory, sizeof(*iter) + SEQUENCE_TUPLE_BUF_SIZE,
"malloc", "sequence_data_iterator");
@@ -377,8 +374,8 @@ sequence_get_value(struct sequence *seq, int64_t *result)
diag_set(ClientError, ER_SEQUENCE_NOT_STARTED, seq->def->name);
return -1;
}
- struct sequence_data data = light_sequence_get(&sequence_data_index,
- pos);
+ struct sequence_data data =
+ light_sequence_get(&sequence_data_index, pos);
*result = data.value;
return 0;
}
diff --git a/src/box/service_engine.c b/src/box/service_engine.c
index 5a33a73..e59fc1a 100644
--- a/src/box/service_engine.c
+++ b/src/box/service_engine.c
@@ -66,12 +66,10 @@ service_engine_create_space(struct engine *engine, struct space_def *def,
free(space);
return NULL;
}
- struct tuple_format *format =
- tuple_format_new(&tuple_format_runtime->vtab, NULL, keys,
- key_count, def->fields, def->field_count,
- def->exact_field_count, def->dict,
- def->opts.is_temporary,
- def->opts.is_ephemeral);
+ struct tuple_format *format = tuple_format_new(
+ &tuple_format_runtime->vtab, NULL, keys, key_count, def->fields,
+ def->field_count, def->exact_field_count, def->dict,
+ def->opts.is_temporary, def->opts.is_ephemeral);
if (format == NULL) {
free(space);
return NULL;
diff --git a/src/box/session.cc b/src/box/session.cc
index 7ba7235..98714f5 100644
--- a/src/box/session.cc
+++ b/src/box/session.cc
@@ -39,12 +39,7 @@
#include "sql_stmt_cache.h"
const char *session_type_strs[] = {
- "background",
- "binary",
- "console",
- "repl",
- "applier",
- "unknown",
+ "background", "binary", "console", "repl", "applier", "unknown",
};
static struct session_vtab generic_session_vtab = {
@@ -98,8 +93,8 @@ session_on_stop(struct trigger *trigger, void * /* event */)
static int
closed_session_push(struct session *session, struct port *port)
{
- (void) session;
- (void) port;
+ (void)session;
+ (void)port;
diag_set(ClientError, ER_SESSION_CLOSED);
return -1;
}
@@ -129,7 +124,7 @@ struct session *
session_create(enum session_type type)
{
struct session *session =
- (struct session *) mempool_alloc(&session_pool);
+ (struct session *)mempool_alloc(&session_pool);
if (session == NULL) {
diag_set(OutOfMemory, session_pool.objsize, "mempool",
"new slab");
@@ -168,9 +163,8 @@ session_create_on_demand(void)
struct session *s = session_create(SESSION_TYPE_BACKGROUND);
if (s == NULL)
return NULL;
- s->fiber_on_stop = {
- RLIST_LINK_INITIALIZER, session_on_stop, NULL, NULL
- };
+ s->fiber_on_stop = { RLIST_LINK_INITIALIZER, session_on_stop, NULL,
+ NULL };
/* Add a trigger to destroy session on fiber stop */
trigger_add(&fiber()->on_stop, &s->fiber_on_stop);
credentials_reset(&s->credentials, admin_user);
@@ -270,8 +264,7 @@ session_find(uint64_t sid)
mh_int_t k = mh_i64ptr_find(session_registry, sid, NULL);
if (k == mh_end(session_registry))
return NULL;
- return (struct session *)
- mh_i64ptr_node(session_registry, k)->val;
+ return (struct session *)mh_i64ptr_node(session_registry, k)->val;
}
extern "C" void
@@ -305,8 +298,7 @@ access_check_session(struct user *user)
*/
if (!(universe.access[user->auth_token].effective & PRIV_S)) {
diag_set(AccessDeniedError, priv_name(PRIV_S),
- schema_object_name(SC_UNIVERSE), "",
- user->def->name);
+ schema_object_name(SC_UNIVERSE), "", user->def->name);
return -1;
}
return 0;
@@ -325,12 +317,12 @@ access_check_universe_object(user_access_t access,
* The user may not exist already, if deleted
* from a different connection.
*/
- int denied_access = access & ((credentials->universal_access
- & access) ^ access);
+ int denied_access =
+ access &
+ ((credentials->universal_access & access) ^ access);
struct user *user = user_find(credentials->uid);
if (user != NULL) {
- diag_set(AccessDeniedError,
- priv_name(denied_access),
+ diag_set(AccessDeniedError, priv_name(denied_access),
schema_object_name(object_type), object_name,
user->def->name);
} else {
@@ -355,7 +347,7 @@ access_check_universe(user_access_t access)
int
generic_session_push(struct session *session, struct port *port)
{
- (void) port;
+ (void)port;
const char *name =
tt_sprintf("Session '%s'", session_type_strs[session->type]);
diag_set(ClientError, ER_UNSUPPORTED, name, "push()");
@@ -365,13 +357,13 @@ generic_session_push(struct session *session, struct port *port)
int
generic_session_fd(struct session *session)
{
- (void) session;
+ (void)session;
return -1;
}
int64_t
generic_session_sync(struct session *session)
{
- (void) session;
+ (void)session;
return 0;
}
diff --git a/src/box/session.h b/src/box/session.h
index 833a457..69a1e85 100644
--- a/src/box/session.h
+++ b/src/box/session.h
@@ -135,16 +135,14 @@ struct session_vtab {
* @retval 0 Success.
* @retval -1 Error.
*/
- int
- (*push)(struct session *session, struct port *port);
+ int (*push)(struct session *session, struct port *port);
/**
* Get session file descriptor if exists.
* @param session Session to get descriptor from.
* @retval -1 No fd.
* @retval >=0 Found fd.
*/
- int
- (*fd)(struct session *session);
+ int (*fd)(struct session *session);
/**
* For iproto requests, we set sync to the value of packet
* sync. Since the session may be reused between many
@@ -152,8 +150,7 @@ struct session_vtab {
* of the request, and gets distorted after the first
* yield. For other sessions it is 0.
*/
- int64_t
- (*sync)(struct session *session);
+ int64_t (*sync)(struct session *session);
};
extern struct session_vtab session_vtab_registry[];
diff --git a/src/box/session_settings.c b/src/box/session_settings.c
index dbbbf24..5bef498 100644
--- a/src/box/session_settings.c
+++ b/src/box/session_settings.c
@@ -42,16 +42,11 @@ struct session_setting session_settings[SESSION_SETTING_COUNT] = {};
/** Corresponding names of session settings. */
const char *session_setting_strs[SESSION_SETTING_COUNT] = {
- "error_marshaling_enabled",
- "sql_default_engine",
- "sql_defer_foreign_keys",
- "sql_full_column_names",
- "sql_full_metadata",
- "sql_parser_debug",
- "sql_recursive_triggers",
- "sql_reverse_unordered_selects",
- "sql_select_debug",
- "sql_vdbe_debug",
+ "error_marshaling_enabled", "sql_default_engine",
+ "sql_defer_foreign_keys", "sql_full_column_names",
+ "sql_full_metadata", "sql_parser_debug",
+ "sql_recursive_triggers", "sql_reverse_unordered_selects",
+ "sql_select_debug", "sql_vdbe_debug",
};
struct session_settings_index {
@@ -105,8 +100,7 @@ session_settings_next(int *sid, const char *key, bool is_eq, bool is_including)
for (; i < SESSION_SETTING_COUNT; ++i) {
const char *name = session_setting_strs[i];
int cmp = strcmp(name, key);
- if ((cmp == 0 && is_including) ||
- (cmp > 0 && !is_eq)) {
+ if ((cmp == 0 && is_including) || (cmp > 0 && !is_eq)) {
*sid = i;
return 0;
}
@@ -128,8 +122,7 @@ session_settings_prev(int *sid, const char *key, bool is_eq, bool is_including)
for (; i >= 0; --i) {
const char *name = session_setting_strs[i];
int cmp = strcmp(name, key);
- if ((cmp == 0 && is_including) ||
- (cmp < 0 && !is_eq)) {
+ if ((cmp == 0 && is_including) || (cmp < 0 && !is_eq)) {
*sid = i;
return 0;
}
@@ -238,9 +231,9 @@ session_settings_index_get(struct index *base, const char *key,
uint32_t part_count, struct tuple **result)
{
struct session_settings_index *index =
- (struct session_settings_index *) base;
+ (struct session_settings_index *)base;
assert(part_count == 1);
- (void) part_count;
+ (void)part_count;
uint32_t len;
key = mp_decode_str(&key, &len);
key = tt_cstr(key, len);
@@ -265,7 +258,7 @@ static const struct index_vtab session_settings_index_vtab = {
/* .update_def = */ generic_index_update_def,
/* .depends_on_pk = */ generic_index_depends_on_pk,
/* .def_change_requires_rebuild = */
- generic_index_def_change_requires_rebuild,
+ generic_index_def_change_requires_rebuild,
/* .size = */ generic_index_size,
/* .bsize = */ generic_index_bsize,
/* .min = */ generic_index_min,
@@ -276,7 +269,7 @@ static const struct index_vtab session_settings_index_vtab = {
/* .replace = */ generic_index_replace,
/* .create_iterator = */ session_settings_index_create_iterator,
/* .create_snapshot_iterator = */
- generic_index_create_snapshot_iterator,
+ generic_index_create_snapshot_iterator,
/* .stat = */ generic_index_stat,
/* .compact = */ generic_index_compact,
/* .reset_stat = */ generic_index_reset_stat,
@@ -443,7 +436,8 @@ const struct space_vtab session_settings_space_vtab = {
};
int
-session_setting_find(const char *name) {
+session_setting_find(const char *name)
+{
int sid = 0;
if (session_settings_next(&sid, name, true, true) == 0)
return sid;
@@ -465,7 +459,7 @@ session_setting_error_marshaling_enabled_get(int id, const char **mp_pair,
size_t size = mp_sizeof_array(2) + mp_sizeof_str(name_len) +
mp_sizeof_bool(value);
- char *pos = (char*)static_alloc(size);
+ char *pos = (char *)static_alloc(size);
assert(pos != NULL);
char *pos_end = mp_encode_array(pos, 2);
pos_end = mp_encode_str(pos_end, name, name_len);
diff --git a/src/box/space.c b/src/box/space.c
index 6d1d771..54435fd 100644
--- a/src/box/space.c
+++ b/src/box/space.c
@@ -73,8 +73,8 @@ access_check_space(struct space *space, user_access_t access)
/* Check for missing USAGE access, ignore owner rights. */
(space_access & PRIV_U ||
/* Check for missing specific access, respect owner rights. */
- (space->def->uid != cr->uid &&
- space_access & ~space->access[cr->auth_token].effective))) {
+ (space->def->uid != cr->uid &&
+ space_access & ~space->access[cr->auth_token].effective))) {
/*
* Report access violation. Throw "no such user"
* error if there is no user with this id.
@@ -84,13 +84,11 @@ access_check_space(struct space *space, user_access_t access)
struct user *user = user_find(cr->uid);
if (user != NULL) {
if (!(cr->universal_access & PRIV_U)) {
- diag_set(AccessDeniedError,
- priv_name(PRIV_U),
+ diag_set(AccessDeniedError, priv_name(PRIV_U),
schema_object_name(SC_UNIVERSE), "",
user->def->name);
} else {
- diag_set(AccessDeniedError,
- priv_name(access),
+ diag_set(AccessDeniedError, priv_name(access),
schema_object_name(SC_SPACE),
space->def->name, user->def->name);
}
@@ -120,8 +118,8 @@ space_create(struct space *space, struct engine *engine,
{
if (!rlist_empty(key_list)) {
/* Primary key must go first. */
- struct index_def *pk = rlist_first_entry(key_list,
- struct index_def, link);
+ struct index_def *pk =
+ rlist_first_entry(key_list, struct index_def, link);
assert(pk->iid == 0);
(void)pk;
}
@@ -152,11 +150,13 @@ space_create(struct space *space, struct engine *engine,
goto fail;
/* Create indexes and fill the index map. */
- space->index_map = (struct index **)
- calloc(index_count + index_id_max + 1, sizeof(struct index *));
+ space->index_map = (struct index **)calloc(
+ index_count + index_id_max + 1, sizeof(struct index *));
if (space->index_map == NULL) {
- diag_set(OutOfMemory, (index_count + index_id_max + 1) *
- sizeof(struct index *), "malloc", "index_map");
+ diag_set(OutOfMemory,
+ (index_count + index_id_max + 1) *
+ sizeof(struct index *),
+ "malloc", "index_map");
goto fail;
}
space->index = space->index_map + index_id_max + 1;
@@ -195,8 +195,9 @@ space_create(struct space *space, struct engine *engine,
continue;
for (int j = 0; j < (int)space->index_count; j++) {
struct index *other = space->index[j];
- if (i != j && bit_test(space->check_unique_constraint_map,
- other->def->iid) &&
+ if (i != j &&
+ bit_test(space->check_unique_constraint_map,
+ other->def->iid) &&
key_def_contains(index->def->key_def,
other->def->key_def)) {
bit_clear(space->check_unique_constraint_map,
@@ -412,9 +413,9 @@ after_old_tuple_lookup:;
old_data_end = old_data + old_size;
new_data = xrow_update_execute(request->tuple,
request->tuple_end, old_data,
- old_data_end,
- space->format, &new_size,
- request->index_base, NULL);
+ old_data_end, space->format,
+ &new_size, request->index_base,
+ NULL);
if (new_data == NULL)
return -1;
new_data_end = new_data + new_size;
@@ -434,10 +435,9 @@ after_old_tuple_lookup:;
*/
new_data = request->tuple;
new_data_end = request->tuple_end;
- if (xrow_update_check_ops(request->ops,
- request->ops_end,
- space->format,
- request->index_base) != 0)
+ if (xrow_update_check_ops(
+ request->ops, request->ops_end,
+ space->format, request->index_base) != 0)
return -1;
break;
}
@@ -456,8 +456,8 @@ after_old_tuple_lookup:;
struct tuple *new_tuple = NULL;
if (new_data != NULL) {
- new_tuple = tuple_new(tuple_format_runtime,
- new_data, new_data_end);
+ new_tuple =
+ tuple_new(tuple_format_runtime, new_data, new_data_end);
if (new_tuple == NULL)
return -1;
tuple_ref(new_tuple);
@@ -511,12 +511,12 @@ after_old_tuple_lookup:;
* We don't allow to change the value of the primary key
* in the same statement.
*/
- if (pk != NULL && request_changed &&
- old_tuple != NULL && new_tuple != NULL &&
+ if (pk != NULL && request_changed && old_tuple != NULL &&
+ new_tuple != NULL &&
tuple_compare(old_tuple, HINT_NONE, new_tuple, HINT_NONE,
pk->def->key_def) != 0) {
- diag_set(ClientError, ER_CANT_UPDATE_PRIMARY_KEY,
- pk->def->name, space->def->name);
+ diag_set(ClientError, ER_CANT_UPDATE_PRIMARY_KEY, pk->def->name,
+ space->def->name);
rc = -1;
goto out;
}
@@ -526,8 +526,8 @@ after_old_tuple_lookup:;
* Fix the request to conform.
*/
if (request_changed)
- rc = request_create_from_tuple(request, space,
- old_tuple, new_tuple);
+ rc = request_create_from_tuple(request, space, old_tuple,
+ new_tuple);
out:
if (new_tuple != NULL)
tuple_unref(new_tuple);
@@ -535,8 +535,8 @@ out:
}
int
-space_execute_dml(struct space *space, struct txn *txn,
- struct request *request, struct tuple **result)
+space_execute_dml(struct space *space, struct txn *txn, struct request *request,
+ struct tuple **result)
{
if (unlikely(space->sequence != NULL) &&
(request->type == IPROTO_INSERT ||
@@ -565,13 +565,13 @@ space_execute_dml(struct space *space, struct txn *txn,
switch (request->type) {
case IPROTO_INSERT:
case IPROTO_REPLACE:
- if (space->vtab->execute_replace(space, txn,
- request, result) != 0)
+ if (space->vtab->execute_replace(space, txn, request, result) !=
+ 0)
return -1;
break;
case IPROTO_UPDATE:
- if (space->vtab->execute_update(space, txn,
- request, result) != 0)
+ if (space->vtab->execute_update(space, txn, request, result) !=
+ 0)
return -1;
if (*result != NULL && request->index_id != 0) {
/*
@@ -583,8 +583,8 @@ space_execute_dml(struct space *space, struct txn *txn,
}
break;
case IPROTO_DELETE:
- if (space->vtab->execute_delete(space, txn,
- request, result) != 0)
+ if (space->vtab->execute_delete(space, txn, request, result) !=
+ 0)
return -1;
if (*result != NULL && request->index_id != 0)
request_rebind_to_primary_key(request, space, *result);
@@ -606,14 +606,14 @@ space_add_ck_constraint(struct space *space, struct ck_constraint *ck)
rlist_add_entry(&space->ck_constraint, ck, link);
if (space->ck_constraint_trigger == NULL) {
struct trigger *ck_trigger =
- (struct trigger *) malloc(sizeof(*ck_trigger));
+ (struct trigger *)malloc(sizeof(*ck_trigger));
if (ck_trigger == NULL) {
diag_set(OutOfMemory, sizeof(*ck_trigger), "malloc",
"ck_trigger");
return -1;
}
trigger_create(ck_trigger, ck_constraint_on_replace_trigger,
- NULL, (trigger_f0) free);
+ NULL, (trigger_f0)free);
trigger_add(&space->on_replace, ck_trigger);
space->ck_constraint_trigger = ck_trigger;
}
@@ -640,7 +640,7 @@ space_find_constraint_id(struct space *space, const char *name)
mh_int_t pos = mh_strnptr_find_inp(ids, name, len);
if (pos == mh_end(ids))
return NULL;
- return (struct constraint_id *) mh_strnptr_node(ids, pos)->val;
+ return (struct constraint_id *)mh_strnptr_node(ids, pos)->val;
}
int
@@ -650,7 +650,7 @@ space_add_constraint_id(struct space *space, struct constraint_id *id)
struct mh_strnptr_t *ids = space->constraint_ids;
uint32_t len = strlen(id->name);
uint32_t hash = mh_strn_hash(id->name, len);
- const struct mh_strnptr_node_t name_node = {id->name, len, hash, id};
+ const struct mh_strnptr_node_t name_node = { id->name, len, hash, id };
if (mh_strnptr_put(ids, &name_node, NULL, NULL) == mh_end(ids)) {
diag_set(OutOfMemory, sizeof(name_node), "malloc", "node");
return -1;
@@ -665,8 +665,8 @@ space_pop_constraint_id(struct space *space, const char *name)
uint32_t len = strlen(name);
mh_int_t pos = mh_strnptr_find_inp(ids, name, len);
assert(pos != mh_end(ids));
- struct constraint_id *id = (struct constraint_id *)
- mh_strnptr_node(ids, pos)->val;
+ struct constraint_id *id =
+ (struct constraint_id *)mh_strnptr_node(ids, pos)->val;
mh_strnptr_del(ids, pos, NULL);
return id;
}
diff --git a/src/box/space.h b/src/box/space.h
index 7cfba65..a74c900 100644
--- a/src/box/space.h
+++ b/src/box/space.h
@@ -60,12 +60,12 @@ struct space_vtab {
/** Return binary size of a space. */
size_t (*bsize)(struct space *);
- int (*execute_replace)(struct space *, struct txn *,
- struct request *, struct tuple **result);
- int (*execute_delete)(struct space *, struct txn *,
- struct request *, struct tuple **result);
- int (*execute_update)(struct space *, struct txn *,
- struct request *, struct tuple **result);
+ int (*execute_replace)(struct space *, struct txn *, struct request *,
+ struct tuple **result);
+ int (*execute_delete)(struct space *, struct txn *, struct request *,
+ struct tuple **result);
+ int (*execute_update)(struct space *, struct txn *, struct request *,
+ struct tuple **result);
int (*execute_upsert)(struct space *, struct txn *, struct request *);
int (*ephemeral_replace)(struct space *, const char *, const char *);
@@ -140,8 +140,7 @@ struct space_vtab {
* Notify the engine about the changed space,
* before it's done, to prepare 'new_space' object.
*/
- int (*prepare_alter)(struct space *old_space,
- struct space *new_space);
+ int (*prepare_alter)(struct space *old_space, struct space *new_space);
/**
* Called right after removing a space from the cache.
* The engine should abort all transactions involving
@@ -253,7 +252,10 @@ space_create(struct space *space, struct engine *engine,
/** Get space ordinal number. */
static inline uint32_t
-space_id(struct space *space) { return space->def->id; }
+space_id(struct space *space)
+{
+ return space->def->id;
+}
/** Get space name. */
static inline const char *
@@ -302,7 +304,7 @@ space_index(struct space *space, uint32_t id)
static inline struct index *
space_index_by_name(struct space *space, const char *index_name)
{
- for(uint32_t i = 0; i < space->index_count; i++) {
+ for (uint32_t i = 0; i < space->index_count; i++) {
struct index *index = space->index[i];
if (strcmp(index_name, index->def->name) == 0)
return index;
@@ -391,8 +393,8 @@ access_check_space(struct space *space, user_access_t access);
* Execute a DML request on the given space.
*/
int
-space_execute_dml(struct space *space, struct txn *txn,
- struct request *request, struct tuple **result);
+space_execute_dml(struct space *space, struct txn *txn, struct request *request,
+ struct tuple **result);
static inline int
space_ephemeral_replace(struct space *space, const char *tuple,
@@ -466,8 +468,8 @@ space_swap_index(struct space *old_space, struct space *new_space,
uint32_t old_index_id, uint32_t new_index_id)
{
assert(old_space->vtab == new_space->vtab);
- return new_space->vtab->swap_index(old_space, new_space,
- old_index_id, new_index_id);
+ return new_space->vtab->swap_index(old_space, new_space, old_index_id,
+ new_index_id);
}
static inline int
@@ -484,11 +486,17 @@ space_invalidate(struct space *space)
}
static inline bool
-space_is_memtx(struct space *space) { return space->engine->id == 0; }
+space_is_memtx(struct space *space)
+{
+ return space->engine->id == 0;
+}
/** Return true if space is run under vinyl engine. */
static inline bool
-space_is_vinyl(struct space *space) { return strcmp(space->engine->name, "vinyl") == 0; }
+space_is_vinyl(struct space *space)
+{
+ return strcmp(space->engine->name, "vinyl") == 0;
+}
struct field_def;
/**
@@ -566,20 +574,33 @@ space_pop_constraint_id(struct space *space, const char *name);
/*
* Virtual method stubs.
*/
-size_t generic_space_bsize(struct space *);
-int generic_space_ephemeral_replace(struct space *, const char *, const char *);
-int generic_space_ephemeral_delete(struct space *, const char *);
-int generic_space_ephemeral_rowid_next(struct space *, uint64_t *);
-void generic_init_system_space(struct space *);
-void generic_init_ephemeral_space(struct space *);
-int generic_space_check_index_def(struct space *, struct index_def *);
-int generic_space_add_primary_key(struct space *space);
-void generic_space_drop_primary_key(struct space *space);
-int generic_space_check_format(struct space *, struct tuple_format *);
-int generic_space_build_index(struct space *, struct index *,
- struct tuple_format *, bool);
-int generic_space_prepare_alter(struct space *, struct space *);
-void generic_space_invalidate(struct space *);
+size_t
+generic_space_bsize(struct space *);
+int
+generic_space_ephemeral_replace(struct space *, const char *, const char *);
+int
+generic_space_ephemeral_delete(struct space *, const char *);
+int
+generic_space_ephemeral_rowid_next(struct space *, uint64_t *);
+void
+generic_init_system_space(struct space *);
+void
+generic_init_ephemeral_space(struct space *);
+int
+generic_space_check_index_def(struct space *, struct index_def *);
+int
+generic_space_add_primary_key(struct space *space);
+void
+generic_space_drop_primary_key(struct space *space);
+int
+generic_space_check_format(struct space *, struct tuple_format *);
+int
+generic_space_build_index(struct space *, struct index *, struct tuple_format *,
+ bool);
+int
+generic_space_prepare_alter(struct space *, struct space *);
+void
+generic_space_invalidate(struct space *);
#if defined(__cplusplus)
} /* extern "C" */
@@ -629,9 +650,9 @@ index_find_unique_xc(struct space *space, uint32_t index_id)
static inline struct index *
index_find_system_xc(struct space *space, uint32_t index_id)
{
- if (! space_is_memtx(space)) {
- tnt_raise(ClientError, ER_UNSUPPORTED,
- space->engine->name, "system data");
+ if (!space_is_memtx(space)) {
+ tnt_raise(ClientError, ER_UNSUPPORTED, space->engine->name,
+ "system data");
}
return index_find_xc(space, index_id);
}
diff --git a/src/box/space_def.c b/src/box/space_def.c
index 83566bf..ec45955 100644
--- a/src/box/space_def.c
+++ b/src/box/space_def.c
@@ -76,8 +76,8 @@ space_def_sizeof(uint32_t name_len, const struct field_def *fields,
*fields_offset = small_align(sizeof(struct space_def) + name_len + 1,
alignof(typeof(fields[0])));
*names_offset = *fields_offset + field_count * sizeof(struct field_def);
- *def_expr_offset = small_align(*names_offset + field_strs_size,
- alignof(uint64_t));
+ *def_expr_offset =
+ small_align(*names_offset + field_strs_size, alignof(uint64_t));
return *def_expr_offset + def_exprs_size;
}
@@ -110,7 +110,7 @@ space_def_dup(const struct space_def *src)
size_t size = space_def_sizeof(strlen(src->name), src->fields,
src->field_count, &strs_offset,
&fields_offset, &def_expr_offset);
- struct space_def *ret = (struct space_def *) malloc(size);
+ struct space_def *ret = (struct space_def *)malloc(size);
if (ret == NULL) {
diag_set(OutOfMemory, size, "malloc", "ret");
return NULL;
@@ -131,7 +131,7 @@ space_def_dup(const struct space_def *src)
struct Expr *e = src->fields[i].default_value_expr;
if (e != NULL) {
char *expr_pos_old = expr_pos;
- (void) expr_pos_old;
+ (void)expr_pos_old;
e = sql_expr_dup(sql_get(), e, 0, &expr_pos);
assert(e != NULL);
/* Note: due to SQL legacy
@@ -156,16 +156,15 @@ space_def_dup(const struct space_def *src)
struct space_def *
space_def_new(uint32_t id, uint32_t uid, uint32_t exact_field_count,
- const char *name, uint32_t name_len,
- const char *engine_name, uint32_t engine_len,
- const struct space_opts *opts, const struct field_def *fields,
- uint32_t field_count)
+ const char *name, uint32_t name_len, const char *engine_name,
+ uint32_t engine_len, const struct space_opts *opts,
+ const struct field_def *fields, uint32_t field_count)
{
uint32_t strs_offset, fields_offset, def_expr_offset;
size_t size = space_def_sizeof(name_len, fields, field_count,
&strs_offset, &fields_offset,
&def_expr_offset);
- struct space_def *def = (struct space_def *) malloc(size);
+ struct space_def *def = (struct space_def *)malloc(size);
if (def == NULL) {
diag_set(OutOfMemory, size, "malloc", "def");
return NULL;
@@ -212,7 +211,7 @@ space_def_new(uint32_t id, uint32_t uid, uint32_t exact_field_count,
struct Expr *e = def->fields[i].default_value_expr;
if (e != NULL) {
char *expr_pos_old = expr_pos;
- (void) expr_pos_old;
+ (void)expr_pos_old;
e = sql_expr_dup(sql_get(), e, 0, &expr_pos);
assert(e != NULL);
/* Note: due to SQL legacy
@@ -235,7 +234,7 @@ space_def_new(uint32_t id, uint32_t uid, uint32_t exact_field_count,
return def;
}
-struct space_def*
+struct space_def *
space_def_new_ephemeral(uint32_t exact_field_count, struct field_def *fields)
{
struct space_opts opts = space_opts_default;
@@ -246,11 +245,9 @@ space_def_new_ephemeral(uint32_t exact_field_count, struct field_def *fields)
fields = (struct field_def *)&field_def_default;
field_count = 0;
}
- struct space_def *space_def = space_def_new(0, 0, exact_field_count,
- "ephemeral",
- strlen("ephemeral"),
- "memtx", strlen("memtx"),
- &opts, fields, field_count);
+ struct space_def *space_def = space_def_new(
+ 0, 0, exact_field_count, "ephemeral", strlen("ephemeral"),
+ "memtx", strlen("memtx"), &opts, fields, field_count);
return space_def;
}
diff --git a/src/box/space_def.h b/src/box/space_def.h
index 198242d..2fa28d1 100644
--- a/src/box/space_def.h
+++ b/src/box/space_def.h
@@ -47,7 +47,7 @@ struct space_opts {
* made to a space are replicated.
*/
uint32_t group_id;
- /**
+ /**
* The space is a temporary:
* - it is empty at server start
* - changes are not written to WAL
@@ -171,10 +171,9 @@ space_def_dup(const struct space_def *src);
*/
struct space_def *
space_def_new(uint32_t id, uint32_t uid, uint32_t exact_field_count,
- const char *name, uint32_t name_len,
- const char *engine_name, uint32_t engine_len,
- const struct space_opts *opts, const struct field_def *fields,
- uint32_t field_count);
+ const char *name, uint32_t name_len, const char *engine_name,
+ uint32_t engine_len, const struct space_opts *opts,
+ const struct field_def *fields, uint32_t field_count);
/**
* Create a new ephemeral space definition.
@@ -220,10 +219,9 @@ space_def_dup_xc(const struct space_def *src)
static inline struct space_def *
space_def_new_xc(uint32_t id, uint32_t uid, uint32_t exact_field_count,
- const char *name, uint32_t name_len,
- const char *engine_name, uint32_t engine_len,
- const struct space_opts *opts, const struct field_def *fields,
- uint32_t field_count)
+ const char *name, uint32_t name_len, const char *engine_name,
+ uint32_t engine_len, const struct space_opts *opts,
+ const struct field_def *fields, uint32_t field_count)
{
struct space_def *ret = space_def_new(id, uid, exact_field_count, name,
name_len, engine_name, engine_len,
diff --git a/src/box/sysview.c b/src/box/sysview.c
index 00c320b..9f2fc40 100644
--- a/src/box/sysview.c
+++ b/src/box/sysview.c
@@ -74,7 +74,7 @@ struct sysview_iterator {
static inline struct sysview_iterator *
sysview_iterator(struct iterator *ptr)
{
- return (struct sysview_iterator *) ptr;
+ return (struct sysview_iterator *)ptr;
}
static void
@@ -150,8 +150,8 @@ sysview_index_create_iterator(struct index *base, enum iterator_type type,
}
static int
-sysview_index_get(struct index *base, const char *key,
- uint32_t part_count, struct tuple **result)
+sysview_index_get(struct index *base, const char *key, uint32_t part_count,
+ struct tuple **result)
{
struct sysview_index *index = (struct sysview_index *)base;
struct space *source = space_cache_find(index->source_space_id);
@@ -185,7 +185,7 @@ static const struct index_vtab sysview_index_vtab = {
/* .update_def = */ generic_index_update_def,
/* .depends_on_pk = */ generic_index_depends_on_pk,
/* .def_change_requires_rebuild = */
- generic_index_def_change_requires_rebuild,
+ generic_index_def_change_requires_rebuild,
/* .size = */ generic_index_size,
/* .bsize = */ generic_index_bsize,
/* .min = */ generic_index_min,
@@ -196,7 +196,7 @@ static const struct index_vtab sysview_index_vtab = {
/* .replace = */ generic_index_replace,
/* .create_iterator = */ sysview_index_create_iterator,
/* .create_snapshot_iterator = */
- generic_index_create_snapshot_iterator,
+ generic_index_create_snapshot_iterator,
/* .stat = */ generic_index_stat,
/* .compact = */ generic_index_compact,
/* .reset_stat = */ generic_index_reset_stat,
@@ -295,8 +295,7 @@ vspace_filter(struct space *source, struct tuple *tuple)
* Allow access for space owners and users with any
* privilege for the space.
*/
- return (PRIV_WRDA & effective ||
- space->def->uid == cr->uid);
+ return (PRIV_WRDA & effective || space->def->uid == cr->uid);
}
static bool
@@ -363,15 +362,14 @@ vfunc_filter(struct space *source, struct tuple *tuple)
return true; /* read access to _func space */
uint32_t name_len;
- const char *name = tuple_field_str(tuple, BOX_FUNC_FIELD_NAME,
- &name_len);
+ const char *name =
+ tuple_field_str(tuple, BOX_FUNC_FIELD_NAME, &name_len);
if (name == NULL)
return false;
struct func *func = func_by_name(name, name_len);
assert(func != NULL);
user_access_t effective = func->access[cr->auth_token].effective;
- return func->def->uid == cr->uid ||
- ((PRIV_WRDA | PRIV_X) & effective);
+ return func->def->uid == cr->uid || ((PRIV_WRDA | PRIV_X) & effective);
}
static bool
@@ -405,8 +403,8 @@ vsequence_filter(struct space *source, struct tuple *tuple)
static bool
vcollation_filter(struct space *source, struct tuple *tuple)
{
- (void) source;
- (void) tuple;
+ (void)source;
+ (void)tuple;
return true;
}
@@ -462,17 +460,16 @@ sysview_space_create_index(struct space *space, struct index_def *def)
filter = vcollation_filter;
break;
default:
- diag_set(ClientError, ER_MODIFY_INDEX,
- def->name, space_name(space),
- "unknown space for system view");
+ diag_set(ClientError, ER_MODIFY_INDEX, def->name,
+ space_name(space), "unknown space for system view");
return NULL;
}
struct sysview_index *index =
(struct sysview_index *)calloc(1, sizeof(*index));
if (index == NULL) {
- diag_set(OutOfMemory, sizeof(*index),
- "malloc", "struct sysview_index");
+ diag_set(OutOfMemory, sizeof(*index), "malloc",
+ "struct sysview_index");
return NULL;
}
if (index_create(&index->base, (struct engine *)sysview,
@@ -525,8 +522,7 @@ sysview_engine_create_space(struct engine *engine, struct space_def *def,
{
struct space *space = (struct space *)calloc(1, sizeof(*space));
if (space == NULL) {
- diag_set(OutOfMemory, sizeof(*space),
- "malloc", "struct space");
+ diag_set(OutOfMemory, sizeof(*space), "malloc", "struct space");
return NULL;
}
int key_count = 0;
@@ -542,18 +538,17 @@ sysview_engine_create_space(struct engine *engine, struct space_def *def,
free(space);
return NULL;
}
- struct tuple_format *format =
- tuple_format_new(NULL, NULL, keys, key_count, def->fields,
- def->field_count, def->exact_field_count,
- def->dict, def->opts.is_temporary,
- def->opts.is_ephemeral);
+ struct tuple_format *format = tuple_format_new(
+ NULL, NULL, keys, key_count, def->fields, def->field_count,
+ def->exact_field_count, def->dict, def->opts.is_temporary,
+ def->opts.is_ephemeral);
if (format == NULL) {
free(space);
return NULL;
}
tuple_format_ref(format);
- if (space_create(space, engine, &sysview_space_vtab,
- def, key_list, format) != 0) {
+ if (space_create(space, engine, &sysview_space_vtab, def, key_list,
+ format) != 0) {
free(space);
return NULL;
}
@@ -595,8 +590,8 @@ sysview_engine_new(void)
{
struct sysview_engine *sysview = calloc(1, sizeof(*sysview));
if (sysview == NULL) {
- diag_set(OutOfMemory, sizeof(*sysview),
- "malloc", "struct sysview_engine");
+ diag_set(OutOfMemory, sizeof(*sysview), "malloc",
+ "struct sysview_engine");
return NULL;
}
diff --git a/src/box/tuple.c b/src/box/tuple.c
index f396547..88b9e32 100644
--- a/src/box/tuple.c
+++ b/src/box/tuple.c
@@ -60,7 +60,8 @@ static void
runtime_tuple_delete(struct tuple_format *format, struct tuple *tuple);
static struct tuple *
-runtime_tuple_new(struct tuple_format *format, const char *data, const char *end);
+runtime_tuple_new(struct tuple_format *format, const char *data,
+ const char *end);
/** A virtual method table for tuple_format_runtime */
static struct tuple_format_vtab tuple_format_runtime_vtab = {
@@ -71,9 +72,11 @@ static struct tuple_format_vtab tuple_format_runtime_vtab = {
};
static struct tuple *
-runtime_tuple_new(struct tuple_format *format, const char *data, const char *end)
+runtime_tuple_new(struct tuple_format *format, const char *data,
+ const char *end)
{
- assert(format->vtab.tuple_delete == tuple_format_runtime_vtab.tuple_delete);
+ assert(format->vtab.tuple_delete ==
+ tuple_format_runtime_vtab.tuple_delete);
mp_tuple_assert(data, end);
struct tuple *tuple = NULL;
@@ -93,10 +96,9 @@ runtime_tuple_new(struct tuple_format *format, const char *data, const char *end
size_t data_len = end - data;
size_t total = sizeof(struct tuple) + field_map_size + data_len;
- tuple = (struct tuple *) smalloc(&runtime_alloc, total);
+ tuple = (struct tuple *)smalloc(&runtime_alloc, total);
if (tuple == NULL) {
- diag_set(OutOfMemory, (unsigned) total,
- "malloc", "tuple");
+ diag_set(OutOfMemory, (unsigned)total, "malloc", "tuple");
goto end;
}
@@ -106,7 +108,7 @@ runtime_tuple_new(struct tuple_format *format, const char *data, const char *end
tuple_format_ref(format);
tuple->data_offset = data_offset;
tuple->is_dirty = false;
- char *raw = (char *) tuple + data_offset;
+ char *raw = (char *)tuple + data_offset;
field_map_build(&builder, raw - field_map_size);
memcpy(raw, data, data_len);
say_debug("%s(%zu) = %p", __func__, data_len, tuple);
@@ -118,7 +120,8 @@ end:
static void
runtime_tuple_delete(struct tuple_format *format, struct tuple *tuple)
{
- assert(format->vtab.tuple_delete == tuple_format_runtime_vtab.tuple_delete);
+ assert(format->vtab.tuple_delete ==
+ tuple_format_runtime_vtab.tuple_delete);
say_debug("%s(%p)", __func__, tuple);
assert(tuple->refs == 0);
size_t total = tuple_size(tuple);
@@ -239,10 +242,11 @@ bigref_list_increase_capacity(void)
capacity = MIN(capacity * BIGREF_FACTOR, BIGREF_MAX_CAPACITY);
else
panic("Too many big references");
- refs = (uint32_t *) realloc(refs, capacity * sizeof(*refs));
+ refs = (uint32_t *)realloc(refs, capacity * sizeof(*refs));
if (refs == NULL) {
- panic("failed to reallocate %zu bytes: Cannot allocate "\
- "memory.", capacity * sizeof(*refs));
+ panic("failed to reallocate %zu bytes: Cannot allocate "
+ "memory.",
+ capacity * sizeof(*refs));
}
for (uint16_t i = bigref_list.capacity; i < capacity; ++i)
refs[i] = i + 1;
@@ -269,7 +273,7 @@ void
tuple_ref_slow(struct tuple *tuple)
{
assert(tuple->is_bigref || tuple->refs == TUPLE_REF_MAX);
- if (! tuple->is_bigref) {
+ if (!tuple->is_bigref) {
tuple->ref_index = bigref_list_new_index();
tuple->is_bigref = true;
bigref_list.refs[tuple->ref_index] = TUPLE_REF_MAX;
@@ -284,7 +288,7 @@ tuple_unref_slow(struct tuple *tuple)
{
assert(tuple->is_bigref &&
bigref_list.refs[tuple->ref_index] > TUPLE_REF_MAX);
- if(--bigref_list.refs[tuple->ref_index] == TUPLE_REF_MAX) {
+ if (--bigref_list.refs[tuple->ref_index] == TUPLE_REF_MAX) {
bigref_list.refs[tuple->ref_index] = bigref_list.vacant_index;
bigref_list.vacant_index = tuple->ref_index;
tuple->ref_index = TUPLE_REF_MAX;
@@ -304,9 +308,9 @@ tuple_init(field_name_hash_f hash)
/*
* Create a format for runtime tuples
*/
- tuple_format_runtime = tuple_format_new(&tuple_format_runtime_vtab, NULL,
- NULL, 0, NULL, 0, 0, NULL, false,
- false);
+ tuple_format_runtime = tuple_format_new(&tuple_format_runtime_vtab,
+ NULL, NULL, 0, NULL, 0, 0, NULL,
+ false, false);
if (tuple_format_runtime == NULL)
return -1;
@@ -331,8 +335,8 @@ tuple_init(field_name_hash_f hash)
void
tuple_arena_create(struct slab_arena *arena, struct quota *quota,
- uint64_t arena_max_size, uint32_t slab_size,
- bool dontdump, const char *arena_name)
+ uint64_t arena_max_size, uint32_t slab_size, bool dontdump,
+ const char *arena_name)
{
/*
* Ensure that quota is a multiple of slab_size, to
@@ -340,24 +344,25 @@ tuple_arena_create(struct slab_arena *arena, struct quota *quota,
*/
size_t prealloc = small_align(arena_max_size, slab_size);
- /*
+ /*
* Skip from coredump if requested.
*/
- int flags = SLAB_ARENA_PRIVATE;
- if (dontdump)
- flags |= SLAB_ARENA_DONTDUMP;
+ int flags = SLAB_ARENA_PRIVATE;
+ if (dontdump)
+ flags |= SLAB_ARENA_DONTDUMP;
say_info("mapping %zu bytes for %s tuple arena...", prealloc,
arena_name);
if (slab_arena_create(arena, quota, prealloc, slab_size, flags) != 0) {
if (errno == ENOMEM) {
- panic("failed to preallocate %zu bytes: Cannot "\
- "allocate memory, check option '%s_memory' in box.cfg(..)", prealloc,
- arena_name);
+ panic("failed to preallocate %zu bytes: Cannot "
+ "allocate memory, check option '%s_memory' in box.cfg(..)",
+ prealloc, arena_name);
} else {
- panic_syserror("failed to preallocate %zu bytes for %s"\
- " tuple arena", prealloc, arena_name);
+ panic_syserror("failed to preallocate %zu bytes for %s"
+ " tuple arena",
+ prealloc, arena_name);
}
}
@@ -506,7 +511,7 @@ tuple_field_raw_by_full_path(struct tuple_format *format, const char *tuple,
json_lexer_create(&lexer, path, path_len, TUPLE_INDEX_BASE);
if (json_lexer_next_token(&lexer, &token) != 0)
return NULL;
- switch(token.type) {
+ switch (token.type) {
case JSON_TOKEN_NUM: {
fieldno = token.num;
break;
@@ -514,7 +519,7 @@ tuple_field_raw_by_full_path(struct tuple_format *format, const char *tuple,
case JSON_TOKEN_STR: {
/* First part of a path is a field name. */
uint32_t name_hash;
- if (path_len == (uint32_t) token.len) {
+ if (path_len == (uint32_t)token.len) {
name_hash = path_hash;
} else {
/*
@@ -537,22 +542,19 @@ tuple_field_raw_by_full_path(struct tuple_format *format, const char *tuple,
}
return tuple_field_raw_by_path(format, tuple, field_map, fieldno,
path + lexer.offset,
- path_len - lexer.offset,
- NULL, MULTIKEY_NONE);
+ path_len - lexer.offset, NULL,
+ MULTIKEY_NONE);
}
uint32_t
tuple_raw_multikey_count(struct tuple_format *format, const char *data,
- const uint32_t *field_map,
- struct key_def *key_def)
+ const uint32_t *field_map, struct key_def *key_def)
{
assert(key_def->is_multikey);
- const char *array_raw =
- tuple_field_raw_by_path(format, data, field_map,
- key_def->multikey_fieldno,
- key_def->multikey_path,
- key_def->multikey_path_len,
- NULL, MULTIKEY_NONE);
+ const char *array_raw = tuple_field_raw_by_path(
+ format, data, field_map, key_def->multikey_fieldno,
+ key_def->multikey_path, key_def->multikey_path_len, NULL,
+ MULTIKEY_NONE);
if (array_raw == NULL)
return 0;
enum mp_type type = mp_typeof(*array_raw);
@@ -576,9 +578,8 @@ box_tuple_format_t *
box_tuple_format_new(struct key_def **keys, uint16_t key_count)
{
box_tuple_format_t *format =
- tuple_format_new(&tuple_format_runtime_vtab, NULL,
- keys, key_count, NULL, 0, 0, NULL, false,
- false);
+ tuple_format_new(&tuple_format_runtime_vtab, NULL, keys,
+ key_count, NULL, 0, 0, NULL, false, false);
if (format != NULL)
tuple_format_ref(format);
return format;
@@ -651,11 +652,11 @@ box_tuple_iterator_t *
box_tuple_iterator(box_tuple_t *tuple)
{
assert(tuple != NULL);
- struct tuple_iterator *it = (struct tuple_iterator *)
- mempool_alloc(&tuple_iterator_pool);
+ struct tuple_iterator *it =
+ (struct tuple_iterator *)mempool_alloc(&tuple_iterator_pool);
if (it == NULL) {
- diag_set(OutOfMemory, tuple_iterator_pool.objsize,
- "mempool", "new slab");
+ diag_set(OutOfMemory, tuple_iterator_pool.objsize, "mempool",
+ "new slab");
return NULL;
}
tuple_ref(tuple);
@@ -702,9 +703,9 @@ box_tuple_update(box_tuple_t *tuple, const char *expr, const char *expr_end)
struct region *region = &fiber()->gc;
size_t used = region_used(region);
struct tuple_format *format = tuple_format(tuple);
- const char *new_data =
- xrow_update_execute(expr, expr_end, old_data, old_data + bsize,
- format, &new_size, 1, NULL);
+ const char *new_data = xrow_update_execute(expr, expr_end, old_data,
+ old_data + bsize, format,
+ &new_size, 1, NULL);
if (new_data == NULL) {
region_truncate(region, used);
return NULL;
@@ -724,9 +725,9 @@ box_tuple_upsert(box_tuple_t *tuple, const char *expr, const char *expr_end)
struct region *region = &fiber()->gc;
size_t used = region_used(region);
struct tuple_format *format = tuple_format(tuple);
- const char *new_data =
- xrow_upsert_execute(expr, expr_end, old_data, old_data + bsize,
- format, &new_size, 1, false, NULL);
+ const char *new_data = xrow_upsert_execute(expr, expr_end, old_data,
+ old_data + bsize, format,
+ &new_size, 1, false, NULL);
if (new_data == NULL) {
region_truncate(region, used);
return NULL;
diff --git a/src/box/tuple.h b/src/box/tuple.h
index 53ae690..0d1dbac 100644
--- a/src/box/tuple.h
+++ b/src/box/tuple.h
@@ -69,8 +69,8 @@ tuple_free(void);
*/
void
tuple_arena_create(struct slab_arena *arena, struct quota *quota,
- uint64_t arena_max_size, uint32_t slab_size,
- bool dontdump, const char *arena_name);
+ uint64_t arena_max_size, uint32_t slab_size, bool dontdump,
+ const char *arena_name);
void
tuple_arena_destroy(struct slab_arena *arena);
@@ -297,8 +297,7 @@ box_tuple_upsert(box_tuple_t *tuple, const char *expr, const char *expr_end);
*
* Each 'off_i' is the offset to the i-th indexed field.
*/
-struct PACKED tuple
-{
+struct PACKED tuple {
union {
/** Reference counter. */
uint16_t refs;
@@ -349,7 +348,7 @@ tuple_size(struct tuple *tuple)
static inline const char *
tuple_data(struct tuple *tuple)
{
- return (const char *) tuple + tuple->data_offset;
+ return (const char *)tuple + tuple->data_offset;
}
/**
@@ -371,7 +370,7 @@ static inline const char *
tuple_data_range(struct tuple *tuple, uint32_t *p_size)
{
*p_size = tuple->bsize;
- return (const char *) tuple + tuple->data_offset;
+ return (const char *)tuple + tuple->data_offset;
}
/**
@@ -524,7 +523,7 @@ tuple_validate(struct tuple_format *format, struct tuple *tuple)
static inline const uint32_t *
tuple_field_map(struct tuple *tuple)
{
- return (const uint32_t *) ((const char *) tuple + tuple->data_offset);
+ return (const uint32_t *)((const char *)tuple + tuple->data_offset);
}
/**
@@ -647,7 +646,7 @@ tuple_field_raw_by_path(struct tuple_format *format, const char *tuple,
*/
goto parse;
}
-offset_slot_access:
+ offset_slot_access:
/* Indexed field */
offset = field_map_get_offset(field_map, offset_slot,
multikey_idx);
@@ -656,7 +655,7 @@ offset_slot_access:
tuple += offset;
} else {
uint32_t field_count;
-parse:
+ parse:
ERROR_INJECT(ERRINJ_TUPLE_FIELD, return NULL);
field_count = mp_decode_array(&tuple);
if (unlikely(fieldno >= field_count))
@@ -686,8 +685,8 @@ static inline const char *
tuple_field_raw(struct tuple_format *format, const char *tuple,
const uint32_t *field_map, uint32_t field_no)
{
- return tuple_field_raw_by_path(format, tuple, field_map, field_no,
- NULL, 0, NULL, MULTIKEY_NONE);
+ return tuple_field_raw_by_path(format, tuple, field_map, field_no, NULL,
+ 0, NULL, MULTIKEY_NONE);
}
/**
@@ -737,8 +736,8 @@ tuple_field_raw_by_full_path(struct tuple_format *format, const char *tuple,
*/
static inline const char *
tuple_field_raw_by_part(struct tuple_format *format, const char *data,
- const uint32_t *field_map,
- struct key_part *part, int multikey_idx)
+ const uint32_t *field_map, struct key_part *part,
+ int multikey_idx)
{
if (unlikely(part->format_epoch != format->epoch)) {
assert(format->epoch != 0);
@@ -837,7 +836,7 @@ tuple_rewind(struct tuple_iterator *it, struct tuple *tuple)
uint32_t bsize;
const char *data = tuple_data_range(tuple, &bsize);
it->pos = data;
- (void) mp_decode_array(&it->pos); /* Skip array header */
+ (void)mp_decode_array(&it->pos); /* Skip array header */
it->fieldno = 0;
it->end = data + bsize;
}
@@ -921,8 +920,8 @@ mp_tuple_assert(const char *tuple, const char *tuple_end)
mp_next(&tuple);
#endif
assert(tuple == tuple_end);
- (void) tuple;
- (void) tuple_end;
+ (void)tuple;
+ (void)tuple_end;
}
static inline const char *
@@ -1158,4 +1157,3 @@ tuple_field_u32_xc(struct tuple *tuple, uint32_t fieldno)
#endif /* defined(__cplusplus) */
#endif /* TARANTOOL_BOX_TUPLE_H_INCLUDED */
-
diff --git a/src/box/tuple_bloom.c b/src/box/tuple_bloom.c
index 420a7c6..88dc6ed 100644
--- a/src/box/tuple_bloom.c
+++ b/src/box/tuple_bloom.c
@@ -51,7 +51,7 @@ struct tuple_bloom_builder *
tuple_bloom_builder_new(uint32_t part_count)
{
size_t size = sizeof(struct tuple_bloom_builder) +
- part_count * sizeof(struct tuple_hash_array);
+ part_count * sizeof(struct tuple_hash_array);
struct tuple_bloom_builder *builder = malloc(size);
if (builder == NULL) {
diag_set(OutOfMemory, size, "malloc", "tuple bloom builder");
@@ -89,8 +89,8 @@ tuple_hash_array_add(struct tuple_hash_array *hash_arr, uint32_t hash)
}
if (hash_arr->count >= hash_arr->capacity) {
uint32_t capacity = MAX(hash_arr->capacity * 2, 1024U);
- uint32_t *values = realloc(hash_arr->values,
- capacity * sizeof(*values));
+ uint32_t *values =
+ realloc(hash_arr->values, capacity * sizeof(*values));
if (values == NULL) {
diag_set(OutOfMemory, capacity * sizeof(*values),
"malloc", "tuple hash array");
@@ -116,9 +116,8 @@ tuple_bloom_builder_add(struct tuple_bloom_builder *builder,
uint32_t total_size = 0;
for (uint32_t i = 0; i < key_def->part_count; i++) {
- total_size += tuple_hash_key_part(&h, &carry, tuple,
- &key_def->parts[i],
- multikey_idx);
+ total_size += tuple_hash_key_part(
+ &h, &carry, tuple, &key_def->parts[i], multikey_idx);
uint32_t hash = PMurHash32_Result(h, carry, total_size);
if (tuple_hash_array_add(&builder->parts[i], hash) != 0)
return -1;
@@ -153,8 +152,8 @@ struct tuple_bloom *
tuple_bloom_new(struct tuple_bloom_builder *builder, double fpr)
{
uint32_t part_count = builder->part_count;
- size_t size = sizeof(struct tuple_bloom) +
- part_count * sizeof(struct bloom);
+ size_t size =
+ sizeof(struct tuple_bloom) + part_count * sizeof(struct bloom);
struct tuple_bloom *bloom = malloc(size);
if (bloom == NULL) {
diag_set(OutOfMemory, size, "malloc", "tuple bloom");
@@ -218,9 +217,8 @@ tuple_bloom_maybe_has(const struct tuple_bloom *bloom, struct tuple *tuple,
uint32_t total_size = 0;
for (uint32_t i = 0; i < key_def->part_count; i++) {
- total_size += tuple_hash_key_part(&h, &carry, tuple,
- &key_def->parts[i],
- multikey_idx);
+ total_size += tuple_hash_key_part(
+ &h, &carry, tuple, &key_def->parts[i], multikey_idx);
uint32_t hash = PMurHash32_Result(h, carry, total_size);
if (!bloom_maybe_has(&bloom->parts[i], hash))
return false;
@@ -229,9 +227,8 @@ tuple_bloom_maybe_has(const struct tuple_bloom *bloom, struct tuple *tuple,
}
bool
-tuple_bloom_maybe_has_key(const struct tuple_bloom *bloom,
- const char *key, uint32_t part_count,
- struct key_def *key_def)
+tuple_bloom_maybe_has_key(const struct tuple_bloom *bloom, const char *key,
+ uint32_t part_count, struct key_def *key_def)
{
if (bloom->is_legacy) {
if (part_count < key_def->part_count)
@@ -321,11 +318,11 @@ struct tuple_bloom *
tuple_bloom_decode(const char **data)
{
uint32_t part_count = mp_decode_array(data);
- struct tuple_bloom *bloom = malloc(sizeof(*bloom) +
- part_count * sizeof(*bloom->parts));
+ struct tuple_bloom *bloom =
+ malloc(sizeof(*bloom) + part_count * sizeof(*bloom->parts));
if (bloom == NULL) {
- diag_set(OutOfMemory, sizeof(*bloom) +
- part_count * sizeof(*bloom->parts),
+ diag_set(OutOfMemory,
+ sizeof(*bloom) + part_count * sizeof(*bloom->parts),
"malloc", "tuple bloom");
return NULL;
}
@@ -346,8 +343,8 @@ tuple_bloom_decode(const char **data)
struct tuple_bloom *
tuple_bloom_decode_legacy(const char **data)
{
- struct tuple_bloom *bloom = malloc(sizeof(*bloom) +
- sizeof(*bloom->parts));
+ struct tuple_bloom *bloom =
+ malloc(sizeof(*bloom) + sizeof(*bloom->parts));
if (bloom == NULL) {
diag_set(OutOfMemory, sizeof(*bloom) + sizeof(*bloom->parts),
"malloc", "tuple bloom");
diff --git a/src/box/tuple_bloom.h b/src/box/tuple_bloom.h
index 1b7e4ac..ca6bfc4 100644
--- a/src/box/tuple_bloom.h
+++ b/src/box/tuple_bloom.h
@@ -189,9 +189,8 @@ tuple_bloom_maybe_has(const struct tuple_bloom *bloom, struct tuple *tuple,
* the bloom, false if there is definitely no such tuple
*/
bool
-tuple_bloom_maybe_has_key(const struct tuple_bloom *bloom,
- const char *key, uint32_t part_count,
- struct key_def *key_def);
+tuple_bloom_maybe_has_key(const struct tuple_bloom *bloom, const char *key,
+ uint32_t part_count, struct key_def *key_def);
/**
* Return the size of a tuple bloom filter when encoded.
diff --git a/src/box/tuple_compare.cc b/src/box/tuple_compare.cc
index bb786cc..4cc20c2 100644
--- a/src/box/tuple_compare.cc
+++ b/src/box/tuple_compare.cc
@@ -175,8 +175,8 @@ mp_compare_double_any_int(double lhs, const char *rhs, enum mp_type rhs_type,
}
static int
-mp_compare_double_any_number(double lhs, const char *rhs,
- enum mp_type rhs_type, int k)
+mp_compare_double_any_number(double lhs, const char *rhs, enum mp_type rhs_type,
+ int k)
{
double v;
if (rhs_type == MP_FLOAT)
@@ -225,7 +225,6 @@ mp_compare_decimal(const char *lhs, const char *rhs)
assert(ret != NULL);
(void)ret;
return decimal_compare(&lhs_dec, &rhs_dec);
-
}
static int
@@ -234,32 +233,27 @@ mp_compare_decimal_any_number(decimal_t *lhs, const char *rhs,
{
decimal_t rhs_dec;
switch (rhs_type) {
- case MP_FLOAT:
- {
+ case MP_FLOAT: {
double d = mp_decode_float(&rhs);
decimal_from_double(&rhs_dec, d);
break;
}
- case MP_DOUBLE:
- {
+ case MP_DOUBLE: {
double d = mp_decode_double(&rhs);
decimal_from_double(&rhs_dec, d);
break;
}
- case MP_INT:
- {
+ case MP_INT: {
int64_t num = mp_decode_int(&rhs);
decimal_from_int64(&rhs_dec, num);
break;
}
- case MP_UINT:
- {
+ case MP_UINT: {
uint64_t num = mp_decode_uint(&rhs);
decimal_from_uint64(&rhs_dec, num);
break;
}
- case MP_EXT:
- {
+ case MP_EXT: {
int8_t ext_type;
uint32_t len = mp_decode_extl(&rhs, &ext_type);
switch (ext_type) {
@@ -297,8 +291,8 @@ mp_compare_number_with_type(const char *lhs, enum mp_type lhs_type,
switch (ext_type) {
case MP_DECIMAL:
return mp_compare_decimal_any_number(
- decimal_unpack(&rhs, len, &dec), lhs, lhs_type, -1
- );
+ decimal_unpack(&rhs, len, &dec), lhs, lhs_type,
+ -1);
default:
unreachable();
}
@@ -309,32 +303,28 @@ mp_compare_number_with_type(const char *lhs, enum mp_type lhs_type,
switch (ext_type) {
case MP_DECIMAL:
return mp_compare_decimal_any_number(
- decimal_unpack(&lhs, len, &dec), rhs, rhs_type, 1
- );
+ decimal_unpack(&lhs, len, &dec), rhs, rhs_type,
+ 1);
default:
unreachable();
}
}
if (rhs_type == MP_FLOAT) {
- return mp_compare_double_any_number(
- mp_decode_float(&rhs), lhs, lhs_type, -1
- );
+ return mp_compare_double_any_number(mp_decode_float(&rhs), lhs,
+ lhs_type, -1);
}
if (rhs_type == MP_DOUBLE) {
- return mp_compare_double_any_number(
- mp_decode_double(&rhs), lhs, lhs_type, -1
- );
+ return mp_compare_double_any_number(mp_decode_double(&rhs), lhs,
+ lhs_type, -1);
}
assert(rhs_type == MP_INT || rhs_type == MP_UINT);
if (lhs_type == MP_FLOAT) {
- return mp_compare_double_any_int(
- mp_decode_float(&lhs), rhs, rhs_type, 1
- );
+ return mp_compare_double_any_int(mp_decode_float(&lhs), rhs,
+ rhs_type, 1);
}
if (lhs_type == MP_DOUBLE) {
- return mp_compare_double_any_int(
- mp_decode_double(&lhs), rhs, rhs_type, 1
- );
+ return mp_compare_double_any_int(mp_decode_double(&lhs), rhs,
+ rhs_type, 1);
}
assert(lhs_type == MP_INT || lhs_type == MP_UINT);
return mp_compare_integer_with_type(lhs, lhs_type, rhs, rhs_type);
@@ -343,8 +333,8 @@ mp_compare_number_with_type(const char *lhs, enum mp_type lhs_type,
static inline int
mp_compare_number(const char *lhs, const char *rhs)
{
- return mp_compare_number_with_type(lhs, mp_typeof(*lhs),
- rhs, mp_typeof(*rhs));
+ return mp_compare_number_with_type(lhs, mp_typeof(*lhs), rhs,
+ mp_typeof(*rhs));
}
static inline int
@@ -407,11 +397,11 @@ mp_compare_scalar_with_type(const char *field_a, enum mp_type a_type,
const char *field_b, enum mp_type b_type)
{
enum mp_class a_class = mp_classof(a_type) < mp_class_max ?
- mp_classof(a_type) :
- mp_extension_class(field_a);
+ mp_classof(a_type) :
+ mp_extension_class(field_a);
enum mp_class b_class = mp_classof(b_type) < mp_class_max ?
- mp_classof(b_type) :
- mp_extension_class(field_b);
+ mp_classof(b_type) :
+ mp_extension_class(field_b);
if (a_class != b_class)
return COMPARE_RESULT(a_class, b_class);
mp_compare_f cmp = mp_class_comparators[a_class];
@@ -447,16 +437,16 @@ mp_compare_scalar_coll(const char *field_a, const char *field_b,
* @retval >0 if field_a > field_b
*/
static int
-tuple_compare_field(const char *field_a, const char *field_b,
- int8_t type, struct coll *coll)
+tuple_compare_field(const char *field_a, const char *field_b, int8_t type,
+ struct coll *coll)
{
switch (type) {
case FIELD_TYPE_UNSIGNED:
return mp_compare_uint(field_a, field_b);
case FIELD_TYPE_STRING:
return coll != NULL ?
- mp_compare_str_coll(field_a, field_b, coll) :
- mp_compare_str(field_a, field_b);
+ mp_compare_str_coll(field_a, field_b, coll) :
+ mp_compare_str(field_a, field_b);
case FIELD_TYPE_INTEGER:
return mp_compare_integer_with_type(field_a,
mp_typeof(*field_a),
@@ -472,8 +462,8 @@ tuple_compare_field(const char *field_a, const char *field_b,
return mp_compare_bin(field_a, field_b);
case FIELD_TYPE_SCALAR:
return coll != NULL ?
- mp_compare_scalar_coll(field_a, field_b, coll) :
- mp_compare_scalar(field_a, field_b);
+ mp_compare_scalar_coll(field_a, field_b, coll) :
+ mp_compare_scalar(field_a, field_b);
case FIELD_TYPE_DECIMAL:
return mp_compare_decimal(field_a, field_b);
case FIELD_TYPE_UUID:
@@ -494,14 +484,14 @@ tuple_compare_field_with_type(const char *field_a, enum mp_type a_type,
return mp_compare_uint(field_a, field_b);
case FIELD_TYPE_STRING:
return coll != NULL ?
- mp_compare_str_coll(field_a, field_b, coll) :
- mp_compare_str(field_a, field_b);
+ mp_compare_str_coll(field_a, field_b, coll) :
+ mp_compare_str(field_a, field_b);
case FIELD_TYPE_INTEGER:
- return mp_compare_integer_with_type(field_a, a_type,
- field_b, b_type);
+ return mp_compare_integer_with_type(field_a, a_type, field_b,
+ b_type);
case FIELD_TYPE_NUMBER:
- return mp_compare_number_with_type(field_a, a_type,
- field_b, b_type);
+ return mp_compare_number_with_type(field_a, a_type, field_b,
+ b_type);
case FIELD_TYPE_DOUBLE:
return mp_compare_double(field_a, field_b);
case FIELD_TYPE_BOOLEAN:
@@ -510,12 +500,12 @@ tuple_compare_field_with_type(const char *field_a, enum mp_type a_type,
return mp_compare_bin(field_a, field_b);
case FIELD_TYPE_SCALAR:
return coll != NULL ?
- mp_compare_scalar_coll(field_a, field_b, coll) :
- mp_compare_scalar_with_type(field_a, a_type,
- field_b, b_type);
+ mp_compare_scalar_coll(field_a, field_b, coll) :
+ mp_compare_scalar_with_type(field_a, a_type,
+ field_b, b_type);
case FIELD_TYPE_DECIMAL:
- return mp_compare_number_with_type(field_a, a_type,
- field_b, b_type);
+ return mp_compare_number_with_type(field_a, a_type, field_b,
+ b_type);
case FIELD_TYPE_UUID:
return mp_compare_uuid(field_a, field_b);
default:
@@ -524,8 +514,8 @@ tuple_compare_field_with_type(const char *field_a, enum mp_type a_type,
}
}
-template<bool is_nullable, bool has_optional_parts, bool has_json_paths,
- bool is_multikey>
+template <bool is_nullable, bool has_optional_parts, bool has_json_paths,
+ bool is_multikey>
static inline int
tuple_compare_slowpath(struct tuple *tuple_a, hint_t tuple_a_hint,
struct tuple *tuple_b, hint_t tuple_b_hint,
@@ -536,8 +526,8 @@ tuple_compare_slowpath(struct tuple *tuple_a, hint_t tuple_a_hint,
assert(is_nullable == key_def->is_nullable);
assert(has_optional_parts == key_def->has_optional_parts);
assert(key_def->is_multikey == is_multikey);
- assert(!is_multikey || (tuple_a_hint != HINT_NONE &&
- tuple_b_hint != HINT_NONE));
+ assert(!is_multikey ||
+ (tuple_a_hint != HINT_NONE && tuple_b_hint != HINT_NONE));
int rc = 0;
if (!is_multikey && (rc = hint_cmp(tuple_a_hint, tuple_b_hint)) != 0)
return rc;
@@ -553,7 +543,7 @@ tuple_compare_slowpath(struct tuple *tuple_a, hint_t tuple_a_hint,
assert(!has_optional_parts);
mp_decode_array(&tuple_a_raw);
mp_decode_array(&tuple_b_raw);
- if (! is_nullable) {
+ if (!is_nullable) {
return tuple_compare_field(tuple_a_raw, tuple_b_raw,
part->type, part->coll);
}
@@ -563,7 +553,7 @@ tuple_compare_slowpath(struct tuple *tuple_a, hint_t tuple_a_hint,
return b_type == MP_NIL ? 0 : -1;
else if (b_type == MP_NIL)
return 1;
- return tuple_compare_field_with_type(tuple_a_raw, a_type,
+ return tuple_compare_field_with_type(tuple_a_raw, a_type,
tuple_b_raw, b_type,
part->type, part->coll);
}
@@ -604,7 +594,7 @@ tuple_compare_slowpath(struct tuple *tuple_a, hint_t tuple_a_hint,
}
assert(has_optional_parts ||
(field_a != NULL && field_b != NULL));
- if (! is_nullable) {
+ if (!is_nullable) {
rc = tuple_compare_field(field_a, field_b, part->type,
part->coll);
if (rc != 0)
@@ -681,8 +671,8 @@ tuple_compare_slowpath(struct tuple *tuple_a, hint_t tuple_a_hint,
return 0;
}
-template<bool is_nullable, bool has_optional_parts, bool has_json_paths,
- bool is_multikey>
+template <bool is_nullable, bool has_optional_parts, bool has_json_paths,
+ bool is_multikey>
static inline int
tuple_compare_with_key_slowpath(struct tuple *tuple, hint_t tuple_hint,
const char *key, uint32_t part_count,
@@ -695,8 +685,8 @@ tuple_compare_with_key_slowpath(struct tuple *tuple, hint_t tuple_hint,
assert(key != NULL || part_count == 0);
assert(part_count <= key_def->part_count);
assert(key_def->is_multikey == is_multikey);
- assert(!is_multikey || (tuple_hint != HINT_NONE &&
- key_hint == HINT_NONE));
+ assert(!is_multikey ||
+ (tuple_hint != HINT_NONE && key_hint == HINT_NONE));
int rc = 0;
if (!is_multikey && (rc = hint_cmp(tuple_hint, key_hint)) != 0)
return rc;
@@ -719,7 +709,7 @@ tuple_compare_with_key_slowpath(struct tuple *tuple, hint_t tuple_hint,
field = tuple_field_raw(format, tuple_raw, field_map,
part->fieldno);
}
- if (! is_nullable) {
+ if (!is_nullable) {
return tuple_compare_field(field, key, part->type,
part->coll);
}
@@ -754,7 +744,7 @@ tuple_compare_with_key_slowpath(struct tuple *tuple, hint_t tuple_hint,
field = tuple_field_raw(format, tuple_raw, field_map,
part->fieldno);
}
- if (! is_nullable) {
+ if (!is_nullable) {
rc = tuple_compare_field(field, key, part->type,
part->coll);
if (rc != 0)
@@ -783,7 +773,7 @@ tuple_compare_with_key_slowpath(struct tuple *tuple, hint_t tuple_hint,
return 0;
}
-template<bool is_nullable>
+template <bool is_nullable>
static inline int
key_compare_parts(const char *key_a, const char *key_b, uint32_t part_count,
struct key_def *key_def)
@@ -792,7 +782,7 @@ key_compare_parts(const char *key_a, const char *key_b, uint32_t part_count,
assert((key_a != NULL && key_b != NULL) || part_count == 0);
struct key_part *part = key_def->parts;
if (likely(part_count == 1)) {
- if (! is_nullable) {
+ if (!is_nullable) {
return tuple_compare_field(key_a, key_b, part->type,
part->coll);
}
@@ -813,7 +803,7 @@ key_compare_parts(const char *key_a, const char *key_b, uint32_t part_count,
struct key_part *end = part + part_count;
int rc;
for (; part < end; ++part, mp_next(&key_a), mp_next(&key_b)) {
- if (! is_nullable) {
+ if (!is_nullable) {
rc = tuple_compare_field(key_a, key_b, part->type,
part->coll);
if (rc != 0)
@@ -839,7 +829,7 @@ key_compare_parts(const char *key_a, const char *key_b, uint32_t part_count,
return 0;
}
-template<bool is_nullable, bool has_optional_parts>
+template <bool is_nullable, bool has_optional_parts>
static inline int
tuple_compare_with_key_sequential(struct tuple *tuple, hint_t tuple_hint,
const char *key, uint32_t part_count,
@@ -885,8 +875,8 @@ tuple_compare_with_key_sequential(struct tuple *tuple, hint_t tuple_hint,
}
int
-key_compare(const char *key_a, hint_t key_a_hint,
- const char *key_b, hint_t key_b_hint, struct key_def *key_def)
+key_compare(const char *key_a, hint_t key_a_hint, const char *key_b,
+ hint_t key_b_hint, struct key_def *key_def)
{
int rc = hint_cmp(key_a_hint, key_b_hint);
if (rc != 0)
@@ -897,7 +887,7 @@ key_compare(const char *key_a, hint_t key_a_hint,
assert(part_count_b <= key_def->part_count);
uint32_t part_count = MIN(part_count_a, part_count_b);
assert(part_count <= key_def->part_count);
- if (! key_def->is_nullable) {
+ if (!key_def->is_nullable) {
return key_compare_parts<false>(key_a, key_b, part_count,
key_def);
} else {
@@ -960,7 +950,7 @@ tuple_compare_sequential(struct tuple *tuple_a, hint_t tuple_a_hint,
if (!has_optional_parts || i < fc_b)
mp_next(&key_b);
}
- if (! was_null_met)
+ if (!was_null_met)
return 0;
end = key_def->parts + key_def->part_count;
for (; part < end; ++part, ++i, mp_next(&key_a), mp_next(&key_b)) {
@@ -970,8 +960,7 @@ tuple_compare_sequential(struct tuple *tuple_a, hint_t tuple_a_hint,
* not be absent or be null.
*/
assert(i < fc_a && i < fc_b);
- rc = tuple_compare_field(key_a, key_b, part->type,
- part->coll);
+ rc = tuple_compare_field(key_a, key_b, part->type, part->coll);
if (rc != 0)
return rc;
}
@@ -1036,20 +1025,17 @@ field_compare_and_next<FIELD_TYPE_STRING>(const char **field_a,
/* Tuple comparator */
namespace /* local symbols */ {
-template <int IDX, int TYPE, int ...MORE_TYPES> struct FieldCompare { };
+template <int IDX, int TYPE, int... MORE_TYPES> struct FieldCompare {};
/**
* Common case.
*/
-template <int IDX, int TYPE, int IDX2, int TYPE2, int ...MORE_TYPES>
-struct FieldCompare<IDX, TYPE, IDX2, TYPE2, MORE_TYPES...>
-{
- inline static int compare(struct tuple *tuple_a,
- struct tuple *tuple_b,
+template <int IDX, int TYPE, int IDX2, int TYPE2, int... MORE_TYPES>
+struct FieldCompare<IDX, TYPE, IDX2, TYPE2, MORE_TYPES...> {
+ inline static int compare(struct tuple *tuple_a, struct tuple *tuple_b,
struct tuple_format *format_a,
struct tuple_format *format_b,
- const char *field_a,
- const char *field_b)
+ const char *field_a, const char *field_b)
{
int r;
/* static if */
@@ -1067,21 +1053,15 @@ struct FieldCompare<IDX, TYPE, IDX2, TYPE2, MORE_TYPES...>
tuple_field_map(tuple_b),
IDX2);
}
- return FieldCompare<IDX2, TYPE2, MORE_TYPES...>::
- compare(tuple_a, tuple_b, format_a,
- format_b, field_a, field_b);
+ return FieldCompare<IDX2, TYPE2, MORE_TYPES...>::compare(
+ tuple_a, tuple_b, format_a, format_b, field_a, field_b);
}
};
-template <int IDX, int TYPE>
-struct FieldCompare<IDX, TYPE>
-{
- inline static int compare(struct tuple *,
- struct tuple *,
- struct tuple_format *,
- struct tuple_format *,
- const char *field_a,
- const char *field_b)
+template <int IDX, int TYPE> struct FieldCompare<IDX, TYPE> {
+ inline static int compare(struct tuple *, struct tuple *,
+ struct tuple_format *, struct tuple_format *,
+ const char *field_a, const char *field_b)
{
return field_compare<TYPE>(&field_a, &field_b);
}
@@ -1090,9 +1070,7 @@ struct FieldCompare<IDX, TYPE>
/**
* header
*/
-template <int IDX, int TYPE, int ...MORE_TYPES>
-struct TupleCompare
-{
+template <int IDX, int TYPE, int... MORE_TYPES> struct TupleCompare {
static int compare(struct tuple *tuple_a, hint_t tuple_a_hint,
struct tuple *tuple_b, hint_t tuple_b_hint,
struct key_def *)
@@ -1107,13 +1085,12 @@ struct TupleCompare
tuple_field_map(tuple_a), IDX);
field_b = tuple_field_raw(format_b, tuple_data(tuple_b),
tuple_field_map(tuple_b), IDX);
- return FieldCompare<IDX, TYPE, MORE_TYPES...>::
- compare(tuple_a, tuple_b, format_a,
- format_b, field_a, field_b);
+ return FieldCompare<IDX, TYPE, MORE_TYPES...>::compare(
+ tuple_a, tuple_b, format_a, format_b, field_a, field_b);
}
};
-template <int TYPE, int ...MORE_TYPES>
+template <int TYPE, int... MORE_TYPES>
struct TupleCompare<0, TYPE, MORE_TYPES...> {
static int compare(struct tuple *tuple_a, hint_t tuple_a_hint,
struct tuple *tuple_b, hint_t tuple_b_hint,
@@ -1128,8 +1105,8 @@ struct TupleCompare<0, TYPE, MORE_TYPES...> {
const char *field_b = tuple_data(tuple_b);
mp_decode_array(&field_a);
mp_decode_array(&field_b);
- return FieldCompare<0, TYPE, MORE_TYPES...>::compare(tuple_a, tuple_b,
- format_a, format_b, field_a, field_b);
+ return FieldCompare<0, TYPE, MORE_TYPES...>::compare(
+ tuple_a, tuple_b, format_a, format_b, field_a, field_b);
}
};
} /* end of anonymous namespace */
@@ -1170,11 +1147,13 @@ static const comparator_signature cmp_arr[] = {
/* {{{ tuple_compare_with_key */
template <int TYPE>
-static inline int field_compare_with_key(const char **field, const char **key);
+static inline int
+field_compare_with_key(const char **field, const char **key);
template <>
inline int
-field_compare_with_key<FIELD_TYPE_UNSIGNED>(const char **field, const char **key)
+field_compare_with_key<FIELD_TYPE_UNSIGNED>(const char **field,
+ const char **key)
{
return mp_compare_uint(*field, *key);
}
@@ -1210,7 +1189,7 @@ field_compare_with_key_and_next<FIELD_TYPE_UNSIGNED>(const char **field_a,
template <>
inline int
field_compare_with_key_and_next<FIELD_TYPE_STRING>(const char **field_a,
- const char **field_b)
+ const char **field_b)
{
uint32_t size_a, size_b;
size_a = mp_decode_strl(field_a);
@@ -1226,18 +1205,17 @@ field_compare_with_key_and_next<FIELD_TYPE_STRING>(const char **field_a,
/* Tuple with key comparator */
namespace /* local symbols */ {
-template <int FLD_ID, int IDX, int TYPE, int ...MORE_TYPES>
+template <int FLD_ID, int IDX, int TYPE, int... MORE_TYPES>
struct FieldCompareWithKey {};
/**
* common
*/
-template <int FLD_ID, int IDX, int TYPE, int IDX2, int TYPE2, int ...MORE_TYPES>
-struct FieldCompareWithKey<FLD_ID, IDX, TYPE, IDX2, TYPE2, MORE_TYPES...>
-{
- inline static int
- compare(struct tuple *tuple, const char *key, uint32_t part_count,
- struct key_def *key_def, struct tuple_format *format,
- const char *field)
+template <int FLD_ID, int IDX, int TYPE, int IDX2, int TYPE2, int... MORE_TYPES>
+struct FieldCompareWithKey<FLD_ID, IDX, TYPE, IDX2, TYPE2, MORE_TYPES...> {
+ inline static int compare(struct tuple *tuple, const char *key,
+ uint32_t part_count, struct key_def *key_def,
+ struct tuple_format *format,
+ const char *field)
{
int r;
/* static if */
@@ -1253,19 +1231,19 @@ struct FieldCompareWithKey<FLD_ID, IDX, TYPE, IDX2, TYPE2, MORE_TYPES...>
tuple_field_map(tuple), IDX2);
mp_next(&key);
}
- return FieldCompareWithKey<FLD_ID + 1, IDX2, TYPE2, MORE_TYPES...>::
- compare(tuple, key, part_count,
- key_def, format, field);
+ return FieldCompareWithKey<FLD_ID + 1, IDX2, TYPE2,
+ MORE_TYPES...>::compare(tuple, key,
+ part_count,
+ key_def,
+ format,
+ field);
}
};
template <int FLD_ID, int IDX, int TYPE>
struct FieldCompareWithKey<FLD_ID, IDX, TYPE> {
- inline static int compare(struct tuple *,
- const char *key,
- uint32_t,
- struct key_def *,
- struct tuple_format *,
+ inline static int compare(struct tuple *, const char *key, uint32_t,
+ struct key_def *, struct tuple_format *,
const char *field)
{
return field_compare_with_key<TYPE>(&field, &key);
@@ -1275,13 +1253,11 @@ struct FieldCompareWithKey<FLD_ID, IDX, TYPE> {
/**
* header
*/
-template <int FLD_ID, int IDX, int TYPE, int ...MORE_TYPES>
-struct TupleCompareWithKey
-{
- static int
- compare(struct tuple *tuple, hint_t tuple_hint,
- const char *key, uint32_t part_count,
- hint_t key_hint, struct key_def *key_def)
+template <int FLD_ID, int IDX, int TYPE, int... MORE_TYPES>
+struct TupleCompareWithKey {
+ static int compare(struct tuple *tuple, hint_t tuple_hint,
+ const char *key, uint32_t part_count,
+ hint_t key_hint, struct key_def *key_def)
{
/* Part count can be 0 in wildcard searches. */
if (part_count == 0)
@@ -1290,18 +1266,19 @@ struct TupleCompareWithKey
if (rc != 0)
return rc;
struct tuple_format *format = tuple_format(tuple);
- const char *field = tuple_field_raw(format, tuple_data(tuple),
- tuple_field_map(tuple),
- IDX);
- return FieldCompareWithKey<FLD_ID, IDX, TYPE, MORE_TYPES...>::
- compare(tuple, key, part_count,
- key_def, format, field);
+ const char *field = tuple_field_raw(
+ format, tuple_data(tuple), tuple_field_map(tuple), IDX);
+ return FieldCompareWithKey<FLD_ID, IDX, TYPE,
+ MORE_TYPES...>::compare(tuple, key,
+ part_count,
+ key_def,
+ format,
+ field);
}
};
-template <int TYPE, int ...MORE_TYPES>
-struct TupleCompareWithKey<0, 0, TYPE, MORE_TYPES...>
-{
+template <int TYPE, int... MORE_TYPES>
+struct TupleCompareWithKey<0, 0, TYPE, MORE_TYPES...> {
static int compare(struct tuple *tuple, hint_t tuple_hint,
const char *key, uint32_t part_count,
hint_t key_hint, struct key_def *key_def)
@@ -1315,16 +1292,14 @@ struct TupleCompareWithKey<0, 0, TYPE, MORE_TYPES...>
struct tuple_format *format = tuple_format(tuple);
const char *field = tuple_data(tuple);
mp_decode_array(&field);
- return FieldCompareWithKey<0, 0, TYPE, MORE_TYPES...>::
- compare(tuple, key, part_count,
- key_def, format, field);
+ return FieldCompareWithKey<0, 0, TYPE, MORE_TYPES...>::compare(
+ tuple, key, part_count, key_def, format, field);
}
};
} /* end of anonymous namespace */
-struct comparator_with_key_signature
-{
+struct comparator_with_key_signature {
tuple_compare_with_key_t f;
uint32_t p[64];
};
@@ -1360,7 +1335,7 @@ static const comparator_with_key_signature cmp_wk_arr[] = {
* and the primary key. So its tail parts are taken from primary
* index key definition.
*/
-template<bool is_nullable>
+template <bool is_nullable>
static inline int
func_index_compare(struct tuple *tuple_a, hint_t tuple_a_hint,
struct tuple *tuple_b, hint_t tuple_b_hint,
@@ -1418,13 +1393,14 @@ func_index_compare(struct tuple *tuple_a, hint_t tuple_a_hint,
* functional key memory and is compared with the given key by
* using the functional index key definition.
*/
-template<bool is_nullable>
+template <bool is_nullable>
static inline int
func_index_compare_with_key(struct tuple *tuple, hint_t tuple_hint,
const char *key, uint32_t part_count,
hint_t key_hint, struct key_def *key_def)
{
- (void)tuple; (void)key_hint;
+ (void)tuple;
+ (void)key_hint;
assert(key_def->for_func_index);
assert(is_nullable == key_def->is_nullable);
const char *tuple_key = (const char *)tuple_hint;
@@ -1497,30 +1473,30 @@ func_index_compare_with_key(struct tuple *tuple, hint_t tuple_hint,
* Note: comparison hint only makes sense for non-multikey
* indexes.
*/
-#define HINT_BITS (sizeof(hint_t) * CHAR_BIT)
-#define HINT_CLASS_BITS 4
-#define HINT_VALUE_BITS (HINT_BITS - HINT_CLASS_BITS)
+#define HINT_BITS (sizeof(hint_t) * CHAR_BIT)
+#define HINT_CLASS_BITS 4
+#define HINT_VALUE_BITS (HINT_BITS - HINT_CLASS_BITS)
/** Number of bytes that fit in a hint value. */
-#define HINT_VALUE_BYTES (HINT_VALUE_BITS / CHAR_BIT)
+#define HINT_VALUE_BYTES (HINT_VALUE_BITS / CHAR_BIT)
/** Max unsigned integer that can be stored in a hint value. */
-#define HINT_VALUE_MAX ((1ULL << HINT_VALUE_BITS) - 1)
+#define HINT_VALUE_MAX ((1ULL << HINT_VALUE_BITS) - 1)
/**
* Max and min signed integer numbers that fit in a hint value.
* For numbers > MAX and < MIN we store MAX and MIN, respectively.
*/
-#define HINT_VALUE_INT_MAX ((1LL << (HINT_VALUE_BITS - 1)) - 1)
-#define HINT_VALUE_INT_MIN (-(1LL << (HINT_VALUE_BITS - 1)))
+#define HINT_VALUE_INT_MAX ((1LL << (HINT_VALUE_BITS - 1)) - 1)
+#define HINT_VALUE_INT_MIN (-(1LL << (HINT_VALUE_BITS - 1)))
/**
* Max and min floating point numbers whose integral parts fit
* in a hint value. Note, we can't compare a floating point number
* with HINT_VALUE_INT_{MIN,MAX} because of rounding errors.
*/
-#define HINT_VALUE_DOUBLE_MAX (exp2(HINT_VALUE_BITS - 1) - 1)
-#define HINT_VALUE_DOUBLE_MIN (-exp2(HINT_VALUE_BITS - 1))
+#define HINT_VALUE_DOUBLE_MAX (exp2(HINT_VALUE_BITS - 1) - 1)
+#define HINT_VALUE_DOUBLE_MIN (-exp2(HINT_VALUE_BITS - 1))
/*
* HINT_CLASS_BITS should be big enough to store any mp_class value.
@@ -1552,7 +1528,8 @@ static inline hint_t
hint_uint(uint64_t u)
{
uint64_t val = (u >= (uint64_t)HINT_VALUE_INT_MAX ?
- HINT_VALUE_MAX : u - HINT_VALUE_INT_MIN);
+ HINT_VALUE_MAX :
+ u - HINT_VALUE_INT_MIN);
return hint_create(MP_CLASS_NUMBER, val);
}
@@ -1586,8 +1563,8 @@ hint_decimal(decimal_t *dec)
{
uint64_t val = 0;
int64_t num;
- if (decimal_to_int64(dec, &num) &&
- num >= HINT_VALUE_INT_MIN && num <= HINT_VALUE_INT_MAX) {
+ if (decimal_to_int64(dec, &num) && num >= HINT_VALUE_INT_MIN &&
+ num <= HINT_VALUE_INT_MAX) {
val = num - HINT_VALUE_INT_MIN;
} else if (!(dec->bits & DECNEG)) {
val = HINT_VALUE_MAX;
@@ -1697,13 +1674,11 @@ field_hint_number(const char *field)
return hint_double(mp_decode_float(&field));
case MP_DOUBLE:
return hint_double(mp_decode_double(&field));
- case MP_EXT:
- {
+ case MP_EXT: {
int8_t ext_type;
uint32_t len = mp_decode_extl(&field, &ext_type);
switch (ext_type) {
- case MP_DECIMAL:
- {
+ case MP_DECIMAL: {
decimal_t dec;
return hint_decimal(decimal_unpack(&field, len, &dec));
}
@@ -1724,8 +1699,7 @@ field_hint_decimal(const char *field)
int8_t ext_type;
uint32_t len = mp_decode_extl(&field, &ext_type);
switch (ext_type) {
- case MP_DECIMAL:
- {
+ case MP_DECIMAL: {
decimal_t dec;
return hint_decimal(decimal_unpack(&field, len, &dec));
}
@@ -1751,7 +1725,7 @@ field_hint_string(const char *field, struct coll *coll)
assert(mp_typeof(*field) == MP_STR);
uint32_t len = mp_decode_strl(&field);
return coll == NULL ? hint_str(field, len) :
- hint_str_coll(field, len, coll);
+ hint_str_coll(field, len, coll);
}
static inline hint_t
@@ -1766,7 +1740,7 @@ static inline hint_t
field_hint_scalar(const char *field, struct coll *coll)
{
uint32_t len;
- switch(mp_typeof(*field)) {
+ switch (mp_typeof(*field)) {
case MP_BOOL:
return hint_bool(mp_decode_bool(&field));
case MP_UINT:
@@ -1780,17 +1754,15 @@ field_hint_scalar(const char *field, struct coll *coll)
case MP_STR:
len = mp_decode_strl(&field);
return coll == NULL ? hint_str(field, len) :
- hint_str_coll(field, len, coll);
+ hint_str_coll(field, len, coll);
case MP_BIN:
len = mp_decode_binl(&field);
return hint_bin(field, len);
- case MP_EXT:
- {
+ case MP_EXT: {
int8_t ext_type;
uint32_t len = mp_decode_extl(&field, &ext_type);
switch (ext_type) {
- case MP_DECIMAL:
- {
+ case MP_DECIMAL: {
decimal_t dec;
return hint_decimal(decimal_unpack(&field, len, &dec));
}
@@ -1852,8 +1824,8 @@ static hint_t
tuple_hint(struct tuple *tuple, struct key_def *key_def)
{
assert(!key_def->is_multikey);
- const char *field = tuple_field_by_part(tuple, key_def->parts,
- MULTIKEY_NONE);
+ const char *field =
+ tuple_field_by_part(tuple, key_def->parts, MULTIKEY_NONE);
if (is_nullable && field == NULL)
return hint_nil();
return field_hint<type, is_nullable>(field, key_def->parts->coll);
@@ -1862,9 +1834,9 @@ tuple_hint(struct tuple *tuple, struct key_def *key_def)
static hint_t
key_hint_stub(const char *key, uint32_t part_count, struct key_def *key_def)
{
- (void) key;
- (void) part_count;
- (void) key_def;
+ (void)key;
+ (void)part_count;
+ (void)key_def;
/*
* Multikey hint for tuple is an index of the key in
* array, it always must be defined. While
@@ -1882,13 +1854,13 @@ key_hint_stub(const char *key, uint32_t part_count, struct key_def *key_def)
static hint_t
key_hint_stub(struct tuple *tuple, struct key_def *key_def)
{
- (void) tuple;
- (void) key_def;
+ (void)tuple;
+ (void)key_def;
unreachable();
return HINT_NONE;
}
-template<enum field_type type, bool is_nullable>
+template <enum field_type type, bool is_nullable>
static void
key_def_set_hint_func(struct key_def *def)
{
@@ -1896,7 +1868,7 @@ key_def_set_hint_func(struct key_def *def)
def->tuple_hint = tuple_hint<type, is_nullable>;
}
-template<enum field_type type>
+template <enum field_type type>
static void
key_def_set_hint_func(struct key_def *def)
{
@@ -1996,57 +1968,63 @@ key_def_set_compare_func_fast(struct key_def *def)
}
if (cmp == NULL) {
cmp = is_sequential ?
- tuple_compare_sequential<false, false> :
- tuple_compare_slowpath<false, false, false, false>;
+ tuple_compare_sequential<false, false> :
+ tuple_compare_slowpath<false, false, false, false>;
}
if (cmp_wk == NULL) {
- cmp_wk = is_sequential ?
- tuple_compare_with_key_sequential<false, false> :
- tuple_compare_with_key_slowpath<false, false,
- false, false>;
+ cmp_wk =
+ is_sequential ?
+ tuple_compare_with_key_sequential<false, false> :
+ tuple_compare_with_key_slowpath<false, false,
+ false, false>;
}
def->tuple_compare = cmp;
def->tuple_compare_with_key = cmp_wk;
}
-template<bool is_nullable, bool has_optional_parts>
+template <bool is_nullable, bool has_optional_parts>
static void
key_def_set_compare_func_plain(struct key_def *def)
{
assert(!def->has_json_paths);
if (key_def_is_sequential(def)) {
- def->tuple_compare = tuple_compare_sequential
- <is_nullable, has_optional_parts>;
- def->tuple_compare_with_key = tuple_compare_with_key_sequential
- <is_nullable, has_optional_parts>;
+ def->tuple_compare =
+ tuple_compare_sequential<is_nullable,
+ has_optional_parts>;
+ def->tuple_compare_with_key =
+ tuple_compare_with_key_sequential<is_nullable,
+ has_optional_parts>;
} else {
- def->tuple_compare = tuple_compare_slowpath
- <is_nullable, has_optional_parts, false, false>;
- def->tuple_compare_with_key = tuple_compare_with_key_slowpath
- <is_nullable, has_optional_parts, false, false>;
+ def->tuple_compare =
+ tuple_compare_slowpath<is_nullable, has_optional_parts,
+ false, false>;
+ def->tuple_compare_with_key = tuple_compare_with_key_slowpath<
+ is_nullable, has_optional_parts, false, false>;
}
}
-template<bool is_nullable, bool has_optional_parts>
+template <bool is_nullable, bool has_optional_parts>
static void
key_def_set_compare_func_json(struct key_def *def)
{
assert(def->has_json_paths);
if (def->is_multikey) {
- def->tuple_compare = tuple_compare_slowpath
- <is_nullable, has_optional_parts, true, true>;
- def->tuple_compare_with_key = tuple_compare_with_key_slowpath
- <is_nullable, has_optional_parts, true, true>;
+ def->tuple_compare =
+ tuple_compare_slowpath<is_nullable, has_optional_parts,
+ true, true>;
+ def->tuple_compare_with_key = tuple_compare_with_key_slowpath<
+ is_nullable, has_optional_parts, true, true>;
} else {
- def->tuple_compare = tuple_compare_slowpath
- <is_nullable, has_optional_parts, true, false>;
- def->tuple_compare_with_key = tuple_compare_with_key_slowpath
- <is_nullable, has_optional_parts, true, false>;
+ def->tuple_compare =
+ tuple_compare_slowpath<is_nullable, has_optional_parts,
+ true, false>;
+ def->tuple_compare_with_key = tuple_compare_with_key_slowpath<
+ is_nullable, has_optional_parts, true, false>;
}
}
-template<bool is_nullable>
+template <bool is_nullable>
static void
key_def_set_compare_func_for_func_index(struct key_def *def)
{
@@ -2063,8 +2041,8 @@ key_def_set_compare_func(struct key_def *def)
key_def_set_compare_func_for_func_index<true>(def);
else
key_def_set_compare_func_for_func_index<false>(def);
- } else if (!key_def_has_collation(def) &&
- !def->is_nullable && !def->has_json_paths) {
+ } else if (!key_def_has_collation(def) && !def->is_nullable &&
+ !def->has_json_paths) {
key_def_set_compare_func_fast(def);
} else if (!def->has_json_paths) {
if (def->is_nullable && def->has_optional_parts) {
diff --git a/src/box/tuple_convert.c b/src/box/tuple_convert.c
index 5cc268a..256a3d6 100644
--- a/src/box/tuple_convert.c
+++ b/src/box/tuple_convert.c
@@ -52,10 +52,10 @@ tuple_to_obuf(struct tuple *tuple, struct obuf *buf)
int
append_output(void *arg, unsigned char *buf, size_t len)
{
- (void) arg;
+ (void)arg;
char *buf_out = region_alloc(&fiber()->gc, len + 1);
if (!buf_out) {
- diag_set(OutOfMemory, len , "region", "tuple_to_yaml");
+ diag_set(OutOfMemory, len, "region", "tuple_to_yaml");
return 0;
}
memcpy(buf_out, buf, len);
@@ -71,8 +71,9 @@ encode_table(yaml_emitter_t *emitter, const char **data)
{
yaml_event_t ev;
yaml_mapping_style_t yaml_style = YAML_FLOW_MAPPING_STYLE;
- if (!yaml_mapping_start_event_initialize(&ev, NULL, NULL, 0, yaml_style)
- || !yaml_emitter_emit(emitter, &ev)) {
+ if (!yaml_mapping_start_event_initialize(&ev, NULL, NULL, 0,
+ yaml_style) ||
+ !yaml_emitter_emit(emitter, &ev)) {
diag_set(SystemError, "failed to init event libyaml");
return 0;
}
@@ -94,15 +95,14 @@ encode_table(yaml_emitter_t *emitter, const char **data)
return 1;
}
-
static int
encode_array(yaml_emitter_t *emitter, const char **data)
{
yaml_event_t ev;
yaml_sequence_style_t yaml_style = YAML_FLOW_SEQUENCE_STYLE;
if (!yaml_sequence_start_event_initialize(&ev, NULL, NULL, 0,
- yaml_style) ||
- !yaml_emitter_emit(emitter, &ev)) {
+ yaml_style) ||
+ !yaml_emitter_emit(emitter, &ev)) {
diag_set(SystemError, "failed to init event libyaml");
return 0;
}
@@ -110,7 +110,7 @@ encode_array(yaml_emitter_t *emitter, const char **data)
uint32_t size = mp_decode_array(data);
for (uint32_t i = 0; i < size; i++) {
if (!encode_node(emitter, data))
- return 0;
+ return 0;
}
if (!yaml_sequence_end_event_initialize(&ev) ||
@@ -136,16 +136,16 @@ encode_node(yaml_emitter_t *emitter, const char **data)
yaml_scalar_style_t style = YAML_PLAIN_SCALAR_STYLE;
char buf[FPCONV_G_FMT_BUFSIZE];
int type = mp_typeof(**data);
- switch(type) {
+ switch (type) {
case MP_UINT:
len = snprintf(buf, sizeof(buf), "%llu",
- (unsigned long long) mp_decode_uint(data));
+ (unsigned long long)mp_decode_uint(data));
buf[len] = 0;
str = buf;
break;
case MP_INT:
len = snprintf(buf, sizeof(buf), "%lld",
- (long long) mp_decode_int(data));
+ (long long)mp_decode_int(data));
buf[len] = 0;
str = buf;
break;
@@ -177,7 +177,7 @@ encode_node(yaml_emitter_t *emitter, const char **data)
style = YAML_ANY_SCALAR_STYLE;
/* Binary or not UTF8 */
binlen = base64_bufsize(len, 0);
- bin = (char *) malloc(binlen);
+ bin = (char *)malloc(binlen);
if (bin == NULL) {
diag_set(OutOfMemory, binlen, "malloc",
"tuple_to_yaml");
@@ -186,7 +186,7 @@ encode_node(yaml_emitter_t *emitter, const char **data)
binlen = base64_encode(str, len, bin, binlen, 0);
str = bin;
len = binlen;
- tag = (yaml_char_t *) LUAYAML_TAG_PREFIX "binary";
+ tag = (yaml_char_t *)LUAYAML_TAG_PREFIX "binary";
break;
case MP_BOOL:
if (mp_decode_bool(data)) {
@@ -266,7 +266,7 @@ tuple_to_yaml(struct tuple *tuple)
yaml_emitter_delete(&emitter);
size_t total_len = region_used(&fiber()->gc) - used;
- char *buf = (char *) region_join(&fiber()->gc, total_len);
+ char *buf = (char *)region_join(&fiber()->gc, total_len);
if (buf == NULL) {
diag_set(OutOfMemory, total_len, "region", "tuple_to_yaml");
return NULL;
diff --git a/src/box/tuple_dictionary.c b/src/box/tuple_dictionary.c
index a8ea13a..4998bac 100644
--- a/src/box/tuple_dictionary.c
+++ b/src/box/tuple_dictionary.c
@@ -52,8 +52,8 @@ struct mh_strnu32_node_t {
#define mh_arg_t void *
#define mh_hash(a, arg) ((a)->hash)
#define mh_hash_key(a, arg) mh_hash(a, arg)
-#define mh_cmp(a, b, arg) ((a)->len != (b)->len || \
- memcmp((a)->str, (b)->str, (a)->len))
+#define mh_cmp(a, b, arg) \
+ ((a)->len != (b)->len || memcmp((a)->str, (b)->str, (a)->len))
#define mh_cmp_key(a, b, arg) mh_cmp(a, b, arg)
#define MH_SOURCE 1
#include "salad/mhash.h" /* Create mh_strnu32_t hash. */
@@ -100,22 +100,18 @@ tuple_dictionary_set_name(struct tuple_dictionary *dict, const char *name,
{
assert(fieldno < dict->name_count);
uint32_t name_hash = field_name_hash(name, name_len);
- struct mh_strnu32_key_t key = {
- name, name_len, name_hash
- };
+ struct mh_strnu32_key_t key = { name, name_len, name_hash };
mh_int_t rc = mh_strnu32_find(dict->hash, &key, NULL);
if (rc != mh_end(dict->hash)) {
- diag_set(ClientError, ER_SPACE_FIELD_IS_DUPLICATE,
- name);
+ diag_set(ClientError, ER_SPACE_FIELD_IS_DUPLICATE, name);
return -1;
}
- struct mh_strnu32_node_t name_node = {
- name, name_len, name_hash, fieldno
- };
+ struct mh_strnu32_node_t name_node = { name, name_len, name_hash,
+ fieldno };
rc = mh_strnu32_put(dict->hash, &name_node, NULL, NULL);
/* Memory was reserved in new(). */
assert(rc != mh_end(dict->hash));
- (void) rc;
+ (void)rc;
return 0;
}
@@ -125,8 +121,7 @@ tuple_dictionary_new(const struct field_def *fields, uint32_t field_count)
struct tuple_dictionary *dict =
(struct tuple_dictionary *)calloc(1, sizeof(*dict));
if (dict == NULL) {
- diag_set(OutOfMemory, sizeof(*dict), "malloc",
- "dict");
+ diag_set(OutOfMemory, sizeof(*dict), "malloc", "dict");
return NULL;
}
dict->refs = 1;
@@ -137,24 +132,24 @@ tuple_dictionary_new(const struct field_def *fields, uint32_t field_count)
uint32_t total = names_offset;
for (uint32_t i = 0; i < field_count; ++i)
total += strlen(fields[i].name) + 1;
- dict->names = (char **) malloc(total);
+ dict->names = (char **)malloc(total);
if (dict->names == NULL) {
diag_set(OutOfMemory, total, "malloc", "dict->names");
goto err_memory;
}
dict->hash = mh_strnu32_new();
if (dict->hash == NULL) {
- diag_set(OutOfMemory, sizeof(*dict->hash),
- "mh_strnu32_new", "dict->hash");
+ diag_set(OutOfMemory, sizeof(*dict->hash), "mh_strnu32_new",
+ "dict->hash");
goto err_hash;
}
if (mh_strnu32_reserve(dict->hash, field_count, NULL) != 0) {
- diag_set(OutOfMemory, field_count *
- sizeof(struct mh_strnu32_node_t), "mh_strnu32_reserve",
- "dict->hash");
+ diag_set(OutOfMemory,
+ field_count * sizeof(struct mh_strnu32_node_t),
+ "mh_strnu32_reserve", "dict->hash");
goto err_name;
}
- char *pos = (char *) dict->names + names_offset;
+ char *pos = (char *)dict->names + names_offset;
for (uint32_t i = 0; i < field_count; ++i) {
int len = strlen(fields[i].name);
memcpy(pos, fields[i].name, len);
@@ -208,7 +203,7 @@ tuple_fieldno_by_name(struct tuple_dictionary *dict, const char *name,
struct mh_strnu32_t *hash = dict->hash;
if (hash == NULL)
return -1;
- struct mh_strnu32_key_t key = {name, name_len, name_hash};
+ struct mh_strnu32_key_t key = { name, name_len, name_hash };
mh_int_t rc = mh_strnu32_find(hash, &key, NULL);
if (rc == mh_end(hash))
return -1;
--git a/src/box/tuple_extract_key.cc b/src/box/tuple_extract_key.cc
index c1ad392..7bb48a8 100644
--- a/src/box/tuple_extract_key.cc
+++ b/src/box/tuple_extract_key.cc
@@ -64,10 +64,10 @@ tuple_extract_key_sequential_raw(const char *data, const char *data_end,
assert(field_end - field_start <= data_end - data);
bsize += field_end - field_start;
- char *key = (char *) region_alloc(&fiber()->gc, bsize);
+ char *key = (char *)region_alloc(&fiber()->gc, bsize);
if (key == NULL) {
diag_set(OutOfMemory, bsize, "region",
- "tuple_extract_key_raw_sequential");
+ "tuple_extract_key_raw_sequential");
return NULL;
}
char *key_buf = mp_encode_array(key, key_def->part_count);
@@ -96,11 +96,8 @@ tuple_extract_key_sequential(struct tuple *tuple, struct key_def *key_def,
assert(has_optional_parts == key_def->has_optional_parts);
const char *data = tuple_data(tuple);
const char *data_end = data + tuple->bsize;
- return tuple_extract_key_sequential_raw<has_optional_parts>(data,
- data_end,
- key_def,
- multikey_idx,
- key_size);
+ return tuple_extract_key_sequential_raw<has_optional_parts>(
+ data, data_end, key_def, multikey_idx, key_size);
}
/**
@@ -156,8 +153,8 @@ tuple_extract_key_slowpath(struct tuple *tuple, struct key_def *key_def,
* minimize tuple_field_raw() calls.
*/
for (; i < part_count - 1; i++) {
- if (!key_def_parts_are_sequential
- <has_json_paths>(key_def, i)) {
+ if (!key_def_parts_are_sequential<
+ has_json_paths>(key_def, i)) {
/*
* End of sequential part.
*/
@@ -176,7 +173,7 @@ tuple_extract_key_slowpath(struct tuple *tuple, struct key_def *key_def,
bsize += end - field;
}
- char *key = (char *) region_alloc(&fiber()->gc, bsize);
+ char *key = (char *)region_alloc(&fiber()->gc, bsize);
if (key == NULL) {
diag_set(OutOfMemory, bsize, "region", "tuple_extract_key");
return NULL;
@@ -208,8 +205,8 @@ tuple_extract_key_slowpath(struct tuple *tuple, struct key_def *key_def,
* minimize tuple_field_raw() calls.
*/
for (; i < part_count - 1; i++) {
- if (!key_def_parts_are_sequential
- <has_json_paths>(key_def, i)) {
+ if (!key_def_parts_are_sequential<
+ has_json_paths>(key_def, i)) {
/*
* End of sequential part.
*/
@@ -255,7 +252,7 @@ tuple_extract_key_slowpath_raw(const char *data, const char *data_end,
assert(!key_def->for_func_index);
assert(mp_sizeof_nil() == 1);
/* allocate buffer with maximal possible size */
- char *key = (char *) region_alloc(&fiber()->gc, data_end - data);
+ char *key = (char *)region_alloc(&fiber()->gc, data_end - data);
if (key == NULL) {
diag_set(OutOfMemory, data_end - data, "region",
"tuple_extract_key_raw");
@@ -268,7 +265,7 @@ tuple_extract_key_slowpath_raw(const char *data, const char *data_end,
* A tuple can not be empty - at least a pk always exists.
*/
assert(field_count > 0);
- (void) field_count;
+ (void)field_count;
const char *field0_end = field0;
mp_next(&field0_end);
const char *field = field0;
@@ -278,8 +275,8 @@ tuple_extract_key_slowpath_raw(const char *data, const char *data_end,
uint32_t fieldno = key_def->parts[i].fieldno;
uint32_t null_count = 0;
for (; i < key_def->part_count - 1; i++) {
- if (!key_def_parts_are_sequential
- <has_json_paths>(key_def, i))
+ if (!key_def_parts_are_sequential<has_json_paths>(
+ key_def, i))
break;
}
const struct key_part *part = &key_def->parts[i];
@@ -363,7 +360,7 @@ tuple_extract_key_slowpath_raw(const char *data, const char *data_end,
/**
* Initialize tuple_extract_key() and tuple_extract_key_raw()
*/
-template<bool contains_sequential_parts, bool has_optional_parts>
+template <bool contains_sequential_parts, bool has_optional_parts>
static void
key_def_set_extract_func_plain(struct key_def *def)
{
@@ -372,43 +369,50 @@ key_def_set_extract_func_plain(struct key_def *def)
assert(!def->for_func_index);
if (key_def_is_sequential(def)) {
assert(contains_sequential_parts || def->part_count == 1);
- def->tuple_extract_key = tuple_extract_key_sequential
- <has_optional_parts>;
- def->tuple_extract_key_raw = tuple_extract_key_sequential_raw
- <has_optional_parts>;
+ def->tuple_extract_key =
+ tuple_extract_key_sequential<has_optional_parts>;
+ def->tuple_extract_key_raw =
+ tuple_extract_key_sequential_raw<has_optional_parts>;
} else {
- def->tuple_extract_key = tuple_extract_key_slowpath
- <contains_sequential_parts,
- has_optional_parts, false, false>;
- def->tuple_extract_key_raw = tuple_extract_key_slowpath_raw
- <has_optional_parts, false>;
+ def->tuple_extract_key =
+ tuple_extract_key_slowpath<contains_sequential_parts,
+ has_optional_parts, false,
+ false>;
+ def->tuple_extract_key_raw =
+ tuple_extract_key_slowpath_raw<has_optional_parts,
+ false>;
}
}
-template<bool contains_sequential_parts, bool has_optional_parts>
+template <bool contains_sequential_parts, bool has_optional_parts>
static void
key_def_set_extract_func_json(struct key_def *def)
{
assert(def->has_json_paths);
assert(!def->for_func_index);
if (def->is_multikey) {
- def->tuple_extract_key = tuple_extract_key_slowpath
- <contains_sequential_parts,
- has_optional_parts, true, true>;
+ def->tuple_extract_key =
+ tuple_extract_key_slowpath<contains_sequential_parts,
+ has_optional_parts, true,
+ true>;
} else {
- def->tuple_extract_key = tuple_extract_key_slowpath
- <contains_sequential_parts,
- has_optional_parts, true, false>;
+ def->tuple_extract_key =
+ tuple_extract_key_slowpath<contains_sequential_parts,
+ has_optional_parts, true,
+ false>;
}
- def->tuple_extract_key_raw = tuple_extract_key_slowpath_raw
- <has_optional_parts, true>;
+ def->tuple_extract_key_raw =
+ tuple_extract_key_slowpath_raw<has_optional_parts, true>;
}
static char *
tuple_extract_key_stub(struct tuple *tuple, struct key_def *key_def,
- int multikey_idx, uint32_t *key_size)
+ int multikey_idx, uint32_t *key_size)
{
- (void)tuple; (void)key_def; (void)multikey_idx; (void)key_size;
+ (void)tuple;
+ (void)key_def;
+ (void)multikey_idx;
+ (void)key_size;
unreachable();
return NULL;
}
@@ -418,8 +422,11 @@ tuple_extract_key_raw_stub(const char *data, const char *data_end,
struct key_def *key_def, int multikey_idx,
uint32_t *key_size)
{
- (void)data; (void)data_end;
- (void)key_def; (void)multikey_idx; (void)key_size;
+ (void)data;
+ (void)data_end;
+ (void)key_def;
+ (void)multikey_idx;
+ (void)key_size;
unreachable();
return NULL;
}
@@ -467,9 +474,8 @@ tuple_key_contains_null(struct tuple *tuple, struct key_def *def,
const uint32_t *field_map = tuple_field_map(tuple);
for (struct key_part *part = def->parts, *end = part + def->part_count;
part < end; ++part) {
- const char *field = tuple_field_raw_by_part(format, data,
- field_map, part,
- multikey_idx);
+ const char *field = tuple_field_raw_by_part(
+ format, data, field_map, part, multikey_idx);
if (field == NULL || mp_typeof(*field) == MP_NIL)
return true;
}
@@ -482,8 +488,8 @@ tuple_validate_key_parts(struct key_def *key_def, struct tuple *tuple)
assert(!key_def->is_multikey);
for (uint32_t idx = 0; idx < key_def->part_count; idx++) {
struct key_part *part = &key_def->parts[idx];
- const char *field = tuple_field_by_part(tuple, part,
- MULTIKEY_NONE);
+ const char *field =
+ tuple_field_by_part(tuple, part, MULTIKEY_NONE);
if (field == NULL) {
if (key_part_is_nullable(part))
continue;
diff --git a/src/box/tuple_format.c b/src/box/tuple_format.c
index 9b817d3..7545315 100644
--- a/src/box/tuple_format.c
+++ b/src/box/tuple_format.c
@@ -55,18 +55,16 @@ tuple_format1_field_by_format2_field(struct tuple_format *format1,
{
struct region *region = &fiber()->gc;
size_t region_svp = region_used(region);
- uint32_t path_len = json_tree_snprint_path(NULL, 0,
- &format2_field->token, TUPLE_INDEX_BASE);
+ uint32_t path_len = json_tree_snprint_path(
+ NULL, 0, &format2_field->token, TUPLE_INDEX_BASE);
char *path = region_alloc(region, path_len + 1);
if (path == NULL)
panic("Can not allocate memory for path");
json_tree_snprint_path(path, path_len + 1, &format2_field->token,
TUPLE_INDEX_BASE);
- struct tuple_field *format1_field =
- json_tree_lookup_path_entry(&format1->fields,
- &format1->fields.root, path,
- path_len, TUPLE_INDEX_BASE,
- struct tuple_field, token);
+ struct tuple_field *format1_field = json_tree_lookup_path_entry(
+ &format1->fields, &format1->fields.root, path, path_len,
+ TUPLE_INDEX_BASE, struct tuple_field, token);
region_truncate(region, region_svp);
return format1_field;
}
@@ -84,7 +82,8 @@ tuple_format_cmp(const struct tuple_format *format1,
struct tuple_field *field_a;
json_tree_foreach_entry_preorder(field_a, &a->fields.root,
- struct tuple_field, token) {
+ struct tuple_field, token)
+ {
struct tuple_field *field_b =
tuple_format1_field_by_format2_field(b, field_a);
if (field_a->type != field_b->type)
@@ -93,10 +92,10 @@ tuple_format_cmp(const struct tuple_format *format1,
return (int)field_a->coll_id - (int)field_b->coll_id;
if (field_a->nullable_action != field_b->nullable_action)
return (int)field_a->nullable_action -
- (int)field_b->nullable_action;
+ (int)field_b->nullable_action;
if (field_a->is_key_part != field_b->is_key_part)
return (int)field_a->is_key_part -
- (int)field_b->is_key_part;
+ (int)field_b->is_key_part;
}
return 0;
@@ -105,9 +104,8 @@ tuple_format_cmp(const struct tuple_format *format1,
static uint32_t
tuple_format_hash(struct tuple_format *format)
{
-#define TUPLE_FIELD_MEMBER_HASH(field, member, h, carry, size) \
- PMurHash32_Process(&h, &carry, &field->member, \
- sizeof(field->member)); \
+#define TUPLE_FIELD_MEMBER_HASH(field, member, h, carry, size) \
+ PMurHash32_Process(&h, &carry, &field->member, sizeof(field->member)); \
size += sizeof(field->member);
uint32_t h = 13;
@@ -115,7 +113,8 @@ tuple_format_hash(struct tuple_format *format)
uint32_t size = 0;
struct tuple_field *f;
json_tree_foreach_entry_preorder(f, &format->fields.root,
- struct tuple_field, token) {
+ struct tuple_field, token)
+ {
TUPLE_FIELD_MEMBER_HASH(f, type, h, carry, size)
TUPLE_FIELD_MEMBER_HASH(f, coll_id, h, carry, size)
TUPLE_FIELD_MEMBER_HASH(f, nullable_action, h, carry, size)
@@ -190,7 +189,8 @@ tuple_format_field_by_id(struct tuple_format *format, uint32_t id)
{
struct tuple_field *field;
json_tree_foreach_entry_preorder(field, &format->fields.root,
- struct tuple_field, token) {
+ struct tuple_field, token)
+ {
if (field->id == id)
return field;
}
@@ -205,9 +205,9 @@ static int
tuple_field_ensure_child_compatibility(struct tuple_field *parent,
struct tuple_field *child)
{
- enum field_type expected_type =
- child->token.type == JSON_TOKEN_STR ?
- FIELD_TYPE_MAP : FIELD_TYPE_ARRAY;
+ enum field_type expected_type = child->token.type == JSON_TOKEN_STR ?
+ FIELD_TYPE_MAP :
+ FIELD_TYPE_ARRAY;
if (field_type1_contains_type2(parent->type, expected_type)) {
parent->type = expected_type;
} else {
@@ -285,10 +285,9 @@ tuple_format_add_field(struct tuple_format *format, uint32_t fieldno,
field->token.type != JSON_TOKEN_END) {
if (tuple_field_ensure_child_compatibility(parent, field) != 0)
goto fail;
- struct tuple_field *next =
- json_tree_lookup_entry(tree, &parent->token,
- &field->token,
- struct tuple_field, token);
+ struct tuple_field *next = json_tree_lookup_entry(
+ tree, &parent->token, &field->token, struct tuple_field,
+ token);
if (next == NULL) {
field->id = format->total_field_count++;
rc = json_tree_add(tree, &parent->token, &field->token);
@@ -355,10 +354,9 @@ tuple_format_use_key_part(struct tuple_format *format, uint32_t field_count,
int *current_slot, char **path_pool)
{
assert(part->fieldno < tuple_format_field_count(format));
- struct tuple_field *field =
- tuple_format_add_field(format, part->fieldno, part->path,
- part->path_len, is_sequential,
- current_slot, path_pool);
+ struct tuple_field *field = tuple_format_add_field(
+ format, part->fieldno, part->path, part->path_len,
+ is_sequential, current_slot, path_pool);
if (field == NULL)
return -1;
/*
@@ -398,11 +396,9 @@ tuple_format_use_key_part(struct tuple_format *format, uint32_t field_count,
* with field's one, then the part type is more strict
* and the part type must be used in tuple_format.
*/
- if (field_type1_contains_type2(field->type,
- part->type)) {
+ if (field_type1_contains_type2(field->type, part->type)) {
field->type = part->type;
- } else if (!field_type1_contains_type2(part->type,
- field->type)) {
+ } else if (!field_type1_contains_type2(part->type, field->type)) {
int errcode;
if (!field->is_key_part)
errcode = ER_FORMAT_MISMATCH_INDEX_PART;
@@ -422,13 +418,12 @@ tuple_format_use_key_part(struct tuple_format *format, uint32_t field_count,
* definitions.
*/
static int
-tuple_format_create(struct tuple_format *format, struct key_def * const *keys,
+tuple_format_create(struct tuple_format *format, struct key_def *const *keys,
uint16_t key_count, const struct field_def *fields,
uint32_t field_count)
{
- format->min_field_count =
- tuple_format_min_field_count(keys, key_count, fields,
- field_count);
+ format->min_field_count = tuple_format_min_field_count(
+ keys, key_count, fields, field_count);
if (tuple_format_field_count(format) == 0) {
format->field_map_size = 0;
return 0;
@@ -443,8 +438,9 @@ tuple_format_create(struct tuple_format *format, struct key_def * const *keys,
if (cid != COLL_NONE) {
struct coll_id *coll_id = coll_by_id(cid);
if (coll_id == NULL) {
- diag_set(ClientError,ER_WRONG_COLLATION_OPTIONS,
- i + 1, "collation was not found by ID");
+ diag_set(ClientError,
+ ER_WRONG_COLLATION_OPTIONS, i + 1,
+ "collation was not found by ID");
return -1;
}
coll = coll_id->coll;
@@ -470,16 +466,16 @@ tuple_format_create(struct tuple_format *format, struct key_def * const *keys,
const struct key_part *parts_end = part + key_def->part_count;
for (; part < parts_end; part++) {
- if (tuple_format_use_key_part(format, field_count, part,
- is_sequential,
- ¤t_slot,
- &path_pool) != 0)
+ if (tuple_format_use_key_part(
+ format, field_count, part, is_sequential,
+ ¤t_slot, &path_pool) != 0)
return -1;
}
}
- assert(tuple_format_field(format, 0)->offset_slot == TUPLE_OFFSET_SLOT_NIL
- || json_token_is_multikey(&tuple_format_field(format, 0)->token));
+ assert(tuple_format_field(format, 0)->offset_slot ==
+ TUPLE_OFFSET_SLOT_NIL ||
+ json_token_is_multikey(&tuple_format_field(format, 0)->token));
size_t field_map_size = -current_slot * sizeof(uint32_t);
if (field_map_size > INT16_MAX) {
/** tuple->data_offset is 15 bits */
@@ -492,14 +488,15 @@ tuple_format_create(struct tuple_format *format, struct key_def * const *keys,
size_t required_fields_sz = bitmap_size(format->total_field_count);
format->required_fields = calloc(1, required_fields_sz);
if (format->required_fields == NULL) {
- diag_set(OutOfMemory, required_fields_sz,
- "malloc", "required field bitmap");
+ diag_set(OutOfMemory, required_fields_sz, "malloc",
+ "required field bitmap");
return -1;
}
struct tuple_field *field;
uint32_t *required_fields = format->required_fields;
json_tree_foreach_entry_preorder(field, &format->fields.root,
- struct tuple_field, token) {
+ struct tuple_field, token)
+ {
/*
* In the case of the multikey index,
* required_fields is overridden with local for
@@ -521,7 +518,7 @@ tuple_format_create(struct tuple_format *format, struct key_def * const *keys,
calloc(1, required_fields_sz);
if (multikey_required_fields == NULL) {
diag_set(OutOfMemory, required_fields_sz,
- "malloc", "required field bitmap");
+ "malloc", "required field bitmap");
return -1;
}
field->multikey_required_fields =
@@ -545,17 +542,17 @@ static int
tuple_format_register(struct tuple_format *format)
{
if (recycled_format_ids != FORMAT_ID_NIL) {
-
- format->id = (uint16_t) recycled_format_ids;
- recycled_format_ids = (intptr_t) tuple_formats[recycled_format_ids];
+ format->id = (uint16_t)recycled_format_ids;
+ recycled_format_ids =
+ (intptr_t)tuple_formats[recycled_format_ids];
} else {
if (formats_size == formats_capacity) {
- uint32_t new_capacity = formats_capacity ?
- formats_capacity * 2 : 16;
+ uint32_t new_capacity =
+ formats_capacity ? formats_capacity * 2 : 16;
struct tuple_format **formats;
- formats = (struct tuple_format **)
- realloc(tuple_formats, new_capacity *
- sizeof(tuple_formats[0]));
+ formats = (struct tuple_format **)realloc(
+ tuple_formats,
+ new_capacity * sizeof(tuple_formats[0]));
if (formats == NULL) {
diag_set(OutOfMemory,
sizeof(struct tuple_format), "malloc",
@@ -567,13 +564,13 @@ tuple_format_register(struct tuple_format *format)
tuple_formats = formats;
}
uint32_t formats_size_max = FORMAT_ID_MAX + 1;
- struct errinj *inj = errinj(ERRINJ_TUPLE_FORMAT_COUNT,
- ERRINJ_INT);
+ struct errinj *inj =
+ errinj(ERRINJ_TUPLE_FORMAT_COUNT, ERRINJ_INT);
if (inj != NULL && inj->iparam > 0)
formats_size_max = inj->iparam;
if (formats_size >= formats_size_max) {
diag_set(ClientError, ER_TUPLE_FORMAT_LIMIT,
- (unsigned) formats_capacity);
+ (unsigned)formats_capacity);
return -1;
}
format->id = formats_size++;
@@ -587,7 +584,7 @@ tuple_format_deregister(struct tuple_format *format)
{
if (format->id == FORMAT_ID_NIL)
return;
- tuple_formats[format->id] = (struct tuple_format *) recycled_format_ids;
+ tuple_formats[format->id] = (struct tuple_format *)recycled_format_ids;
recycled_format_ids = format->id;
format->id = FORMAT_ID_NIL;
}
@@ -601,7 +598,8 @@ tuple_format_destroy_fields(struct tuple_format *format)
{
struct tuple_field *field, *tmp;
json_tree_foreach_entry_safe(field, &format->fields.root,
- struct tuple_field, token, tmp) {
+ struct tuple_field, token, tmp)
+ {
json_tree_del(&format->fields, &field->token);
tuple_field_delete(field);
}
@@ -609,7 +607,7 @@ tuple_format_destroy_fields(struct tuple_format *format)
}
static struct tuple_format *
-tuple_format_alloc(struct key_def * const *keys, uint16_t key_count,
+tuple_format_alloc(struct key_def *const *keys, uint16_t key_count,
uint32_t space_field_count, struct tuple_dictionary *dict)
{
/* Size of area to store JSON paths data. */
@@ -623,8 +621,8 @@ tuple_format_alloc(struct key_def * const *keys, uint16_t key_count,
const struct key_part *part = key_def->parts;
const struct key_part *pend = part + key_def->part_count;
for (; part < pend; part++) {
- index_field_count = MAX(index_field_count,
- part->fieldno + 1);
+ index_field_count =
+ MAX(index_field_count, part->fieldno + 1);
path_pool_size += part->path_len;
}
}
@@ -717,11 +715,10 @@ tuple_format_reuse(struct tuple_format **p_format)
struct tuple_format *format = *p_format;
assert(format->is_ephemeral);
assert(format->is_temporary);
- mh_int_t key = mh_tuple_format_find(tuple_formats_hash, format,
- NULL);
+ mh_int_t key = mh_tuple_format_find(tuple_formats_hash, format, NULL);
if (key != mh_end(tuple_formats_hash)) {
- struct tuple_format **entry = mh_tuple_format_node(
- tuple_formats_hash, key);
+ struct tuple_format **entry =
+ mh_tuple_format_node(tuple_formats_hash, key);
tuple_format_destroy(format);
free(format);
*p_format = *entry;
@@ -741,9 +738,9 @@ tuple_format_add_to_hash(struct tuple_format *format)
{
assert(format->is_ephemeral);
assert(format->is_temporary);
- mh_int_t key = mh_tuple_format_put(tuple_formats_hash,
- (const struct tuple_format **)&format,
- NULL, NULL);
+ mh_int_t key = mh_tuple_format_put(
+ tuple_formats_hash, (const struct tuple_format **)&format, NULL,
+ NULL);
if (key == mh_end(tuple_formats_hash)) {
diag_set(OutOfMemory, 0, "tuple_format_add_to_hash",
"tuple formats hash entry");
@@ -771,7 +768,7 @@ tuple_format_delete(struct tuple_format *format)
struct tuple_format *
tuple_format_new(struct tuple_format_vtab *vtab, void *engine,
- struct key_def * const *keys, uint16_t key_count,
+ struct key_def *const *keys, uint16_t key_count,
const struct field_def *space_fields,
uint32_t space_field_count, uint32_t exact_field_count,
struct tuple_dictionary *dict, bool is_temporary,
@@ -816,7 +813,8 @@ tuple_format1_can_store_format2_tuples(struct tuple_format *format1,
return false;
struct tuple_field *field1;
json_tree_foreach_entry_preorder(field1, &format1->fields.root,
- struct tuple_field, token) {
+ struct tuple_field, token)
+ {
struct tuple_field *field2 =
tuple_format1_field_by_format2_field(format2, field1);
/*
@@ -839,7 +837,7 @@ tuple_format1_can_store_format2_tuples(struct tuple_format *format1,
else
return false;
}
- if (! field_type1_contains_type2(field1->type, field2->type))
+ if (!field_type1_contains_type2(field1->type, field2->type))
return false;
/*
* Do not allow transition from nullable to non-nullable:
@@ -858,8 +856,8 @@ tuple_field_map_create(struct tuple_format *format, const char *tuple,
bool validate, struct field_map_builder *builder)
{
struct region *region = &fiber()->gc;
- if (field_map_builder_create(builder, format->field_map_size,
- region) != 0)
+ if (field_map_builder_create(builder, format->field_map_size, region) !=
+ 0)
return -1;
if (tuple_format_field_count(format) == 0)
return 0; /* Nothing to initialize */
@@ -876,22 +874,23 @@ tuple_field_map_create(struct tuple_format *format, const char *tuple,
if (entry.field == NULL)
continue;
if (entry.field->offset_slot != TUPLE_OFFSET_SLOT_NIL &&
- field_map_builder_set_slot(builder, entry.field->offset_slot,
- entry.data - tuple, entry.multikey_idx,
- entry.multikey_count, region) != 0)
+ field_map_builder_set_slot(
+ builder, entry.field->offset_slot,
+ entry.data - tuple, entry.multikey_idx,
+ entry.multikey_count, region) != 0)
return -1;
}
return entry.data == NULL ? 0 : -1;
}
uint32_t
-tuple_format_min_field_count(struct key_def * const *keys, uint16_t key_count,
+tuple_format_min_field_count(struct key_def *const *keys, uint16_t key_count,
const struct field_def *space_fields,
uint32_t space_field_count)
{
uint32_t min_field_count = 0;
for (uint32_t i = 0; i < space_field_count; ++i) {
- if (! space_fields[i].is_nullable)
+ if (!space_fields[i].is_nullable)
min_field_count = i + 1;
}
for (uint32_t i = 0; i < key_count; ++i) {
@@ -911,8 +910,8 @@ tuple_format_init()
{
tuple_formats_hash = mh_tuple_format_new();
if (tuple_formats_hash == NULL) {
- diag_set(OutOfMemory, sizeof(struct mh_tuple_format_t), "malloc",
- "tuple format hash");
+ diag_set(OutOfMemory, sizeof(struct mh_tuple_format_t),
+ "malloc", "tuple format hash");
return -1;
}
return 0;
@@ -924,8 +923,8 @@ tuple_format_free()
{
/* Clear recycled ids. */
while (recycled_format_ids != FORMAT_ID_NIL) {
- uint16_t id = (uint16_t) recycled_format_ids;
- recycled_format_ids = (intptr_t) tuple_formats[id];
+ uint16_t id = (uint16_t)recycled_format_ids;
+ recycled_format_ids = (intptr_t)tuple_formats[id];
tuple_formats[id] = NULL;
}
for (struct tuple_format **format = tuple_formats;
@@ -964,8 +963,8 @@ tuple_format_iterator_create(struct tuple_format_iterator *it,
if (validate && format->exact_field_count > 0 &&
format->exact_field_count != *defined_field_count) {
diag_set(ClientError, ER_EXACT_FIELD_COUNT,
- (unsigned) *defined_field_count,
- (unsigned) format->exact_field_count);
+ (unsigned)*defined_field_count,
+ (unsigned)format->exact_field_count);
return -1;
}
it->parent = &format->fields.root;
@@ -981,19 +980,20 @@ tuple_format_iterator_create(struct tuple_format_iterator *it,
if (validate)
it->required_fields_sz = bitmap_size(format->total_field_count);
uint32_t total_sz = frames_sz + 2 * it->required_fields_sz;
- struct mp_frame *frames = region_aligned_alloc(region, total_sz,
- alignof(frames[0]));
+ struct mp_frame *frames =
+ region_aligned_alloc(region, total_sz, alignof(frames[0]));
if (frames == NULL) {
diag_set(OutOfMemory, total_sz, "region",
"tuple_format_iterator");
return -1;
}
mp_stack_create(&it->stack, format->fields_depth, frames);
- bool key_parts_only =
- (flags & TUPLE_FORMAT_ITERATOR_KEY_PARTS_ONLY) != 0;
- *defined_field_count = MIN(*defined_field_count, key_parts_only ?
- format->index_field_count :
- tuple_format_field_count(format));
+ bool key_parts_only = (flags & TUPLE_FORMAT_ITERATOR_KEY_PARTS_ONLY) !=
+ 0;
+ *defined_field_count = MIN(*defined_field_count,
+ key_parts_only ?
+ format->index_field_count :
+ tuple_format_field_count(format));
mp_stack_push(&it->stack, MP_ARRAY, *defined_field_count);
if (validate) {
@@ -1066,15 +1066,16 @@ tuple_format_iterator_next(struct tuple_format_iterator *it,
* all required fields are present.
*/
if (it->flags & TUPLE_FORMAT_ITERATOR_VALIDATE &&
- tuple_format_required_fields_validate(it->format,
- it->multikey_required_fields,
- it->required_fields_sz) != 0)
+ tuple_format_required_fields_validate(
+ it->format, it->multikey_required_fields,
+ it->required_fields_sz) != 0)
return -1;
}
}
entry->parent =
it->parent != &it->format->fields.root ?
- json_tree_entry(it->parent, struct tuple_field, token) : NULL;
+ json_tree_entry(it->parent, struct tuple_field, token) :
+ NULL;
/*
* Use the top frame of the stack and the
* current data offset to prepare the JSON token
@@ -1105,8 +1106,8 @@ tuple_format_iterator_next(struct tuple_format_iterator *it,
struct tuple_field *field =
json_tree_lookup_entry(&it->format->fields, it->parent, &token,
struct tuple_field, token);
- if (it->flags & TUPLE_FORMAT_ITERATOR_KEY_PARTS_ONLY &&
- field != NULL && !field->is_key_part)
+ if (it->flags & TUPLE_FORMAT_ITERATOR_KEY_PARTS_ONLY && field != NULL &&
+ !field->is_key_part)
field = NULL;
entry->field = field;
entry->data = it->pos;
@@ -1127,9 +1128,8 @@ tuple_format_iterator_next(struct tuple_format_iterator *it,
enum mp_type type = mp_typeof(*it->pos);
if ((type == MP_ARRAY || type == MP_MAP) &&
!mp_stack_is_full(&it->stack) && field != NULL) {
- uint32_t size = type == MP_ARRAY ?
- mp_decode_array(&it->pos) :
- mp_decode_map(&it->pos);
+ uint32_t size = type == MP_ARRAY ? mp_decode_array(&it->pos) :
+ mp_decode_map(&it->pos);
entry->count = size;
mp_stack_push(&it->stack, type, size);
if (json_token_is_multikey(&field->token)) {
@@ -1169,19 +1169,21 @@ tuple_format_iterator_next(struct tuple_format_iterator *it,
* defined in format.
*/
bool is_nullable = tuple_field_is_nullable(field);
- if (!field_mp_type_is_compatible(field->type, entry->data, is_nullable) != 0) {
- diag_set(ClientError, ER_FIELD_TYPE,
- tuple_field_path(field),
+ if (!field_mp_type_is_compatible(field->type, entry->data,
+ is_nullable) != 0) {
+ diag_set(ClientError, ER_FIELD_TYPE, tuple_field_path(field),
field_type_strs[field->type]);
return -1;
}
- bit_clear(it->multikey_frame != NULL ?
- it->multikey_required_fields : it->required_fields, field->id);
+ bit_clear(it->multikey_frame != NULL ? it->multikey_required_fields :
+ it->required_fields,
+ field->id);
return 0;
eof:
if (it->flags & TUPLE_FORMAT_ITERATOR_VALIDATE &&
tuple_format_required_fields_validate(it->format,
- it->required_fields, it->required_fields_sz) != 0)
+ it->required_fields,
+ it->required_fields_sz) != 0)
return -1;
entry->data = NULL;
return 0;
diff --git a/src/box/tuple_format.h b/src/box/tuple_format.h
index 021072d..a25132f 100644
--- a/src/box/tuple_format.h
+++ b/src/box/tuple_format.h
@@ -49,7 +49,7 @@ void
tuple_format_free();
enum { FORMAT_ID_MAX = UINT16_MAX - 1, FORMAT_ID_NIL = UINT16_MAX };
-enum { FORMAT_REF_MAX = INT32_MAX};
+enum { FORMAT_REF_MAX = INT32_MAX };
/*
* We don't pass TUPLE_INDEX_BASE around dynamically all the time,
@@ -74,29 +74,26 @@ struct tuple_format_vtab {
* Free allocated tuple using engine-specific
* memory allocator.
*/
- void
- (*tuple_delete)(struct tuple_format *format, struct tuple *tuple);
+ void (*tuple_delete)(struct tuple_format *format, struct tuple *tuple);
/**
* Allocates a new tuple on the same allocator
* and with the same format.
*/
- struct tuple*
- (*tuple_new)(struct tuple_format *format, const char *data,
- const char *end);
+ struct tuple *(*tuple_new)(struct tuple_format *format,
+ const char *data, const char *end);
/**
* Free a tuple_chunk allocated for given tuple and
* data.
*/
- void
- (*tuple_chunk_delete)(struct tuple_format *format,
- const char *data);
+ void (*tuple_chunk_delete)(struct tuple_format *format,
+ const char *data);
/**
* Allocate a new tuple_chunk for given tuple and data and
* return a pointer to it's data section.
*/
- const char *
- (*tuple_chunk_new)(struct tuple_format *format, struct tuple *tuple,
- const char *data, uint32_t data_sz);
+ const char *(*tuple_chunk_new)(struct tuple_format *format,
+ struct tuple *tuple, const char *data,
+ uint32_t data_sz);
};
/** Tuple field meta information for tuple_format. */
@@ -272,8 +269,8 @@ tuple_format_field_by_path(struct tuple_format *format, uint32_t fieldno,
assert(root != NULL);
if (path == NULL)
return root;
- return json_tree_lookup_path_entry(&format->fields, &root->token,
- path, path_len, TUPLE_INDEX_BASE,
+ return json_tree_lookup_path_entry(&format->fields, &root->token, path,
+ path_len, TUPLE_INDEX_BASE,
struct tuple_field, token);
}
@@ -338,7 +335,7 @@ tuple_format_unref(struct tuple_format *format)
*/
struct tuple_format *
tuple_format_new(struct tuple_format_vtab *vtab, void *engine,
- struct key_def * const *keys, uint16_t key_count,
+ struct key_def *const *keys, uint16_t key_count,
const struct field_def *space_fields,
uint32_t space_field_count, uint32_t exact_field_count,
struct tuple_dictionary *dict, bool is_temporary,
@@ -370,7 +367,7 @@ tuple_format1_can_store_format2_tuples(struct tuple_format *format1,
* @retval Minimal field count.
*/
uint32_t
-tuple_format_min_field_count(struct key_def * const *keys, uint16_t key_count,
+tuple_format_min_field_count(struct key_def *const *keys, uint16_t key_count,
const struct field_def *space_fields,
uint32_t space_field_count);
@@ -436,20 +433,19 @@ tuple_field_map_create(struct tuple_format *format, const char *tuple,
int
tuple_format_init();
-
/** Tuple format iterator flags to configure parse mode. */
enum {
/**
* This flag is set for iterator that should perform tuple
* validation to conform the specified format.
*/
- TUPLE_FORMAT_ITERATOR_VALIDATE = 1 << 0,
+ TUPLE_FORMAT_ITERATOR_VALIDATE = 1 << 0,
/**
* This flag is set for iterator that should skip the
* tuple fields that are not marked as key_parts in
* format::fields tree.
*/
- TUPLE_FORMAT_ITERATOR_KEY_PARTS_ONLY = 1 << 1,
+ TUPLE_FORMAT_ITERATOR_KEY_PARTS_ONLY = 1 << 1,
};
/**
diff --git a/src/box/tuple_hash.cc b/src/box/tuple_hash.cc
index 39f89a6..3d26dd0 100644
--- a/src/box/tuple_hash.cc
+++ b/src/box/tuple_hash.cc
@@ -38,9 +38,7 @@
/* Tuple and key hasher */
namespace {
-enum {
- HASH_SEED = 13U
-};
+enum { HASH_SEED = 13U };
template <int TYPE>
static inline uint32_t
@@ -57,7 +55,7 @@ field_hash(uint32_t *ph, uint32_t *pcarry, const char **field)
const char *f = *field;
uint32_t size;
mp_next(field);
- size = *field - f; /* calculate the size of field */
+ size = *field - f; /* calculate the size of field */
assert(size < INT32_MAX);
PMurHash32_Process(ph, pcarry, f, size);
return size;
@@ -81,30 +79,28 @@ field_hash<FIELD_TYPE_STRING>(uint32_t *ph, uint32_t *pcarry,
return size;
}
-template <int TYPE, int ...MORE_TYPES> struct KeyFieldHash {};
+template <int TYPE, int... MORE_TYPES> struct KeyFieldHash {};
-template <int TYPE, int TYPE2, int ...MORE_TYPES>
+template <int TYPE, int TYPE2, int... MORE_TYPES>
struct KeyFieldHash<TYPE, TYPE2, MORE_TYPES...> {
- static void hash(uint32_t *ph, uint32_t *pcarry,
- const char **pfield, uint32_t *ptotal_size)
+ static void hash(uint32_t *ph, uint32_t *pcarry, const char **pfield,
+ uint32_t *ptotal_size)
{
*ptotal_size += field_hash<TYPE>(ph, pcarry, pfield);
- KeyFieldHash<TYPE2, MORE_TYPES...>::
- hash(ph, pcarry, pfield, ptotal_size);
+ KeyFieldHash<TYPE2, MORE_TYPES...>::hash(ph, pcarry, pfield,
+ ptotal_size);
}
};
-template <int TYPE>
-struct KeyFieldHash<TYPE> {
- static void hash(uint32_t *ph, uint32_t *pcarry,
- const char **pfield, uint32_t *ptotal_size)
+template <int TYPE> struct KeyFieldHash<TYPE> {
+ static void hash(uint32_t *ph, uint32_t *pcarry, const char **pfield,
+ uint32_t *ptotal_size)
{
*ptotal_size += field_hash<TYPE>(ph, pcarry, pfield);
}
};
-template <int TYPE, int ...MORE_TYPES>
-struct KeyHash {
+template <int TYPE, int... MORE_TYPES> struct KeyHash {
static uint32_t hash(const char *key, struct key_def *)
{
uint32_t h = HASH_SEED;
@@ -116,33 +112,31 @@ struct KeyHash {
}
};
-template <>
-struct KeyHash<FIELD_TYPE_UNSIGNED> {
+template <> struct KeyHash<FIELD_TYPE_UNSIGNED> {
static uint32_t hash(const char *key, struct key_def *key_def)
{
uint64_t val = mp_decode_uint(&key);
- (void) key_def;
+ (void)key_def;
if (likely(val <= UINT32_MAX))
return val;
- return ((uint32_t)((val)>>33^(val)^(val)<<11));
+ return ((uint32_t)((val) >> 33 ^ (val) ^ (val) << 11));
}
};
-template <int TYPE, int ...MORE_TYPES> struct TupleFieldHash { };
+template <int TYPE, int... MORE_TYPES> struct TupleFieldHash {};
-template <int TYPE, int TYPE2, int ...MORE_TYPES>
+template <int TYPE, int TYPE2, int... MORE_TYPES>
struct TupleFieldHash<TYPE, TYPE2, MORE_TYPES...> {
static void hash(const char **pfield, uint32_t *ph, uint32_t *pcarry,
uint32_t *ptotal_size)
{
*ptotal_size += field_hash<TYPE>(ph, pcarry, pfield);
- TupleFieldHash<TYPE2, MORE_TYPES...>::
- hash(pfield, ph, pcarry, ptotal_size);
+ TupleFieldHash<TYPE2, MORE_TYPES...>::hash(pfield, ph, pcarry,
+ ptotal_size);
}
};
-template <int TYPE>
-struct TupleFieldHash<TYPE> {
+template <int TYPE> struct TupleFieldHash<TYPE> {
static void hash(const char **pfield, uint32_t *ph, uint32_t *pcarry,
uint32_t *ptotal_size)
{
@@ -150,44 +144,40 @@ struct TupleFieldHash<TYPE> {
}
};
-template <int TYPE, int ...MORE_TYPES>
-struct TupleHash
-{
+template <int TYPE, int... MORE_TYPES> struct TupleHash {
static uint32_t hash(struct tuple *tuple, struct key_def *key_def)
{
assert(!key_def->is_multikey);
uint32_t h = HASH_SEED;
uint32_t carry = 0;
uint32_t total_size = 0;
- const char *field = tuple_field_by_part(tuple,
- key_def->parts,
- MULTIKEY_NONE);
- TupleFieldHash<TYPE, MORE_TYPES...>::
- hash(&field, &h, &carry, &total_size);
+ const char *field = tuple_field_by_part(tuple, key_def->parts,
+ MULTIKEY_NONE);
+ TupleFieldHash<TYPE, MORE_TYPES...>::hash(&field, &h, &carry,
+ &total_size);
return PMurHash32_Result(h, carry, total_size);
}
};
-template <>
-struct TupleHash<FIELD_TYPE_UNSIGNED> {
- static uint32_t hash(struct tuple *tuple, struct key_def *key_def)
+template <> struct TupleHash<FIELD_TYPE_UNSIGNED> {
+ static uint32_t hash(struct tuple *tuple, struct key_def *key_def)
{
assert(!key_def->is_multikey);
- const char *field = tuple_field_by_part(tuple,
- key_def->parts,
- MULTIKEY_NONE);
+ const char *field = tuple_field_by_part(tuple, key_def->parts,
+ MULTIKEY_NONE);
uint64_t val = mp_decode_uint(&field);
if (likely(val <= UINT32_MAX))
return val;
- return ((uint32_t)((val)>>33^(val)^(val)<<11));
+ return ((uint32_t)((val) >> 33 ^ (val) ^ (val) << 11));
}
};
-}; /* namespace { */
+}; // namespace
-#define HASHER(...) \
- { KeyHash<__VA_ARGS__>::hash, TupleHash<__VA_ARGS__>::hash, \
- { __VA_ARGS__, UINT32_MAX } },
+#define HASHER(...) \
+ { KeyHash<__VA_ARGS__>::hash, \
+ TupleHash<__VA_ARGS__>::hash, \
+ { __VA_ARGS__, UINT32_MAX } },
struct hasher_signature {
key_hash_t kf;
@@ -199,20 +189,32 @@ struct hasher_signature {
* field1 type, field2 type, ...
*/
static const hasher_signature hash_arr[] = {
- HASHER(FIELD_TYPE_UNSIGNED)
- HASHER(FIELD_TYPE_STRING)
- HASHER(FIELD_TYPE_UNSIGNED, FIELD_TYPE_UNSIGNED)
- HASHER(FIELD_TYPE_STRING , FIELD_TYPE_UNSIGNED)
- HASHER(FIELD_TYPE_UNSIGNED, FIELD_TYPE_STRING)
- HASHER(FIELD_TYPE_STRING , FIELD_TYPE_STRING)
- HASHER(FIELD_TYPE_UNSIGNED, FIELD_TYPE_UNSIGNED, FIELD_TYPE_UNSIGNED)
- HASHER(FIELD_TYPE_STRING , FIELD_TYPE_UNSIGNED, FIELD_TYPE_UNSIGNED)
- HASHER(FIELD_TYPE_UNSIGNED, FIELD_TYPE_STRING , FIELD_TYPE_UNSIGNED)
- HASHER(FIELD_TYPE_STRING , FIELD_TYPE_STRING , FIELD_TYPE_UNSIGNED)
- HASHER(FIELD_TYPE_UNSIGNED, FIELD_TYPE_UNSIGNED, FIELD_TYPE_STRING)
- HASHER(FIELD_TYPE_STRING , FIELD_TYPE_UNSIGNED, FIELD_TYPE_STRING)
- HASHER(FIELD_TYPE_UNSIGNED, FIELD_TYPE_STRING , FIELD_TYPE_STRING)
- HASHER(FIELD_TYPE_STRING , FIELD_TYPE_STRING , FIELD_TYPE_STRING)
+ HASHER(FIELD_TYPE_UNSIGNED) HASHER(FIELD_TYPE_STRING) HASHER(
+ FIELD_TYPE_UNSIGNED,
+ FIELD_TYPE_UNSIGNED) HASHER(FIELD_TYPE_STRING,
+ FIELD_TYPE_UNSIGNED)
+ HASHER(FIELD_TYPE_UNSIGNED, FIELD_TYPE_STRING) HASHER(
+ FIELD_TYPE_STRING,
+ FIELD_TYPE_STRING) HASHER(FIELD_TYPE_UNSIGNED,
+ FIELD_TYPE_UNSIGNED,
+ FIELD_TYPE_UNSIGNED)
+ HASHER(FIELD_TYPE_STRING, FIELD_TYPE_UNSIGNED,
+ FIELD_TYPE_UNSIGNED) HASHER(FIELD_TYPE_UNSIGNED,
+ FIELD_TYPE_STRING,
+ FIELD_TYPE_UNSIGNED)
+ HASHER(FIELD_TYPE_STRING, FIELD_TYPE_STRING,
+ FIELD_TYPE_UNSIGNED) HASHER(FIELD_TYPE_UNSIGNED,
+ FIELD_TYPE_UNSIGNED,
+ FIELD_TYPE_STRING)
+ HASHER(FIELD_TYPE_STRING,
+ FIELD_TYPE_UNSIGNED,
+ FIELD_TYPE_STRING)
+ HASHER(FIELD_TYPE_UNSIGNED,
+ FIELD_TYPE_STRING,
+ FIELD_TYPE_STRING)
+ HASHER(FIELD_TYPE_STRING,
+ FIELD_TYPE_STRING,
+ FIELD_TYPE_STRING)
};
#undef HASHER
@@ -225,7 +227,8 @@ uint32_t
key_hash_slowpath(const char *key, struct key_def *key_def);
void
-key_def_set_hash_func(struct key_def *key_def) {
+key_def_set_hash_func(struct key_def *key_def)
+{
if (key_def->is_nullable || key_def->has_json_paths)
goto slowpath;
/*
@@ -252,7 +255,8 @@ key_def_set_hash_func(struct key_def *key_def) {
break;
}
}
- if (i == key_def->part_count && hash_arr[k].p[i] == UINT32_MAX){
+ if (i == key_def->part_count &&
+ hash_arr[k].p[i] == UINT32_MAX) {
key_def->tuple_hash = hash_arr[k].tf;
key_def->key_hash = hash_arr[k].kf;
return;
@@ -304,8 +308,8 @@ tuple_hash_field(uint32_t *ph1, uint32_t *pcarry, const char **field,
*/
double iptr;
double val = mp_typeof(**field) == MP_FLOAT ?
- mp_decode_float(field) :
- mp_decode_double(field);
+ mp_decode_float(field) :
+ mp_decode_double(field);
if (!isfinite(val) || modf(val, &iptr) != 0 ||
val < -exp2(63) || val >= exp2(64)) {
size = *field - f;
@@ -323,7 +327,7 @@ tuple_hash_field(uint32_t *ph1, uint32_t *pcarry, const char **field,
}
default:
mp_next(field);
- size = *field - f; /* calculate the size of field */
+ size = *field - f; /* calculate the size of field */
/*
* (!) All other fields hashed **including** MsgPack format
* identifier (e.g. 0xcc). This was done **intentionally**
@@ -396,12 +400,14 @@ tuple_hash_slowpath(struct tuple *tuple, struct key_def *key_def)
if (prev_fieldno + 1 != key_def->parts[part_id].fieldno) {
struct key_part *part = &key_def->parts[part_id];
if (has_json_paths) {
- field = tuple_field_raw_by_part(format, tuple_raw,
+ field = tuple_field_raw_by_part(format,
+ tuple_raw,
field_map, part,
MULTIKEY_NONE);
} else {
- field = tuple_field_raw(format, tuple_raw, field_map,
- part->fieldno);
+ field = tuple_field_raw(format, tuple_raw,
+ field_map,
+ part->fieldno);
}
}
if (has_optional_parts && (field == NULL || field >= end)) {
diff --git a/src/box/txn.c b/src/box/txn.c
index 4f5484e..5f27e3b 100644
--- a/src/box/txn.c
+++ b/src/box/txn.c
@@ -45,7 +45,7 @@ double too_long_threshold;
int64_t txn_last_psn = 0;
/* Txn cache. */
-static struct stailq txn_cache = {NULL, &txn_cache.first};
+static struct stailq txn_cache = { NULL, &txn_cache.first };
static int
txn_on_stop(struct trigger *trigger, void *event);
@@ -159,7 +159,8 @@ txn_rollback_to_svp(struct txn *txn, struct stailq_entry *svp)
struct stailq rollback;
stailq_cut_tail(&txn->stmts, svp, &rollback);
stailq_reverse(&rollback);
- stailq_foreach_entry(stmt, &rollback, next) {
+ stailq_foreach_entry(stmt, &rollback, next)
+ {
txn_rollback_one_stmt(txn, stmt);
if (stmt->row != NULL && stmt->row->replica_id == 0) {
assert(txn->n_new_rows > 0);
@@ -213,16 +214,15 @@ inline static void
txn_free(struct txn *txn)
{
struct tx_read_tracker *tracker, *tmp;
- rlist_foreach_entry_safe(tracker, &txn->read_set,
- in_read_set, tmp) {
+ rlist_foreach_entry_safe(tracker, &txn->read_set, in_read_set, tmp) {
rlist_del(&tracker->in_reader_list);
rlist_del(&tracker->in_read_set);
}
assert(rlist_empty(&txn->read_set));
struct tx_conflict_tracker *entry, *next;
- rlist_foreach_entry_safe(entry, &txn->conflict_list,
- in_conflict_list, next) {
+ rlist_foreach_entry_safe(entry, &txn->conflict_list, in_conflict_list,
+ next) {
rlist_del(&entry->in_conflict_list);
rlist_del(&entry->in_conflicted_by_list);
}
@@ -237,8 +237,7 @@ txn_free(struct txn *txn)
rlist_del(&txn->in_read_view_txs);
struct txn_stmt *stmt;
- stailq_foreach_entry(stmt, &txn->stmts, next)
- txn_stmt_destroy(stmt);
+ stailq_foreach_entry(stmt, &txn->stmts, next) txn_stmt_destroy(stmt);
/* Truncate region up to struct txn size. */
region_truncate(&txn->region, sizeof(struct txn));
@@ -249,7 +248,7 @@ struct txn *
txn_begin(void)
{
static int64_t tsn = 0;
- assert(! in_txn());
+ assert(!in_txn());
struct txn *txn = txn_new();
if (txn == NULL)
return NULL;
@@ -419,7 +418,7 @@ txn_commit_stmt(struct txn *txn, struct request *request)
stmt->does_require_old_tuple = true;
int rc = 0;
- if(!space_is_temporary(stmt->space)) {
+ if (!space_is_temporary(stmt->space)) {
rc = trigger_run(&stmt->space->on_replace, txn);
} else {
/*
@@ -545,7 +544,8 @@ txn_complete(struct txn *txn)
if (stop_tm - txn->start_tm > too_long_threshold) {
int n_rows = txn->n_new_rows + txn->n_applier_rows;
say_warn_ratelimited("too long WAL write: %d rows at "
- "LSN %lld: %.3f sec", n_rows,
+ "LSN %lld: %.3f sec",
+ n_rows,
txn->signature - n_rows + 1,
stop_tm - txn->start_tm);
}
@@ -622,7 +622,8 @@ txn_journal_entry_new(struct txn *txn)
struct xrow_header **local_row = req->rows + txn->n_applier_rows;
bool is_sync = false;
- stailq_foreach_entry(stmt, &txn->stmts, next) {
+ stailq_foreach_entry(stmt, &txn->stmts, next)
+ {
if (stmt->has_triggers) {
txn_init_triggers(txn);
rlist_splice(&txn->on_commit, &stmt->on_commit);
@@ -737,8 +738,8 @@ txn_prepare(struct txn *txn)
struct tx_conflict_tracker *entry, *next;
/* Handle conflicts. */
- rlist_foreach_entry_safe(entry, &txn->conflict_list,
- in_conflict_list, next) {
+ rlist_foreach_entry_safe(entry, &txn->conflict_list, in_conflict_list,
+ next) {
assert(entry->breaker == txn);
memtx_tx_handle_conflict(txn, entry->victim);
rlist_del(&entry->in_conflict_list);
@@ -786,12 +787,12 @@ txn_commit_nop(struct txn *txn)
static int
txn_limbo_on_rollback(struct trigger *trig, void *event)
{
- (void) event;
- struct txn *txn = (struct txn *) event;
+ (void)event;
+ struct txn *txn = (struct txn *)event;
/* Check whether limbo has performed the cleanup. */
if (txn->signature != TXN_SIGNATURE_ROLLBACK)
return 0;
- struct txn_limbo_entry *entry = (struct txn_limbo_entry *) trig->data;
+ struct txn_limbo_entry *entry = (struct txn_limbo_entry *)trig->data;
txn_limbo_abort(&txn_limbo, entry);
return 0;
}
@@ -861,8 +862,7 @@ txn_commit_async(struct txn *txn)
* Set a trigger to abort waiting for confirm on
* WAL write failure.
*/
- trigger_create(trig, txn_limbo_on_rollback,
- limbo_entry, NULL);
+ trigger_create(trig, txn_limbo_on_rollback, limbo_entry, NULL);
txn_on_rollback(txn, trig);
}
@@ -1049,7 +1049,7 @@ box_txn_commit(void)
* Do nothing if transaction is not started,
* it's the same as BEGIN + COMMIT.
*/
- if (! txn)
+ if (!txn)
return 0;
if (txn->in_sub_stmt) {
diag_set(ClientError, ER_COMMIT_IN_SUB_STMT);
@@ -1090,7 +1090,7 @@ box_txn_alloc(size_t size)
long l;
};
return region_aligned_alloc(&txn->region, size,
- alignof(union natural_align));
+ alignof(union natural_align));
}
struct txn_savepoint *
@@ -1160,8 +1160,10 @@ box_txn_rollback_to_savepoint(box_txn_savepoint_t *svp)
diag_set(ClientError, ER_NO_TRANSACTION);
return -1;
}
- struct txn_stmt *stmt = svp->stmt == NULL ? NULL :
- stailq_entry(svp->stmt, struct txn_stmt, next);
+ struct txn_stmt *stmt =
+ svp->stmt == NULL ?
+ NULL :
+ stailq_entry(svp->stmt, struct txn_stmt, next);
if (stmt != NULL && stmt->space == NULL && stmt->row == NULL) {
/*
* The statement at which this savepoint was
@@ -1188,10 +1190,12 @@ txn_savepoint_release(struct txn_savepoint *svp)
struct txn *txn = in_txn();
assert(txn != NULL);
/* Make sure that savepoint hasn't been released yet. */
- struct txn_stmt *stmt = svp->stmt == NULL ? NULL :
- stailq_entry(svp->stmt, struct txn_stmt, next);
+ struct txn_stmt *stmt =
+ svp->stmt == NULL ?
+ NULL :
+ stailq_entry(svp->stmt, struct txn_stmt, next);
assert(stmt == NULL || (stmt->space != NULL && stmt->row != NULL));
- (void) stmt;
+ (void)stmt;
/*
* Discard current savepoint alongside with all
* created after it savepoints.
@@ -1203,9 +1207,9 @@ txn_savepoint_release(struct txn_savepoint *svp)
static int
txn_on_stop(struct trigger *trigger, void *event)
{
- (void) trigger;
- (void) event;
- txn_rollback(in_txn()); /* doesn't yield or fail */
+ (void)trigger;
+ (void)event;
+ txn_rollback(in_txn()); /* doesn't yield or fail */
fiber_gc();
return 0;
}
@@ -1230,8 +1234,8 @@ txn_on_stop(struct trigger *trigger, void *event)
static int
txn_on_yield(struct trigger *trigger, void *event)
{
- (void) trigger;
- (void) event;
+ (void)trigger;
+ (void)event;
struct txn *txn = in_txn();
assert(txn != NULL);
assert(!txn_has_flag(txn, TXN_CAN_YIELD));
diff --git a/src/box/txn_limbo.c b/src/box/txn_limbo.c
index 3655338..24475cc 100644
--- a/src/box/txn_limbo.c
+++ b/src/box/txn_limbo.c
@@ -83,8 +83,8 @@ txn_limbo_append(struct txn_limbo *limbo, uint32_t id, struct txn *txn)
}
}
size_t size;
- struct txn_limbo_entry *e = region_alloc_object(&txn->region,
- typeof(*e), &size);
+ struct txn_limbo_entry *e =
+ region_alloc_object(&txn->region, typeof(*e), &size);
if (e == NULL) {
diag_set(OutOfMemory, size, "region_alloc_object", "e");
return NULL;
@@ -103,7 +103,7 @@ txn_limbo_remove(struct txn_limbo *limbo, struct txn_limbo_entry *entry)
{
assert(!rlist_empty(&entry->in_queue));
assert(txn_limbo_first_entry(limbo) == entry);
- (void) limbo;
+ (void)limbo;
rlist_del_entry(entry, in_queue);
}
@@ -140,7 +140,7 @@ txn_limbo_assign_remote_lsn(struct txn_limbo *limbo,
assert(entry->lsn == -1);
assert(lsn > 0);
assert(txn_has_flag(entry->txn, TXN_WAIT_ACK));
- (void) limbo;
+ (void)limbo;
entry->lsn = lsn;
}
@@ -164,8 +164,7 @@ txn_limbo_assign_local_lsn(struct txn_limbo *limbo,
struct vclock_iterator iter;
vclock_iterator_init(&iter, &limbo->vclock);
int ack_count = 0;
- vclock_foreach(&iter, vc)
- ack_count += vc.lsn >= lsn;
+ vclock_foreach(&iter, vc) ack_count += vc.lsn >= lsn;
assert(ack_count >= entry->ack_count);
entry->ack_count = ack_count;
}
@@ -233,8 +232,8 @@ txn_limbo_wait_complete(struct txn_limbo *limbo, struct txn_limbo_entry *entry)
txn_limbo_write_rollback(limbo, entry->lsn);
struct txn_limbo_entry *e, *tmp;
- rlist_foreach_entry_safe_reverse(e, &limbo->queue,
- in_queue, tmp) {
+ rlist_foreach_entry_safe_reverse(e, &limbo->queue, in_queue, tmp)
+ {
e->txn->signature = TXN_SIGNATURE_QUORUM_TIMEOUT;
txn_limbo_abort(limbo, e);
txn_clear_flag(e->txn, TXN_WAIT_SYNC);
@@ -291,9 +290,9 @@ txn_limbo_write_synchro(struct txn_limbo *limbo, uint32_t type, int64_t lsn)
assert(lsn > 0);
struct synchro_request req = {
- .type = type,
- .replica_id = limbo->instance_id,
- .lsn = lsn,
+ .type = type,
+ .replica_id = limbo->instance_id,
+ .lsn = lsn,
};
/*
@@ -302,8 +301,7 @@ txn_limbo_write_synchro(struct txn_limbo *limbo, uint32_t type, int64_t lsn)
*/
struct synchro_body_bin body;
struct xrow_header row;
- char buf[sizeof(struct journal_entry) +
- sizeof(struct xrow_header *)];
+ char buf[sizeof(struct journal_entry) + sizeof(struct xrow_header *)];
struct journal_entry *entry = (struct journal_entry *)buf;
entry->rows[0] = &row;
@@ -325,8 +323,8 @@ txn_limbo_write_synchro(struct txn_limbo *limbo, uint32_t type, int64_t lsn)
* Or retry automatically with some period.
*/
panic("Could not write a synchro request to WAL: "
- "lsn = %lld, type = %s\n", lsn,
- iproto_type_name(type));
+ "lsn = %lld, type = %s\n",
+ lsn, iproto_type_name(type));
}
}
@@ -404,7 +402,8 @@ txn_limbo_read_rollback(struct txn_limbo *limbo, int64_t lsn)
assert(limbo->instance_id != REPLICA_ID_NIL);
struct txn_limbo_entry *e, *tmp;
struct txn_limbo_entry *last_rollback = NULL;
- rlist_foreach_entry_reverse(e, &limbo->queue, in_queue) {
+ rlist_foreach_entry_reverse(e, &limbo->queue, in_queue)
+ {
if (!txn_has_flag(e->txn, TXN_WAIT_ACK))
continue;
if (e->lsn < lsn)
@@ -413,7 +412,8 @@ txn_limbo_read_rollback(struct txn_limbo *limbo, int64_t lsn)
}
if (last_rollback == NULL)
return;
- rlist_foreach_entry_safe_reverse(e, &limbo->queue, in_queue, tmp) {
+ rlist_foreach_entry_safe_reverse(e, &limbo->queue, in_queue, tmp)
+ {
txn_limbo_abort(limbo, e);
txn_clear_flag(e->txn, TXN_WAIT_SYNC);
txn_clear_flag(e->txn, TXN_WAIT_ACK);
diff --git a/src/box/user.cc b/src/box/user.cc
index 5042fb1..fbb526d 100644
--- a/src/box/user.cc
+++ b/src/box/user.cc
@@ -62,14 +62,13 @@ user_map_calc_idx(uint8_t auth_token, uint8_t *bit_no)
return auth_token / UMAP_INT_BITS;
}
-
/** Set a bit in the user map - add a user. */
static inline void
user_map_set(struct user_map *map, uint8_t auth_token)
{
uint8_t bit_no;
int idx = user_map_calc_idx(auth_token, &bit_no);
- map->m[idx] |= ((umap_int_t) 1) << bit_no;
+ map->m[idx] |= ((umap_int_t)1) << bit_no;
}
/** Clear a bit in the user map - remove a user. */
@@ -78,7 +77,7 @@ user_map_clear(struct user_map *map, uint8_t auth_token)
{
uint8_t bit_no;
int idx = user_map_calc_idx(auth_token, &bit_no);
- map->m[idx] &= ~(((umap_int_t) 1) << bit_no);
+ map->m[idx] &= ~(((umap_int_t)1) << bit_no);
}
/* Check if a bit is set in the user map. */
@@ -87,7 +86,7 @@ user_map_is_set(struct user_map *map, uint8_t auth_token)
{
uint8_t bit_no;
int idx = user_map_calc_idx(auth_token, &bit_no);
- return map->m[idx] & (((umap_int_t) 1) << bit_no);
+ return map->m[idx] & (((umap_int_t)1) << bit_no);
}
/**
@@ -112,16 +111,15 @@ user_map_minus(struct user_map *lhs, struct user_map *rhs)
}
/** Iterate over users in the set of users. */
-struct user_map_iterator
-{
+struct user_map_iterator {
struct bit_iterator it;
};
static void
user_map_iterator_init(struct user_map_iterator *it, struct user_map *map)
{
- bit_iterator_init(&it->it, map->m,
- USER_MAP_SIZE * sizeof(umap_int_t), true);
+ bit_iterator_init(&it->it, map->m, USER_MAP_SIZE * sizeof(umap_int_t),
+ true);
}
static struct user *
@@ -216,66 +214,55 @@ access_find(enum schema_object_type object_type, uint32_t object_id)
{
struct access *access = NULL;
switch (object_type) {
- case SC_UNIVERSE:
- {
+ case SC_UNIVERSE: {
access = universe.access;
break;
}
- case SC_ENTITY_SPACE:
- {
+ case SC_ENTITY_SPACE: {
access = entity_access.space;
break;
}
- case SC_ENTITY_FUNCTION:
- {
+ case SC_ENTITY_FUNCTION: {
access = entity_access.function;
break;
}
- case SC_ENTITY_USER:
- {
+ case SC_ENTITY_USER: {
access = entity_access.user;
break;
}
- case SC_ENTITY_ROLE:
- {
+ case SC_ENTITY_ROLE: {
access = entity_access.role;
break;
}
- case SC_ENTITY_SEQUENCE:
- {
+ case SC_ENTITY_SEQUENCE: {
access = entity_access.sequence;
break;
}
- case SC_SPACE:
- {
+ case SC_SPACE: {
struct space *space = space_by_id(object_id);
if (space)
access = space->access;
break;
}
- case SC_FUNCTION:
- {
+ case SC_FUNCTION: {
struct func *func = func_by_id(object_id);
if (func)
access = func->access;
break;
}
- case SC_USER:
- {
+ case SC_USER: {
struct user *user = user_by_id(object_id);
if (user)
access = user->access;
break;
}
- case SC_ROLE:
- {
+ case SC_ROLE: {
struct user *role = user_by_id(object_id);
if (role)
access = role->access;
break;
}
- case SC_SEQUENCE:
- {
+ case SC_SEQUENCE: {
struct sequence *seq = sequence_by_id(object_id);
if (seq)
access = seq->access;
@@ -287,7 +274,6 @@ access_find(enum schema_object_type object_type, uint32_t object_id)
return access;
}
-
/**
* Reset effective access of the user in the
* corresponding objects.
@@ -299,9 +285,9 @@ user_set_effective_access(struct user *user)
privset_ifirst(&user->privs, &it);
struct priv_def *priv;
while ((priv = privset_inext(&it)) != NULL) {
- struct access *object = access_find(priv->object_type,
- priv->object_id);
- /* Protect against a concurrent drop. */
+ struct access *object =
+ access_find(priv->object_type, priv->object_id);
+ /* Protect against a concurrent drop. */
if (object == NULL)
continue;
struct access *access = &object[user->auth_token];
@@ -340,7 +326,7 @@ user_reload_privs(struct user *user)
/** Primary key - by user id */
if (!space_is_memtx(space)) {
diag_set(ClientError, ER_UNSUPPORTED,
- space->engine->name, "system data");
+ space->engine->name, "system data");
return -1;
}
struct index *index = index_find(space, 0);
@@ -348,8 +334,8 @@ user_reload_privs(struct user *user)
return -1;
mp_encode_uint(key, user->def->uid);
- struct iterator *it = index_create_iterator(index, ITER_EQ,
- key, 1);
+ struct iterator *it =
+ index_create_iterator(index, ITER_EQ, key, 1);
if (it == NULL)
return -1;
IteratorGuard iter_guard(it);
@@ -365,7 +351,8 @@ user_reload_privs(struct user *user)
* Skip role grants, we're only
* interested in real objects.
*/
- if (priv.object_type != SC_ROLE || !(priv.access & PRIV_X))
+ if (priv.object_type != SC_ROLE ||
+ !(priv.access & PRIV_X))
if (user_grant_priv(user, &priv) != 0)
return -1;
if (iterator_next(it, &tuple) != 0)
@@ -418,11 +405,11 @@ auth_token_get(void)
{
uint8_t bit_no = 0;
while (min_token_idx < USER_MAP_SIZE) {
- bit_no = __builtin_ffs(tokens[min_token_idx]);
+ bit_no = __builtin_ffs(tokens[min_token_idx]);
if (bit_no)
break;
min_token_idx++;
- }
+ }
if (bit_no == 0 || bit_no > BOX_USER_MAX) {
/* A cap on the number of users was reached.
* Check for BOX_USER_MAX to cover case when
@@ -430,12 +417,12 @@ auth_token_get(void)
*/
tnt_raise(LoggedError, ER_USER_MAX, BOX_USER_MAX);
}
- /*
+ /*
* find-first-set returns bit index starting from 1,
* or 0 if no bit is set. Rebase the index to offset 0.
*/
bit_no--;
- tokens[min_token_idx] ^= ((umap_int_t) 1) << bit_no;
+ tokens[min_token_idx] ^= ((umap_int_t)1) << bit_no;
int auth_token = min_token_idx * UMAP_INT_BITS + bit_no;
assert(auth_token < UINT8_MAX);
return auth_token;
@@ -450,7 +437,7 @@ auth_token_put(uint8_t auth_token)
{
uint8_t bit_no;
int idx = user_map_calc_idx(auth_token, &bit_no);
- tokens[idx] |= ((umap_int_t) 1) << bit_no;
+ tokens[idx] |= ((umap_int_t)1) << bit_no;
if (idx < min_token_idx)
min_token_idx = idx;
}
@@ -481,8 +468,8 @@ user_cache_delete(uint32_t uid)
{
mh_int_t k = mh_i32ptr_find(user_registry, uid, NULL);
if (k != mh_end(user_registry)) {
- struct user *user = (struct user *)
- mh_i32ptr_node(user_registry, k)->val;
+ struct user *user =
+ (struct user *)mh_i32ptr_node(user_registry, k)->val;
assert(user->auth_token > ADMIN);
auth_token_put(user->auth_token);
assert(user_map_is_empty(&user->roles));
@@ -505,7 +492,7 @@ user_by_id(uint32_t uid)
mh_int_t k = mh_i32ptr_find(user_registry, uid, NULL);
if (k == mh_end(user_registry))
return NULL;
- return (struct user *) mh_i32ptr_node(user_registry, k)->val;
+ return (struct user *)mh_i32ptr_node(user_registry, k)->val;
}
struct user *
@@ -521,7 +508,7 @@ user_find(uint32_t uid)
struct user *
user_find_by_token(uint8_t auth_token)
{
- return &users[auth_token];
+ return &users[auth_token];
}
/** Find user by name. */
@@ -537,7 +524,7 @@ user_find_by_name(const char *name, uint32_t len)
return user;
}
diag_set(ClientError, ER_NO_SUCH_USER,
- tt_cstr(name, MIN((uint32_t) BOX_INVALID_NAME_MAX, len)));
+ tt_cstr(name, MIN((uint32_t)BOX_INVALID_NAME_MAX, len)));
return NULL;
}
@@ -557,7 +544,7 @@ user_cache_init(void)
*/
size_t name_len = strlen("guest");
size_t sz = user_def_sizeof(name_len);
- struct user_def *def = (struct user_def *) calloc(1, sz);
+ struct user_def *def = (struct user_def *)calloc(1, sz);
if (def == NULL)
tnt_raise(OutOfMemory, sz, "malloc", "def");
/* Free def in a case of exception. */
@@ -570,11 +557,11 @@ user_cache_init(void)
guest_def_guard.is_active = false;
/* 0 is the auth token and user id by default. */
assert(user->def->uid == GUEST && user->auth_token == GUEST);
- (void) user;
+ (void)user;
name_len = strlen("admin");
sz = user_def_sizeof(name_len);
- def = (struct user_def *) calloc(1, sz);
+ def = (struct user_def *)calloc(1, sz);
if (def == NULL)
tnt_raise(OutOfMemory, sz, "malloc", "def");
auto admin_def_guard = make_scoped_guard([=] { free(def); });
@@ -628,7 +615,7 @@ role_check(struct user *grantee, struct user *role)
struct user_map transitive_closure = user_map_nil;
user_map_set(&transitive_closure, grantee->auth_token);
struct user_map current_layer = transitive_closure;
- while (! user_map_is_empty(¤t_layer)) {
+ while (!user_map_is_empty(¤t_layer)) {
/*
* As long as we're traversing a directed
* acyclic graph, we're bound to end at some
@@ -647,10 +634,9 @@ role_check(struct user *grantee, struct user *role)
* Check if the role is in the list of roles to which the
* grantee is granted.
*/
- if (user_map_is_set(&transitive_closure,
- role->auth_token)) {
- diag_set(ClientError, ER_ROLE_LOOP,
- role->def->name, grantee->def->name);
+ if (user_map_is_set(&transitive_closure, role->auth_token)) {
+ diag_set(ClientError, ER_ROLE_LOOP, role->def->name,
+ grantee->def->name);
return -1;
}
return 0;
@@ -704,7 +690,7 @@ rebuild_effective_grants(struct user *grantee)
* Propagate effective privileges from the nodes
* with no incoming edges to the remaining nodes.
*/
- while (! user_map_is_empty(¤t_layer)) {
+ while (!user_map_is_empty(¤t_layer)) {
struct user_map postponed = user_map_nil;
struct user_map next_layer = user_map_nil;
user_map_iterator_init(&it, ¤t_layer);
@@ -737,7 +723,6 @@ rebuild_effective_grants(struct user *grantee)
return 0;
}
-
/**
* Update verges in the graph of dependencies.
* Grant all effective privileges of the role to whoever
diff --git a/src/box/user.h b/src/box/user.h
index 9ed52c4..4600ba8 100644
--- a/src/box/user.h
+++ b/src/box/user.h
@@ -51,7 +51,7 @@ extern struct universe universe;
typedef unsigned int umap_int_t;
enum {
UMAP_INT_BITS = CHAR_BIT * sizeof(umap_int_t),
- USER_MAP_SIZE = (BOX_USER_MAX + UMAP_INT_BITS - 1)/UMAP_INT_BITS
+ USER_MAP_SIZE = (BOX_USER_MAX + UMAP_INT_BITS - 1) / UMAP_INT_BITS
};
struct user_map {
@@ -70,8 +70,7 @@ user_map_is_empty(struct user_map *map)
typedef rb_tree(struct priv_def) privset_t;
rb_proto(, privset_, privset_t, struct priv_def);
-struct user
-{
+struct user {
struct user_def *def;
/**
* An id in privileges array to quickly find a
diff --git a/src/box/user_def.c b/src/box/user_def.c
index 4d9821a..ef1dd70 100644
--- a/src/box/user_def.c
+++ b/src/box/user_def.c
@@ -36,25 +36,12 @@ const char *
priv_name(user_access_t access)
{
static const char *priv_name_strs[] = {
- "Read",
- "Write",
- "Execute",
- "Session",
- "Usage",
- "Create",
- "Drop",
- "Alter",
- "Reference",
- "Trigger",
- "Insert",
- "Update",
- "Delete",
- "Grant",
- "Revoke",
+ "Read", "Write", "Execute", "Session", "Usage",
+ "Create", "Drop", "Alter", "Reference", "Trigger",
+ "Insert", "Update", "Delete", "Grant", "Revoke",
};
- int bit_no = __builtin_ffs((int) access);
- if (bit_no > 0 && bit_no <= (int) lengthof(priv_name_strs))
+ int bit_no = __builtin_ffs((int)access);
+ if (bit_no > 0 && bit_no <= (int)lengthof(priv_name_strs))
return priv_name_strs[bit_no - 1];
return "Any";
}
-
diff --git a/src/box/user_def.h b/src/box/user_def.h
index 486a4ae..a82d5f3 100644
--- a/src/box/user_def.h
+++ b/src/box/user_def.h
@@ -31,7 +31,7 @@
* SUCH DAMAGE.
*/
#include "schema_def.h" /* for SCHEMA_OBJECT_TYPE */
-#include "scramble.h" /* for SCRAMBLE_SIZE */
+#include "scramble.h" /* for SCRAMBLE_SIZE */
#define RB_COMPACT 1
#include "small/rb.h"
#include "small/rlist.h"
@@ -102,7 +102,7 @@ enum priv_type {
/* Never granted, but used internally. */
PRIV_REVOKE = 16384,
/* all bits */
- PRIV_ALL = ~((user_access_t) 0),
+ PRIV_ALL = ~((user_access_t)0),
};
/**
@@ -180,7 +180,7 @@ user_def_sizeof(uint32_t name_len)
enum {
BOX_SYSTEM_USER_ID_MIN = 0,
GUEST = 0,
- ADMIN = 1,
+ ADMIN = 1,
PUBLIC = 2, /* role */
SUPER = 31, /* role */
BOX_SYSTEM_USER_ID_MAX = PUBLIC
diff --git a/src/box/vclock.c b/src/box/vclock.c
index 90ae275..00e4563 100644
--- a/src/box/vclock.c
+++ b/src/box/vclock.c
@@ -60,9 +60,10 @@ vclock_snprint(char *buf, int size, const struct vclock *vclock)
const char *sep = "";
struct vclock_iterator it;
vclock_iterator_init(&it, vclock);
- vclock_foreach(&it, replica) {
- SNPRINT(total, snprintf, buf, size, "%s%u: %lld",
- sep, (unsigned)replica.id, (long long)replica.lsn);
+ vclock_foreach(&it, replica)
+ {
+ SNPRINT(total, snprintf, buf, size, "%s%u: %lld", sep,
+ (unsigned)replica.id, (long long)replica.lsn);
sep = ", ";
}
@@ -86,78 +87,78 @@ vclock_from_string(struct vclock *vclock, const char *str)
long long lsn;
const char *p = str;
- begin:
- if (*p == '{') {
- ++p;
- goto key;
- } else if (isblank(*p)) {
- ++p;
- goto begin;
- }
- goto error;
- key:
- if (isdigit(*p)) {
- errno = 0;
- replica_id = strtol(p, (char **) &p, 10);
- if (errno != 0 || replica_id < 0 || replica_id >= VCLOCK_MAX)
- goto error;
- goto sep;
- } else if (*p == '}') {
- ++p;
- goto end;
- } else if (isblank(*p)) {
- ++p;
- goto key;
- }
- goto error;
- sep:
- if (*p == ':') {
- ++p;
- goto val;
- } else if (isblank(*p)) {
- ++p;
- goto sep;
- }
- goto error;
- val:
- if (isblank(*p)) {
- ++p;
- goto val;
- } else if (isdigit(*p)) {
- errno = 0;
- lsn = strtoll(p, (char **) &p, 10);
- if (errno != 0 || lsn < 0 || lsn > INT64_MAX ||
- replica_id >= VCLOCK_MAX ||
- vclock_get(vclock, replica_id) > 0)
- goto error;
- vclock->map |= 1 << replica_id;
- vclock->lsn[replica_id] = lsn;
- goto comma;
- }
- goto error;
- comma:
- if (isspace(*p)) {
- ++p;
- goto comma;
- } else if (*p == '}') {
- ++p;
- goto end;
- } else if (*p == ',') {
- ++p;
- goto key;
- }
- goto error;
- end:
- if (*p == '\0') {
- vclock->signature = vclock_calc_sum(vclock);
- return 0;
- } else if (isblank(*p)) {
- ++p;
- goto end;
- }
- /* goto error; */
- error:
- return p - str + 1; /* error */
+begin:
+ if (*p == '{') {
+ ++p;
+ goto key;
+ } else if (isblank(*p)) {
+ ++p;
+ goto begin;
+ }
+ goto error;
+key:
+ if (isdigit(*p)) {
+ errno = 0;
+ replica_id = strtol(p, (char **)&p, 10);
+ if (errno != 0 || replica_id < 0 || replica_id >= VCLOCK_MAX)
+ goto error;
+ goto sep;
+ } else if (*p == '}') {
+ ++p;
+ goto end;
+ } else if (isblank(*p)) {
+ ++p;
+ goto key;
+ }
+ goto error;
+sep:
+ if (*p == ':') {
+ ++p;
+ goto val;
+ } else if (isblank(*p)) {
+ ++p;
+ goto sep;
+ }
+ goto error;
+val:
+ if (isblank(*p)) {
+ ++p;
+ goto val;
+ } else if (isdigit(*p)) {
+ errno = 0;
+ lsn = strtoll(p, (char **)&p, 10);
+ if (errno != 0 || lsn < 0 || lsn > INT64_MAX ||
+ replica_id >= VCLOCK_MAX ||
+ vclock_get(vclock, replica_id) > 0)
+ goto error;
+ vclock->map |= 1 << replica_id;
+ vclock->lsn[replica_id] = lsn;
+ goto comma;
+ }
+ goto error;
+comma:
+ if (isspace(*p)) {
+ ++p;
+ goto comma;
+ } else if (*p == '}') {
+ ++p;
+ goto end;
+ } else if (*p == ',') {
+ ++p;
+ goto key;
+ }
+ goto error;
+end:
+ if (*p == '\0') {
+ vclock->signature = vclock_calc_sum(vclock);
+ return 0;
+ } else if (isblank(*p)) {
+ ++p;
+ goto end;
+ }
+ /* goto error; */
+error:
+ return p - str + 1; /* error */
}
static int
diff --git a/src/box/vclock.h b/src/box/vclock.h
index 5865f74..bb7eac8 100644
--- a/src/box/vclock.h
+++ b/src/box/vclock.h
@@ -98,8 +98,7 @@ struct vclock_c {
int64_t lsn;
};
-struct vclock_iterator
-{
+struct vclock_iterator {
struct bit_iterator it;
const struct vclock *vclock;
};
@@ -116,14 +115,13 @@ vclock_iterator_next(struct vclock_iterator *it)
{
struct vclock_c c = { 0, 0 };
size_t id = bit_iterator_next(&it->it);
- c.id = id == SIZE_MAX ? (int) VCLOCK_MAX : id;
+ c.id = id == SIZE_MAX ? (int)VCLOCK_MAX : id;
if (c.id < VCLOCK_MAX)
c.lsn = it->vclock->lsn[c.id];
return c;
}
-
-#define vclock_foreach(it, var) \
+#define vclock_foreach(it, var) \
for (struct vclock_c var = vclock_iterator_next(it); \
(var).id < VCLOCK_MAX; (var) = vclock_iterator_next(it))
@@ -215,8 +213,8 @@ vclock_copy(struct vclock *dst, const struct vclock *src)
* undefined result if zero passed.
*/
unsigned int max_pos = VCLOCK_MAX - bit_clz_u32(src->map | 0x01);
- memcpy(dst, src, offsetof(struct vclock, lsn) +
- sizeof(*dst->lsn) * max_pos);
+ memcpy(dst, src,
+ offsetof(struct vclock, lsn) + sizeof(*dst->lsn) * max_pos);
}
static inline uint32_t
@@ -237,8 +235,7 @@ vclock_calc_sum(const struct vclock *vclock)
int64_t sum = 0;
struct vclock_iterator it;
vclock_iterator_init(&it, vclock);
- vclock_foreach(&it, replica)
- sum += replica.lsn;
+ vclock_foreach(&it, replica) sum += replica.lsn;
return sum;
}
@@ -268,8 +265,8 @@ vclock_merge(struct vclock *dst, struct vclock *diff)
{
struct vclock_iterator it;
vclock_iterator_init(&it, diff);
- vclock_foreach(&it, item)
- vclock_follow(dst, item.id, vclock_get(dst, item.id) + item.lsn);
+ vclock_foreach(&it, item) vclock_follow(
+ dst, item.id, vclock_get(dst, item.id) + item.lsn);
vclock_create(diff);
}
@@ -377,8 +374,8 @@ vclock_lex_compare(const struct vclock *a, const struct vclock *b)
vclock_map_t map = a->map | b->map;
struct bit_iterator it;
bit_iterator_init(&it, &map, sizeof(map), true);
- for(size_t replica_id = bit_iterator_next(&it); replica_id < VCLOCK_MAX;
- replica_id = bit_iterator_next(&it)) {
+ for (size_t replica_id = bit_iterator_next(&it);
+ replica_id < VCLOCK_MAX; replica_id = bit_iterator_next(&it)) {
int64_t lsn_a = vclock_get(a, replica_id);
int64_t lsn_b = vclock_get(b, replica_id);
if (lsn_a < lsn_b)
@@ -406,7 +403,7 @@ vclock_min_ignore0(struct vclock *a, const struct vclock *b)
if (replica_id == 0)
replica_id = bit_iterator_next(&it);
- for( ; replica_id < VCLOCK_MAX; replica_id = bit_iterator_next(&it)) {
+ for (; replica_id < VCLOCK_MAX; replica_id = bit_iterator_next(&it)) {
int64_t lsn_a = vclock_get(a, replica_id);
int64_t lsn_b = vclock_get(b, replica_id);
if (lsn_a <= lsn_b)
diff --git a/src/box/vinyl.c b/src/box/vinyl.c
index 7e56299..0035581 100644
--- a/src/box/vinyl.c
+++ b/src/box/vinyl.c
@@ -104,7 +104,7 @@ struct vy_env {
/** Memory pool for index iterator. */
struct mempool iterator_pool;
/** Memory quota */
- struct vy_quota quota;
+ struct vy_quota quota;
/** Statement environment. */
struct vy_stmt_env stmt_env;
/** Common LSM tree environment. */
@@ -134,14 +134,14 @@ struct vy_env {
/** Mask passed to vy_gc(). */
enum {
/** Delete incomplete runs. */
- VY_GC_INCOMPLETE = 1 << 0,
+ VY_GC_INCOMPLETE = 1 << 0,
/** Delete dropped runs. */
- VY_GC_DROPPED = 1 << 1,
+ VY_GC_DROPPED = 1 << 1,
};
static void
-vy_gc(struct vy_env *env, struct vy_recovery *recovery,
- unsigned int gc_mask, int64_t gc_lsn);
+vy_gc(struct vy_env *env, struct vy_recovery *recovery, unsigned int gc_mask,
+ int64_t gc_lsn);
struct vinyl_iterator {
struct iterator base;
@@ -259,8 +259,9 @@ vy_info_append_regulator(struct vy_env *env, struct info_handler *h)
info_append_int(h, "write_rate", r->write_rate);
info_append_int(h, "dump_bandwidth", r->dump_bandwidth);
info_append_int(h, "dump_watermark", r->dump_watermark);
- info_append_int(h, "rate_limit", vy_quota_get_rate_limit(r->quota,
- VY_QUOTA_CONSUMER_TX));
+ info_append_int(h, "rate_limit",
+ vy_quota_get_rate_limit(r->quota,
+ VY_QUOTA_CONSUMER_TX));
info_table_end(h); /* regulator */
}
@@ -419,9 +420,12 @@ vinyl_index_stat(struct index *index, struct info_handler *h)
info_table_begin(h, "compaction");
info_append_int(h, "count", stat->disk.compaction.count);
info_append_double(h, "time", stat->disk.compaction.time);
- vy_info_append_disk_stmt_counter(h, "input", &stat->disk.compaction.input);
- vy_info_append_disk_stmt_counter(h, "output", &stat->disk.compaction.output);
- vy_info_append_disk_stmt_counter(h, "queue", &stat->disk.compaction.queue);
+ vy_info_append_disk_stmt_counter(h, "input",
+ &stat->disk.compaction.input);
+ vy_info_append_disk_stmt_counter(h, "output",
+ &stat->disk.compaction.output);
+ vy_info_append_disk_stmt_counter(h, "queue",
+ &stat->disk.compaction.queue);
info_table_end(h); /* compaction */
info_append_int(h, "index_size", lsm->page_index_size);
info_append_int(h, "bloom_size", lsm->bloom_size);
@@ -504,7 +508,7 @@ vinyl_engine_memory_stat(struct engine *engine, struct engine_memory_stat *stat)
struct vy_env *env = vy_env(engine);
stat->data += lsregion_used(&env->mem_env.allocator) -
- env->mem_env.tree_extent_size;
+ env->mem_env.tree_extent_size;
stat->index += env->mem_env.tree_extent_size;
stat->index += env->lsm_env.bloom_size;
stat->index += env->lsm_env.page_index_size;
@@ -563,7 +567,7 @@ vy_lsm_find(struct space *space, uint32_t iid)
* Wrapper around vy_lsm_find() which ensures that
* the found index is unique.
*/
-static struct vy_lsm *
+static struct vy_lsm *
vy_lsm_find_unique(struct space *space, uint32_t index_id)
{
struct vy_lsm *lsm = vy_lsm_find(space, index_id);
@@ -578,8 +582,8 @@ static int
vinyl_engine_check_space_def(struct space_def *def)
{
if (def->opts.is_temporary) {
- diag_set(ClientError, ER_ALTER_SPACE,
- def->name, "engine does not support temporary flag");
+ diag_set(ClientError, ER_ALTER_SPACE, def->name,
+ "engine does not support temporary flag");
return -1;
}
return 0;
@@ -592,8 +596,7 @@ vinyl_engine_create_space(struct engine *engine, struct space_def *def,
struct vy_env *env = vy_env(engine);
struct space *space = malloc(sizeof(*space));
if (space == NULL) {
- diag_set(OutOfMemory, sizeof(*space),
- "malloc", "struct space");
+ diag_set(OutOfMemory, sizeof(*space), "malloc", "struct space");
return NULL;
}
@@ -625,8 +628,8 @@ vinyl_engine_create_space(struct engine *engine, struct space_def *def,
}
tuple_format_ref(format);
- if (space_create(space, engine, &vinyl_space_vtab,
- def, key_list, format) != 0) {
+ if (space_create(space, engine, &vinyl_space_vtab, def, key_list,
+ format) != 0) {
tuple_format_unref(format);
free(space);
return NULL;
@@ -647,8 +650,8 @@ static int
vinyl_space_check_index_def(struct space *space, struct index_def *index_def)
{
if (index_def->type != TREE) {
- diag_set(ClientError, ER_INDEX_TYPE,
- index_def->name, space_name(space));
+ diag_set(ClientError, ER_INDEX_TYPE, index_def->name,
+ space_name(space));
return -1;
}
@@ -663,8 +666,8 @@ vinyl_space_check_index_def(struct space *space, struct index_def *index_def)
struct key_part *part = &key_def->parts[i];
if (part->type <= FIELD_TYPE_ANY ||
part->type >= FIELD_TYPE_ARRAY) {
- diag_set(ClientError, ER_MODIFY_INDEX,
- index_def->name, space_name(space),
+ diag_set(ClientError, ER_MODIFY_INDEX, index_def->name,
+ space_name(space),
tt_sprintf("field type '%s' is not supported",
field_type_strs[part->type]));
return -1;
@@ -694,8 +697,8 @@ vinyl_space_create_index(struct space *space, struct index_def *index_def)
if (lsm == NULL)
return NULL;
- if (index_create(&lsm->base, &env->base,
- &vinyl_index_vtab, index_def) != 0) {
+ if (index_create(&lsm->base, &env->base, &vinyl_index_vtab,
+ index_def) != 0) {
vy_lsm_delete(lsm);
return NULL;
}
@@ -1143,8 +1146,8 @@ vinyl_space_swap_index(struct space *old_space, struct space *new_space,
* Swap the two indexes between the two spaces,
* but leave tuple formats.
*/
- generic_space_swap_index(old_space, new_space,
- old_index_id, new_index_id);
+ generic_space_swap_index(old_space, new_space, old_index_id,
+ new_index_id);
SWAP(old_lsm, new_lsm);
SWAP(old_lsm->mem_format, new_lsm->mem_format);
@@ -1201,8 +1204,8 @@ vinyl_index_bsize(struct index *index)
* they are only needed for building the index.
*/
struct vy_lsm *lsm = vy_lsm(index);
- ssize_t bsize = vy_lsm_mem_tree_size(lsm) +
- lsm->page_index_size + lsm->bloom_size;
+ ssize_t bsize = vy_lsm_mem_tree_size(lsm) + lsm->page_index_size +
+ lsm->bloom_size;
if (lsm->index_id > 0)
bsize += lsm->stat.disk.count.bytes;
return bsize;
@@ -1266,8 +1269,8 @@ vy_is_committed(struct vy_env *env, struct vy_lsm *lsm)
*/
static int
vy_get_by_secondary_tuple(struct vy_lsm *lsm, struct vy_tx *tx,
- const struct vy_read_view **rv,
- struct vy_entry entry, struct vy_entry *result)
+ const struct vy_read_view **rv, struct vy_entry entry,
+ struct vy_entry *result)
{
int rc = 0;
assert(lsm->index_id > 0);
@@ -1306,8 +1309,8 @@ vy_get_by_secondary_tuple(struct vy_lsm *lsm, struct vy_tx *tx,
struct vy_entry full_entry;
if (pk_entry.stmt != NULL) {
vy_stmt_foreach_entry(full_entry, pk_entry.stmt, lsm->cmp_def) {
- if (vy_entry_compare(full_entry, entry,
- lsm->cmp_def) == 0) {
+ if (vy_entry_compare(full_entry, entry, lsm->cmp_def) ==
+ 0) {
match = true;
break;
}
@@ -1354,8 +1357,8 @@ vy_get_by_secondary_tuple(struct vy_lsm *lsm, struct vy_tx *tx,
}
if ((*rv)->vlsn == INT64_MAX) {
- vy_cache_add(&lsm->pk->cache, pk_entry,
- vy_entry_none(), key, ITER_EQ);
+ vy_cache_add(&lsm->pk->cache, pk_entry, vy_entry_none(), key,
+ ITER_EQ);
}
vy_stmt_counter_acct_tuple(&lsm->pk->stat.get, pk_entry.stmt);
@@ -1378,8 +1381,7 @@ out:
* @param -1 Memory error or read error.
*/
static int
-vy_get(struct vy_lsm *lsm, struct vy_tx *tx,
- const struct vy_read_view **rv,
+vy_get(struct vy_lsm *lsm, struct vy_tx *tx, const struct vy_read_view **rv,
struct tuple *key_stmt, struct tuple **result)
{
double start_time = ev_monotonic_now(loop());
@@ -1407,8 +1409,8 @@ vy_get(struct vy_lsm *lsm, struct vy_tx *tx,
if (vy_point_lookup(lsm, tx, rv, key, &partial) != 0)
return -1;
if (lsm->index_id > 0 && partial.stmt != NULL) {
- rc = vy_get_by_secondary_tuple(lsm, tx, rv,
- partial, &entry);
+ rc = vy_get_by_secondary_tuple(lsm, tx, rv, partial,
+ &entry);
tuple_unref(partial.stmt);
if (rc != 0)
return -1;
@@ -1416,8 +1418,8 @@ vy_get(struct vy_lsm *lsm, struct vy_tx *tx,
entry = partial;
}
if ((*rv)->vlsn == INT64_MAX) {
- vy_cache_add(&lsm->cache, entry,
- vy_entry_none(), key, ITER_EQ);
+ vy_cache_add(&lsm->cache, entry, vy_entry_none(), key,
+ ITER_EQ);
}
goto out;
}
@@ -1472,12 +1474,11 @@ out:
*/
static int
vy_get_by_raw_key(struct vy_lsm *lsm, struct vy_tx *tx,
- const struct vy_read_view **rv,
- const char *key_raw, uint32_t part_count,
- struct tuple **result)
+ const struct vy_read_view **rv, const char *key_raw,
+ uint32_t part_count, struct tuple **result)
{
- struct tuple *key = vy_key_new(lsm->env->key_format,
- key_raw, part_count);
+ struct tuple *key =
+ vy_key_new(lsm->env->key_format, key_raw, part_count);
if (key == NULL)
return -1;
int rc = vy_get(lsm, tx, rv, key, result);
@@ -1512,15 +1513,15 @@ vy_check_is_unique_primary(struct vy_tx *tx, const struct vy_read_view **rv,
return -1;
if (found != NULL) {
tuple_unref(found);
- diag_set(ClientError, ER_TUPLE_FOUND,
- index_name, space_name);
+ diag_set(ClientError, ER_TUPLE_FOUND, index_name, space_name);
return -1;
}
return 0;
}
static int
-vy_check_is_unique_secondary_one(struct vy_tx *tx, const struct vy_read_view **rv,
+vy_check_is_unique_secondary_one(struct vy_tx *tx,
+ const struct vy_read_view **rv,
const char *space_name, const char *index_name,
struct vy_lsm *lsm, struct tuple *stmt,
int multikey_idx)
@@ -1532,9 +1533,8 @@ vy_check_is_unique_secondary_one(struct vy_tx *tx, const struct vy_read_view **r
if (lsm->key_def->is_nullable &&
tuple_key_contains_null(stmt, lsm->key_def, multikey_idx))
return 0;
- struct tuple *key = vy_stmt_extract_key(stmt, lsm->key_def,
- lsm->env->key_format,
- multikey_idx);
+ struct tuple *key = vy_stmt_extract_key(
+ stmt, lsm->key_def, lsm->env->key_format, multikey_idx);
if (key == NULL)
return -1;
struct tuple *found;
@@ -1560,8 +1560,7 @@ vy_check_is_unique_secondary_one(struct vy_tx *tx, const struct vy_read_view **r
}
if (found != NULL) {
tuple_unref(found);
- diag_set(ClientError, ER_TUPLE_FOUND,
- index_name, space_name);
+ diag_set(ClientError, ER_TUPLE_FOUND, index_name, space_name);
return -1;
}
return 0;
@@ -1587,14 +1586,14 @@ vy_check_is_unique_secondary(struct vy_tx *tx, const struct vy_read_view **rv,
{
assert(lsm->opts.is_unique);
if (!lsm->cmp_def->is_multikey) {
- return vy_check_is_unique_secondary_one(tx, rv,
- space_name, index_name, lsm, stmt,
- MULTIKEY_NONE);
+ return vy_check_is_unique_secondary_one(tx, rv, space_name,
+ index_name, lsm, stmt,
+ MULTIKEY_NONE);
}
int count = tuple_multikey_count(stmt, lsm->cmp_def);
for (int i = 0; i < count; ++i) {
- if (vy_check_is_unique_secondary_one(tx, rv,
- space_name, index_name, lsm, stmt, i) != 0)
+ if (vy_check_is_unique_secondary_one(
+ tx, rv, space_name, index_name, lsm, stmt, i) != 0)
return -1;
}
return 0;
@@ -1616,9 +1615,8 @@ vy_check_is_unique_secondary(struct vy_tx *tx, const struct vy_read_view **rv,
* @retval -1 Duplicate is found or read error occurred.
*/
static int
-vy_check_is_unique(struct vy_env *env, struct vy_tx *tx,
- struct space *space, struct tuple *stmt,
- uint64_t column_mask)
+vy_check_is_unique(struct vy_env *env, struct vy_tx *tx, struct space *space,
+ struct tuple *stmt, uint64_t column_mask)
{
assert(space->index_count > 0);
assert(vy_stmt_type(stmt) == IPROTO_INSERT ||
@@ -1641,8 +1639,8 @@ vy_check_is_unique(struct vy_env *env, struct vy_tx *tx,
vy_stmt_type(stmt) == IPROTO_INSERT) {
struct vy_lsm *lsm = vy_lsm(space->index[0]);
if (vy_check_is_unique_primary(tx, rv, space_name(space),
- index_name_by_id(space, 0),
- lsm, stmt) != 0)
+ index_name_by_id(space, 0), lsm,
+ stmt) != 0)
return -1;
}
@@ -1678,8 +1676,7 @@ vy_check_is_unique(struct vy_env *env, struct vy_tx *tx,
* in the diagnostics area.
*/
static inline int
-vy_unique_key_validate(struct vy_lsm *lsm, const char *key,
- uint32_t part_count)
+vy_unique_key_validate(struct vy_lsm *lsm, const char *key, uint32_t part_count)
{
assert(lsm->opts.is_unique);
assert(key != NULL || part_count == 0);
@@ -1694,8 +1691,8 @@ vy_unique_key_validate(struct vy_lsm *lsm, const char *key,
*/
uint32_t original_part_count = lsm->key_def->part_count;
if (original_part_count != part_count) {
- diag_set(ClientError, ER_EXACT_MATCH,
- original_part_count, part_count);
+ diag_set(ClientError, ER_EXACT_MATCH, original_part_count,
+ part_count);
return -1;
}
const char *key_end;
@@ -1740,8 +1737,8 @@ vy_delete(struct vy_env *env, struct vy_tx *tx, struct txn_stmt *stmt,
* - if deletion is done by a secondary index.
*/
if (lsm->index_id > 0 || !rlist_empty(&space->on_replace)) {
- if (vy_get_by_raw_key(lsm, tx, vy_tx_read_view(tx),
- key, part_count, &stmt->old_tuple) != 0)
+ if (vy_get_by_raw_key(lsm, tx, vy_tx_read_view(tx), key,
+ part_count, &stmt->old_tuple) != 0)
return -1;
if (stmt->old_tuple == NULL)
return 0;
@@ -1763,8 +1760,8 @@ vy_delete(struct vy_env *env, struct vy_tx *tx, struct txn_stmt *stmt,
}
} else {
assert(lsm->index_id == 0);
- delete = vy_stmt_new_delete(pk->env->key_format,
- request->key, request->key_end);
+ delete = vy_stmt_new_delete(pk->env->key_format, request->key,
+ request->key_end);
if (delete == NULL)
return -1;
if (space->index_count > 1)
@@ -1797,8 +1794,8 @@ vy_check_update(struct space *space, const struct vy_lsm *pk,
uint64_t column_mask)
{
if (!key_update_can_be_skipped(pk->key_def->column_mask, column_mask) &&
- vy_stmt_compare(old_tuple, HINT_NONE, new_tuple,
- HINT_NONE, pk->key_def) != 0) {
+ vy_stmt_compare(old_tuple, HINT_NONE, new_tuple, HINT_NONE,
+ pk->key_def) != 0) {
diag_set(ClientError, ER_CANT_UPDATE_PRIMARY_KEY,
index_name_by_id(space, pk->index_id),
space_name(space));
@@ -1821,8 +1818,8 @@ vy_perform_update(struct vy_env *env, struct vy_tx *tx, struct txn_stmt *stmt,
assert(stmt->old_tuple != NULL);
assert(stmt->new_tuple != NULL);
- if (vy_check_is_unique(env, tx, space, stmt->new_tuple,
- column_mask) != 0)
+ if (vy_check_is_unique(env, tx, space, stmt->new_tuple, column_mask) !=
+ 0)
return -1;
vy_stmt_set_flags(stmt->new_tuple, VY_STMT_UPDATE);
@@ -1832,8 +1829,8 @@ vy_perform_update(struct vy_env *env, struct vy_tx *tx, struct txn_stmt *stmt,
if (space->index_count == 1)
return 0;
- struct tuple *delete = vy_stmt_new_surrogate_delete(pk->mem_format,
- stmt->old_tuple);
+ struct tuple *delete =
+ vy_stmt_new_surrogate_delete(pk->mem_format, stmt->old_tuple);
if (delete == NULL)
return -1;
@@ -1884,8 +1881,8 @@ vy_update(struct vy_env *env, struct vy_tx *tx, struct txn_stmt *stmt,
if (vy_unique_key_validate(lsm, key, part_count))
return -1;
- if (vy_get_by_raw_key(lsm, tx, vy_tx_read_view(tx),
- key, part_count, &stmt->old_tuple) != 0)
+ if (vy_get_by_raw_key(lsm, tx, vy_tx_read_view(tx), key, part_count,
+ &stmt->old_tuple) != 0)
return -1;
/* Nothing to update. */
if (stmt->old_tuple == NULL)
@@ -1910,8 +1907,8 @@ vy_update(struct vy_env *env, struct vy_tx *tx, struct txn_stmt *stmt,
*/
if (tuple_validate_raw(pk->mem_format, new_tuple))
return -1;
- stmt->new_tuple = vy_stmt_new_replace(pk->mem_format, new_tuple,
- new_tuple_end);
+ stmt->new_tuple =
+ vy_stmt_new_replace(pk->mem_format, new_tuple, new_tuple_end);
if (stmt->new_tuple == NULL)
return -1;
if (vy_check_update(space, pk, stmt->old_tuple, stmt->new_tuple,
@@ -1966,9 +1963,8 @@ vy_insert_first_upsert(struct vy_env *env, struct vy_tx *tx,
* @retval -1 Memory error.
*/
static int
-vy_lsm_upsert(struct vy_tx *tx, struct vy_lsm *lsm,
- const char *tuple, const char *tuple_end,
- const char *expr, const char *expr_end)
+vy_lsm_upsert(struct vy_tx *tx, struct vy_lsm *lsm, const char *tuple,
+ const char *tuple_end, const char *expr, const char *expr_end)
{
assert(tx == NULL || tx->state == VINYL_TX_READY);
struct tuple *vystmt;
@@ -2006,7 +2002,7 @@ request_normalize_ops(struct request *request)
ops_end = mp_encode_array(ops_end, op_len);
uint32_t op_name_len;
- const char *op_name = mp_decode_str(&pos, &op_name_len);
+ const char *op_name = mp_decode_str(&pos, &op_name_len);
ops_end = mp_encode_str(ops_end, op_name, op_name_len);
int field_no;
@@ -2122,9 +2118,9 @@ vy_upsert(struct vy_env *env, struct vy_tx *tx, struct txn_stmt *stmt,
* to delete old tuples from secondary indexes.
*/
/* Find the old tuple using the primary key. */
- struct tuple *key = vy_stmt_extract_key_raw(tuple, tuple_end,
- pk->key_def, pk->env->key_format,
- MULTIKEY_NONE);
+ struct tuple *key =
+ vy_stmt_extract_key_raw(tuple, tuple_end, pk->key_def,
+ pk->env->key_format, MULTIKEY_NONE);
if (key == NULL)
return -1;
int rc = vy_get(pk, tx, vy_tx_read_view(tx), key, &stmt->old_tuple);
@@ -2136,8 +2132,8 @@ vy_upsert(struct vy_env *env, struct vy_tx *tx, struct txn_stmt *stmt,
* turns into INSERT.
*/
if (stmt->old_tuple == NULL) {
- stmt->new_tuple = vy_stmt_new_insert(pk->mem_format,
- tuple, tuple_end);
+ stmt->new_tuple =
+ vy_stmt_new_insert(pk->mem_format, tuple, tuple_end);
if (stmt->new_tuple == NULL)
return -1;
return vy_insert_first_upsert(env, tx, space, stmt->new_tuple);
@@ -2159,8 +2155,8 @@ vy_upsert(struct vy_env *env, struct vy_tx *tx, struct txn_stmt *stmt,
if (tuple_validate_raw(pk->mem_format, new_tuple))
return -1;
new_tuple_end = new_tuple + new_size;
- stmt->new_tuple = vy_stmt_new_replace(pk->mem_format, new_tuple,
- new_tuple_end);
+ stmt->new_tuple =
+ vy_stmt_new_replace(pk->mem_format, new_tuple, new_tuple_end);
if (stmt->new_tuple == NULL)
return -1;
if (vy_check_update(space, pk, stmt->old_tuple, stmt->new_tuple,
@@ -2263,8 +2259,8 @@ vy_replace(struct vy_env *env, struct vy_tx *tx, struct txn_stmt *stmt,
* need to pass the old tuple to trigger callbacks.
*/
if (!rlist_empty(&space->on_replace)) {
- if (vy_get(pk, tx, vy_tx_read_view(tx),
- stmt->new_tuple, &stmt->old_tuple) != 0)
+ if (vy_get(pk, tx, vy_tx_read_view(tx), stmt->new_tuple,
+ &stmt->old_tuple) != 0)
return -1;
if (stmt->old_tuple == NULL) {
/*
@@ -2365,7 +2361,7 @@ vinyl_space_execute_update(struct space *space, struct txn *txn,
static int
vinyl_space_execute_upsert(struct space *space, struct txn *txn,
- struct request *request)
+ struct request *request)
{
struct vy_env *env = vy_env(space->engine);
struct vy_tx *tx = txn->engine_tx;
@@ -2391,8 +2387,7 @@ vinyl_engine_prepare(struct engine *engine, struct txn *txn)
struct vy_tx *tx = txn->engine_tx;
assert(tx != NULL);
- if (tx->write_size > 0 &&
- vinyl_check_wal(env, "DML") != 0)
+ if (tx->write_size > 0 && vinyl_check_wal(env, "DML") != 0)
return -1;
/*
@@ -2401,16 +2396,16 @@ vinyl_engine_prepare(struct engine *engine, struct txn *txn)
* available for the admin to track the lag so let the applier
* wait as long as necessary for memory dump to complete.
*/
- double timeout = (tx->is_applier_session ?
- TIMEOUT_INFINITY : env->timeout);
+ double timeout =
+ (tx->is_applier_session ? TIMEOUT_INFINITY : env->timeout);
/*
* Reserve quota needed by the transaction before allocating
* memory. Since this may yield, which opens a time window for
* the transaction to be sent to read view or aborted, we call
* it before checking for conflicts.
*/
- if (vy_quota_use(&env->quota, VY_QUOTA_CONSUMER_TX,
- tx->write_size, timeout) != 0)
+ if (vy_quota_use(&env->quota, VY_QUOTA_CONSUMER_TX, tx->write_size,
+ timeout) != 0)
return -1;
size_t mem_used_before = lsregion_used(&env->mem_env.allocator);
@@ -2419,8 +2414,8 @@ vinyl_engine_prepare(struct engine *engine, struct txn *txn)
size_t mem_used_after = lsregion_used(&env->mem_env.allocator);
assert(mem_used_after >= mem_used_before);
- vy_quota_adjust(&env->quota, VY_QUOTA_CONSUMER_TX,
- tx->write_size, mem_used_after - mem_used_before);
+ vy_quota_adjust(&env->quota, VY_QUOTA_CONSUMER_TX, tx->write_size,
+ mem_used_after - mem_used_before);
vy_regulator_check_dump_watermark(&env->regulator);
return rc;
}
@@ -2522,8 +2517,8 @@ vy_env_trigger_dump_cb(struct vy_regulator *regulator)
}
static void
-vy_env_dump_complete_cb(struct vy_scheduler *scheduler,
- int64_t dump_generation, double dump_duration)
+vy_env_dump_complete_cb(struct vy_scheduler *scheduler, int64_t dump_generation,
+ double dump_duration)
{
struct vy_env *env = container_of(scheduler, struct vy_env, scheduler);
@@ -2557,8 +2552,8 @@ vy_squash_schedule(struct vy_lsm *lsm, struct vy_entry entry,
void /* struct vy_env */ *arg);
static struct vy_env *
-vy_env_new(const char *path, size_t memory,
- int read_threads, int write_threads, bool force_recovery)
+vy_env_new(const char *path, size_t memory, int read_threads, int write_threads,
+ bool force_recovery)
{
struct vy_env *e = malloc(sizeof(*e));
if (unlikely(e == NULL)) {
@@ -2571,8 +2566,7 @@ vy_env_new(const char *path, size_t memory,
e->force_recovery = force_recovery;
e->path = strdup(path);
if (e->path == NULL) {
- diag_set(OutOfMemory, strlen(path),
- "malloc", "env->path");
+ diag_set(OutOfMemory, strlen(path), "malloc", "env->path");
goto error_path;
}
@@ -2586,22 +2580,20 @@ vy_env_new(const char *path, size_t memory,
vy_stmt_env_create(&e->stmt_env);
vy_mem_env_create(&e->mem_env, memory);
vy_scheduler_create(&e->scheduler, write_threads,
- vy_env_dump_complete_cb,
- &e->run_env, &e->xm->read_views);
+ vy_env_dump_complete_cb, &e->run_env,
+ &e->xm->read_views);
- if (vy_lsm_env_create(&e->lsm_env, e->path,
- &e->scheduler.generation,
- e->stmt_env.key_format,
- vy_squash_schedule, e) != 0)
+ if (vy_lsm_env_create(&e->lsm_env, e->path, &e->scheduler.generation,
+ e->stmt_env.key_format, vy_squash_schedule,
+ e) != 0)
goto error_lsm_env;
vy_quota_create(&e->quota, memory, vy_env_quota_exceeded_cb);
- vy_regulator_create(&e->regulator, &e->quota,
- vy_env_trigger_dump_cb);
+ vy_regulator_create(&e->regulator, &e->quota, vy_env_trigger_dump_cb);
struct slab_cache *slab_cache = cord_slab_cache();
mempool_create(&e->iterator_pool, slab_cache,
- sizeof(struct vinyl_iterator));
+ sizeof(struct vinyl_iterator));
vy_cache_env_create(&e->cache_env, slab_cache);
vy_run_env_create(&e->run_env, read_threads);
vy_log_init(e->path);
@@ -2655,8 +2647,8 @@ vy_env_complete_recovery(struct vy_env *env)
}
struct engine *
-vinyl_engine_new(const char *dir, size_t memory,
- int read_threads, int write_threads, bool force_recovery)
+vinyl_engine_new(const char *dir, size_t memory, int read_threads,
+ int write_threads, bool force_recovery)
{
struct vy_env *env = vy_env_new(dir, memory, read_threads,
write_threads, force_recovery);
@@ -2734,7 +2726,7 @@ vinyl_engine_set_snap_io_rate_limit(struct engine *engine, double limit)
static int
vinyl_engine_begin_checkpoint(struct engine *engine, bool is_scheduled)
{
- (void) is_scheduled;
+ (void)is_scheduled;
struct vy_env *env = vy_env(engine);
assert(env->status == VINYL_ONLINE);
/*
@@ -2750,8 +2742,7 @@ vinyl_engine_begin_checkpoint(struct engine *engine, bool is_scheduled)
}
static int
-vinyl_engine_wait_checkpoint(struct engine *engine,
- const struct vclock *vclock)
+vinyl_engine_wait_checkpoint(struct engine *engine, const struct vclock *vclock)
{
struct vy_env *env = vy_env(engine);
assert(env->status == VINYL_ONLINE);
@@ -2941,8 +2932,8 @@ vy_join_add_space(struct space *space, void *arg)
return 0;
struct vy_join_entry *entry = malloc(sizeof(*entry));
if (entry == NULL) {
- diag_set(OutOfMemory, sizeof(*entry),
- "malloc", "struct vy_join_entry");
+ diag_set(OutOfMemory, sizeof(*entry), "malloc",
+ "struct vy_join_entry");
return -1;
}
entry->space_id = space_id(space);
@@ -2961,8 +2952,8 @@ vinyl_engine_prepare_join(struct engine *engine, void **arg)
(void)engine;
struct vy_join_ctx *ctx = malloc(sizeof(*ctx));
if (ctx == NULL) {
- diag_set(OutOfMemory, sizeof(*ctx),
- "malloc", "struct vy_join_ctx");
+ diag_set(OutOfMemory, sizeof(*ctx), "malloc",
+ "struct vy_join_ctx");
return -1;
}
rlist_create(&ctx->entries);
@@ -2975,8 +2966,8 @@ vinyl_engine_prepare_join(struct engine *engine, void **arg)
}
static int
-vy_join_send_tuple(struct xstream *stream, uint32_t space_id,
- const char *data, size_t size)
+vy_join_send_tuple(struct xstream *stream, uint32_t space_id, const char *data,
+ size_t size)
{
struct request_replace_body body;
request_replace_body_create(&body, space_id);
@@ -3007,8 +2998,8 @@ vinyl_engine_join(struct engine *engine, void *arg, struct xstream *stream)
uint32_t size;
const char *data;
while ((rc = it->next(it, &data, &size)) == 0 && data != NULL) {
- if (vy_join_send_tuple(stream, entry->space_id,
- data, size) != 0)
+ if (vy_join_send_tuple(stream, entry->space_id, data,
+ size) != 0)
return -1;
}
if (rc != 0)
@@ -3043,8 +3034,7 @@ vinyl_engine_complete_join(struct engine *engine, void *arg)
* next log rotation.
*/
static void
-vy_gc_run(struct vy_env *env,
- struct vy_lsm_recovery_info *lsm_info,
+vy_gc_run(struct vy_env *env, struct vy_lsm_recovery_info *lsm_info,
struct vy_run_recovery_info *run_info)
{
/* Try to delete files. */
@@ -3072,8 +3062,7 @@ vy_gc_run(struct vy_env *env,
static void
vy_gc_lsm(struct vy_lsm_recovery_info *lsm_info)
{
- assert(lsm_info->drop_lsn >= 0 ||
- lsm_info->create_lsn < 0);
+ assert(lsm_info->drop_lsn >= 0 || lsm_info->create_lsn < 0);
vy_log_tx_begin();
if (lsm_info->drop_lsn < 0) {
@@ -3097,8 +3086,7 @@ vy_gc_lsm(struct vy_lsm_recovery_info *lsm_info)
vy_log_drop_run(run_info->id, run_info->gc_lsn);
}
}
- if (rlist_empty(&lsm_info->ranges) &&
- rlist_empty(&lsm_info->runs))
+ if (rlist_empty(&lsm_info->ranges) && rlist_empty(&lsm_info->runs))
vy_log_forget_lsm(lsm_info->id);
vy_log_tx_try_commit();
}
@@ -3111,8 +3099,8 @@ vy_gc_lsm(struct vy_lsm_recovery_info *lsm_info)
* @param gc_lsn LSN of the oldest checkpoint to save.
*/
static void
-vy_gc(struct vy_env *env, struct vy_recovery *recovery,
- unsigned int gc_mask, int64_t gc_lsn)
+vy_gc(struct vy_env *env, struct vy_recovery *recovery, unsigned int gc_mask,
+ int64_t gc_lsn)
{
int loops = 0;
struct vy_lsm_recovery_info *lsm_info;
@@ -3246,8 +3234,8 @@ struct vy_squash_queue {
};
static struct vy_squash *
-vy_squash_new(struct mempool *pool, struct vy_env *env,
- struct vy_lsm *lsm, struct vy_entry entry)
+vy_squash_new(struct mempool *pool, struct vy_env *env, struct vy_lsm *lsm,
+ struct vy_entry entry)
{
struct vy_squash *squash;
squash = mempool_alloc(pool);
@@ -3319,7 +3307,8 @@ vy_squash_process(struct vy_squash *squash)
uint8_t n_upserts = 0;
while (!vy_mem_tree_iterator_is_invalid(&mem_itr)) {
struct vy_entry mem_entry;
- mem_entry = *vy_mem_tree_iterator_get_elem(&mem->tree, &mem_itr);
+ mem_entry =
+ *vy_mem_tree_iterator_get_elem(&mem->tree, &mem_itr);
if (vy_entry_compare(result, mem_entry, lsm->cmp_def) != 0 ||
vy_stmt_type(mem_entry.stmt) != IPROTO_UPSERT)
break;
@@ -3367,8 +3356,7 @@ vy_squash_queue_new(void)
sq->fiber = NULL;
fiber_cond_create(&sq->cond);
stailq_create(&sq->queue);
- mempool_create(&sq->pool, cord_slab_cache(),
- sizeof(struct vy_squash));
+ mempool_create(&sq->pool, cord_slab_cache(), sizeof(struct vy_squash));
return sq;
}
@@ -3414,8 +3402,8 @@ vy_squash_schedule(struct vy_lsm *lsm, struct vy_entry entry, void *arg)
struct vy_env *env = arg;
struct vy_squash_queue *sq = env->squash_queue;
- say_verbose("%s: schedule upsert optimization for %s",
- vy_lsm_name(lsm), vy_stmt_str(entry.stmt));
+ say_verbose("%s: schedule upsert optimization for %s", vy_lsm_name(lsm),
+ vy_stmt_str(entry.stmt));
/* Start the upsert squashing fiber on demand. */
if (sq->fiber == NULL) {
@@ -3443,8 +3431,8 @@ static int
vinyl_iterator_on_tx_destroy(struct trigger *trigger, void *event)
{
(void)event;
- struct vinyl_iterator *it = container_of(trigger,
- struct vinyl_iterator, on_tx_destroy);
+ struct vinyl_iterator *it =
+ container_of(trigger, struct vinyl_iterator, on_tx_destroy);
it->tx = NULL;
return 0;
}
@@ -3647,12 +3635,12 @@ vinyl_index_create_iterator(struct index *base, enum iterator_type type,
struct vinyl_iterator *it = mempool_alloc(&env->iterator_pool);
if (it == NULL) {
- diag_set(OutOfMemory, sizeof(struct vinyl_iterator),
- "mempool", "struct vinyl_iterator");
+ diag_set(OutOfMemory, sizeof(struct vinyl_iterator), "mempool",
+ "struct vinyl_iterator");
return NULL;
}
- it->key = vy_entry_key_new(lsm->env->key_format, lsm->cmp_def,
- key, part_count);
+ it->key = vy_entry_key_new(lsm->env->key_format, lsm->cmp_def, key,
+ part_count);
if (it->key.stmt == NULL) {
mempool_free(&env->iterator_pool, it);
return NULL;
@@ -3671,8 +3659,8 @@ vinyl_index_create_iterator(struct index *base, enum iterator_type type,
* Register a trigger that will abort this iterator
* when the transaction ends.
*/
- trigger_create(&it->on_tx_destroy,
- vinyl_iterator_on_tx_destroy, NULL, NULL);
+ trigger_create(&it->on_tx_destroy, vinyl_iterator_on_tx_destroy,
+ NULL, NULL);
trigger_add(&tx->on_destroy, &it->on_tx_destroy);
} else {
tx = &it->tx_autocommit;
@@ -3687,8 +3675,8 @@ vinyl_index_create_iterator(struct index *base, enum iterator_type type,
}
static int
-vinyl_snapshot_iterator_next(struct snapshot_iterator *base,
- const char **data, uint32_t *size)
+vinyl_snapshot_iterator_next(struct snapshot_iterator *base, const char **data,
+ uint32_t *size)
{
assert(base->next == vinyl_snapshot_iterator_next);
struct vinyl_snapshot_iterator *it =
@@ -3734,8 +3722,8 @@ vinyl_index_create_snapshot_iterator(struct index *base)
free(it);
return NULL;
}
- vy_read_iterator_open(&it->iterator, lsm, NULL,
- ITER_ALL, lsm->env->empty_key,
+ vy_read_iterator_open(&it->iterator, lsm, NULL, ITER_ALL,
+ lsm->env->empty_key,
(const struct vy_read_view **)&it->rv);
/*
* The index may be dropped while we are reading it.
@@ -3747,8 +3735,8 @@ vinyl_index_create_snapshot_iterator(struct index *base)
}
static int
-vinyl_index_get(struct index *index, const char *key,
- uint32_t part_count, struct tuple **ret)
+vinyl_index_get(struct index *index, const char *key, uint32_t part_count,
+ struct tuple **ret)
{
assert(index->def->opts.is_unique);
assert(index->def->key_def->part_count == part_count);
@@ -3756,8 +3744,9 @@ vinyl_index_get(struct index *index, const char *key,
struct vy_lsm *lsm = vy_lsm(index);
struct vy_env *env = vy_env(index->engine);
struct vy_tx *tx = in_txn() ? in_txn()->engine_tx : NULL;
- const struct vy_read_view **rv = (tx != NULL ? vy_tx_read_view(tx) :
- &env->xm->p_global_read_view);
+ const struct vy_read_view **rv = (tx != NULL ?
+ vy_tx_read_view(tx) :
+ &env->xm->p_global_read_view);
if (tx != NULL && tx->state == VINYL_TX_ABORT) {
diag_set(ClientError, ER_TRANSACTION_CONFLICT);
@@ -3832,14 +3821,14 @@ vy_build_on_replace(struct trigger *trigger, void *event)
/* Check key uniqueness if necessary. */
if (ctx->check_unique_constraint && stmt->new_tuple != NULL &&
vy_check_is_unique_secondary(tx, vy_tx_read_view(tx),
- ctx->space_name, ctx->index_name,
- lsm, stmt->new_tuple) != 0)
+ ctx->space_name, ctx->index_name, lsm,
+ stmt->new_tuple) != 0)
goto err;
/* Forward the statement to the new LSM tree. */
if (stmt->old_tuple != NULL) {
- struct tuple *delete = vy_stmt_new_surrogate_delete(format,
- stmt->old_tuple);
+ struct tuple *delete =
+ vy_stmt_new_surrogate_delete(format, stmt->old_tuple);
if (delete == NULL)
goto err;
int rc = vy_tx_set(tx, lsm, delete);
@@ -3850,8 +3839,8 @@ vy_build_on_replace(struct trigger *trigger, void *event)
if (stmt->new_tuple != NULL) {
uint32_t data_len;
const char *data = tuple_data_range(stmt->new_tuple, &data_len);
- struct tuple *insert = vy_stmt_new_insert(format, data,
- data + data_len);
+ struct tuple *insert =
+ vy_stmt_new_insert(format, data, data + data_len);
if (insert == NULL)
goto err;
int rc = vy_tx_set(tx, lsm, insert);
@@ -3880,11 +3869,11 @@ err:
* being built.
*/
static int
-vy_build_insert_stmt(struct vy_lsm *lsm, struct vy_mem *mem,
- struct tuple *stmt, int64_t lsn)
+vy_build_insert_stmt(struct vy_lsm *lsm, struct vy_mem *mem, struct tuple *stmt,
+ int64_t lsn)
{
- struct tuple *region_stmt = vy_stmt_dup_lsregion(stmt,
- &mem->env->allocator, mem->generation);
+ struct tuple *region_stmt = vy_stmt_dup_lsregion(
+ stmt, &mem->env->allocator, mem->generation);
if (region_stmt == NULL)
return -1;
vy_stmt_set_lsn(region_stmt, lsn);
@@ -3920,8 +3909,8 @@ vy_build_insert_tuple(struct vy_env *env, struct vy_lsm *lsm,
/* Reallocate the new tuple using the new space format. */
uint32_t data_len;
const char *data = tuple_data_range(tuple, &data_len);
- struct tuple *stmt = vy_stmt_new_replace(new_format, data,
- data + data_len);
+ struct tuple *stmt =
+ vy_stmt_new_replace(new_format, data, data + data_len);
if (stmt == NULL)
return -1;
@@ -3946,9 +3935,9 @@ vy_build_insert_tuple(struct vy_env *env, struct vy_lsm *lsm,
*/
if (check_unique_constraint) {
vy_mem_pin(mem);
- rc = vy_check_is_unique_secondary(NULL,
- &env->xm->p_committed_read_view,
- space_name, index_name, lsm, stmt);
+ rc = vy_check_is_unique_secondary(
+ NULL, &env->xm->p_committed_read_view, space_name,
+ index_name, lsm, stmt);
vy_mem_unpin(mem);
if (rc != 0) {
tuple_unref(stmt);
@@ -4023,19 +4012,19 @@ vy_build_recover_stmt(struct vy_lsm *lsm, struct vy_lsm *pk,
if (type == IPROTO_REPLACE || type == IPROTO_INSERT) {
uint32_t data_len;
const char *data = tuple_data_range(mem_stmt, &data_len);
- insert = vy_stmt_new_insert(lsm->mem_format,
- data, data + data_len);
+ insert = vy_stmt_new_insert(lsm->mem_format, data,
+ data + data_len);
if (insert == NULL)
goto err;
} else if (type == IPROTO_UPSERT) {
- struct tuple *new_tuple = vy_apply_upsert(mem_stmt, old_tuple,
- pk->cmp_def, true);
+ struct tuple *new_tuple =
+ vy_apply_upsert(mem_stmt, old_tuple, pk->cmp_def, true);
if (new_tuple == NULL)
goto err;
uint32_t data_len;
const char *data = tuple_data_range(new_tuple, &data_len);
- insert = vy_stmt_new_insert(lsm->mem_format,
- data, data + data_len);
+ insert = vy_stmt_new_insert(lsm->mem_format, data,
+ data + data_len);
tuple_unref(new_tuple);
if (insert == NULL)
goto err;
@@ -4105,7 +4094,8 @@ vy_build_recover(struct vy_env *env, struct vy_lsm *lsm, struct vy_lsm *pk)
size_t mem_used_before, mem_used_after;
mem_used_before = lsregion_used(&env->mem_env.allocator);
- rlist_foreach_entry_reverse(mem, &pk->sealed, in_sealed) {
+ rlist_foreach_entry_reverse(mem, &pk->sealed, in_sealed)
+ {
rc = vy_build_recover_mem(lsm, pk, mem);
if (rc != 0)
break;
@@ -4236,12 +4226,10 @@ vinyl_space_build_index(struct space *src_space, struct index *new_index,
* in which case we would insert an outdated tuple.
*/
if (vy_stmt_lsn(tuple) <= build_lsn) {
- rc = vy_build_insert_tuple(env, new_lsm,
- space_name(src_space),
- new_index->def->name,
- new_format,
- check_unique_constraint,
- tuple);
+ rc = vy_build_insert_tuple(
+ env, new_lsm, space_name(src_space),
+ new_index->def->name, new_format,
+ check_unique_constraint, tuple);
if (rc != 0)
break;
}
@@ -4455,18 +4443,16 @@ vy_deferred_delete_on_replace(struct trigger *trigger, void *event)
* the LSM tree.
*/
size_t size;
- struct trigger *on_commit =
- region_alloc_object(&txn->region, typeof(*on_commit),
- &size);
+ struct trigger *on_commit = region_alloc_object(
+ &txn->region, typeof(*on_commit), &size);
if (on_commit == NULL) {
diag_set(OutOfMemory, size, "region_alloc_object",
"on_commit");
rc = -1;
break;
}
- struct trigger *on_rollback =
- region_alloc_object(&txn->region, typeof(*on_rollback),
- &size);
+ struct trigger *on_rollback = region_alloc_object(
+ &txn->region, typeof(*on_rollback), &size);
if (on_rollback == NULL) {
diag_set(OutOfMemory, size, "region_alloc_object",
"on_rollback");
@@ -4474,8 +4460,10 @@ vy_deferred_delete_on_replace(struct trigger *trigger, void *event)
break;
}
vy_mem_pin(mem);
- trigger_create(on_commit, vy_deferred_delete_on_commit, mem, NULL);
- trigger_create(on_rollback, vy_deferred_delete_on_rollback, mem, NULL);
+ trigger_create(on_commit, vy_deferred_delete_on_commit, mem,
+ NULL);
+ trigger_create(on_rollback, vy_deferred_delete_on_rollback, mem,
+ NULL);
txn_on_commit(txn, on_commit);
txn_on_rollback(txn, on_rollback);
}
@@ -4557,7 +4545,7 @@ static const struct index_vtab vinyl_index_vtab = {
/* .update_def = */ vinyl_index_update_def,
/* .depends_on_pk = */ vinyl_index_depends_on_pk,
/* .def_change_requires_rebuild = */
- vinyl_index_def_change_requires_rebuild,
+ vinyl_index_def_change_requires_rebuild,
/* .size = */ vinyl_index_size,
/* .bsize = */ vinyl_index_bsize,
/* .min = */ generic_index_min,
@@ -4568,7 +4556,7 @@ static const struct index_vtab vinyl_index_vtab = {
/* .replace = */ generic_index_replace,
/* .create_iterator = */ vinyl_index_create_iterator,
/* .create_snapshot_iterator = */
- vinyl_index_create_snapshot_iterator,
+ vinyl_index_create_snapshot_iterator,
/* .stat = */ vinyl_index_stat,
/* .compact = */ vinyl_index_compact,
/* .reset_stat = */ vinyl_index_reset_stat,
diff --git a/src/box/vinyl.h b/src/box/vinyl.h
index 2a3e8f1..715c49b 100644
--- a/src/box/vinyl.h
+++ b/src/box/vinyl.h
@@ -42,8 +42,8 @@ struct info_handler;
struct engine;
struct engine *
-vinyl_engine_new(const char *dir, size_t memory,
- int read_threads, int write_threads, bool force_recovery);
+vinyl_engine_new(const char *dir, size_t memory, int read_threads,
+ int write_threads, bool force_recovery);
/**
* Vinyl engine statistics (box.stat.vinyl()).
@@ -94,12 +94,12 @@ vinyl_engine_set_snap_io_rate_limit(struct engine *engine, double limit);
#include "diag.h"
static inline struct engine *
-vinyl_engine_new_xc(const char *dir, size_t memory,
- int read_threads, int write_threads, bool force_recovery)
+vinyl_engine_new_xc(const char *dir, size_t memory, int read_threads,
+ int write_threads, bool force_recovery)
{
struct engine *vinyl;
- vinyl = vinyl_engine_new(dir, memory, read_threads,
- write_threads, force_recovery);
+ vinyl = vinyl_engine_new(dir, memory, read_threads, write_threads,
+ force_recovery);
if (vinyl == NULL)
diag_raise();
return vinyl;
diff --git a/src/box/vy_cache.c b/src/box/vy_cache.c
index 7007d0e..e0bcafd 100644
--- a/src/box/vy_cache.c
+++ b/src/box/vy_cache.c
@@ -35,7 +35,7 @@
#include "vy_history.h"
#ifndef CT_ASSERT_G
-#define CT_ASSERT_G(e) typedef char CONCAT(__ct_assert_, __LINE__)[(e) ? 1 :-1]
+#define CT_ASSERT_G(e) typedef char CONCAT(__ct_assert_, __LINE__)[(e) ? 1 : -1]
#endif
CT_ASSERT_G(BOX_INDEX_PART_MAX <= UINT8_MAX);
@@ -143,8 +143,8 @@ vy_cache_create(struct vy_cache *cache, struct vy_cache_env *env,
cache->is_primary = is_primary;
cache->version = 1;
vy_cache_tree_create(&cache->cache_tree, cmp_def,
- vy_cache_tree_page_alloc,
- vy_cache_tree_page_free, env);
+ vy_cache_tree_page_alloc, vy_cache_tree_page_free,
+ env);
}
void
@@ -153,9 +153,8 @@ vy_cache_destroy(struct vy_cache *cache)
struct vy_cache_tree_iterator itr =
vy_cache_tree_iterator_first(&cache->cache_tree);
while (!vy_cache_tree_iterator_is_invalid(&itr)) {
- struct vy_cache_node **node =
- vy_cache_tree_iterator_get_elem(&cache->cache_tree,
- &itr);
+ struct vy_cache_node **node = vy_cache_tree_iterator_get_elem(
+ &cache->cache_tree, &itr);
assert(node != NULL && *node != NULL);
vy_cache_node_delete(cache->env, *node);
vy_cache_tree_iterator_next(&cache->cache_tree, &itr);
@@ -186,8 +185,7 @@ vy_cache_gc_step(struct vy_cache_env *env)
}
if (node->flags & VY_CACHE_RIGHT_LINKED) {
struct vy_cache_tree_iterator next = itr;
- vy_cache_tree_iterator_next(&cache->cache_tree,
- &next);
+ vy_cache_tree_iterator_next(&cache->cache_tree, &next);
struct vy_cache_node **next_node =
vy_cache_tree_iterator_get_elem(tree, &next);
assert((*next_node)->flags & VY_CACHE_LEFT_LINKED);
@@ -225,9 +223,8 @@ vy_cache_env_set_quota(struct vy_cache_env *env, size_t quota)
}
void
-vy_cache_add(struct vy_cache *cache, struct vy_entry curr,
- struct vy_entry prev, struct vy_entry key,
- enum iterator_type order)
+vy_cache_add(struct vy_cache *cache, struct vy_entry curr, struct vy_entry prev,
+ struct vy_entry key, enum iterator_type order)
{
if (cache->env->mem_quota == 0) {
/* Cache is disabled. */
@@ -269,14 +266,14 @@ vy_cache_add(struct vy_cache *cache, struct vy_entry curr,
* sequence of statements that is equal to the key.
*/
boundary_level = vy_stmt_key_part_count(key.stmt,
- cache->cmp_def);
+ cache->cmp_def);
}
} else {
assert(prev.stmt != NULL);
if (order == ITER_EQ || order == ITER_REQ) {
/* that is the last statement that is equal to key */
boundary_level = vy_stmt_key_part_count(key.stmt,
- cache->cmp_def);
+ cache->cmp_def);
} else {
/* that is the last statement */
boundary_level = 0;
@@ -295,14 +292,12 @@ vy_cache_add(struct vy_cache *cache, struct vy_entry curr,
assert(vy_stmt_type(curr.stmt) == IPROTO_INSERT ||
vy_stmt_type(curr.stmt) == IPROTO_REPLACE);
- assert(prev.stmt == NULL ||
- vy_stmt_type(prev.stmt) == IPROTO_INSERT ||
+ assert(prev.stmt == NULL || vy_stmt_type(prev.stmt) == IPROTO_INSERT ||
vy_stmt_type(prev.stmt) == IPROTO_REPLACE);
cache->version++;
/* Insert/replace new node to the tree */
- struct vy_cache_node *node =
- vy_cache_node_new(cache->env, cache, curr);
+ struct vy_cache_node *node = vy_cache_node_new(cache->env, cache, curr);
if (node == NULL) {
/* memory error, let's live without a cache */
return;
@@ -335,7 +330,7 @@ vy_cache_add(struct vy_cache *cache, struct vy_entry curr,
/* The flag that must be set in the inserted chain node */
uint32_t flag = direction > 0 ? VY_CACHE_LEFT_LINKED :
- VY_CACHE_RIGHT_LINKED;
+ VY_CACHE_RIGHT_LINKED;
#ifndef NDEBUG
/**
@@ -405,15 +400,16 @@ vy_cache_add(struct vy_cache *cache, struct vy_entry curr,
if (replaced != NULL) {
prev_node->flags = replaced->flags;
prev_node->left_boundary_level = replaced->left_boundary_level;
- prev_node->right_boundary_level = replaced->right_boundary_level;
+ prev_node->right_boundary_level =
+ replaced->right_boundary_level;
vy_cache_node_delete(cache->env, replaced);
}
/* Set proper flags */
node->flags |= flag;
/* Set inverted flag in the previous node */
- prev_node->flags |= (VY_CACHE_LEFT_LINKED |
- VY_CACHE_RIGHT_LINKED) ^ flag;
+ prev_node->flags |= (VY_CACHE_LEFT_LINKED | VY_CACHE_RIGHT_LINKED) ^
+ flag;
}
struct vy_entry
@@ -526,8 +522,8 @@ static inline bool
vy_cache_iterator_is_stop(struct vy_cache_iterator *itr,
struct vy_cache_node *node)
{
- uint8_t key_level = vy_stmt_key_part_count(itr->key.stmt,
- itr->cache->cmp_def);
+ uint8_t key_level =
+ vy_stmt_key_part_count(itr->key.stmt, itr->cache->cmp_def);
/* select{} is actually an EQ iterator with part_count == 0 */
bool iter_is_eq = itr->iterator_type == ITER_EQ || key_level == 0;
if (iterator_direction(itr->iterator_type) > 0) {
@@ -556,8 +552,8 @@ static inline bool
vy_cache_iterator_is_end_stop(struct vy_cache_iterator *itr,
struct vy_cache_node *last_node)
{
- uint8_t key_level = vy_stmt_key_part_count(itr->key.stmt,
- itr->cache->cmp_def);
+ uint8_t key_level =
+ vy_stmt_key_part_count(itr->key.stmt, itr->cache->cmp_def);
/* select{} is actually an EQ iterator with part_count == 0 */
bool iter_is_eq = itr->iterator_type == ITER_EQ || key_level == 0;
if (iterator_direction(itr->iterator_type) > 0) {
@@ -658,16 +654,17 @@ vy_cache_iterator_seek(struct vy_cache_iterator *itr, struct vy_entry last)
if (last.stmt != NULL) {
key = last;
iterator_type = iterator_direction(itr->iterator_type) > 0 ?
- ITER_GT : ITER_LT;
+ ITER_GT :
+ ITER_LT;
}
bool exact = false;
if (!vy_stmt_is_empty_key(key.stmt)) {
- itr->curr_pos = iterator_type == ITER_EQ ||
- iterator_type == ITER_GE ||
- iterator_type == ITER_LT ?
- vy_cache_tree_lower_bound(tree, key, &exact) :
- vy_cache_tree_upper_bound(tree, key, &exact);
+ itr->curr_pos =
+ iterator_type == ITER_EQ || iterator_type == ITER_GE ||
+ iterator_type == ITER_LT ?
+ vy_cache_tree_lower_bound(tree, key, &exact) :
+ vy_cache_tree_upper_bound(tree, key, &exact);
} else if (iterator_type == ITER_LE) {
itr->curr_pos = vy_cache_tree_invalid_iterator();
} else {
@@ -734,7 +731,9 @@ vy_cache_iterator_skip(struct vy_cache_iterator *itr, struct vy_entry last,
if (itr->search_started &&
(itr->curr.stmt == NULL || last.stmt == NULL ||
iterator_direction(itr->iterator_type) *
- vy_entry_compare(itr->curr, last, itr->cache->cmp_def) > 0))
+ vy_entry_compare(itr->curr, last,
+ itr->cache->cmp_def) >
+ 0))
return 0;
vy_history_cleanup(history);
@@ -806,7 +805,7 @@ vy_cache_iterator_restore(struct vy_cache_iterator *itr, struct vy_entry last,
if (cmp < 0 || (cmp == 0 && !key_belongs))
break;
if (vy_stmt_lsn(node->entry.stmt) <=
- (**itr->read_view).vlsn) {
+ (**itr->read_view).vlsn) {
itr->curr_pos = pos;
if (itr->curr.stmt != NULL)
tuple_unref(itr->curr.stmt);
diff --git a/src/box/vy_cache.h b/src/box/vy_cache.h
index 9f08391..76547ab 100644
--- a/src/box/vy_cache.h
+++ b/src/box/vy_cache.h
@@ -71,8 +71,8 @@ struct vy_cache_node {
* Internal comparator (1) for BPS tree.
*/
static inline int
-vy_cache_tree_cmp(struct vy_cache_node *a,
- struct vy_cache_node *b, struct key_def *cmp_def)
+vy_cache_tree_cmp(struct vy_cache_node *a, struct vy_cache_node *b,
+ struct key_def *cmp_def)
{
return vy_entry_compare(a->entry, b->entry, cmp_def);
}
@@ -201,9 +201,8 @@ vy_cache_destroy(struct vy_cache *cache);
* +1 - forward, -1 - backward.
*/
void
-vy_cache_add(struct vy_cache *cache, struct vy_entry curr,
- struct vy_entry prev, struct vy_entry key,
- enum iterator_type order);
+vy_cache_add(struct vy_cache *cache, struct vy_entry curr, struct vy_entry prev,
+ struct vy_entry key, enum iterator_type order);
/**
* Find value in cache.
@@ -223,7 +222,6 @@ void
vy_cache_on_write(struct vy_cache *cache, struct vy_entry entry,
struct vy_entry *deleted);
-
/**
* Cache iterator
*/
diff --git a/src/box/vy_history.c b/src/box/vy_history.c
index c3717b2..235af51 100644
--- a/src/box/vy_history.c
+++ b/src/box/vy_history.c
@@ -82,8 +82,8 @@ vy_history_apply(struct vy_history *history, struct key_def *cmp_def,
return 0;
struct vy_entry curr = vy_entry_none();
- struct vy_history_node *node = rlist_last_entry(&history->stmts,
- struct vy_history_node, link);
+ struct vy_history_node *node =
+ rlist_last_entry(&history->stmts, struct vy_history_node, link);
if (vy_history_is_terminal(history)) {
if (!keep_delete &&
vy_stmt_type(node->entry.stmt) == IPROTO_DELETE) {
@@ -103,8 +103,8 @@ vy_history_apply(struct vy_history *history, struct key_def *cmp_def,
node = rlist_prev_entry_safe(node, &history->stmts, link);
}
while (node != NULL) {
- struct vy_entry entry = vy_entry_apply_upsert(node->entry, curr,
- cmp_def, true);
+ struct vy_entry entry =
+ vy_entry_apply_upsert(node->entry, curr, cmp_def, true);
++*upserts_applied;
if (curr.stmt != NULL)
tuple_unref(curr.stmt);
diff --git a/src/box/vy_history.h b/src/box/vy_history.h
index b25c27f..a7f33e9 100644
--- a/src/box/vy_history.h
+++ b/src/box/vy_history.h
@@ -97,8 +97,8 @@ vy_history_is_terminal(struct vy_history *history)
{
if (rlist_empty(&history->stmts))
return false;
- struct vy_history_node *node = rlist_last_entry(&history->stmts,
- struct vy_history_node, link);
+ struct vy_history_node *node =
+ rlist_last_entry(&history->stmts, struct vy_history_node, link);
assert(vy_stmt_type(node->entry.stmt) == IPROTO_REPLACE ||
vy_stmt_type(node->entry.stmt) == IPROTO_DELETE ||
vy_stmt_type(node->entry.stmt) == IPROTO_INSERT ||
@@ -116,8 +116,8 @@ vy_history_last_stmt(struct vy_history *history)
if (rlist_empty(&history->stmts))
return vy_entry_none();
/* Newest statement is at the head of the list. */
- struct vy_history_node *node = rlist_first_entry(&history->stmts,
- struct vy_history_node, link);
+ struct vy_history_node *node = rlist_first_entry(
+ &history->stmts, struct vy_history_node, link);
return node->entry;
}
diff --git a/src/box/vy_log.c b/src/box/vy_log.c
index 06b2596..c7da9c0 100644
--- a/src/box/vy_log.c
+++ b/src/box/vy_log.c
@@ -250,70 +250,64 @@ vy_log_record_snprint(char *buf, int size, const struct vy_log_record *record)
SNPRINT(total, snprintf, buf, size, "%s{",
vy_log_type_name[record->type]);
if (record->lsm_id > 0)
- SNPRINT(total, snprintf, buf, size, "%s=%"PRIi64", ",
+ SNPRINT(total, snprintf, buf, size, "%s=%" PRIi64 ", ",
vy_log_key_name[VY_LOG_KEY_LSM_ID], record->lsm_id);
if (record->range_id > 0)
- SNPRINT(total, snprintf, buf, size, "%s=%"PRIi64", ",
- vy_log_key_name[VY_LOG_KEY_RANGE_ID],
- record->range_id);
+ SNPRINT(total, snprintf, buf, size, "%s=%" PRIi64 ", ",
+ vy_log_key_name[VY_LOG_KEY_RANGE_ID], record->range_id);
if (record->run_id > 0)
- SNPRINT(total, snprintf, buf, size, "%s=%"PRIi64", ",
- vy_log_key_name[VY_LOG_KEY_RUN_ID],
- record->run_id);
+ SNPRINT(total, snprintf, buf, size, "%s=%" PRIi64 ", ",
+ vy_log_key_name[VY_LOG_KEY_RUN_ID], record->run_id);
if (record->begin != NULL) {
- SNPRINT(total, snprintf, buf, size, "%s=",
- vy_log_key_name[VY_LOG_KEY_BEGIN]);
+ SNPRINT(total, snprintf, buf, size,
+ "%s=", vy_log_key_name[VY_LOG_KEY_BEGIN]);
SNPRINT(total, mp_snprint, buf, size, record->begin);
SNPRINT(total, snprintf, buf, size, ", ");
}
if (record->end != NULL) {
- SNPRINT(total, snprintf, buf, size, "%s=",
- vy_log_key_name[VY_LOG_KEY_END]);
+ SNPRINT(total, snprintf, buf, size,
+ "%s=", vy_log_key_name[VY_LOG_KEY_END]);
SNPRINT(total, mp_snprint, buf, size, record->end);
SNPRINT(total, snprintf, buf, size, ", ");
}
if (record->index_id > 0)
- SNPRINT(total, snprintf, buf, size, "%s=%"PRIu32", ",
+ SNPRINT(total, snprintf, buf, size, "%s=%" PRIu32 ", ",
vy_log_key_name[VY_LOG_KEY_INDEX_ID], record->index_id);
if (record->space_id > 0)
- SNPRINT(total, snprintf, buf, size, "%s=%"PRIu32", ",
+ SNPRINT(total, snprintf, buf, size, "%s=%" PRIu32 ", ",
vy_log_key_name[VY_LOG_KEY_SPACE_ID], record->space_id);
if (record->group_id > 0)
- SNPRINT(total, snprintf, buf, size, "%s=%"PRIu32", ",
+ SNPRINT(total, snprintf, buf, size, "%s=%" PRIu32 ", ",
vy_log_key_name[VY_LOG_KEY_GROUP_ID], record->group_id);
if (record->key_parts != NULL) {
- SNPRINT(total, snprintf, buf, size, "%s=",
- vy_log_key_name[VY_LOG_KEY_DEF]);
+ SNPRINT(total, snprintf, buf, size,
+ "%s=", vy_log_key_name[VY_LOG_KEY_DEF]);
SNPRINT(total, key_def_snprint_parts, buf, size,
record->key_parts, record->key_part_count);
SNPRINT(total, snprintf, buf, size, ", ");
}
if (record->slice_id > 0)
- SNPRINT(total, snprintf, buf, size, "%s=%"PRIi64", ",
- vy_log_key_name[VY_LOG_KEY_SLICE_ID],
- record->slice_id);
+ SNPRINT(total, snprintf, buf, size, "%s=%" PRIi64 ", ",
+ vy_log_key_name[VY_LOG_KEY_SLICE_ID], record->slice_id);
if (record->create_lsn > 0)
- SNPRINT(total, snprintf, buf, size, "%s=%"PRIi64", ",
+ SNPRINT(total, snprintf, buf, size, "%s=%" PRIi64 ", ",
vy_log_key_name[VY_LOG_KEY_CREATE_LSN],
record->create_lsn);
if (record->modify_lsn > 0)
- SNPRINT(total, snprintf, buf, size, "%s=%"PRIi64", ",
+ SNPRINT(total, snprintf, buf, size, "%s=%" PRIi64 ", ",
vy_log_key_name[VY_LOG_KEY_MODIFY_LSN],
record->modify_lsn);
if (record->drop_lsn > 0)
- SNPRINT(total, snprintf, buf, size, "%s=%"PRIi64", ",
- vy_log_key_name[VY_LOG_KEY_DROP_LSN],
- record->drop_lsn);
+ SNPRINT(total, snprintf, buf, size, "%s=%" PRIi64 ", ",
+ vy_log_key_name[VY_LOG_KEY_DROP_LSN], record->drop_lsn);
if (record->dump_lsn > 0)
- SNPRINT(total, snprintf, buf, size, "%s=%"PRIi64", ",
- vy_log_key_name[VY_LOG_KEY_DUMP_LSN],
- record->dump_lsn);
+ SNPRINT(total, snprintf, buf, size, "%s=%" PRIi64 ", ",
+ vy_log_key_name[VY_LOG_KEY_DUMP_LSN], record->dump_lsn);
if (record->gc_lsn > 0)
- SNPRINT(total, snprintf, buf, size, "%s=%"PRIi64", ",
- vy_log_key_name[VY_LOG_KEY_GC_LSN],
- record->gc_lsn);
+ SNPRINT(total, snprintf, buf, size, "%s=%" PRIi64 ", ",
+ vy_log_key_name[VY_LOG_KEY_GC_LSN], record->gc_lsn);
if (record->dump_count > 0)
- SNPRINT(total, snprintf, buf, size, "%s=%"PRIu32", ",
+ SNPRINT(total, snprintf, buf, size, "%s=%" PRIu32 ", ",
vy_log_key_name[VY_LOG_KEY_DUMP_COUNT],
record->dump_count);
SNPRINT(total, snprintf, buf, size, "}");
@@ -554,8 +548,7 @@ vy_log_record_encode(const struct vy_log_record *record,
* Return 0 on success, -1 on failure.
*/
static int
-vy_log_record_decode(struct vy_log_record *record,
- struct xrow_header *row)
+vy_log_record_decode(struct vy_log_record *record, struct xrow_header *row)
{
char *buf;
@@ -624,16 +617,15 @@ vy_log_record_decode(struct vy_log_record *record,
struct region *region = &fiber()->gc;
uint32_t part_count = mp_decode_array(&pos);
size_t size;
- struct key_part_def *parts =
- region_alloc_array(region, typeof(parts[0]),
- part_count, &size);
+ struct key_part_def *parts = region_alloc_array(
+ region, typeof(parts[0]), part_count, &size);
if (parts == NULL) {
diag_set(OutOfMemory, size,
"region_alloc_array", "parts");
return -1;
}
- if (key_def_decode_parts(parts, part_count, &pos,
- NULL, 0, region) != 0) {
+ if (key_def_decode_parts(parts, part_count, &pos, NULL,
+ 0, region) != 0) {
diag_log();
diag_set(ClientError, ER_INVALID_VYLOG_FILE,
"Bad record: failed to decode "
@@ -705,8 +697,8 @@ vy_log_record_dup(struct region *pool, const struct vy_log_record *src)
{
size_t used = region_used(pool);
size_t size;
- struct vy_log_record *dst = region_alloc_object(pool, typeof(*dst),
- &size);
+ struct vy_log_record *dst =
+ region_alloc_object(pool, typeof(*dst), &size);
if (dst == NULL) {
diag_set(OutOfMemory, size, "region_alloc_object", "dst");
goto err;
@@ -769,8 +761,7 @@ vy_log_init(const char *dir)
diag_create(&vy_log.tx_diag);
wal_init_vy_log();
fiber_cond_create(&vy_log.flusher_cond);
- vy_log.flusher = fiber_new("vinyl.vylog_flusher",
- vy_log_flusher_f);
+ vy_log.flusher = fiber_new("vinyl.vylog_flusher", vy_log_flusher_f);
if (vy_log.flusher == NULL)
panic("failed to allocate vylog flusher fiber");
fiber_wakeup(vy_log.flusher);
@@ -814,8 +805,7 @@ vy_log_tx_flush(struct vy_log_tx *tx)
int tx_size = 0;
struct vy_log_record *record;
- stailq_foreach_entry(record, &tx->records, in_tx)
- tx_size++;
+ stailq_foreach_entry(record, &tx->records, in_tx) tx_size++;
size_t used = region_used(&fiber()->gc);
@@ -835,7 +825,8 @@ vy_log_tx_flush(struct vy_log_tx *tx)
* Encode buffered records.
*/
int i = 0;
- stailq_foreach_entry(record, &tx->records, in_tx) {
+ stailq_foreach_entry(record, &tx->records, in_tx)
+ {
if (record->gc_lsn == VY_LOG_GC_LSN_CURRENT)
record->gc_lsn = vy_log_signature();
assert(i < tx_size);
@@ -880,8 +871,8 @@ vy_log_flush(void)
int rc = 0;
while (!stailq_empty(&pending)) {
- struct vy_log_tx *tx = stailq_first_entry(&pending,
- struct vy_log_tx, in_pending);
+ struct vy_log_tx *tx = stailq_first_entry(
+ &pending, struct vy_log_tx, in_pending);
rc = vy_log_tx_flush(tx);
if (rc != 0)
break;
@@ -952,8 +943,7 @@ vy_log_open(struct xlog *xlog)
goto fail;
}
- if (xdir_create_xlog(&vy_log.dir, xlog,
- &vy_log.last_checkpoint) < 0)
+ if (xdir_create_xlog(&vy_log.dir, xlog, &vy_log.last_checkpoint) < 0)
goto fail;
struct xrow_header row;
@@ -1113,7 +1103,8 @@ vy_log_end_recovery(void)
* recovery - we will need them for garbage collection.
*/
struct vy_log_tx *tx;
- stailq_foreach_entry(tx, &vy_log.pending_tx, in_pending) {
+ stailq_foreach_entry(tx, &vy_log.pending_tx, in_pending)
+ {
struct vy_log_record *record;
stailq_foreach_entry(record, &tx->records, in_tx)
vy_recovery_process_record(vy_log.recovery, record);
@@ -1156,8 +1147,8 @@ vy_log_rotate(const struct vclock *vclock)
return 0;
assert(signature > prev_signature);
- say_verbose("rotating vylog %lld => %lld",
- (long long)prev_signature, (long long)signature);
+ say_verbose("rotating vylog %lld => %lld", (long long)prev_signature,
+ (long long)signature);
/*
* Lock out all concurrent log writers while we are rotating it.
@@ -1321,8 +1312,8 @@ vy_log_write(const struct vy_log_record *record)
return;
assert(vy_log.tx != NULL);
- struct vy_log_record *tx_record = vy_log_record_dup(&vy_log.tx->region,
- record);
+ struct vy_log_record *tx_record =
+ vy_log_record_dup(&vy_log.tx->region, record);
if (tx_record == NULL) {
diag_move(diag_get(), &vy_log.tx_diag);
vy_log.tx_failed = true;
@@ -1343,8 +1334,8 @@ vy_recovery_index_id_hash(uint32_t space_id, uint32_t index_id)
/** Lookup an LSM tree in vy_recovery::index_id_hash map. */
struct vy_lsm_recovery_info *
-vy_recovery_lsm_by_index_id(struct vy_recovery *recovery,
- uint32_t space_id, uint32_t index_id)
+vy_recovery_lsm_by_index_id(struct vy_recovery *recovery, uint32_t space_id,
+ uint32_t index_id)
{
int64_t key = vy_recovery_index_id_hash(space_id, index_id);
struct mh_i64ptr_t *h = recovery->index_id_hash;
@@ -1413,7 +1404,8 @@ vy_recovery_alloc_key_parts(const struct key_part_def *key_parts,
uint32_t new_parts_sz = sizeof(*key_parts) * key_part_count;
for (uint32_t i = 0; i < key_part_count; i++) {
new_parts_sz += key_parts[i].path != NULL ?
- strlen(key_parts[i].path) + 1 : 0;
+ strlen(key_parts[i].path) + 1 :
+ 0;
}
struct key_part_def *new_parts = malloc(new_parts_sz);
if (new_parts == NULL) {
@@ -1459,8 +1451,8 @@ vy_recovery_do_create_lsm(struct vy_recovery *recovery, int64_t id,
}
struct vy_lsm_recovery_info *lsm = malloc(sizeof(*lsm));
if (lsm == NULL) {
- diag_set(OutOfMemory, sizeof(*lsm),
- "malloc", "struct vy_lsm_recovery_info");
+ diag_set(OutOfMemory, sizeof(*lsm), "malloc",
+ "struct vy_lsm_recovery_info");
return NULL;
}
lsm->key_parts = vy_recovery_alloc_key_parts(key_parts, key_part_count);
@@ -1521,8 +1513,7 @@ vy_recovery_do_create_lsm(struct vy_recovery *recovery, int64_t id,
*/
static int
vy_recovery_prepare_lsm(struct vy_recovery *recovery, int64_t id,
- uint32_t space_id, uint32_t index_id,
- uint32_t group_id,
+ uint32_t space_id, uint32_t index_id, uint32_t group_id,
const struct key_part_def *key_parts,
uint32_t key_part_count)
{
@@ -1570,8 +1561,8 @@ vy_recovery_create_lsm(struct vy_recovery *recovery, int64_t id,
}
} else {
lsm = vy_recovery_do_create_lsm(recovery, id, space_id,
- index_id, group_id,
- key_parts, key_part_count);
+ index_id, group_id, key_parts,
+ key_part_count);
if (lsm == NULL)
return -1;
lsm->dump_lsn = dump_lsn;
@@ -1656,9 +1647,10 @@ vy_recovery_forget_lsm(struct vy_recovery *recovery, int64_t id)
struct mh_i64ptr_t *h = recovery->lsm_hash;
mh_int_t k = mh_i64ptr_find(h, id, NULL);
if (k == mh_end(h)) {
- diag_set(ClientError, ER_INVALID_VYLOG_FILE,
- tt_sprintf("LSM tree %lld forgotten but not registered",
- (long long)id));
+ diag_set(
+ ClientError, ER_INVALID_VYLOG_FILE,
+ tt_sprintf("LSM tree %lld forgotten but not registered",
+ (long long)id));
return -1;
}
struct vy_lsm_recovery_info *lsm = mh_i64ptr_node(h, k)->val;
@@ -1682,8 +1674,7 @@ vy_recovery_forget_lsm(struct vy_recovery *recovery, int64_t id)
* Returns 0 on success, -1 if ID not found or LSM tree is dropped.
*/
static int
-vy_recovery_dump_lsm(struct vy_recovery *recovery,
- int64_t id, int64_t dump_lsn)
+vy_recovery_dump_lsm(struct vy_recovery *recovery, int64_t id, int64_t dump_lsn)
{
struct vy_lsm_recovery_info *lsm;
lsm = vy_recovery_lookup_lsm(recovery, id);
@@ -1706,8 +1697,8 @@ vy_recovery_do_create_run(struct vy_recovery *recovery, int64_t run_id)
{
struct vy_run_recovery_info *run = malloc(sizeof(*run));
if (run == NULL) {
- diag_set(OutOfMemory, sizeof(*run),
- "malloc", "struct vy_run_recovery_info");
+ diag_set(OutOfMemory, sizeof(*run), "malloc",
+ "struct vy_run_recovery_info");
return NULL;
}
struct mh_i64ptr_t *h = recovery->run_hash;
@@ -1748,8 +1739,8 @@ vy_recovery_prepare_run(struct vy_recovery *recovery, int64_t lsm_id,
if (lsm == NULL) {
diag_set(ClientError, ER_INVALID_VYLOG_FILE,
tt_sprintf("Run %lld created for unregistered "
- "LSM tree %lld", (long long)run_id,
- (long long)lsm_id));
+ "LSM tree %lld",
+ (long long)run_id, (long long)lsm_id));
return -1;
}
if (vy_recovery_lookup_run(recovery, run_id) != NULL) {
@@ -1784,8 +1775,8 @@ vy_recovery_create_run(struct vy_recovery *recovery, int64_t lsm_id,
if (lsm == NULL) {
diag_set(ClientError, ER_INVALID_VYLOG_FILE,
tt_sprintf("Run %lld created for unregistered "
- "LSM tree %lld", (long long)run_id,
- (long long)lsm_id));
+ "LSM tree %lld",
+ (long long)run_id, (long long)lsm_id));
return -1;
}
struct vy_run_recovery_info *run;
@@ -1883,8 +1874,8 @@ vy_recovery_insert_range(struct vy_recovery *recovery, int64_t lsm_id,
if (lsm == NULL) {
diag_set(ClientError, ER_INVALID_VYLOG_FILE,
tt_sprintf("Range %lld created for unregistered "
- "LSM tree %lld", (long long)range_id,
- (long long)lsm_id));
+ "LSM tree %lld",
+ (long long)range_id, (long long)lsm_id));
return -1;
}
@@ -1903,8 +1894,8 @@ vy_recovery_insert_range(struct vy_recovery *recovery, int64_t lsm_id,
struct vy_range_recovery_info *range = malloc(size);
if (range == NULL) {
- diag_set(OutOfMemory, size,
- "malloc", "struct vy_range_recovery_info");
+ diag_set(OutOfMemory, size, "malloc",
+ "struct vy_range_recovery_info");
return -1;
}
struct mh_i64ptr_t *h = recovery->range_hash;
@@ -1971,8 +1962,8 @@ vy_recovery_delete_range(struct vy_recovery *recovery, int64_t range_id)
*/
static int
vy_recovery_insert_slice(struct vy_recovery *recovery, int64_t range_id,
- int64_t run_id, int64_t slice_id,
- const char *begin, const char *end)
+ int64_t run_id, int64_t slice_id, const char *begin,
+ const char *end)
{
if (vy_recovery_lookup_slice(recovery, slice_id) != NULL) {
diag_set(ClientError, ER_INVALID_VYLOG_FILE,
@@ -1985,8 +1976,8 @@ vy_recovery_insert_slice(struct vy_recovery *recovery, int64_t range_id,
if (range == NULL) {
diag_set(ClientError, ER_INVALID_VYLOG_FILE,
tt_sprintf("Slice %lld created for unregistered "
- "range %lld", (long long)slice_id,
- (long long)range_id));
+ "range %lld",
+ (long long)slice_id, (long long)range_id));
return -1;
}
struct vy_run_recovery_info *run;
@@ -1994,8 +1985,8 @@ vy_recovery_insert_slice(struct vy_recovery *recovery, int64_t range_id,
if (run == NULL) {
diag_set(ClientError, ER_INVALID_VYLOG_FILE,
tt_sprintf("Slice %lld created for unregistered "
- "run %lld", (long long)slice_id,
- (long long)run_id));
+ "run %lld",
+ (long long)slice_id, (long long)run_id));
return -1;
}
@@ -2014,8 +2005,8 @@ vy_recovery_insert_slice(struct vy_recovery *recovery, int64_t range_id,
struct vy_slice_recovery_info *slice = malloc(size);
if (slice == NULL) {
- diag_set(OutOfMemory, size,
- "malloc", "struct vy_slice_recovery_info");
+ diag_set(OutOfMemory, size, "malloc",
+ "struct vy_slice_recovery_info");
return -1;
}
struct mh_i64ptr_t *h = recovery->slice_hash;
@@ -2128,21 +2119,23 @@ vy_recovery_process_record(struct vy_recovery *recovery,
switch (record->type) {
case VY_LOG_PREPARE_LSM:
rc = vy_recovery_prepare_lsm(recovery, record->lsm_id,
- record->space_id, record->index_id,
- record->group_id, record->key_parts,
- record->key_part_count);
+ record->space_id, record->index_id,
+ record->group_id,
+ record->key_parts,
+ record->key_part_count);
break;
case VY_LOG_CREATE_LSM:
- rc = vy_recovery_create_lsm(recovery, record->lsm_id,
- record->space_id, record->index_id,
- record->group_id, record->key_parts,
- record->key_part_count, record->create_lsn,
- record->modify_lsn, record->dump_lsn);
+ rc = vy_recovery_create_lsm(
+ recovery, record->lsm_id, record->space_id,
+ record->index_id, record->group_id, record->key_parts,
+ record->key_part_count, record->create_lsn,
+ record->modify_lsn, record->dump_lsn);
break;
case VY_LOG_MODIFY_LSM:
rc = vy_recovery_modify_lsm(recovery, record->lsm_id,
- record->key_parts, record->key_part_count,
- record->modify_lsn);
+ record->key_parts,
+ record->key_part_count,
+ record->modify_lsn);
break;
case VY_LOG_DROP_LSM:
rc = vy_recovery_drop_lsm(recovery, record->lsm_id,
@@ -2153,7 +2146,8 @@ vy_recovery_process_record(struct vy_recovery *recovery,
break;
case VY_LOG_INSERT_RANGE:
rc = vy_recovery_insert_range(recovery, record->lsm_id,
- record->range_id, record->begin, record->end);
+ record->range_id, record->begin,
+ record->end);
break;
case VY_LOG_DELETE_RANGE:
rc = vy_recovery_delete_range(recovery, record->range_id);
@@ -2184,7 +2178,7 @@ vy_recovery_process_record(struct vy_recovery *recovery,
break;
case VY_LOG_DUMP_LSM:
rc = vy_recovery_dump_lsm(recovery, record->lsm_id,
- record->dump_lsn);
+ record->dump_lsn);
break;
case VY_LOG_TRUNCATE_LSM:
/* Not used anymore, ignore. */
@@ -2247,8 +2241,8 @@ vy_recovery_build_index_id_hash(struct vy_recovery *recovery)
uint32_t space_id = lsm->space_id;
uint32_t index_id = lsm->index_id;
struct vy_lsm_recovery_info *hashed_lsm;
- hashed_lsm = vy_recovery_lsm_by_index_id(recovery,
- space_id, index_id);
+ hashed_lsm = vy_recovery_lsm_by_index_id(recovery, space_id,
+ index_id);
/*
* If there's no LSM tree for these space_id/index_id
* or it was dropped, simply replace it with the latest
@@ -2257,7 +2251,8 @@ vy_recovery_build_index_id_hash(struct vy_recovery *recovery)
if (hashed_lsm == NULL ||
(hashed_lsm->drop_lsn >= 0 && lsm->create_lsn >= 0)) {
struct mh_i64ptr_node_t node;
- node.key = vy_recovery_index_id_hash(space_id, index_id);
+ node.key =
+ vy_recovery_index_id_hash(space_id, index_id);
node.val = lsm;
if (mh_i64ptr_put(h, &node, NULL, NULL) == mh_end(h)) {
diag_set(OutOfMemory, 0, "mh_i64ptr_put",
@@ -2305,8 +2300,8 @@ vy_recovery_new_f(va_list ap)
struct vy_recovery *recovery = malloc(sizeof(*recovery));
if (recovery == NULL) {
- diag_set(OutOfMemory, sizeof(*recovery),
- "malloc", "struct vy_recovery");
+ diag_set(OutOfMemory, sizeof(*recovery), "malloc",
+ "struct vy_recovery");
goto fail;
}
@@ -2324,10 +2319,8 @@ vy_recovery_new_f(va_list ap)
recovery->range_hash = mh_i64ptr_new();
recovery->run_hash = mh_i64ptr_new();
recovery->slice_hash = mh_i64ptr_new();
- if (recovery->index_id_hash == NULL ||
- recovery->lsm_hash == NULL ||
- recovery->range_hash == NULL ||
- recovery->run_hash == NULL ||
+ if (recovery->index_id_hash == NULL || recovery->lsm_hash == NULL ||
+ recovery->range_hash == NULL || recovery->run_hash == NULL ||
recovery->slice_hash == NULL) {
diag_set(OutOfMemory, 0, "mh_i64ptr_new", "mh_i64ptr_t");
goto fail_free;
@@ -2445,8 +2438,8 @@ vy_recovery_delete(struct vy_recovery *recovery)
struct vy_run_recovery_info *run, *next_run;
rlist_foreach_entry_safe(lsm, &recovery->lsms, in_recovery, next_lsm) {
- rlist_foreach_entry_safe(range, &lsm->ranges,
- in_lsm, next_range) {
+ rlist_foreach_entry_safe(range, &lsm->ranges, in_lsm,
+ next_range) {
rlist_foreach_entry_safe(slice, &range->slices,
in_range, next_slice)
free(slice);
@@ -2495,8 +2488,8 @@ vy_log_append_lsm(struct xlog *xlog, struct vy_lsm_recovery_info *lsm)
struct vy_log_record record;
vy_log_record_init(&record);
- record.type = lsm->create_lsn < 0 ?
- VY_LOG_PREPARE_LSM : VY_LOG_CREATE_LSM;
+ record.type = lsm->create_lsn < 0 ? VY_LOG_PREPARE_LSM :
+ VY_LOG_CREATE_LSM;
record.lsm_id = lsm->id;
record.index_id = lsm->index_id;
record.space_id = lsm->space_id;
@@ -2548,7 +2541,8 @@ vy_log_append_lsm(struct xlog *xlog, struct vy_lsm_recovery_info *lsm)
* while we are supposed to return slices in chronological
* order, so use reverse iterator.
*/
- rlist_foreach_entry_reverse(slice, &range->slices, in_range) {
+ rlist_foreach_entry_reverse(slice, &range->slices, in_range)
+ {
vy_log_record_init(&record);
record.type = VY_LOG_INSERT_SLICE;
record.range_id = range->id;
@@ -2612,8 +2606,7 @@ vy_log_create(const struct vclock *vclock, struct vy_recovery *recovery)
});
/* Finalize the new xlog. */
- if (xlog_flush(&xlog) < 0 ||
- xlog_sync(&xlog) < 0 ||
+ if (xlog_flush(&xlog) < 0 || xlog_sync(&xlog) < 0 ||
xlog_rename(&xlog) < 0)
goto err_write_xlog;
@@ -2626,8 +2619,7 @@ err_write_xlog:
/* Delete the unfinished xlog. */
assert(xlog_is_open(&xlog));
if (unlink(xlog.filename) < 0)
- say_syserror("failed to delete file '%s'",
- xlog.filename);
+ say_syserror("failed to delete file '%s'", xlog.filename);
xlog_close(&xlog, false);
err_create_xlog:
diff --git a/src/box/vy_log.h b/src/box/vy_log.h
index 298a8ed..bf1fd42 100644
--- a/src/box/vy_log.h
+++ b/src/box/vy_log.h
@@ -69,22 +69,22 @@ enum vy_log_record_type {
* After rotation, it also stores space_id, index_id, group_id,
* key_def, create_lsn, modify_lsn, dump_lsn.
*/
- VY_LOG_CREATE_LSM = 0,
+ VY_LOG_CREATE_LSM = 0,
/**
* Drop an LSM tree.
* Requires vy_log_record::lsm_id, drop_lsn.
*/
- VY_LOG_DROP_LSM = 1,
+ VY_LOG_DROP_LSM = 1,
/**
* Insert a new range into an LSM tree.
* Requires vy_log_record::lsm_id, range_id, begin, end.
*/
- VY_LOG_INSERT_RANGE = 2,
+ VY_LOG_INSERT_RANGE = 2,
/**
* Delete a vinyl range and all its runs.
* Requires vy_log_record::range_id.
*/
- VY_LOG_DELETE_RANGE = 3,
+ VY_LOG_DELETE_RANGE = 3,
/**
* Prepare a vinyl run file.
* Requires vy_log_record::lsm_id, run_id.
@@ -93,14 +93,14 @@ enum vy_log_record_type {
* It is needed to keep track of unfinished due to errors run
* files so that we could remove them after recovery.
*/
- VY_LOG_PREPARE_RUN = 4,
+ VY_LOG_PREPARE_RUN = 4,
/**
* Commit a vinyl run file creation.
* Requires vy_log_record::lsm_id, run_id, dump_lsn, dump_count.
*
* Written after a run file was successfully created.
*/
- VY_LOG_CREATE_RUN = 5,
+ VY_LOG_CREATE_RUN = 5,
/**
* Drop a vinyl run.
* Requires vy_log_record::run_id, gc_lsn.
@@ -113,7 +113,7 @@ enum vy_log_record_type {
* deleted, but not "forgotten" are not expunged from the log
* on rotation.
*/
- VY_LOG_DROP_RUN = 6,
+ VY_LOG_DROP_RUN = 6,
/**
* Forget a vinyl run.
* Requires vy_log_record::run_id.
@@ -124,22 +124,22 @@ enum vy_log_record_type {
* run. Information about "forgotten" runs is not included in
* the new log on rotation.
*/
- VY_LOG_FORGET_RUN = 7,
+ VY_LOG_FORGET_RUN = 7,
/**
* Insert a run slice into a range.
* Requires vy_log_record::range_id, run_id, slice_id, begin, end.
*/
- VY_LOG_INSERT_SLICE = 8,
+ VY_LOG_INSERT_SLICE = 8,
/**
* Delete a run slice.
* Requires vy_log_record::slice_id.
*/
- VY_LOG_DELETE_SLICE = 9,
+ VY_LOG_DELETE_SLICE = 9,
/**
* Log LSM tree dump. Used to update max LSN stored on disk.
* Requires vy_log_record::lsm_id, dump_lsn.
*/
- VY_LOG_DUMP_LSM = 10,
+ VY_LOG_DUMP_LSM = 10,
/**
* We don't split vylog into snapshot and log - all records
* are written to the same file. Since we need to load a
@@ -150,7 +150,7 @@ enum vy_log_record_type {
*
* See also: @only_checkpoint argument of vy_recovery_new().
*/
- VY_LOG_SNAPSHOT = 11,
+ VY_LOG_SNAPSHOT = 11,
/**
* When we used LSN for identifying LSM trees in vylog, we
* couldn't simply recreate an LSM tree on space truncation,
@@ -164,12 +164,12 @@ enum vy_log_record_type {
* 'truncate' records - this will result in replay of all
* WAL records written after truncation.
*/
- VY_LOG_TRUNCATE_LSM = 12,
+ VY_LOG_TRUNCATE_LSM = 12,
/**
* Modify key definition of an LSM tree.
* Requires vy_log_record::lsm_id, key_def, modify_lsn.
*/
- VY_LOG_MODIFY_LSM = 13,
+ VY_LOG_MODIFY_LSM = 13,
/**
* Forget an LSM tree.
* Requires vy_log_record::lsm_id.
@@ -179,7 +179,7 @@ enum vy_log_record_type {
* so the LSM tree is not needed any longer and can be removed
* from vylog on the next rotation.
*/
- VY_LOG_FORGET_LSM = 14,
+ VY_LOG_FORGET_LSM = 14,
/**
* Prepare a new LSM tree for building.
* Requires vy_log_record::lsm_id, index_id, space_id, group_id,
@@ -195,7 +195,7 @@ enum vy_log_record_type {
* for building. Once the index has been built, we write
* a VY_LOG_CREATE_LSM record to commit it.
*/
- VY_LOG_PREPARE_LSM = 15,
+ VY_LOG_PREPARE_LSM = 15,
/**
* This record denotes the beginning of a rebootstrap section.
* A rebootstrap section ends either by another record of this
@@ -211,12 +211,12 @@ enum vy_log_record_type {
* record as dropped in the rotated vylog. If rebootstrap fails,
* we write VY_LOG_ABORT_REBOOTSTRAP on recovery.
*/
- VY_LOG_REBOOTSTRAP = 16,
+ VY_LOG_REBOOTSTRAP = 16,
/**
* This record is written on recovery if rebootstrap failed.
* See also VY_LOG_REBOOTSTRAP.
*/
- VY_LOG_ABORT_REBOOTSTRAP = 17,
+ VY_LOG_ABORT_REBOOTSTRAP = 17,
vy_log_record_type_MAX
};
@@ -577,12 +577,12 @@ enum vy_recovery_flag {
* i.e. get a consistent view of vinyl database at the time
* of the last checkpoint.
*/
- VY_RECOVERY_LOAD_CHECKPOINT = 1 << 0,
+ VY_RECOVERY_LOAD_CHECKPOINT = 1 << 0,
/**
* Consider the last attempt to rebootstrap aborted even if
* there's no VY_LOG_ABORT_REBOOTSTRAP record.
*/
- VY_RECOVERY_ABORT_REBOOTSTRAP = 1 << 1,
+ VY_RECOVERY_ABORT_REBOOTSTRAP = 1 << 1,
};
/**
@@ -608,8 +608,8 @@ vy_recovery_delete(struct vy_recovery *recovery);
* Returns NULL if the LSM tree was not found.
*/
struct vy_lsm_recovery_info *
-vy_recovery_lsm_by_index_id(struct vy_recovery *recovery,
- uint32_t space_id, uint32_t index_id);
+vy_recovery_lsm_by_index_id(struct vy_recovery *recovery, uint32_t space_id,
+ uint32_t index_id);
/**
* Initialize a log record with default values.
@@ -688,8 +688,8 @@ vy_log_drop_lsm(int64_t id, int64_t drop_lsn)
/** Helper to log a vinyl range insertion. */
static inline void
-vy_log_insert_range(int64_t lsm_id, int64_t range_id,
- const char *begin, const char *end)
+vy_log_insert_range(int64_t lsm_id, int64_t range_id, const char *begin,
+ const char *end)
{
struct vy_log_record record;
vy_log_record_init(&record);
@@ -726,8 +726,8 @@ vy_log_prepare_run(int64_t lsm_id, int64_t run_id)
/** Helper to log a vinyl run creation. */
static inline void
-vy_log_create_run(int64_t lsm_id, int64_t run_id,
- int64_t dump_lsn, uint32_t dump_count)
+vy_log_create_run(int64_t lsm_id, int64_t run_id, int64_t dump_lsn,
+ uint32_t dump_count)
{
struct vy_log_record record;
vy_log_record_init(&record);
diff --git a/src/box/vy_lsm.c b/src/box/vy_lsm.c
index 1f67bea..2abe27a 100644
--- a/src/box/vy_lsm.c
+++ b/src/box/vy_lsm.c
@@ -73,8 +73,7 @@ static const int64_t VY_MAX_RANGE_SIZE = 2LL * 1024 * 1024 * 1024;
int
vy_lsm_env_create(struct vy_lsm_env *env, const char *path,
int64_t *p_generation, struct tuple_format *key_format,
- vy_upsert_thresh_cb upsert_thresh_cb,
- void *upsert_thresh_arg)
+ vy_upsert_thresh_cb upsert_thresh_cb, void *upsert_thresh_arg)
{
env->empty_key.hint = HINT_NONE;
env->empty_key.stmt = vy_key_new(key_format, NULL, 0);
@@ -105,8 +104,8 @@ const char *
vy_lsm_name(struct vy_lsm *lsm)
{
char *buf = tt_static_buf();
- snprintf(buf, TT_STATIC_BUF_LEN, "%u/%u",
- (unsigned)lsm->space_id, (unsigned)lsm->index_id);
+ snprintf(buf, TT_STATIC_BUF_LEN, "%u/%u", (unsigned)lsm->space_id,
+ (unsigned)lsm->index_id);
return buf;
}
@@ -134,8 +133,8 @@ vy_lsm_new(struct vy_lsm_env *lsm_env, struct vy_cache_env *cache_env,
struct vy_lsm *lsm = calloc(1, sizeof(struct vy_lsm));
if (lsm == NULL) {
- diag_set(OutOfMemory, sizeof(struct vy_lsm),
- "calloc", "struct vy_lsm");
+ diag_set(OutOfMemory, sizeof(struct vy_lsm), "calloc",
+ "struct vy_lsm");
goto fail;
}
lsm->env = lsm_env;
@@ -167,9 +166,8 @@ vy_lsm_new(struct vy_lsm_env *lsm_env, struct vy_cache_env *cache_env,
*/
lsm->disk_format = lsm_env->key_format;
- lsm->pk_in_cmp_def = key_def_find_pk_in_cmp_def(lsm->cmp_def,
- pk->key_def,
- &fiber()->gc);
+ lsm->pk_in_cmp_def = key_def_find_pk_in_cmp_def(
+ lsm->cmp_def, pk->key_def, &fiber()->gc);
if (lsm->pk_in_cmp_def == NULL)
goto fail_pk_in_cmp_def;
}
@@ -182,8 +180,7 @@ vy_lsm_new(struct vy_lsm_env *lsm_env, struct vy_cache_env *cache_env,
if (lsm->run_hist == NULL)
goto fail_run_hist;
- lsm->mem = vy_mem_new(mem_env, cmp_def, format,
- *lsm->env->p_generation,
+ lsm->mem = vy_mem_new(mem_env, cmp_def, format, *lsm->env->p_generation,
space_cache_version);
if (lsm->mem == NULL)
goto fail_mem;
@@ -255,7 +252,7 @@ vy_lsm_delete(struct vy_lsm *lsm)
lsm->env->lsm_count--;
lsm->env->compaction_queue_size -=
- lsm->stat.disk.compaction.queue.bytes;
+ lsm->stat.disk.compaction.queue.bytes;
if (lsm->index_id == 0)
lsm->env->compacted_data_size -=
lsm->stat.disk.last_level_count.bytes;
@@ -292,8 +289,8 @@ vy_lsm_create(struct vy_lsm *lsm)
/* Make LSM tree directory. */
int rc;
char path[PATH_MAX];
- vy_lsm_snprint_path(path, sizeof(path), lsm->env->path,
- lsm->space_id, lsm->index_id);
+ vy_lsm_snprint_path(path, sizeof(path), lsm->env->path, lsm->space_id,
+ lsm->index_id);
char *path_sep = path;
while (*path_sep == '/') {
/* Don't create root */
@@ -305,7 +302,7 @@ vy_lsm_create(struct vy_lsm *lsm)
rc = mkdir(path, 0777);
if (rc == -1 && errno != EEXIST) {
diag_set(SystemError, "failed to create directory '%s'",
- path);
+ path);
*path_sep = '/';
return -1;
}
@@ -314,8 +311,7 @@ vy_lsm_create(struct vy_lsm *lsm)
}
rc = mkdir(path, 0777);
if (rc == -1 && errno != EEXIST) {
- diag_set(SystemError, "failed to create directory '%s'",
- path);
+ diag_set(SystemError, "failed to create directory '%s'", path);
return -1;
}
@@ -338,8 +334,8 @@ vy_lsm_create(struct vy_lsm *lsm)
/* Write the new LSM tree record to vylog. */
vy_log_tx_begin();
- vy_log_prepare_lsm(id, lsm->space_id, lsm->index_id,
- lsm->group_id, lsm->key_def);
+ vy_log_prepare_lsm(id, lsm->space_id, lsm->index_id, lsm->group_id,
+ lsm->key_def);
vy_log_insert_range(id, range->id, NULL, NULL);
vy_log_tx_try_commit();
@@ -370,9 +366,8 @@ vy_lsm_recover_run(struct vy_lsm *lsm, struct vy_run_recovery_info *run_info,
if (vy_run_recover(run, lsm->env->path, lsm->space_id, lsm->index_id,
lsm->cmp_def) != 0 &&
(!force_recovery ||
- vy_run_rebuild_index(run, lsm->env->path,
- lsm->space_id, lsm->index_id,
- lsm->cmp_def, lsm->key_def,
+ vy_run_rebuild_index(run, lsm->env->path, lsm->space_id,
+ lsm->index_id, lsm->cmp_def, lsm->key_def,
lsm->disk_format, &lsm->opts) != 0)) {
vy_run_unref(run);
return NULL;
@@ -403,16 +398,14 @@ vy_lsm_recover_slice(struct vy_lsm *lsm, struct vy_range *range,
struct vy_run *run;
if (slice_info->begin != NULL) {
- begin = vy_entry_key_from_msgpack(lsm->env->key_format,
- lsm->cmp_def,
- slice_info->begin);
+ begin = vy_entry_key_from_msgpack(
+ lsm->env->key_format, lsm->cmp_def, slice_info->begin);
if (begin.stmt == NULL)
goto out;
}
if (slice_info->end != NULL) {
end = vy_entry_key_from_msgpack(lsm->env->key_format,
- lsm->cmp_def,
- slice_info->end);
+ lsm->cmp_def, slice_info->end);
if (end.stmt == NULL)
goto out;
}
@@ -424,8 +417,7 @@ vy_lsm_recover_slice(struct vy_lsm *lsm, struct vy_range *range,
goto out;
}
- run = vy_lsm_recover_run(lsm, slice_info->run,
- run_env, force_recovery);
+ run = vy_lsm_recover_run(lsm, slice_info->run, run_env, force_recovery);
if (run == NULL)
goto out;
@@ -452,16 +444,14 @@ vy_lsm_recover_range(struct vy_lsm *lsm,
struct vy_range *range = NULL;
if (range_info->begin != NULL) {
- begin = vy_entry_key_from_msgpack(lsm->env->key_format,
- lsm->cmp_def,
- range_info->begin);
+ begin = vy_entry_key_from_msgpack(
+ lsm->env->key_format, lsm->cmp_def, range_info->begin);
if (begin.stmt == NULL)
goto out;
}
if (range_info->end != NULL) {
end = vy_entry_key_from_msgpack(lsm->env->key_format,
- lsm->cmp_def,
- range_info->end);
+ lsm->cmp_def, range_info->end);
if (end.stmt == NULL)
goto out;
}
@@ -483,9 +473,10 @@ vy_lsm_recover_range(struct vy_lsm *lsm,
* order, so use reverse iterator.
*/
struct vy_slice_recovery_info *slice_info;
- rlist_foreach_entry_reverse(slice_info, &range_info->slices, in_range) {
- if (vy_lsm_recover_slice(lsm, range, slice_info,
- run_env, force_recovery) == NULL) {
+ rlist_foreach_entry_reverse(slice_info, &range_info->slices, in_range)
+ {
+ if (vy_lsm_recover_slice(lsm, range, slice_info, run_env,
+ force_recovery) == NULL) {
vy_range_delete(range);
range = NULL;
goto out;
@@ -502,8 +493,8 @@ out:
int
vy_lsm_recover(struct vy_lsm *lsm, struct vy_recovery *recovery,
- struct vy_run_env *run_env, int64_t lsn,
- bool is_checkpoint_recovery, bool force_recovery)
+ struct vy_run_env *run_env, int64_t lsn,
+ bool is_checkpoint_recovery, bool force_recovery)
{
assert(lsm->id < 0);
assert(lsm->commit_lsn < 0);
@@ -523,8 +514,8 @@ vy_lsm_recover(struct vy_lsm *lsm, struct vy_recovery *recovery,
* Look up the last incarnation of the LSM tree in vylog.
*/
struct vy_lsm_recovery_info *lsm_info;
- lsm_info = vy_recovery_lsm_by_index_id(recovery,
- lsm->space_id, lsm->index_id);
+ lsm_info = vy_recovery_lsm_by_index_id(recovery, lsm->space_id,
+ lsm->index_id);
if (is_checkpoint_recovery) {
if (lsm_info == NULL || lsm_info->create_lsn < 0) {
/*
@@ -549,9 +540,9 @@ vy_lsm_recover(struct vy_lsm *lsm, struct vy_recovery *recovery,
}
}
- if (lsm_info == NULL || (lsm_info->prepared == NULL &&
- lsm_info->create_lsn >= 0 &&
- lsn > lsm_info->create_lsn)) {
+ if (lsm_info == NULL ||
+ (lsm_info->prepared == NULL && lsm_info->create_lsn >= 0 &&
+ lsn > lsm_info->create_lsn)) {
/*
* If we failed to log LSM tree creation before restart,
* we won't find it in the log on recovery. This is OK as
@@ -646,7 +637,8 @@ vy_lsm_recover(struct vy_lsm *lsm, struct vy_recovery *recovery,
*/
struct vy_range *range, *prev = NULL;
for (range = vy_range_tree_first(&lsm->range_tree); range != NULL;
- prev = range, range = vy_range_tree_next(&lsm->range_tree, range)) {
+ prev = range,
+ range = vy_range_tree_next(&lsm->range_tree, range)) {
if (prev == NULL && range->begin.stmt != NULL) {
diag_set(ClientError, ER_INVALID_VYLOG_FILE,
tt_sprintf("Range %lld is leftmost but "
@@ -659,12 +651,12 @@ vy_lsm_recover(struct vy_lsm *lsm, struct vy_recovery *recovery,
(prev->end.stmt == NULL || range->begin.stmt == NULL ||
(cmp = vy_entry_compare(prev->end, range->begin,
lsm->cmp_def)) != 0)) {
- const char *errmsg = cmp > 0 ?
- "Nearby ranges %lld and %lld overlap" :
- "Keys between ranges %lld and %lld not spanned";
+ const char *errmsg =
+ cmp > 0 ?
+ "Nearby ranges %lld and %lld overlap" :
+ "Keys between ranges %lld and %lld not spanned";
diag_set(ClientError, ER_INVALID_VYLOG_FILE,
- tt_sprintf(errmsg,
- (long long)prev->id,
+ tt_sprintf(errmsg, (long long)prev->id,
(long long)range->id));
return -1;
}
@@ -690,8 +682,11 @@ vy_lsm_recover(struct vy_lsm *lsm, struct vy_recovery *recovery,
int64_t
vy_lsm_generation(struct vy_lsm *lsm)
{
- struct vy_mem *oldest = rlist_empty(&lsm->sealed) ? lsm->mem :
- rlist_last_entry(&lsm->sealed, struct vy_mem, in_sealed);
+ struct vy_mem *oldest = rlist_empty(&lsm->sealed) ?
+ lsm->mem :
+ rlist_last_entry(&lsm->sealed,
+ struct vy_mem,
+ in_sealed);
return oldest->generation;
}
@@ -722,8 +717,10 @@ vy_lsm_range_size(struct vy_lsm *lsm)
* create four times more than that for better smoothing.
*/
int range_count = 4 * vy_lsm_dumps_per_compaction(lsm);
- int64_t range_size = range_count == 0 ? 0 :
- lsm->stat.disk.last_level_count.bytes / range_count;
+ int64_t range_size =
+ range_count == 0 ?
+ 0 :
+ lsm->stat.disk.last_level_count.bytes / range_count;
range_size = MAX(range_size, VY_MIN_RANGE_SIZE);
range_size = MIN(range_size, VY_MAX_RANGE_SIZE);
return range_size;
@@ -798,7 +795,7 @@ vy_lsm_add_range(struct vy_lsm *lsm, struct vy_range *range)
void
vy_lsm_remove_range(struct vy_lsm *lsm, struct vy_range *range)
{
- assert(! heap_node_is_stray(&range->heap_node));
+ assert(!heap_node_is_stray(&range->heap_node));
vy_range_heap_delete(&lsm->range_heap, range);
vy_range_tree_remove(&lsm->range_tree, range);
lsm->range_count--;
@@ -813,8 +810,8 @@ vy_lsm_acct_range(struct vy_lsm *lsm, struct vy_range *range)
&range->compaction_queue);
lsm->env->compaction_queue_size += range->compaction_queue.bytes;
if (!rlist_empty(&range->slices)) {
- struct vy_slice *slice = rlist_last_entry(&range->slices,
- struct vy_slice, in_range);
+ struct vy_slice *slice = rlist_last_entry(
+ &range->slices, struct vy_slice, in_range);
vy_disk_stmt_counter_add(&lsm->stat.disk.last_level_count,
&slice->count);
if (lsm->index_id == 0)
@@ -831,8 +828,8 @@ vy_lsm_unacct_range(struct vy_lsm *lsm, struct vy_range *range)
&range->compaction_queue);
lsm->env->compaction_queue_size -= range->compaction_queue.bytes;
if (!rlist_empty(&range->slices)) {
- struct vy_slice *slice = rlist_last_entry(&range->slices,
- struct vy_slice, in_range);
+ struct vy_slice *slice = rlist_last_entry(
+ &range->slices, struct vy_slice, in_range);
vy_disk_stmt_counter_sub(&lsm->stat.disk.last_level_count,
&slice->count);
if (lsm->index_id == 0)
@@ -890,8 +887,8 @@ vy_lsm_delete_mem(struct vy_lsm *lsm, struct vy_mem *mem)
}
int
-vy_lsm_set(struct vy_lsm *lsm, struct vy_mem *mem,
- struct vy_entry entry, struct tuple **region_stmt)
+vy_lsm_set(struct vy_lsm *lsm, struct vy_mem *mem, struct vy_entry entry,
+ struct tuple **region_stmt)
{
uint32_t format_id = entry.stmt->format_id;
@@ -907,9 +904,8 @@ vy_lsm_set(struct vy_lsm *lsm, struct vy_mem *mem,
* while other LSM trees still use the old space format.
*/
if (*region_stmt == NULL || (*region_stmt)->format_id != format_id) {
- *region_stmt = vy_stmt_dup_lsregion(entry.stmt,
- &mem->env->allocator,
- mem->generation);
+ *region_stmt = vy_stmt_dup_lsregion(
+ entry.stmt, &mem->env->allocator, mem->generation);
if (*region_stmt == NULL)
return -1;
}
@@ -982,7 +978,8 @@ vy_lsm_commit_upsert(struct vy_lsm *lsm, struct vy_mem *mem,
older = vy_mem_older_lsn(mem, entry);
assert(older.stmt != NULL &&
vy_stmt_type(older.stmt) == IPROTO_UPSERT &&
- vy_stmt_n_upserts(older.stmt) == VY_UPSERT_THRESHOLD - 1);
+ vy_stmt_n_upserts(older.stmt) ==
+ VY_UPSERT_THRESHOLD - 1);
#endif
if (lsm->env->upsert_thresh_cb == NULL) {
/* Squash callback is not installed. */
@@ -994,7 +991,7 @@ vy_lsm_commit_upsert(struct vy_lsm *lsm, struct vy_mem *mem,
dup.stmt = vy_stmt_dup(entry.stmt);
if (dup.stmt != NULL) {
lsm->env->upsert_thresh_cb(lsm, dup,
- lsm->env->upsert_thresh_arg);
+ lsm->env->upsert_thresh_arg);
tuple_unref(dup.stmt);
}
/*
@@ -1015,8 +1012,8 @@ vy_lsm_commit_upsert(struct vy_lsm *lsm, struct vy_mem *mem,
assert(older.stmt == NULL ||
vy_stmt_type(older.stmt) != IPROTO_UPSERT);
struct vy_entry upserted;
- upserted = vy_entry_apply_upsert(entry, older,
- lsm->cmp_def, false);
+ upserted = vy_entry_apply_upsert(entry, older, lsm->cmp_def,
+ false);
lsm->stat.upsert.applied++;
if (upserted.stmt == NULL) {
@@ -1041,10 +1038,8 @@ vy_lsm_commit_upsert(struct vy_lsm *lsm, struct vy_mem *mem,
upserted_lsn != vy_stmt_lsn(older.stmt));
assert(vy_stmt_type(upserted.stmt) == IPROTO_REPLACE);
- struct tuple *region_stmt =
- vy_stmt_dup_lsregion(upserted.stmt,
- &mem->env->allocator,
- mem->generation);
+ struct tuple *region_stmt = vy_stmt_dup_lsregion(
+ upserted.stmt, &mem->env->allocator, mem->generation);
if (region_stmt == NULL) {
/* OOM */
tuple_unref(upserted.stmt);
@@ -1058,7 +1053,8 @@ vy_lsm_commit_upsert(struct vy_lsm *lsm, struct vy_mem *mem,
* now we replacing one statement with another, the
* vy_lsm_set() cannot fail.
*/
- assert(rc == 0); (void)rc;
+ assert(rc == 0);
+ (void)rc;
tuple_unref(upserted.stmt);
upserted.stmt = region_stmt;
vy_mem_commit_stmt(mem, upserted);
@@ -1094,9 +1090,9 @@ vy_lsm_rollback_stmt(struct vy_lsm *lsm, struct vy_mem *mem,
}
int
-vy_lsm_find_range_intersection(struct vy_lsm *lsm,
- const char *min_key, const char *max_key,
- struct vy_range **begin, struct vy_range **end)
+vy_lsm_find_range_intersection(struct vy_lsm *lsm, const char *min_key,
+ const char *max_key, struct vy_range **begin,
+ struct vy_range **end)
{
struct tuple_format *key_format = lsm->env->key_format;
struct vy_entry entry;
@@ -1161,7 +1157,8 @@ vy_lsm_split_range(struct vy_lsm *lsm, struct vy_range *range)
* so to preserve the order of the slices list, we have
* to iterate backward.
*/
- rlist_foreach_entry_reverse(slice, &range->slices, in_range) {
+ rlist_foreach_entry_reverse(slice, &range->slices, in_range)
+ {
if (vy_slice_cut(slice, vy_log_next_id(), part->begin,
part->end, lsm->cmp_def,
&new_slice) != 0)
@@ -1187,9 +1184,10 @@ vy_lsm_split_range(struct vy_lsm *lsm, struct vy_range *range)
tuple_data_or_null(part->begin.stmt),
tuple_data_or_null(part->end.stmt));
rlist_foreach_entry(slice, &part->slices, in_range)
- vy_log_insert_slice(part->id, slice->run->id, slice->id,
- tuple_data_or_null(slice->begin.stmt),
- tuple_data_or_null(slice->end.stmt));
+ vy_log_insert_slice(
+ part->id, slice->run->id, slice->id,
+ tuple_data_or_null(slice->begin.stmt),
+ tuple_data_or_null(slice->end.stmt));
}
if (vy_log_tx_commit() < 0)
goto fail;
@@ -1224,8 +1222,8 @@ fail:
tuple_unref(split_key.stmt);
diag_log();
- say_error("%s: failed to split range %s",
- vy_lsm_name(lsm), vy_range_str(range));
+ say_error("%s: failed to split range %s", vy_lsm_name(lsm),
+ vy_range_str(range));
return false;
}
@@ -1237,8 +1235,8 @@ vy_lsm_coalesce_range(struct vy_lsm *lsm, struct vy_range *range)
vy_lsm_range_size(lsm), &first, &last))
return false;
- struct vy_range *result = vy_range_new(vy_log_next_id(),
- first->begin, last->end, lsm->cmp_def);
+ struct vy_range *result = vy_range_new(vy_log_next_id(), first->begin,
+ last->end, lsm->cmp_def);
if (result == NULL)
goto fail_range;
@@ -1259,9 +1257,10 @@ vy_lsm_coalesce_range(struct vy_lsm *lsm, struct vy_range *range)
vy_log_delete_slice(slice->id);
vy_log_delete_range(it->id);
rlist_foreach_entry(slice, &it->slices, in_range) {
- vy_log_insert_slice(result->id, slice->run->id, slice->id,
- tuple_data_or_null(slice->begin.stmt),
- tuple_data_or_null(slice->end.stmt));
+ vy_log_insert_slice(
+ result->id, slice->run->id, slice->id,
+ tuple_data_or_null(slice->begin.stmt),
+ tuple_data_or_null(slice->end.stmt));
}
}
if (vy_log_tx_commit() < 0)
@@ -1273,7 +1272,8 @@ vy_lsm_coalesce_range(struct vy_lsm *lsm, struct vy_range *range)
*/
it = first;
while (it != end) {
- struct vy_range *next = vy_range_tree_next(&lsm->range_tree, it);
+ struct vy_range *next =
+ vy_range_tree_next(&lsm->range_tree, it);
vy_lsm_unacct_range(lsm, it);
vy_lsm_remove_range(lsm, it);
rlist_splice(&result->slices, &it->slices);
@@ -1295,16 +1295,16 @@ vy_lsm_coalesce_range(struct vy_lsm *lsm, struct vy_range *range)
vy_lsm_add_range(lsm, result);
lsm->range_tree_version++;
- say_info("%s: coalesced ranges %s",
- vy_lsm_name(lsm), vy_range_str(result));
+ say_info("%s: coalesced ranges %s", vy_lsm_name(lsm),
+ vy_range_str(result));
return true;
fail_commit:
vy_range_delete(result);
fail_range:
diag_log();
- say_error("%s: failed to coalesce range %s",
- vy_lsm_name(lsm), vy_range_str(range));
+ say_error("%s: failed to coalesce range %s", vy_lsm_name(lsm),
+ vy_range_str(range));
return false;
}
diff --git a/src/box/vy_lsm.h b/src/box/vy_lsm.h
index 3b553ea..c8e82d9 100644
--- a/src/box/vy_lsm.h
+++ b/src/box/vy_lsm.h
@@ -61,8 +61,8 @@ struct vy_recovery;
struct vy_run;
struct vy_run_env;
-typedef void
-(*vy_upsert_thresh_cb)(struct vy_lsm *lsm, struct vy_entry entry, void *arg);
+typedef void (*vy_upsert_thresh_cb)(struct vy_lsm *lsm, struct vy_entry entry,
+ void *arg);
/** Common LSM tree environment. */
struct vy_lsm_env {
@@ -442,8 +442,8 @@ vy_lsm_create(struct vy_lsm *lsm);
*/
int
vy_lsm_recover(struct vy_lsm *lsm, struct vy_recovery *recovery,
- struct vy_run_env *run_env, int64_t lsn,
- bool is_checkpoint_recovery, bool force_recovery);
+ struct vy_run_env *run_env, int64_t lsn,
+ bool is_checkpoint_recovery, bool force_recovery);
/**
* Return generation of in-memory data stored in an LSM tree
@@ -547,9 +547,9 @@ vy_lsm_delete_mem(struct vy_lsm *lsm, struct vy_mem *mem);
* On memory allocation error returns -1 and sets diag.
*/
int
-vy_lsm_find_range_intersection(struct vy_lsm *lsm,
- const char *min_key, const char *max_key,
- struct vy_range **begin, struct vy_range **end);
+vy_lsm_find_range_intersection(struct vy_lsm *lsm, const char *min_key,
+ const char *max_key, struct vy_range **begin,
+ struct vy_range **end);
/**
* Split a range if it has grown too big, return true if the range
@@ -597,8 +597,8 @@ vy_lsm_force_compaction(struct vy_lsm *lsm);
* @retval -1 Memory error.
*/
int
-vy_lsm_set(struct vy_lsm *lsm, struct vy_mem *mem,
- struct vy_entry entry, struct tuple **region_stmt);
+vy_lsm_set(struct vy_lsm *lsm, struct vy_mem *mem, struct vy_entry entry,
+ struct tuple **region_stmt);
/**
* Confirm that the statement stays in the in-memory index of
diff --git a/src/box/vy_mem.c b/src/box/vy_mem.c
index 98027e7..d9c4f4b 100644
--- a/src/box/vy_mem.c
+++ b/src/box/vy_mem.c
@@ -53,8 +53,8 @@ vy_mem_env_create(struct vy_mem_env *env, size_t memory)
{
/* Vinyl memory is limited by vy_quota. */
quota_init(&env->quota, QUOTA_MAX);
- tuple_arena_create(&env->arena, &env->quota, memory,
- SLAB_SIZE, false, "vinyl");
+ tuple_arena_create(&env->arena, &env->quota, memory, SLAB_SIZE, false,
+ "vinyl");
lsregion_create(&env->allocator, &env->arena);
env->tree_extent_size = 0;
}
@@ -73,7 +73,7 @@ vy_mem_env_destroy(struct vy_mem_env *env)
static void *
vy_mem_tree_extent_alloc(void *ctx)
{
- struct vy_mem *mem = (struct vy_mem *) ctx;
+ struct vy_mem *mem = (struct vy_mem *)ctx;
struct vy_mem_env *env = mem->env;
void *ret = lsregion_aligned_alloc(&env->allocator,
VY_MEM_TREE_EXTENT_SIZE,
@@ -103,8 +103,8 @@ vy_mem_new(struct vy_mem_env *env, struct key_def *cmp_def,
{
struct vy_mem *index = calloc(1, sizeof(*index));
if (!index) {
- diag_set(OutOfMemory, sizeof(*index),
- "malloc", "struct vy_mem");
+ diag_set(OutOfMemory, sizeof(*index), "malloc",
+ "struct vy_mem");
return NULL;
}
index->env = env;
@@ -114,8 +114,7 @@ vy_mem_new(struct vy_mem_env *env, struct key_def *cmp_def,
index->space_cache_version = space_cache_version;
index->format = format;
tuple_format_ref(format);
- vy_mem_tree_create(&index->tree, cmp_def,
- vy_mem_tree_extent_alloc,
+ vy_mem_tree_create(&index->tree, cmp_def, vy_mem_tree_extent_alloc,
vy_mem_tree_extent_free, index);
rlist_create(&index->in_sealed);
fiber_cond_create(&index->pin_cond);
@@ -166,9 +165,9 @@ vy_mem_insert_upsert(struct vy_mem *mem, struct vy_entry entry)
if (vy_mem_tree_insert_get_iterator(&mem->tree, entry, &replaced,
&inserted) != 0)
return -1;
- assert(! vy_mem_tree_iterator_is_invalid(&inserted));
- assert(vy_entry_is_equal(entry,
- *vy_mem_tree_iterator_get_elem(&mem->tree, &inserted)));
+ assert(!vy_mem_tree_iterator_is_invalid(&inserted));
+ assert(vy_entry_is_equal(
+ entry, *vy_mem_tree_iterator_get_elem(&mem->tree, &inserted)));
if (replaced.stmt == NULL)
mem->count.rows++;
mem->count.bytes += size;
@@ -194,8 +193,8 @@ vy_mem_insert_upsert(struct vy_mem *mem, struct vy_entry entry)
* UPSERTs subsequence.
*/
vy_mem_tree_iterator_next(&mem->tree, &inserted);
- struct vy_entry *older = vy_mem_tree_iterator_get_elem(&mem->tree,
- &inserted);
+ struct vy_entry *older =
+ vy_mem_tree_iterator_get_elem(&mem->tree, &inserted);
if (older == NULL || vy_stmt_type(older->stmt) != IPROTO_UPSERT ||
vy_entry_compare(entry, *older, mem->cmp_def) != 0)
return 0;
@@ -265,7 +264,7 @@ vy_mem_rollback_stmt(struct vy_mem *mem, struct vy_entry entry)
assert(!vy_stmt_is_refable(entry.stmt));
int rc = vy_mem_tree_delete(&mem->tree, entry);
assert(rc == 0);
- (void) rc;
+ (void)rc;
/* We can't free memory in case of rollback. */
mem->count.rows--;
mem->version++;
@@ -289,8 +288,8 @@ vy_mem_iterator_step(struct vy_mem_iterator *itr)
vy_mem_tree_iterator_next(&itr->mem->tree, &itr->curr_pos);
if (vy_mem_tree_iterator_is_invalid(&itr->curr_pos))
return 1;
- itr->curr = *vy_mem_tree_iterator_get_elem(&itr->mem->tree,
- &itr->curr_pos);
+ itr->curr =
+ *vy_mem_tree_iterator_get_elem(&itr->mem->tree, &itr->curr_pos);
return 0;
}
@@ -307,9 +306,9 @@ vy_mem_iterator_find_lsn(struct vy_mem_iterator *itr)
{
/* Skip to the first statement visible in the read view. */
assert(!vy_mem_tree_iterator_is_invalid(&itr->curr_pos));
- assert(vy_entry_is_equal(itr->curr,
- *vy_mem_tree_iterator_get_elem(&itr->mem->tree,
- &itr->curr_pos)));
+ assert(vy_entry_is_equal(
+ itr->curr, *vy_mem_tree_iterator_get_elem(&itr->mem->tree,
+ &itr->curr_pos)));
struct key_def *cmp_def = itr->mem->cmp_def;
while (vy_stmt_lsn(itr->curr.stmt) > (**itr->read_view).vlsn ||
vy_stmt_flags(itr->curr.stmt) & VY_STMT_SKIP_READ) {
@@ -353,11 +352,11 @@ vy_mem_iterator_find_lsn(struct vy_mem_iterator *itr)
struct vy_mem_tree_key tree_key;
tree_key.entry = itr->curr;
tree_key.lsn = (**itr->read_view).vlsn;
- itr->curr_pos = vy_mem_tree_lower_bound(&itr->mem->tree,
- &tree_key, NULL);
+ itr->curr_pos =
+ vy_mem_tree_lower_bound(&itr->mem->tree, &tree_key, NULL);
assert(!vy_mem_tree_iterator_is_invalid(&itr->curr_pos));
- itr->curr = *vy_mem_tree_iterator_get_elem(&itr->mem->tree,
- &itr->curr_pos);
+ itr->curr =
+ *vy_mem_tree_iterator_get_elem(&itr->mem->tree, &itr->curr_pos);
/* Skip VY_STMT_SKIP_READ statements, if any. */
while (vy_stmt_flags(itr->curr.stmt) & VY_STMT_SKIP_READ) {
@@ -390,7 +389,8 @@ vy_mem_iterator_seek(struct vy_mem_iterator *itr, struct vy_entry last)
if (last.stmt != NULL) {
key = last;
iterator_type = iterator_direction(itr->iterator_type) > 0 ?
- ITER_GT : ITER_LT;
+ ITER_GT :
+ ITER_LT;
}
bool exact = false;
@@ -400,16 +400,14 @@ vy_mem_iterator_seek(struct vy_mem_iterator *itr, struct vy_entry last)
tree_key.lsn = INT64_MAX - 1;
if (!vy_stmt_is_empty_key(key.stmt)) {
if (iterator_type == ITER_LE || iterator_type == ITER_GT) {
- itr->curr_pos =
- vy_mem_tree_upper_bound(&itr->mem->tree,
- &tree_key, &exact);
+ itr->curr_pos = vy_mem_tree_upper_bound(
+ &itr->mem->tree, &tree_key, &exact);
} else {
assert(iterator_type == ITER_EQ ||
iterator_type == ITER_GE ||
iterator_type == ITER_LT);
- itr->curr_pos =
- vy_mem_tree_lower_bound(&itr->mem->tree,
- &tree_key, &exact);
+ itr->curr_pos = vy_mem_tree_lower_bound(
+ &itr->mem->tree, &tree_key, &exact);
}
} else if (iterator_type == ITER_LE) {
itr->curr_pos = vy_mem_tree_invalid_iterator();
@@ -422,8 +420,8 @@ vy_mem_iterator_seek(struct vy_mem_iterator *itr, struct vy_entry last)
vy_mem_tree_iterator_prev(&itr->mem->tree, &itr->curr_pos);
if (vy_mem_tree_iterator_is_invalid(&itr->curr_pos))
return 1;
- itr->curr = *vy_mem_tree_iterator_get_elem(&itr->mem->tree,
- &itr->curr_pos);
+ itr->curr =
+ *vy_mem_tree_iterator_get_elem(&itr->mem->tree, &itr->curr_pos);
if (itr->iterator_type == ITER_EQ &&
((last.stmt == NULL && !exact) ||
(last.stmt != NULL &&
@@ -439,9 +437,10 @@ vy_mem_iterator_seek(struct vy_mem_iterator *itr, struct vy_entry last)
/* {{{ vy_mem_iterator API implementation */
void
-vy_mem_iterator_open(struct vy_mem_iterator *itr, struct vy_mem_iterator_stat *stat,
- struct vy_mem *mem, enum iterator_type iterator_type,
- struct vy_entry key, const struct vy_read_view **rv)
+vy_mem_iterator_open(struct vy_mem_iterator *itr,
+ struct vy_mem_iterator_stat *stat, struct vy_mem *mem,
+ enum iterator_type iterator_type, struct vy_entry key,
+ const struct vy_read_view **rv)
{
itr->stat = stat;
@@ -472,9 +471,9 @@ vy_mem_iterator_next_key(struct vy_mem_iterator *itr)
return 1;
assert(itr->mem->version == itr->version);
assert(!vy_mem_tree_iterator_is_invalid(&itr->curr_pos));
- assert(vy_entry_is_equal(itr->curr,
- *vy_mem_tree_iterator_get_elem(&itr->mem->tree,
- &itr->curr_pos)));
+ assert(vy_entry_is_equal(
+ itr->curr, *vy_mem_tree_iterator_get_elem(&itr->mem->tree,
+ &itr->curr_pos)));
struct key_def *cmp_def = itr->mem->cmp_def;
struct vy_entry prev = itr->curr;
@@ -512,9 +511,9 @@ vy_mem_iterator_next_lsn(struct vy_mem_iterator *itr)
return 1;
assert(itr->mem->version == itr->version);
assert(!vy_mem_tree_iterator_is_invalid(&itr->curr_pos));
- assert(vy_entry_is_equal(itr->curr,
- *vy_mem_tree_iterator_get_elem(&itr->mem->tree,
- &itr->curr_pos)));
+ assert(vy_entry_is_equal(
+ itr->curr, *vy_mem_tree_iterator_get_elem(&itr->mem->tree,
+ &itr->curr_pos)));
struct key_def *cmp_def = itr->mem->cmp_def;
struct vy_mem_tree_iterator next_pos = itr->curr_pos;
@@ -555,8 +554,7 @@ vy_mem_iterator_get_history(struct vy_mem_iterator *itr,
}
NODISCARD int
-vy_mem_iterator_next(struct vy_mem_iterator *itr,
- struct vy_history *history)
+vy_mem_iterator_next(struct vy_mem_iterator *itr, struct vy_history *history)
{
vy_history_cleanup(history);
if (vy_mem_iterator_next_key(itr) == 0)
@@ -577,7 +575,9 @@ vy_mem_iterator_skip(struct vy_mem_iterator *itr, struct vy_entry last,
if (itr->search_started &&
(itr->curr.stmt == NULL || last.stmt == NULL ||
iterator_direction(itr->iterator_type) *
- vy_entry_compare(itr->curr, last, itr->mem->cmp_def) > 0))
+ vy_entry_compare(itr->curr, last,
+ itr->mem->cmp_def) >
+ 0))
return 0;
vy_history_cleanup(history);
@@ -614,9 +614,8 @@ vy_mem_stream_next(struct vy_stmt_stream *virt_stream, struct vy_entry *ret)
assert(virt_stream->iface->next == vy_mem_stream_next);
struct vy_mem_stream *stream = (struct vy_mem_stream *)virt_stream;
- struct vy_entry *res =
- vy_mem_tree_iterator_get_elem(&stream->mem->tree,
- &stream->curr_pos);
+ struct vy_entry *res = vy_mem_tree_iterator_get_elem(&stream->mem->tree,
+ &stream->curr_pos);
if (res == NULL) {
*ret = vy_entry_none();
} else {
diff --git a/src/box/vy_mem.h b/src/box/vy_mem.h
index 4f06c75..0591ad8 100644
--- a/src/box/vy_mem.h
+++ b/src/box/vy_mem.h
@@ -87,8 +87,7 @@ struct vy_mem_tree_key {
* Internal. Extracted to speed up BPS tree.
*/
static int
-vy_mem_tree_cmp(struct vy_entry a, struct vy_entry b,
- struct key_def *cmp_def)
+vy_mem_tree_cmp(struct vy_entry a, struct vy_entry b, struct key_def *cmp_def)
{
int res = vy_entry_compare(a, b, cmp_def);
if (res)
@@ -370,9 +369,10 @@ struct vy_mem_iterator {
* Open an iterator over in-memory tree.
*/
void
-vy_mem_iterator_open(struct vy_mem_iterator *itr, struct vy_mem_iterator_stat *stat,
- struct vy_mem *mem, enum iterator_type iterator_type,
- struct vy_entry key, const struct vy_read_view **rv);
+vy_mem_iterator_open(struct vy_mem_iterator *itr,
+ struct vy_mem_iterator_stat *stat, struct vy_mem *mem,
+ enum iterator_type iterator_type, struct vy_entry key,
+ const struct vy_read_view **rv);
/**
* Advance a mem iterator to the next key.
@@ -380,8 +380,7 @@ vy_mem_iterator_open(struct vy_mem_iterator *itr, struct vy_mem_iterator_stat *s
* Returns 0 on success, -1 on memory allocation error.
*/
NODISCARD int
-vy_mem_iterator_next(struct vy_mem_iterator *itr,
- struct vy_history *history);
+vy_mem_iterator_next(struct vy_mem_iterator *itr, struct vy_history *history);
/**
* Advance a mem iterator to the key following @last.
diff --git a/src/box/vy_point_lookup.c b/src/box/vy_point_lookup.c
index 80b5c59..51e4645 100644
--- a/src/box/vy_point_lookup.c
+++ b/src/box/vy_point_lookup.c
@@ -58,8 +58,7 @@ vy_point_lookup_scan_txw(struct vy_lsm *lsm, struct vy_tx *tx,
if (tx == NULL)
return 0;
lsm->stat.txw.iterator.lookup++;
- struct txv *txv =
- write_set_search_key(&tx->write_set, lsm, key);
+ struct txv *txv = write_set_search_key(&tx->write_set, lsm, key);
assert(txv == NULL || txv->lsm == lsm);
if (txv == NULL)
return 0;
@@ -92,19 +91,18 @@ vy_point_lookup_scan_cache(struct vy_lsm *lsm, const struct vy_read_view **rv,
*/
static int
vy_point_lookup_scan_mem(struct vy_lsm *lsm, struct vy_mem *mem,
- const struct vy_read_view **rv,
- struct vy_entry key, struct vy_history *history)
+ const struct vy_read_view **rv, struct vy_entry key,
+ struct vy_history *history)
{
struct vy_mem_iterator mem_itr;
- vy_mem_iterator_open(&mem_itr, &lsm->stat.memory.iterator,
- mem, ITER_EQ, key, rv);
+ vy_mem_iterator_open(&mem_itr, &lsm->stat.memory.iterator, mem, ITER_EQ,
+ key, rv);
struct vy_history mem_history;
vy_history_create(&mem_history, &lsm->env->history_node_pool);
int rc = vy_mem_iterator_next(&mem_itr, &mem_history);
vy_history_splice(history, &mem_history);
vy_mem_iterator_close(&mem_itr);
return rc;
-
}
/**
@@ -142,8 +140,8 @@ vy_point_lookup_scan_slice(struct vy_lsm *lsm, struct vy_slice *slice,
* format in vy_mem.
*/
struct vy_run_iterator run_itr;
- vy_run_iterator_open(&run_itr, &lsm->stat.disk.iterator, slice,
- ITER_EQ, key, rv, lsm->cmp_def, lsm->key_def,
+ vy_run_iterator_open(&run_itr, &lsm->stat.disk.iterator, slice, ITER_EQ,
+ key, rv, lsm->cmp_def, lsm->key_def,
lsm->disk_format);
struct vy_history slice_history;
vy_history_create(&slice_history, &lsm->env->history_node_pool);
@@ -163,14 +161,13 @@ static int
vy_point_lookup_scan_slices(struct vy_lsm *lsm, const struct vy_read_view **rv,
struct vy_entry key, struct vy_history *history)
{
- struct vy_range *range = vy_range_tree_find_by_key(&lsm->range_tree,
- ITER_EQ, key);
+ struct vy_range *range =
+ vy_range_tree_find_by_key(&lsm->range_tree, ITER_EQ, key);
assert(range != NULL);
int slice_count = range->slice_count;
size_t size;
- struct vy_slice **slices =
- region_alloc_array(&fiber()->gc, typeof(slices[0]), slice_count,
- &size);
+ struct vy_slice **slices = region_alloc_array(
+ &fiber()->gc, typeof(slices[0]), slice_count, &size);
if (slices == NULL) {
diag_set(OutOfMemory, size, "region_alloc_array", "slices");
return -1;
@@ -185,8 +182,8 @@ vy_point_lookup_scan_slices(struct vy_lsm *lsm, const struct vy_read_view **rv,
int rc = 0;
for (i = 0; i < slice_count; i++) {
if (rc == 0 && !vy_history_is_terminal(history))
- rc = vy_point_lookup_scan_slice(lsm, slices[i],
- rv, key, history);
+ rc = vy_point_lookup_scan_slice(lsm, slices[i], rv, key,
+ history);
vy_slice_unpin(slices[i]);
}
return rc;
@@ -194,8 +191,8 @@ vy_point_lookup_scan_slices(struct vy_lsm *lsm, const struct vy_read_view **rv,
int
vy_point_lookup(struct vy_lsm *lsm, struct vy_tx *tx,
- const struct vy_read_view **rv,
- struct vy_entry key, struct vy_entry *ret)
+ const struct vy_read_view **rv, struct vy_entry key,
+ struct vy_entry *ret)
{
/* All key parts must be set for a point lookup. */
assert(vy_stmt_is_full_key(key.stmt, lsm->cmp_def));
@@ -284,8 +281,8 @@ done:
if (rc == 0) {
int upserts_applied;
- rc = vy_history_apply(&history, lsm->cmp_def,
- false, &upserts_applied, ret);
+ rc = vy_history_apply(&history, lsm->cmp_def, false,
+ &upserts_applied, ret);
lsm->stat.upsert.applied += upserts_applied;
}
vy_history_cleanup(&history);
@@ -319,8 +316,8 @@ vy_point_lookup_mem(struct vy_lsm *lsm, const struct vy_read_view **rv,
done:
if (rc == 0) {
int upserts_applied;
- rc = vy_history_apply(&history, lsm->cmp_def,
- true, &upserts_applied, ret);
+ rc = vy_history_apply(&history, lsm->cmp_def, true,
+ &upserts_applied, ret);
lsm->stat.upsert.applied += upserts_applied;
}
out:
diff --git a/src/box/vy_point_lookup.h b/src/box/vy_point_lookup.h
index b4092ee..b33312c 100644
--- a/src/box/vy_point_lookup.h
+++ b/src/box/vy_point_lookup.h
@@ -67,8 +67,8 @@ struct vy_read_view;
*/
int
vy_point_lookup(struct vy_lsm *lsm, struct vy_tx *tx,
- const struct vy_read_view **rv,
- struct vy_entry key, struct vy_entry *ret);
+ const struct vy_read_view **rv, struct vy_entry key,
+ struct vy_entry *ret);
/**
* Look up a tuple by key in memory.
diff --git a/src/box/vy_quota.c b/src/box/vy_quota.c
index f1ac8dd..3c7078e 100644
--- a/src/box/vy_quota.c
+++ b/src/box/vy_quota.c
@@ -54,8 +54,7 @@ static const double VY_QUOTA_TIMER_PERIOD = 0.1;
/**
* Bit mask of resources used by a particular consumer type.
*/
-static unsigned
-vy_quota_consumer_resource_map[] = {
+static unsigned vy_quota_consumer_resource_map[] = {
/**
* Transaction throttling pursues two goals. First, it is
* capping memory consumption rate so that the hard memory
@@ -100,7 +99,7 @@ vy_rate_limit_is_applicable(enum vy_quota_consumer_type consumer_type,
enum vy_quota_resource_type resource_type)
{
return (vy_quota_consumer_resource_map[consumer_type] &
- (1 << resource_type)) != 0;
+ (1 << resource_type)) != 0;
}
/**
@@ -300,8 +299,8 @@ vy_quota_release(struct vy_quota *q, size_t size)
}
int
-vy_quota_use(struct vy_quota *q, enum vy_quota_consumer_type type,
- size_t size, double timeout)
+vy_quota_use(struct vy_quota *q, enum vy_quota_consumer_type type, size_t size,
+ double timeout)
{
/*
* Fail early if the configured memory limit never allows
@@ -342,8 +341,8 @@ vy_quota_use(struct vy_quota *q, enum vy_quota_consumer_type type,
double wait_time = ev_monotonic_now(loop()) - wait_start;
if (wait_time > q->too_long_threshold) {
say_warn_ratelimited("waited for %zu bytes of vinyl memory "
- "quota for too long: %.3f sec", size,
- wait_time);
+ "quota for too long: %.3f sec",
+ size, wait_time);
}
vy_quota_do_use(q, type, size);
diff --git a/src/box/vy_quota.h b/src/box/vy_quota.h
index bd7d4e0..15507aa 100644
--- a/src/box/vy_quota.h
+++ b/src/box/vy_quota.h
@@ -107,8 +107,7 @@ vy_rate_limit_refill(struct vy_rate_limit *rl, double time)
rl->value = MIN((ssize_t)value, SSIZE_MAX);
}
-typedef void
-(*vy_quota_exceeded_f)(struct vy_quota *quota);
+typedef void (*vy_quota_exceeded_f)(struct vy_quota *quota);
/**
* Apart from memory usage accounting and limiting, vy_quota is
@@ -311,8 +310,8 @@ vy_quota_release(struct vy_quota *q, size_t size);
* account while estimating the size of a memory allocation.
*/
int
-vy_quota_use(struct vy_quota *q, enum vy_quota_consumer_type type,
- size_t size, double timeout);
+vy_quota_use(struct vy_quota *q, enum vy_quota_consumer_type type, size_t size,
+ double timeout);
/**
* Adjust quota after allocating memory.
diff --git a/src/box/vy_range.c b/src/box/vy_range.c
index 4ff8521..586b3ef 100644
--- a/src/box/vy_range.c
+++ b/src/box/vy_range.c
@@ -78,8 +78,7 @@ vy_range_tree_key_cmp(struct vy_entry entry, struct vy_range *range)
struct vy_range *
vy_range_tree_find_by_key(vy_range_tree_t *tree,
- enum iterator_type iterator_type,
- struct vy_entry key)
+ enum iterator_type iterator_type, struct vy_entry key)
{
if (vy_stmt_is_empty_key(key.stmt)) {
switch (iterator_type) {
@@ -180,8 +179,8 @@ vy_range_new(int64_t id, struct vy_entry begin, struct vy_entry end,
{
struct vy_range *range = calloc(1, sizeof(*range));
if (range == NULL) {
- diag_set(OutOfMemory, sizeof(*range),
- "malloc", "struct vy_range");
+ diag_set(OutOfMemory, sizeof(*range), "malloc",
+ "struct vy_range");
return NULL;
}
range->id = id;
@@ -429,8 +428,8 @@ void
vy_range_update_dumps_per_compaction(struct vy_range *range)
{
if (!rlist_empty(&range->slices)) {
- struct vy_slice *slice = rlist_last_entry(&range->slices,
- struct vy_slice, in_range);
+ struct vy_slice *slice = rlist_last_entry(
+ &range->slices, struct vy_slice, in_range);
range->dumps_per_compaction = slice->run->dump_count;
} else {
range->dumps_per_compaction = 0;
@@ -470,12 +469,13 @@ vy_range_needs_split(struct vy_range *range, int64_t range_size,
/* Find the median key in the oldest run (approximately). */
struct vy_page_info *mid_page;
- mid_page = vy_run_page_info(slice->run, slice->first_page_no +
- (slice->last_page_no -
- slice->first_page_no) / 2);
+ mid_page = vy_run_page_info(
+ slice->run,
+ slice->first_page_no +
+ (slice->last_page_no - slice->first_page_no) / 2);
- struct vy_page_info *first_page = vy_run_page_info(slice->run,
- slice->first_page_no);
+ struct vy_page_info *first_page =
+ vy_run_page_info(slice->run, slice->first_page_no);
/* No point in splitting if a new range is going to be empty. */
if (key_compare(first_page->min_key, first_page->min_key_hint,
diff --git a/src/box/vy_range.h b/src/box/vy_range.h
index 2eb843b..0a6d62b 100644
--- a/src/box/vy_range.h
+++ b/src/box/vy_range.h
@@ -171,8 +171,8 @@ vy_range_tree_key_cmp(struct vy_entry entry, struct vy_range *range);
typedef rb_tree(struct vy_range) vy_range_tree_t;
rb_gen_ext_key(MAYBE_UNUSED static inline, vy_range_tree_, vy_range_tree_t,
- struct vy_range, tree_node, vy_range_tree_cmp,
- struct vy_entry, vy_range_tree_key_cmp);
+ struct vy_range, tree_node, vy_range_tree_cmp, struct vy_entry,
+ vy_range_tree_key_cmp);
/**
* Find the first range in which a given key should be looked up.
diff --git a/src/box/vy_read_iterator.c b/src/box/vy_read_iterator.c
index 4097969..b35e90c 100644
--- a/src/box/vy_read_iterator.c
+++ b/src/box/vy_read_iterator.c
@@ -70,8 +70,8 @@ vy_read_iterator_reserve(struct vy_read_iterator *itr, uint32_t capacity)
return 0;
struct vy_read_src *new_src = calloc(capacity, sizeof(*new_src));
if (new_src == NULL) {
- diag_set(OutOfMemory, capacity * sizeof(*new_src),
- "calloc", "new_src");
+ diag_set(OutOfMemory, capacity * sizeof(*new_src), "calloc",
+ "new_src");
return -1;
}
memcpy(new_src, itr->src, itr->src_count * sizeof(*new_src));
@@ -148,15 +148,15 @@ vy_read_iterator_range_is_done(struct vy_read_iterator *itr,
int dir = iterator_direction(itr->iterator_type);
if (dir > 0 && range->end.stmt != NULL &&
- (next.stmt == NULL || vy_entry_compare(next, range->end,
- cmp_def) >= 0) &&
+ (next.stmt == NULL ||
+ vy_entry_compare(next, range->end, cmp_def) >= 0) &&
(itr->iterator_type != ITER_EQ ||
vy_entry_compare(itr->key, range->end, cmp_def) >= 0))
return true;
if (dir < 0 && range->begin.stmt != NULL &&
- (next.stmt == NULL || vy_entry_compare(next, range->begin,
- cmp_def) < 0) &&
+ (next.stmt == NULL ||
+ vy_entry_compare(next, range->begin, cmp_def) < 0) &&
(itr->iterator_type != ITER_REQ ||
vy_entry_compare(itr->key, range->begin, cmp_def) <= 0))
return true;
@@ -175,8 +175,8 @@ vy_read_iterator_range_is_done(struct vy_read_iterator *itr,
* NULL denotes the statement following the last one.
*/
static inline int
-vy_read_iterator_cmp_stmt(struct vy_read_iterator *itr,
- struct vy_entry a, struct vy_entry b)
+vy_read_iterator_cmp_stmt(struct vy_read_iterator *itr, struct vy_entry a,
+ struct vy_entry b)
{
if (a.stmt == NULL && b.stmt != NULL)
return 1;
@@ -185,7 +185,7 @@ vy_read_iterator_cmp_stmt(struct vy_read_iterator *itr,
if (a.stmt == NULL && b.stmt == NULL)
return 0;
return iterator_direction(itr->iterator_type) *
- vy_entry_compare(a, b, itr->lsm->cmp_def);
+ vy_entry_compare(a, b, itr->lsm->cmp_def);
}
/**
@@ -205,10 +205,10 @@ vy_read_iterator_is_exact_match(struct vy_read_iterator *itr,
* in case the key is found in memory.
*/
return itr->last.stmt == NULL && entry.stmt != NULL &&
- (type == ITER_EQ || type == ITER_REQ ||
- type == ITER_GE || type == ITER_LE) &&
- vy_stmt_is_full_key(itr->key.stmt, cmp_def) &&
- vy_entry_compare(entry, itr->key, cmp_def) == 0;
+ (type == ITER_EQ || type == ITER_REQ || type == ITER_GE ||
+ type == ITER_LE) &&
+ vy_stmt_is_full_key(itr->key.stmt, cmp_def) &&
+ vy_entry_compare(entry, itr->key, cmp_def) == 0;
}
/**
@@ -220,8 +220,8 @@ vy_read_iterator_is_exact_match(struct vy_read_iterator *itr,
*/
static void
vy_read_iterator_evaluate_src(struct vy_read_iterator *itr,
- struct vy_read_src *src,
- struct vy_entry *next, bool *stop)
+ struct vy_read_src *src, struct vy_entry *next,
+ bool *stop)
{
uint32_t src_id = src - itr->src;
struct vy_entry entry = vy_history_last_stmt(&src->history);
@@ -268,8 +268,8 @@ vy_read_iterator_evaluate_src(struct vy_read_iterator *itr,
*/
static NODISCARD int
-vy_read_iterator_scan_txw(struct vy_read_iterator *itr,
- struct vy_entry *next, bool *stop)
+vy_read_iterator_scan_txw(struct vy_read_iterator *itr, struct vy_entry *next,
+ bool *stop)
{
struct vy_read_src *src = &itr->src[itr->txw_src];
struct vy_txw_iterator *src_itr = &src->txw_iterator;
@@ -297,19 +297,20 @@ vy_read_iterator_scan_txw(struct vy_read_iterator *itr,
}
static NODISCARD int
-vy_read_iterator_scan_cache(struct vy_read_iterator *itr,
- struct vy_entry *next, bool *stop)
+vy_read_iterator_scan_cache(struct vy_read_iterator *itr, struct vy_entry *next,
+ bool *stop)
{
bool is_interval = false;
struct vy_read_src *src = &itr->src[itr->cache_src];
struct vy_cache_iterator *src_itr = &src->cache_iterator;
- int rc = vy_cache_iterator_restore(src_itr, itr->last,
- &src->history, &is_interval);
+ int rc = vy_cache_iterator_restore(src_itr, itr->last, &src->history,
+ &is_interval);
if (rc == 0) {
if (!src->is_started || itr->cache_src >= itr->skipped_src) {
rc = vy_cache_iterator_skip(src_itr, itr->last,
- &src->history, &is_interval);
+ &src->history,
+ &is_interval);
} else if (src->front_id == itr->prev_front_id) {
rc = vy_cache_iterator_next(src_itr, &src->history,
&is_interval);
@@ -365,8 +366,7 @@ vy_read_iterator_scan_disk(struct vy_read_iterator *itr, uint32_t disk_src,
assert(disk_src >= itr->disk_src && disk_src < itr->src_count);
if (!src->is_started || disk_src >= itr->skipped_src)
- rc = vy_run_iterator_skip(src_itr, itr->last,
- &src->history);
+ rc = vy_run_iterator_skip(src_itr, itr->last, &src->history);
else if (src->front_id == itr->prev_front_id)
rc = vy_run_iterator_next(src_itr, &src->history);
src->is_started = true;
@@ -391,8 +391,8 @@ vy_read_iterator_next_range(struct vy_read_iterator *itr);
static NODISCARD int
vy_read_iterator_advance(struct vy_read_iterator *itr)
{
- if (itr->last.stmt != NULL && (itr->iterator_type == ITER_EQ ||
- itr->iterator_type == ITER_REQ) &&
+ if (itr->last.stmt != NULL &&
+ (itr->iterator_type == ITER_EQ || itr->iterator_type == ITER_REQ) &&
vy_stmt_is_full_key(itr->key.stmt, itr->lsm->cmp_def)) {
/*
* There may be one statement at max satisfying
@@ -507,7 +507,7 @@ done:
* and respects statement order.
*/
if (itr->last.stmt != NULL && next.stmt != NULL) {
- assert(vy_read_iterator_cmp_stmt(itr, next, itr->last) > 0);
+ assert(vy_read_iterator_cmp_stmt(itr, next, itr->last) > 0);
}
#endif
if (itr->need_check_eq && next.stmt != NULL &&
@@ -520,8 +520,8 @@ static void
vy_read_iterator_add_tx(struct vy_read_iterator *itr)
{
assert(itr->tx != NULL);
- enum iterator_type iterator_type = (itr->iterator_type != ITER_REQ ?
- itr->iterator_type : ITER_LE);
+ enum iterator_type iterator_type =
+ (itr->iterator_type != ITER_REQ ? itr->iterator_type : ITER_LE);
struct vy_txw_iterator_stat *stat = &itr->lsm->stat.txw.iterator;
struct vy_read_src *sub_src = vy_read_iterator_add_src(itr);
vy_txw_iterator_open(&sub_src->txw_iterator, stat, itr->tx, itr->lsm,
@@ -531,19 +531,18 @@ vy_read_iterator_add_tx(struct vy_read_iterator *itr)
static void
vy_read_iterator_add_cache(struct vy_read_iterator *itr)
{
- enum iterator_type iterator_type = (itr->iterator_type != ITER_REQ ?
- itr->iterator_type : ITER_LE);
+ enum iterator_type iterator_type =
+ (itr->iterator_type != ITER_REQ ? itr->iterator_type : ITER_LE);
struct vy_read_src *sub_src = vy_read_iterator_add_src(itr);
- vy_cache_iterator_open(&sub_src->cache_iterator,
- &itr->lsm->cache, iterator_type,
- itr->key, itr->read_view);
+ vy_cache_iterator_open(&sub_src->cache_iterator, &itr->lsm->cache,
+ iterator_type, itr->key, itr->read_view);
}
static void
vy_read_iterator_add_mem(struct vy_read_iterator *itr)
{
- enum iterator_type iterator_type = (itr->iterator_type != ITER_REQ ?
- itr->iterator_type : ITER_LE);
+ enum iterator_type iterator_type =
+ (itr->iterator_type != ITER_REQ ? itr->iterator_type : ITER_LE);
struct vy_lsm *lsm = itr->lsm;
struct vy_read_src *sub_src;
@@ -557,9 +556,8 @@ vy_read_iterator_add_mem(struct vy_read_iterator *itr)
rlist_foreach_entry(mem, &lsm->sealed, in_sealed) {
sub_src = vy_read_iterator_add_src(itr);
vy_mem_iterator_open(&sub_src->mem_iterator,
- &lsm->stat.memory.iterator,
- mem, iterator_type, itr->key,
- itr->read_view);
+ &lsm->stat.memory.iterator, mem,
+ iterator_type, itr->key, itr->read_view);
}
}
@@ -567,8 +565,8 @@ static void
vy_read_iterator_add_disk(struct vy_read_iterator *itr)
{
assert(itr->curr_range != NULL);
- enum iterator_type iterator_type = (itr->iterator_type != ITER_REQ ?
- itr->iterator_type : ITER_LE);
+ enum iterator_type iterator_type =
+ (itr->iterator_type != ITER_REQ ? itr->iterator_type : ITER_LE);
struct vy_lsm *lsm = itr->lsm;
struct vy_slice *slice;
/*
@@ -580,9 +578,9 @@ vy_read_iterator_add_disk(struct vy_read_iterator *itr)
struct vy_read_src *sub_src = vy_read_iterator_add_src(itr);
vy_run_iterator_open(&sub_src->run_iterator,
&lsm->stat.disk.iterator, slice,
- iterator_type, itr->key,
- itr->read_view, lsm->cmp_def,
- lsm->key_def, lsm->disk_format);
+ iterator_type, itr->key, itr->read_view,
+ lsm->cmp_def, lsm->key_def,
+ lsm->disk_format);
}
}
@@ -648,7 +646,8 @@ vy_read_iterator_open(struct vy_read_iterator *itr, struct vy_lsm *lsm,
* in this case.
*/
itr->iterator_type = iterator_direction(iterator_type) > 0 ?
- ITER_GE : ITER_LE;
+ ITER_GE :
+ ITER_LE;
}
if (iterator_type == ITER_ALL)
@@ -664,7 +663,6 @@ vy_read_iterator_open(struct vy_read_iterator *itr, struct vy_lsm *lsm,
*/
itr->need_check_eq = true;
}
-
}
/**
@@ -681,10 +679,9 @@ vy_read_iterator_restore(struct vy_read_iterator *itr)
itr->mem_list_version = itr->lsm->mem_list_version;
itr->range_tree_version = itr->lsm->range_tree_version;
- itr->curr_range = vy_range_tree_find_by_key(&itr->lsm->range_tree,
- itr->iterator_type,
- itr->last.stmt != NULL ?
- itr->last : itr->key);
+ itr->curr_range = vy_range_tree_find_by_key(
+ &itr->lsm->range_tree, itr->iterator_type,
+ itr->last.stmt != NULL ? itr->last : itr->key);
itr->range_version = itr->curr_range->version;
if (itr->tx != NULL) {
@@ -714,9 +711,10 @@ vy_read_iterator_next_range(struct vy_read_iterator *itr)
assert(range != NULL);
while (true) {
- range = dir > 0 ?
- vy_range_tree_next(&itr->lsm->range_tree, range) :
- vy_range_tree_prev(&itr->lsm->range_tree, range);
+ range = dir > 0 ? vy_range_tree_next(&itr->lsm->range_tree,
+ range) :
+ vy_range_tree_prev(&itr->lsm->range_tree,
+ range);
assert(range != NULL);
if (itr->last.stmt == NULL)
@@ -725,13 +723,13 @@ vy_read_iterator_next_range(struct vy_read_iterator *itr)
* We could skip an entire range due to the cache.
* Make sure the next statement falls in the range.
*/
- if (dir > 0 && (range->end.stmt == NULL ||
- vy_entry_compare(itr->last, range->end,
- cmp_def) < 0))
+ if (dir > 0 &&
+ (range->end.stmt == NULL ||
+ vy_entry_compare(itr->last, range->end, cmp_def) < 0))
break;
- if (dir < 0 && (range->begin.stmt == NULL ||
- vy_entry_compare(itr->last, range->begin,
- cmp_def) > 0))
+ if (dir < 0 &&
+ (range->begin.stmt == NULL ||
+ vy_entry_compare(itr->last, range->begin, cmp_def) > 0))
break;
}
itr->curr_range = range;
@@ -768,8 +766,8 @@ vy_read_iterator_apply_history(struct vy_read_iterator *itr,
}
int upserts_applied = 0;
- int rc = vy_history_apply(&history, lsm->cmp_def,
- true, &upserts_applied, ret);
+ int rc = vy_history_apply(&history, lsm->cmp_def, true,
+ &upserts_applied, ret);
lsm->stat.upsert.applied += upserts_applied;
vy_history_cleanup(&history);
@@ -787,18 +785,18 @@ vy_read_iterator_track_read(struct vy_read_iterator *itr, struct vy_entry entry)
if (entry.stmt == NULL) {
entry = (itr->iterator_type == ITER_EQ ||
- itr->iterator_type == ITER_REQ ?
- itr->key : itr->lsm->env->empty_key);
+ itr->iterator_type == ITER_REQ ?
+ itr->key :
+ itr->lsm->env->empty_key);
}
int rc;
if (iterator_direction(itr->iterator_type) >= 0) {
rc = vy_tx_track(itr->tx, itr->lsm, itr->key,
- itr->iterator_type != ITER_GT,
- entry, true);
+ itr->iterator_type != ITER_GT, entry, true);
} else {
- rc = vy_tx_track(itr->tx, itr->lsm, entry, true,
- itr->key, itr->iterator_type != ITER_LT);
+ rc = vy_tx_track(itr->tx, itr->lsm, entry, true, itr->key,
+ itr->iterator_type != ITER_LT);
}
return rc;
}
@@ -853,8 +851,8 @@ vy_read_iterator_cache_add(struct vy_read_iterator *itr, struct vy_entry entry)
itr->last_cached = vy_entry_none();
return;
}
- vy_cache_add(&itr->lsm->cache, entry, itr->last_cached,
- itr->key, itr->iterator_type);
+ vy_cache_add(&itr->lsm->cache, entry, itr->last_cached, itr->key,
+ itr->iterator_type);
if (entry.stmt != NULL)
tuple_ref(entry.stmt);
if (itr->last_cached.stmt != NULL)
diff --git a/src/box/vy_read_set.c b/src/box/vy_read_set.c
index 431b24f..58a67c8 100644
--- a/src/box/vy_read_set.c
+++ b/src/box/vy_read_set.c
@@ -118,8 +118,7 @@ vy_tx_conflict_iterator_next(struct vy_tx_conflict_iterator *it)
assert(left == NULL || left->lsm == curr->lsm);
assert(right == NULL || right->lsm == curr->lsm);
- int cmp_right = vy_entry_compare(it->key, last->right,
- cmp_def);
+ int cmp_right = vy_entry_compare(it->key, last->right, cmp_def);
if (cmp_right == 0 && !last->right_belongs)
cmp_right = 1;
@@ -138,8 +137,8 @@ vy_tx_conflict_iterator_next(struct vy_tx_conflict_iterator *it)
/* Optimize comparison out. */
cmp_left = cmp_right;
} else {
- cmp_left = vy_entry_compare(it->key, curr->left,
- cmp_def);
+ cmp_left =
+ vy_entry_compare(it->key, curr->left, cmp_def);
if (cmp_left == 0 && !curr->left_belongs)
cmp_left = -1;
}
@@ -166,8 +165,8 @@ vy_tx_conflict_iterator_next(struct vy_tx_conflict_iterator *it)
/* Optimize comparison out. */
cmp_right = cmp_left;
} else if (curr != last) {
- cmp_right = vy_entry_compare(it->key, curr->right,
- cmp_def);
+ cmp_right =
+ vy_entry_compare(it->key, curr->right, cmp_def);
if (cmp_right == 0 && !curr->right_belongs)
cmp_right = 1;
}
diff --git a/src/box/vy_regulator.c b/src/box/vy_regulator.c
index 8ec7e25..f4b1a4b 100644
--- a/src/box/vy_regulator.c
+++ b/src/box/vy_regulator.c
@@ -105,11 +105,11 @@ vy_regulator_trigger_dump(struct vy_regulator *regulator)
* write_rate dump_bandwidth
*/
struct vy_quota *quota = regulator->quota;
- size_t mem_left = (quota->used < quota->limit ?
- quota->limit - quota->used : 0);
+ size_t mem_left =
+ (quota->used < quota->limit ? quota->limit - quota->used : 0);
size_t mem_used = quota->used;
- size_t max_write_rate = (double)mem_left / (mem_used + 1) *
- regulator->dump_bandwidth;
+ size_t max_write_rate =
+ (double)mem_left / (mem_used + 1) * regulator->dump_bandwidth;
max_write_rate = MIN(max_write_rate, regulator->dump_bandwidth);
vy_quota_set_rate_limit(quota, VY_QUOTA_RESOURCE_MEMORY,
max_write_rate);
@@ -144,8 +144,8 @@ vy_regulator_update_write_rate(struct vy_regulator *regulator)
size_t rate_avg = regulator->write_rate;
size_t rate_curr = (used_curr - used_last) / VY_REGULATOR_TIMER_PERIOD;
- double weight = 1 - exp(-VY_REGULATOR_TIMER_PERIOD /
- VY_WRITE_RATE_AVG_WIN);
+ double weight =
+ 1 - exp(-VY_REGULATOR_TIMER_PERIOD / VY_WRITE_RATE_AVG_WIN);
rate_avg = (1 - weight) * rate_avg + weight * rate_curr;
regulator->write_rate = rate_avg;
@@ -178,15 +178,15 @@ vy_regulator_update_dump_watermark(struct vy_regulator *regulator)
*/
size_t write_rate = regulator->write_rate_max * 3 / 2;
regulator->dump_watermark =
- (double)quota->limit * regulator->dump_bandwidth /
- (regulator->dump_bandwidth + write_rate + 1);
+ (double)quota->limit * regulator->dump_bandwidth /
+ (regulator->dump_bandwidth + write_rate + 1);
/*
* It doesn't make sense to set the watermark below 50%
* of the memory limit because the write rate can exceed
* the dump bandwidth under no circumstances.
*/
- regulator->dump_watermark = MAX(regulator->dump_watermark,
- quota->limit / 2);
+ regulator->dump_watermark =
+ MAX(regulator->dump_watermark, quota->limit / 2);
}
static void
@@ -209,17 +209,17 @@ vy_regulator_create(struct vy_regulator *regulator, struct vy_quota *quota,
enum { KB = 1024, MB = KB * KB };
static int64_t dump_bandwidth_buckets[] = {
100 * KB, 200 * KB, 300 * KB, 400 * KB, 500 * KB, 600 * KB,
- 700 * KB, 800 * KB, 900 * KB, 1 * MB, 2 * MB, 3 * MB,
- 4 * MB, 5 * MB, 6 * MB, 7 * MB, 8 * MB, 9 * MB,
- 10 * MB, 15 * MB, 20 * MB, 25 * MB, 30 * MB, 35 * MB,
- 40 * MB, 45 * MB, 50 * MB, 55 * MB, 60 * MB, 65 * MB,
- 70 * MB, 75 * MB, 80 * MB, 85 * MB, 90 * MB, 95 * MB,
+ 700 * KB, 800 * KB, 900 * KB, 1 * MB, 2 * MB, 3 * MB,
+ 4 * MB, 5 * MB, 6 * MB, 7 * MB, 8 * MB, 9 * MB,
+ 10 * MB, 15 * MB, 20 * MB, 25 * MB, 30 * MB, 35 * MB,
+ 40 * MB, 45 * MB, 50 * MB, 55 * MB, 60 * MB, 65 * MB,
+ 70 * MB, 75 * MB, 80 * MB, 85 * MB, 90 * MB, 95 * MB,
100 * MB, 200 * MB, 300 * MB, 400 * MB, 500 * MB, 600 * MB,
700 * MB, 800 * MB, 900 * MB,
};
memset(regulator, 0, sizeof(*regulator));
- regulator->dump_bandwidth_hist = histogram_new(dump_bandwidth_buckets,
- lengthof(dump_bandwidth_buckets));
+ regulator->dump_bandwidth_hist = histogram_new(
+ dump_bandwidth_buckets, lengthof(dump_bandwidth_buckets));
if (regulator->dump_bandwidth_hist == NULL)
panic("failed to allocate dump bandwidth histogram");
@@ -262,8 +262,8 @@ vy_regulator_check_dump_watermark(struct vy_regulator *regulator)
}
void
-vy_regulator_dump_complete(struct vy_regulator *regulator,
- size_t mem_dumped, double dump_duration)
+vy_regulator_dump_complete(struct vy_regulator *regulator, size_t mem_dumped,
+ double dump_duration)
{
regulator->dump_in_progress = false;
@@ -430,7 +430,7 @@ vy_regulator_update_rate_limit(struct vy_regulator *regulator,
recent->compaction_time += compaction_time;
double rate = 0.75 * compaction_threads * recent->dump_input /
- recent->compaction_time;
+ recent->compaction_time;
/*
* We can't simply use (size_t)MIN(rate, SIZE_MAX) to cast
* the rate from double to size_t here, because on a 64-bit
diff --git a/src/box/vy_regulator.h b/src/box/vy_regulator.h
index 5131ac5..5ceeb34 100644
--- a/src/box/vy_regulator.h
+++ b/src/box/vy_regulator.h
@@ -45,8 +45,7 @@ struct histogram;
struct vy_quota;
struct vy_regulator;
-typedef int
-(*vy_trigger_dump_f)(struct vy_regulator *regulator);
+typedef int (*vy_trigger_dump_f)(struct vy_regulator *regulator);
/**
* The regulator is supposed to keep track of vinyl memory usage
@@ -153,8 +152,8 @@ vy_regulator_quota_exceeded(struct vy_regulator *regulator);
* Notify the regulator about memory dump completion.
*/
void
-vy_regulator_dump_complete(struct vy_regulator *regulator,
- size_t mem_dumped, double dump_duration);
+vy_regulator_dump_complete(struct vy_regulator *regulator, size_t mem_dumped,
+ double dump_duration);
/**
* Set memory limit and update the dump watermark accordingly.
diff --git a/src/box/vy_run.c b/src/box/vy_run.c
index b9822dc..4b6151d 100644
--- a/src/box/vy_run.c
+++ b/src/box/vy_run.c
@@ -45,18 +45,15 @@
#include "xrow.h"
#include "vy_history.h"
-static const uint64_t vy_page_info_key_map = (1 << VY_PAGE_INFO_OFFSET) |
- (1 << VY_PAGE_INFO_SIZE) |
- (1 << VY_PAGE_INFO_UNPACKED_SIZE) |
- (1 << VY_PAGE_INFO_ROW_COUNT) |
- (1 << VY_PAGE_INFO_MIN_KEY) |
- (1 << VY_PAGE_INFO_ROW_INDEX_OFFSET);
-
-static const uint64_t vy_run_info_key_map = (1 << VY_RUN_INFO_MIN_KEY) |
- (1 << VY_RUN_INFO_MAX_KEY) |
- (1 << VY_RUN_INFO_MIN_LSN) |
- (1 << VY_RUN_INFO_MAX_LSN) |
- (1 << VY_RUN_INFO_PAGE_COUNT);
+static const uint64_t vy_page_info_key_map =
+ (1 << VY_PAGE_INFO_OFFSET) | (1 << VY_PAGE_INFO_SIZE) |
+ (1 << VY_PAGE_INFO_UNPACKED_SIZE) | (1 << VY_PAGE_INFO_ROW_COUNT) |
+ (1 << VY_PAGE_INFO_MIN_KEY) | (1 << VY_PAGE_INFO_ROW_INDEX_OFFSET);
+
+static const uint64_t vy_run_info_key_map =
+ (1 << VY_RUN_INFO_MIN_KEY) | (1 << VY_RUN_INFO_MAX_KEY) |
+ (1 << VY_RUN_INFO_MIN_LSN) | (1 << VY_RUN_INFO_MAX_LSN) |
+ (1 << VY_RUN_INFO_PAGE_COUNT);
/** xlog meta type for .run files */
#define XLOG_META_TYPE_RUN "RUN"
@@ -65,10 +62,10 @@ static const uint64_t vy_run_info_key_map = (1 << VY_RUN_INFO_MIN_KEY) |
#define XLOG_META_TYPE_INDEX "INDEX"
const char *vy_file_suffix[] = {
- "index", /* VY_FILE_INDEX */
- "index" inprogress_suffix, /* VY_FILE_INDEX_INPROGRESS */
- "run", /* VY_FILE_RUN */
- "run" inprogress_suffix, /* VY_FILE_RUN_INPROGRESS */
+ "index", /* VY_FILE_INDEX */
+ "index" inprogress_suffix, /* VY_FILE_INDEX_INPROGRESS */
+ "run", /* VY_FILE_RUN */
+ "run" inprogress_suffix, /* VY_FILE_RUN_INPROGRESS */
};
/* sync run and index files very 16 MB */
@@ -127,8 +124,8 @@ vy_run_reader_f(va_list ap)
struct cbus_endpoint endpoint;
cpipe_create(&reader->tx_pipe, "tx_prio");
- cbus_endpoint_create(&endpoint, cord_name(cord()),
- fiber_schedule_cb, fiber());
+ cbus_endpoint_create(&endpoint, cord_name(cord()), fiber_schedule_cb,
+ fiber());
cbus_loop(&endpoint);
cbus_endpoint_destroy(&endpoint, cbus_process);
cpipe_destroy(&reader->tx_pipe);
@@ -142,8 +139,8 @@ vy_run_env_start_readers(struct vy_run_env *env)
assert(env->reader_pool == NULL);
assert(env->reader_pool_size > 0);
- env->reader_pool = calloc(env->reader_pool_size,
- sizeof(*env->reader_pool));
+ env->reader_pool =
+ calloc(env->reader_pool_size, sizeof(*env->reader_pool));
if (env->reader_pool == NULL)
panic("failed to allocate vinyl reader thread pool");
@@ -152,8 +149,8 @@ vy_run_env_start_readers(struct vy_run_env *env)
char name[FIBER_NAME_MAX];
snprintf(name, sizeof(name), "vinyl.reader.%d", i);
- if (cord_costart(&reader->cord, name,
- vy_run_reader_f, reader) != 0)
+ if (cord_costart(&reader->cord, name, vy_run_reader_f,
+ reader) != 0)
panic("failed to start vinyl reader thread");
cpipe_create(&reader->reader_pipe, name);
}
@@ -226,8 +223,8 @@ vy_run_env_coio_call(struct vy_run_env *env, struct cbus_call_msg *msg,
/* Post the task to the reader thread. */
bool cancellable = fiber_set_cancellable(false);
- int rc = cbus_call(&reader->reader_pipe, &reader->tx_pipe,
- msg, func, NULL, TIMEOUT_INFINITY);
+ int rc = cbus_call(&reader->reader_pipe, &reader->tx_pipe, msg, func,
+ NULL, TIMEOUT_INFINITY);
fiber_set_cancellable(cancellable);
if (rc != 0)
return -1;
@@ -353,8 +350,8 @@ vy_page_index_find_page(struct vy_run *run, struct vy_entry key,
{
if (itype == ITER_EQ)
itype = ITER_GE; /* One day it'll become obsolete */
- assert(itype == ITER_GE || itype == ITER_GT ||
- itype == ITER_LE || itype == ITER_LT);
+ assert(itype == ITER_GE || itype == ITER_GT || itype == ITER_LE ||
+ itype == ITER_LT);
int dir = iterator_direction(itype);
*equal_key = false;
@@ -388,9 +385,8 @@ vy_page_index_find_page(struct vy_run *run, struct vy_entry key,
do {
int32_t mid = range[0] + (range[1] - range[0]) / 2;
struct vy_page_info *info = vy_run_page_info(run, mid);
- int cmp = vy_entry_compare_with_raw_key(key, info->min_key,
- info->min_key_hint,
- cmp_def);
+ int cmp = vy_entry_compare_with_raw_key(
+ key, info->min_key, info->min_key_hint, cmp_def);
if (is_lower_bound)
range[cmp <= 0] = mid;
else
@@ -417,8 +413,8 @@ vy_slice_new(int64_t id, struct vy_run *run, struct vy_entry begin,
{
struct vy_slice *slice = malloc(sizeof(*slice));
if (slice == NULL) {
- diag_set(OutOfMemory, sizeof(*slice),
- "malloc", "struct vy_slice");
+ diag_set(OutOfMemory, sizeof(*slice), "malloc",
+ "struct vy_slice");
return NULL;
}
memset(slice, 0, sizeof(*slice));
@@ -444,17 +440,15 @@ vy_slice_new(int64_t id, struct vy_run *run, struct vy_entry begin,
if (slice->begin.stmt == NULL) {
slice->first_page_no = 0;
} else {
- slice->first_page_no =
- vy_page_index_find_page(run, slice->begin, cmp_def,
- ITER_GE, &unused);
+ slice->first_page_no = vy_page_index_find_page(
+ run, slice->begin, cmp_def, ITER_GE, &unused);
assert(slice->first_page_no < run->info.page_count);
}
if (slice->end.stmt == NULL) {
slice->last_page_no = run->info.page_count - 1;
} else {
- slice->last_page_no =
- vy_page_index_find_page(run, slice->end, cmp_def,
- ITER_LT, &unused);
+ slice->last_page_no = vy_page_index_find_page(
+ run, slice->end, cmp_def, ITER_LT, &unused);
if (slice->last_page_no == run->info.page_count) {
/* It's an empty slice */
slice->first_page_no = 0;
@@ -467,10 +461,10 @@ vy_slice_new(int64_t id, struct vy_run *run, struct vy_entry begin,
uint32_t run_pages = run->info.page_count;
uint32_t slice_pages = slice->last_page_no - slice->first_page_no + 1;
slice->count.pages = slice_pages;
- slice->count.rows = DIV_ROUND_UP(run->count.rows *
- slice_pages, run_pages);
- slice->count.bytes = DIV_ROUND_UP(run->count.bytes *
- slice_pages, run_pages);
+ slice->count.rows =
+ DIV_ROUND_UP(run->count.rows * slice_pages, run_pages);
+ slice->count.bytes =
+ DIV_ROUND_UP(run->count.bytes * slice_pages, run_pages);
slice->count.bytes_compressed = DIV_ROUND_UP(
run->count.bytes_compressed * slice_pages, run_pages);
return slice;
@@ -509,14 +503,14 @@ vy_slice_cut(struct vy_slice *slice, int64_t id, struct vy_entry begin,
/* begin = MAX(begin, slice->begin) */
if (slice->begin.stmt != NULL &&
- (begin.stmt == NULL || vy_entry_compare(begin, slice->begin,
- cmp_def) < 0))
+ (begin.stmt == NULL ||
+ vy_entry_compare(begin, slice->begin, cmp_def) < 0))
begin = slice->begin;
/* end = MIN(end, slice->end) */
if (slice->end.stmt != NULL &&
- (end.stmt == NULL || vy_entry_compare(end, slice->end,
- cmp_def) > 0))
+ (end.stmt == NULL ||
+ vy_entry_compare(end, slice->end, cmp_def) > 0))
end = slice->end;
*result = vy_slice_new(id, slice->run, begin, end, cmp_def);
@@ -569,8 +563,8 @@ vy_page_info_decode(struct vy_page_info *page, const struct xrow_header *xrow,
if (page->min_key == NULL)
return -1;
part_count = mp_decode_array(&key_beg);
- page->min_key_hint = key_hint(key_beg, part_count,
- cmp_def);
+ page->min_key_hint =
+ key_hint(key_beg, part_count, cmp_def);
break;
case VY_PAGE_INFO_UNPACKED_SIZE:
page->unpacked_size = mp_decode_uint(&pos);
@@ -633,8 +627,7 @@ vy_stmt_stat_decode(struct vy_stmt_stat *stat, const char **data)
* @retval -1 error (check diag)
*/
int
-vy_run_info_decode(struct vy_run_info *run_info,
- const struct xrow_header *xrow,
+vy_run_info_decode(struct vy_run_info *run_info, const struct xrow_header *xrow,
const char *filename)
{
assert(xrow->type == VY_INDEX_RUN_INFO);
@@ -707,8 +700,7 @@ vy_page_new(const struct vy_page_info *page_info)
{
struct vy_page *page = malloc(sizeof(*page));
if (page == NULL) {
- diag_set(OutOfMemory, sizeof(*page),
- "load_page", "page cache");
+ diag_set(OutOfMemory, sizeof(*page), "load_page", "page cache");
return NULL;
}
page->unpacked_size = page_info->unpacked_size;
@@ -723,8 +715,8 @@ vy_page_new(const struct vy_page_info *page_info)
page->data = (char *)malloc(page_info->unpacked_size);
if (page->data == NULL) {
- diag_set(OutOfMemory, page_info->unpacked_size,
- "malloc", "page->data");
+ diag_set(OutOfMemory, page_info->unpacked_size, "malloc",
+ "page->data");
free(page->row_index);
free(page);
return NULL;
@@ -748,14 +740,14 @@ vy_page_delete(struct vy_page *page)
}
static int
-vy_page_xrow(struct vy_page *page, uint32_t stmt_no,
- struct xrow_header *xrow)
+vy_page_xrow(struct vy_page *page, uint32_t stmt_no, struct xrow_header *xrow)
{
assert(stmt_no < page->row_count);
const char *data = page->data + page->row_index[stmt_no];
- const char *data_end = stmt_no + 1 < page->row_count ?
- page->data + page->row_index[stmt_no + 1] :
- page->data + page->unpacked_size;
+ const char *data_end =
+ stmt_no + 1 < page->row_count ?
+ page->data + page->row_index[stmt_no + 1] :
+ page->data + page->unpacked_size;
return xrow_header_decode(xrow, &data, data_end, false);
}
@@ -772,8 +764,8 @@ vy_page_xrow(struct vy_page *page, uint32_t stmt_no,
* @retval NULL Memory error.
*/
static struct vy_entry
-vy_page_stmt(struct vy_page *page, uint32_t stmt_no,
- struct key_def *cmp_def, struct tuple_format *format)
+vy_page_stmt(struct vy_page *page, uint32_t stmt_no, struct key_def *cmp_def,
+ struct tuple_format *format)
{
struct xrow_header xrow;
if (vy_page_xrow(page, stmt_no, &xrow) != 0)
@@ -802,12 +794,12 @@ vy_page_find_key(struct vy_page *page, struct vy_entry key,
uint32_t end = page->row_count;
*equal_key = false;
/* for upper bound we change zero comparison result to -1 */
- int zero_cmp = (iterator_type == ITER_GT ||
- iterator_type == ITER_LE ? -1 : 0);
+ int zero_cmp =
+ (iterator_type == ITER_GT || iterator_type == ITER_LE ? -1 : 0);
while (beg != end) {
uint32_t mid = beg + (end - beg) / 2;
- struct vy_entry fnd_key = vy_page_stmt(page, mid, cmp_def,
- format);
+ struct vy_entry fnd_key =
+ vy_page_stmt(page, mid, cmp_def, format);
if (fnd_key.stmt == NULL)
return end;
int cmp = vy_entry_compare(fnd_key, key, cmp_def);
@@ -898,11 +890,12 @@ vy_page_read(struct vy_page *page, const struct vy_page_info *page_info,
diag_set(OutOfMemory, page_info->size, "region gc", "page");
return -1;
}
- ssize_t readen = fio_pread(run->fd, data, page_info->size,
- page_info->offset);
+ ssize_t readen =
+ fio_pread(run->fd, data, page_info->size, page_info->offset);
ERROR_INJECT(ERRINJ_VYRUN_DATA_READ, {
readen = -1;
- errno = EIO;});
+ errno = EIO;
+ });
if (readen < 0) {
diag_set(SystemError, "failed to read from file");
goto error;
@@ -944,7 +937,8 @@ vy_page_read(struct vy_page *page, const struct vy_page_info *page_info,
region_truncate(&fiber()->gc, region_svp);
ERROR_INJECT(ERRINJ_VY_READ_PAGE, {
diag_set(ClientError, ER_INJECTION, "vinyl page read");
- return -1;});
+ return -1;
+ });
return 0;
error:
region_truncate(&fiber()->gc, region_svp);
@@ -987,10 +981,9 @@ vy_page_read_cb(struct cbus_call_msg *base)
if (vy_page_read(task->page, task->page_info, task->run, zdctx) != 0)
return -1;
if (task->key.stmt != NULL) {
- task->pos_in_page = vy_page_find_key(task->page, task->key,
- task->cmp_def, task->format,
- task->iterator_type,
- &task->equal_found);
+ task->pos_in_page = vy_page_find_key(
+ task->page, task->key, task->cmp_def, task->format,
+ task->iterator_type, &task->equal_found);
}
return 0;
}
@@ -1013,8 +1006,7 @@ vy_run_iterator_load_page(struct vy_run_iterator *itr, uint32_t page_no,
/* Check cache */
struct vy_page *page = NULL;
- if (itr->curr_page != NULL &&
- itr->curr_page->page_no == page_no) {
+ if (itr->curr_page != NULL && itr->curr_page->page_no == page_no) {
page = itr->curr_page;
} else if (itr->prev_page != NULL &&
itr->prev_page->page_no == page_no) {
@@ -1024,7 +1016,8 @@ vy_run_iterator_load_page(struct vy_run_iterator *itr, uint32_t page_no,
if (page != NULL) {
if (key.stmt != NULL)
*pos_in_page = vy_page_find_key(page, key, itr->cmp_def,
- itr->format, iterator_type,
+ itr->format,
+ iterator_type,
equal_found);
*result = page;
return 0;
@@ -1039,8 +1032,8 @@ vy_run_iterator_load_page(struct vy_run_iterator *itr, uint32_t page_no,
/* Read page data from the disk */
struct vy_page_read_task *task = mempool_alloc(&env->read_task_pool);
if (task == NULL) {
- diag_set(OutOfMemory, sizeof(*task),
- "mempool", "vy_page_read_task");
+ diag_set(OutOfMemory, sizeof(*task), "mempool",
+ "vy_page_read_task");
vy_page_delete(page);
return -1;
}
@@ -1092,8 +1085,7 @@ vy_run_iterator_load_page(struct vy_run_iterator *itr, uint32_t page_no,
*/
static NODISCARD int
vy_run_iterator_read(struct vy_run_iterator *itr,
- struct vy_run_iterator_pos pos,
- struct vy_entry *ret)
+ struct vy_run_iterator_pos pos, struct vy_entry *ret)
{
struct vy_page *page;
bool equal_found;
@@ -1125,9 +1117,8 @@ vy_run_iterator_search(struct vy_run_iterator *itr,
enum iterator_type iterator_type, struct vy_entry key,
struct vy_run_iterator_pos *pos, bool *equal_key)
{
- pos->page_no = vy_page_index_find_page(itr->slice->run, key,
- itr->cmp_def, iterator_type,
- equal_key);
+ pos->page_no = vy_page_index_find_page(
+ itr->slice->run, key, itr->cmp_def, iterator_type, equal_key);
if (pos->page_no == itr->slice->run->info.page_count)
return 1;
bool equal_in_page;
@@ -1287,7 +1278,7 @@ vy_run_iterator_do_seek(struct vy_run_iterator *itr,
enum iterator_type iterator_type, struct vy_entry key)
{
struct vy_run *run = itr->slice->run;
- struct vy_run_iterator_pos end_pos = {run->info.page_count, 0};
+ struct vy_run_iterator_pos end_pos = { run->info.page_count, 0 };
bool equal_found = false;
if (!vy_stmt_is_empty_key(key.stmt)) {
int rc = vy_run_iterator_search(itr, iterator_type, key,
@@ -1372,7 +1363,8 @@ vy_run_iterator_seek(struct vy_run_iterator *itr, struct vy_entry last,
if (iterator_type == ITER_EQ)
check_eq = true;
iterator_type = iterator_direction(iterator_type) > 0 ?
- ITER_GT : ITER_LT;
+ ITER_GT :
+ ITER_LT;
key = last;
}
@@ -1442,8 +1434,8 @@ vy_run_iterator_seek(struct vy_run_iterator *itr, struct vy_entry last,
return -1;
/* Check EQ constraint if necessary. */
- if (check_eq && vy_entry_compare(itr->curr, itr->key,
- itr->cmp_def) != 0)
+ if (check_eq &&
+ vy_entry_compare(itr->curr, itr->key, itr->cmp_def) != 0)
goto not_found;
/* Skip statements invisible from the iterator read view. */
@@ -1462,11 +1454,10 @@ not_found:
void
vy_run_iterator_open(struct vy_run_iterator *itr,
- struct vy_run_iterator_stat *stat,
- struct vy_slice *slice, enum iterator_type iterator_type,
- struct vy_entry key, const struct vy_read_view **rv,
- struct key_def *cmp_def, struct key_def *key_def,
- struct tuple_format *format)
+ struct vy_run_iterator_stat *stat, struct vy_slice *slice,
+ enum iterator_type iterator_type, struct vy_entry key,
+ const struct vy_read_view **rv, struct key_def *cmp_def,
+ struct key_def *key_def, struct tuple_format *format)
{
itr->stat = stat;
itr->cmp_def = cmp_def;
@@ -1581,8 +1572,7 @@ next:
}
NODISCARD int
-vy_run_iterator_next(struct vy_run_iterator *itr,
- struct vy_history *history)
+vy_run_iterator_next(struct vy_run_iterator *itr, struct vy_history *history)
{
vy_history_cleanup(history);
struct vy_entry entry;
@@ -1610,7 +1600,8 @@ vy_run_iterator_skip(struct vy_run_iterator *itr, struct vy_entry last,
if (itr->search_started &&
(itr->curr.stmt == NULL || last.stmt == NULL ||
iterator_direction(itr->iterator_type) *
- vy_entry_compare(itr->curr, last, itr->cmp_def) > 0))
+ vy_entry_compare(itr->curr, last, itr->cmp_def) >
+ 0))
return 0;
vy_history_cleanup(history);
@@ -1656,12 +1647,12 @@ vy_run_acct_page(struct vy_run *run, struct vy_page_info *page)
}
int
-vy_run_recover(struct vy_run *run, const char *dir,
- uint32_t space_id, uint32_t iid, struct key_def *cmp_def)
+vy_run_recover(struct vy_run *run, const char *dir, uint32_t space_id,
+ uint32_t iid, struct key_def *cmp_def)
{
char path[PATH_MAX];
- vy_run_snprint_path(path, sizeof(path), dir,
- space_id, iid, run->id, VY_FILE_INDEX);
+ vy_run_snprint_path(path, sizeof(path), dir, space_id, iid, run->id,
+ VY_FILE_INDEX);
struct xlog_cursor cursor;
ERROR_INJECT_COUNTDOWN(ERRINJ_VY_RUN_OPEN, {
@@ -1685,15 +1676,15 @@ vy_run_recover(struct vy_run *run, const char *dir,
if (rc != 0) {
if (rc > 0)
- diag_set(ClientError, ER_INVALID_INDEX_FILE,
- path, "Unexpected end of file");
+ diag_set(ClientError, ER_INVALID_INDEX_FILE, path,
+ "Unexpected end of file");
goto fail_close;
}
rc = xlog_cursor_next_row(&cursor, &xrow);
if (rc != 0) {
if (rc > 0)
- diag_set(ClientError, ER_INVALID_INDEX_FILE,
- path, "Unexpected end of file");
+ diag_set(ClientError, ER_INVALID_INDEX_FILE, path,
+ "Unexpected end of file");
goto fail_close;
}
@@ -1708,8 +1699,8 @@ vy_run_recover(struct vy_run *run, const char *dir,
goto fail_close;
/* Allocate buffer for page info. */
- run->page_info = calloc(run->info.page_count,
- sizeof(struct vy_page_info));
+ run->page_info =
+ calloc(run->info.page_count, sizeof(struct vy_page_info));
if (run->page_info == NULL) {
diag_set(OutOfMemory,
run->info.page_count * sizeof(struct vy_page_info),
@@ -1756,14 +1747,14 @@ vy_run_recover(struct vy_run *run, const char *dir,
xlog_cursor_close(&cursor, false);
/* Prepare data file for reading. */
- vy_run_snprint_path(path, sizeof(path), dir,
- space_id, iid, run->id, VY_FILE_RUN);
+ vy_run_snprint_path(path, sizeof(path), dir, space_id, iid, run->id,
+ VY_FILE_RUN);
if (xlog_cursor_open(&cursor, path))
goto fail;
meta = &cursor.meta;
if (strcmp(meta->filetype, XLOG_META_TYPE_RUN) != 0) {
- diag_set(ClientError, ER_INVALID_XLOG_TYPE,
- XLOG_META_TYPE_RUN, meta->filetype);
+ diag_set(ClientError, ER_INVALID_XLOG_TYPE, XLOG_META_TYPE_RUN,
+ meta->filetype);
goto fail_close;
}
run->fd = cursor.fd;
@@ -1786,11 +1777,12 @@ vy_run_dump_stmt(struct vy_entry entry, struct xlog *data_xlog,
bool is_primary)
{
struct xrow_header xrow;
- int rc = (is_primary ?
- vy_stmt_encode_primary(entry.stmt, key_def, 0, &xrow) :
- vy_stmt_encode_secondary(entry.stmt, key_def,
- vy_entry_multikey_idx(entry, key_def),
- &xrow));
+ int rc =
+ (is_primary ?
+ vy_stmt_encode_primary(entry.stmt, key_def, 0, &xrow) :
+ vy_stmt_encode_secondary(
+ entry.stmt, key_def,
+ vy_entry_multikey_idx(entry, key_def), &xrow));
if (rc != 0)
return -1;
@@ -1819,8 +1811,7 @@ vy_row_index_encode(const uint32_t *row_index, uint32_t row_count,
memset(xrow, 0, sizeof(*xrow));
xrow->type = VY_RUN_ROW_INDEX;
- size_t size = mp_sizeof_map(1) +
- mp_sizeof_uint(VY_ROW_INDEX_DATA) +
+ size_t size = mp_sizeof_map(1) + mp_sizeof_uint(VY_ROW_INDEX_DATA) +
mp_sizeof_bin(sizeof(uint32_t) * row_count);
char *pos = region_alloc(&fiber()->gc, size);
if (pos == NULL) {
@@ -1845,13 +1836,12 @@ vy_row_index_encode(const uint32_t *row_index, uint32_t row_count,
static inline int
vy_run_alloc_page_info(struct vy_run *run, uint32_t *page_info_capacity)
{
- uint32_t cap = *page_info_capacity > 0 ?
- *page_info_capacity * 2 : 16;
- struct vy_page_info *page_info = realloc(run->page_info,
- cap * sizeof(*page_info));
+ uint32_t cap = *page_info_capacity > 0 ? *page_info_capacity * 2 : 16;
+ struct vy_page_info *page_info =
+ realloc(run->page_info, cap * sizeof(*page_info));
if (page_info == NULL) {
- diag_set(OutOfMemory, cap * sizeof(*page_info),
- "realloc", "struct vy_page_info");
+ diag_set(OutOfMemory, cap * sizeof(*page_info), "realloc",
+ "struct vy_page_info");
return -1;
}
run->page_info = page_info;
@@ -1886,15 +1876,13 @@ vy_page_info_encode(const struct vy_page_info *page_info,
/* calc tuple size */
uint32_t size;
/* 3 items: page offset, size, and map */
- size = mp_sizeof_map(6) +
- mp_sizeof_uint(VY_PAGE_INFO_OFFSET) +
+ size = mp_sizeof_map(6) + mp_sizeof_uint(VY_PAGE_INFO_OFFSET) +
mp_sizeof_uint(page_info->offset) +
mp_sizeof_uint(VY_PAGE_INFO_SIZE) +
mp_sizeof_uint(page_info->size) +
mp_sizeof_uint(VY_PAGE_INFO_ROW_COUNT) +
mp_sizeof_uint(page_info->row_count) +
- mp_sizeof_uint(VY_PAGE_INFO_MIN_KEY) +
- min_key_size +
+ mp_sizeof_uint(VY_PAGE_INFO_MIN_KEY) + min_key_size +
mp_sizeof_uint(VY_PAGE_INFO_UNPACKED_SIZE) +
mp_sizeof_uint(page_info->unpacked_size) +
mp_sizeof_uint(VY_PAGE_INFO_ROW_INDEX_OFFSET) +
@@ -1938,15 +1926,11 @@ vy_page_info_encode(const struct vy_page_info *page_info,
static size_t
vy_stmt_stat_sizeof(const struct vy_stmt_stat *stat)
{
- return mp_sizeof_map(4) +
- mp_sizeof_uint(IPROTO_INSERT) +
- mp_sizeof_uint(IPROTO_REPLACE) +
- mp_sizeof_uint(IPROTO_DELETE) +
- mp_sizeof_uint(IPROTO_UPSERT) +
- mp_sizeof_uint(stat->inserts) +
- mp_sizeof_uint(stat->replaces) +
- mp_sizeof_uint(stat->deletes) +
- mp_sizeof_uint(stat->upserts);
+ return mp_sizeof_map(4) + mp_sizeof_uint(IPROTO_INSERT) +
+ mp_sizeof_uint(IPROTO_REPLACE) + mp_sizeof_uint(IPROTO_DELETE) +
+ mp_sizeof_uint(IPROTO_UPSERT) + mp_sizeof_uint(stat->inserts) +
+ mp_sizeof_uint(stat->replaces) + mp_sizeof_uint(stat->deletes) +
+ mp_sizeof_uint(stat->upserts);
}
/** Encode statement statistics to @buf and return advanced @buf. */
@@ -1976,8 +1960,7 @@ vy_stmt_stat_encode(const struct vy_stmt_stat *stat, char *buf)
* @retval -1 on error, check diag
*/
static int
-vy_run_info_encode(const struct vy_run_info *run_info,
- struct xrow_header *xrow)
+vy_run_info_encode(const struct vy_run_info *run_info, struct xrow_header *xrow)
{
const char *tmp;
tmp = run_info->min_key;
@@ -2045,19 +2028,19 @@ vy_run_info_encode(const struct vy_run_info *run_info,
* Write run index to file.
*/
static int
-vy_run_write_index(struct vy_run *run, const char *dirpath,
- uint32_t space_id, uint32_t iid)
+vy_run_write_index(struct vy_run *run, const char *dirpath, uint32_t space_id,
+ uint32_t iid)
{
char path[PATH_MAX];
- vy_run_snprint_path(path, sizeof(path), dirpath,
- space_id, iid, run->id, VY_FILE_INDEX);
+ vy_run_snprint_path(path, sizeof(path), dirpath, space_id, iid, run->id,
+ VY_FILE_INDEX);
say_info("writing `%s'", path);
struct xlog index_xlog;
struct xlog_meta meta;
- xlog_meta_create(&meta, XLOG_META_TYPE_INDEX, &INSTANCE_UUID,
- NULL, NULL);
+ xlog_meta_create(&meta, XLOG_META_TYPE_INDEX, &INSTANCE_UUID, NULL,
+ NULL);
struct xlog_opts opts = xlog_opts_default;
opts.rate_limit = run->env->snap_io_rate_limit;
opts.sync_interval = VY_RUN_SYNC_INTERVAL;
@@ -2092,8 +2075,7 @@ vy_run_write_index(struct vy_run *run, const char *dirpath,
return -1;
});
- if (xlog_flush(&index_xlog) < 0 ||
- xlog_rename(&index_xlog) < 0)
+ if (xlog_flush(&index_xlog) < 0 || xlog_rename(&index_xlog) < 0)
goto fail;
xlog_close(&index_xlog, false);
@@ -2154,8 +2136,7 @@ vy_run_writer_create_xlog(struct vy_run_writer *writer)
VY_FILE_RUN);
say_info("writing `%s'", path);
struct xlog_meta meta;
- xlog_meta_create(&meta, XLOG_META_TYPE_RUN, &INSTANCE_UUID,
- NULL, NULL);
+ xlog_meta_create(&meta, XLOG_META_TYPE_RUN, &INSTANCE_UUID, NULL, NULL);
struct xlog_opts opts = xlog_opts_default;
opts.rate_limit = writer->run->env->snap_io_rate_limit;
opts.sync_interval = VY_RUN_SYNC_INTERVAL;
@@ -2181,12 +2162,13 @@ vy_run_writer_start_page(struct vy_run_writer *writer,
if (run->info.page_count >= writer->page_info_capacity &&
vy_run_alloc_page_info(run, &writer->page_info_capacity) != 0)
return -1;
- const char *key = vy_stmt_is_key(first_entry.stmt) ?
- tuple_data(first_entry.stmt) :
- tuple_extract_key(first_entry.stmt, writer->cmp_def,
- vy_entry_multikey_idx(first_entry,
- writer->cmp_def),
- NULL);
+ const char *key =
+ vy_stmt_is_key(first_entry.stmt) ?
+ tuple_data(first_entry.stmt) :
+ tuple_extract_key(first_entry.stmt, writer->cmp_def,
+ vy_entry_multikey_idx(
+ first_entry, writer->cmp_def),
+ NULL);
if (key == NULL)
return -1;
if (run->info.page_count == 0) {
@@ -2196,8 +2178,8 @@ vy_run_writer_start_page(struct vy_run_writer *writer,
return -1;
}
struct vy_page_info *page = run->page_info + run->info.page_count;
- if (vy_page_info_create(page, writer->data_xlog.offset,
- key, writer->cmp_def) != 0)
+ if (vy_page_info_create(page, writer->data_xlog.offset, key,
+ writer->cmp_def) != 0)
return -1;
xlog_tx_begin(&writer->data_xlog);
return 0;
@@ -2230,8 +2212,8 @@ vy_run_writer_write_to_page(struct vy_run_writer *writer, struct vy_entry entry)
return -1;
}
*offset = page->unpacked_size;
- if (vy_run_dump_stmt(entry, &writer->data_xlog, page,
- writer->cmp_def, writer->iid == 0) != 0)
+ if (vy_run_dump_stmt(entry, &writer->data_xlog, page, writer->cmp_def,
+ writer->iid == 0) != 0)
return -1;
int64_t lsn = vy_stmt_lsn(entry.stmt);
run->info.min_lsn = MIN(run->info.min_lsn, lsn);
@@ -2336,12 +2318,14 @@ vy_run_writer_commit(struct vy_run_writer *writer)
}
assert(writer->last.stmt != NULL);
- const char *key = vy_stmt_is_key(writer->last.stmt) ?
- tuple_data(writer->last.stmt) :
- tuple_extract_key(writer->last.stmt, writer->cmp_def,
- vy_entry_multikey_idx(writer->last,
- writer->cmp_def),
- NULL);
+ const char *key =
+ vy_stmt_is_key(writer->last.stmt) ?
+ tuple_data(writer->last.stmt) :
+ tuple_extract_key(
+ writer->last.stmt, writer->cmp_def,
+ vy_entry_multikey_idx(writer->last,
+ writer->cmp_def),
+ NULL);
if (key == NULL)
goto out;
@@ -2361,13 +2345,13 @@ vy_run_writer_commit(struct vy_run_writer *writer)
goto out;
if (writer->bloom != NULL) {
- run->info.bloom = tuple_bloom_new(writer->bloom,
- writer->bloom_fpr);
+ run->info.bloom =
+ tuple_bloom_new(writer->bloom, writer->bloom_fpr);
if (run->info.bloom == NULL)
goto out;
}
- if (vy_run_write_index(run, writer->dirpath,
- writer->space_id, writer->iid) != 0)
+ if (vy_run_write_index(run, writer->dirpath, writer->space_id,
+ writer->iid) != 0)
goto out;
run->fd = writer->data_xlog.fd;
@@ -2385,10 +2369,10 @@ vy_run_writer_abort(struct vy_run_writer *writer)
}
int
-vy_run_rebuild_index(struct vy_run *run, const char *dir,
- uint32_t space_id, uint32_t iid,
- struct key_def *cmp_def, struct key_def *key_def,
- struct tuple_format *format, const struct index_opts *opts)
+vy_run_rebuild_index(struct vy_run *run, const char *dir, uint32_t space_id,
+ uint32_t iid, struct key_def *cmp_def,
+ struct key_def *key_def, struct tuple_format *format,
+ const struct index_opts *opts)
{
assert(run->info.bloom == NULL);
assert(run->page_info == NULL);
@@ -2397,8 +2381,8 @@ vy_run_rebuild_index(struct vy_run *run, const char *dir,
struct xlog_cursor cursor;
char path[PATH_MAX];
- vy_run_snprint_path(path, sizeof(path), dir,
- space_id, iid, run->id, VY_FILE_RUN);
+ vy_run_snprint_path(path, sizeof(path), dir, space_id, iid, run->id,
+ VY_FILE_RUN);
say_info("rebuilding index for `%s'", path);
if (xlog_cursor_open(&cursor, path))
@@ -2445,16 +2429,17 @@ vy_run_rebuild_index(struct vy_run *run, const char *dir,
if (tuple == NULL)
goto close_err;
if (bloom_builder != NULL) {
- struct vy_entry entry = {tuple, HINT_NONE};
+ struct vy_entry entry = { tuple, HINT_NONE };
if (vy_bloom_builder_add(bloom_builder, entry,
key_def) != 0) {
tuple_unref(tuple);
goto close_err;
}
}
- key = vy_stmt_is_key(tuple) ? tuple_data(tuple) :
- tuple_extract_key(tuple, cmp_def,
- MULTIKEY_NONE, NULL);
+ key = vy_stmt_is_key(tuple) ?
+ tuple_data(tuple) :
+ tuple_extract_key(tuple, cmp_def,
+ MULTIKEY_NONE, NULL);
if (prev_tuple != NULL)
tuple_unref(prev_tuple);
prev_tuple = tuple;
@@ -2478,8 +2463,8 @@ vy_run_rebuild_index(struct vy_run *run, const char *dir,
}
struct vy_page_info *info;
info = run->page_info + run->info.page_count;
- if (vy_page_info_create(info, page_offset,
- page_min_key, cmp_def) != 0)
+ if (vy_page_info_create(info, page_offset, page_min_key,
+ cmp_def) != 0)
goto close_err;
info->row_count = page_row_count;
info->size = next_page_offset - page_offset;
@@ -2509,8 +2494,8 @@ vy_run_rebuild_index(struct vy_run *run, const char *dir,
xlog_cursor_close(&cursor, true);
if (bloom_builder != NULL) {
- run->info.bloom = tuple_bloom_new(bloom_builder,
- opts->bloom_fpr);
+ run->info.bloom =
+ tuple_bloom_new(bloom_builder, opts->bloom_fpr);
if (run->info.bloom == NULL)
goto close_err;
tuple_bloom_builder_delete(bloom_builder);
@@ -2518,11 +2503,10 @@ vy_run_rebuild_index(struct vy_run *run, const char *dir,
}
/* New run index is ready for write, unlink old file if exists */
- vy_run_snprint_path(path, sizeof(path), dir,
- space_id, iid, run->id, VY_FILE_INDEX);
+ vy_run_snprint_path(path, sizeof(path), dir, space_id, iid, run->id,
+ VY_FILE_INDEX);
if (unlink(path) < 0 && errno != ENOENT) {
- diag_set(SystemError, "failed to unlink file '%s'",
- path);
+ diag_set(SystemError, "failed to unlink file '%s'", path);
goto close_err;
}
if (vy_run_write_index(run, dir, space_id, iid) != 0)
@@ -2543,17 +2527,19 @@ close_err:
}
int
-vy_run_remove_files(const char *dir, uint32_t space_id,
- uint32_t iid, int64_t run_id)
+vy_run_remove_files(const char *dir, uint32_t space_id, uint32_t iid,
+ int64_t run_id)
{
- ERROR_INJECT(ERRINJ_VY_GC,
- {say_error("error injection: vinyl run %lld not deleted",
- (long long)run_id); return -1;});
+ ERROR_INJECT(ERRINJ_VY_GC, {
+ say_error("error injection: vinyl run %lld not deleted",
+ (long long)run_id);
+ return -1;
+ });
int ret = 0;
char path[PATH_MAX];
for (int type = 0; type < vy_file_MAX; type++) {
- vy_run_snprint_path(path, sizeof(path), dir,
- space_id, iid, run_id, type);
+ vy_run_snprint_path(path, sizeof(path), dir, space_id, iid,
+ run_id, type);
if (coio_unlink(path) < 0) {
if (errno != ENOENT) {
say_syserror("error while removing %s", path);
@@ -2619,8 +2605,7 @@ vy_slice_stream_search(struct vy_stmt_stream *virt_stream)
bool unused;
stream->pos_in_page = vy_page_find_key(stream->page,
stream->slice->begin,
- stream->cmp_def,
- stream->format,
+ stream->cmp_def, stream->format,
ITER_GE, &unused);
if (stream->pos_in_page == stream->page->row_count) {
@@ -2679,8 +2664,8 @@ vy_slice_stream_next(struct vy_stmt_stream *virt_stream, struct vy_entry *ret)
stream->pos_in_page++;
/* Check whether the position is out of page */
- struct vy_page_info *page_info = vy_run_page_info(stream->slice->run,
- stream->page_no);
+ struct vy_page_info *page_info =
+ vy_run_page_info(stream->slice->run, stream->page_no);
if (stream->pos_in_page >= page_info->row_count) {
/**
* Out of page. Free page, move the position to the next page
diff --git a/src/box/vy_run.h b/src/box/vy_run.h
index 9618d85..1164faf 100644
--- a/src/box/vy_run.h
+++ b/src/box/vy_run.h
@@ -380,8 +380,8 @@ vy_run_unref(struct vy_run *run)
* @return - 0 on sucess, -1 on fail
*/
int
-vy_run_recover(struct vy_run *run, const char *dir,
- uint32_t space_id, uint32_t iid, struct key_def *cmp_def);
+vy_run_recover(struct vy_run *run, const char *dir, uint32_t space_id,
+ uint32_t iid, struct key_def *cmp_def);
/**
* Rebuild run index
@@ -396,10 +396,9 @@ vy_run_recover(struct vy_run *run, const char *dir,
* @return - 0 on sucess, -1 on fail
*/
int
-vy_run_rebuild_index(struct vy_run *run, const char *dir,
- uint32_t space_id, uint32_t iid,
- struct key_def *cmp_def, struct key_def *key_def,
- struct tuple_format *format,
+vy_run_rebuild_index(struct vy_run *run, const char *dir, uint32_t space_id,
+ uint32_t iid, struct key_def *cmp_def,
+ struct key_def *key_def, struct tuple_format *format,
const struct index_opts *opts);
enum vy_file_type {
@@ -413,29 +412,28 @@ enum vy_file_type {
extern const char *vy_file_suffix[];
static inline int
-vy_lsm_snprint_path(char *buf, int size, const char *dir,
- uint32_t space_id, uint32_t iid)
+vy_lsm_snprint_path(char *buf, int size, const char *dir, uint32_t space_id,
+ uint32_t iid)
{
- return snprintf(buf, size, "%s/%u/%u",
- dir, (unsigned)space_id, (unsigned)iid);
+ return snprintf(buf, size, "%s/%u/%u", dir, (unsigned)space_id,
+ (unsigned)iid);
}
static inline int
vy_run_snprint_filename(char *buf, int size, int64_t run_id,
enum vy_file_type type)
{
- return snprintf(buf, size, "%020lld.%s",
- (long long)run_id, vy_file_suffix[type]);
+ return snprintf(buf, size, "%020lld.%s", (long long)run_id,
+ vy_file_suffix[type]);
}
static inline int
-vy_run_snprint_path(char *buf, int size, const char *dir,
- uint32_t space_id, uint32_t iid,
- int64_t run_id, enum vy_file_type type)
+vy_run_snprint_path(char *buf, int size, const char *dir, uint32_t space_id,
+ uint32_t iid, int64_t run_id, enum vy_file_type type)
{
int total = 0;
- SNPRINT(total, vy_lsm_snprint_path, buf, size,
- dir, (unsigned)space_id, (unsigned)iid);
+ SNPRINT(total, vy_lsm_snprint_path, buf, size, dir, (unsigned)space_id,
+ (unsigned)iid);
SNPRINT(total, snprintf, buf, size, "/");
SNPRINT(total, vy_run_snprint_filename, buf, size, run_id, type);
return total;
@@ -447,8 +445,8 @@ vy_run_snprint_path(char *buf, int size, const char *dir,
* failed.
*/
int
-vy_run_remove_files(const char *dir, uint32_t space_id,
- uint32_t iid, int64_t run_id);
+vy_run_remove_files(const char *dir, uint32_t space_id, uint32_t iid,
+ int64_t run_id);
/**
* Allocate a new run slice.
@@ -518,11 +516,10 @@ vy_slice_cut(struct vy_slice *slice, int64_t id, struct vy_entry begin,
*/
void
vy_run_iterator_open(struct vy_run_iterator *itr,
- struct vy_run_iterator_stat *stat,
- struct vy_slice *slice, enum iterator_type iterator_type,
- struct vy_entry key, const struct vy_read_view **rv,
- struct key_def *cmp_def, struct key_def *key_def,
- struct tuple_format *format);
+ struct vy_run_iterator_stat *stat, struct vy_slice *slice,
+ enum iterator_type iterator_type, struct vy_entry key,
+ const struct vy_read_view **rv, struct key_def *cmp_def,
+ struct key_def *key_def, struct tuple_format *format);
/**
* Advance a run iterator to the next key.
@@ -530,8 +527,7 @@ vy_run_iterator_open(struct vy_run_iterator *itr,
* Returns 0 on success, -1 on memory allocation or IO error.
*/
NODISCARD int
-vy_run_iterator_next(struct vy_run_iterator *itr,
- struct vy_history *history);
+vy_run_iterator_next(struct vy_run_iterator *itr, struct vy_history *history);
/**
* Advance a run iterator to the key following @last.
diff --git a/src/box/vy_scheduler.c b/src/box/vy_scheduler.c
index b641dd9..c2cc463 100644
--- a/src/box/vy_scheduler.c
+++ b/src/box/vy_scheduler.c
@@ -60,15 +60,19 @@
#include "trivia/util.h"
/* Min and max values for vy_scheduler::timeout. */
-#define VY_SCHEDULER_TIMEOUT_MIN 1
-#define VY_SCHEDULER_TIMEOUT_MAX 60
+#define VY_SCHEDULER_TIMEOUT_MIN 1
+#define VY_SCHEDULER_TIMEOUT_MAX 60
static int vy_worker_f(va_list);
static int vy_scheduler_f(va_list);
-static void vy_task_execute_f(struct cmsg *);
-static void vy_task_complete_f(struct cmsg *);
-static void vy_deferred_delete_batch_process_f(struct cmsg *);
-static void vy_deferred_delete_batch_free_f(struct cmsg *);
+static void
+vy_task_execute_f(struct cmsg *);
+static void
+vy_task_complete_f(struct cmsg *);
+static void
+vy_deferred_delete_batch_process_f(struct cmsg *);
+static void
+vy_deferred_delete_batch_free_f(struct cmsg *);
static const struct cmsg_hop vy_task_execute_route[] = {
{ vy_task_execute_f, NULL },
@@ -222,7 +226,7 @@ struct vy_task {
};
static const struct vy_deferred_delete_handler_iface
-vy_task_deferred_delete_iface;
+ vy_task_deferred_delete_iface;
/**
* Allocate a new task to be executed by a worker thread.
@@ -237,8 +241,8 @@ vy_task_new(struct vy_scheduler *scheduler, struct vy_worker *worker,
{
struct vy_task *task = calloc(1, sizeof(*task));
if (task == NULL) {
- diag_set(OutOfMemory, sizeof(*task),
- "malloc", "struct vy_task");
+ diag_set(OutOfMemory, sizeof(*task), "malloc",
+ "struct vy_task");
return NULL;
}
memset(task, 0, sizeof(*task));
@@ -432,8 +436,8 @@ vy_scheduler_create(struct vy_scheduler *scheduler, int write_threads,
scheduler->read_views = read_views;
scheduler->run_env = run_env;
- scheduler->scheduler_fiber = fiber_new("vinyl.scheduler",
- vy_scheduler_f);
+ scheduler->scheduler_fiber =
+ fiber_new("vinyl.scheduler", vy_scheduler_f);
if (scheduler->scheduler_fiber == NULL)
panic("failed to allocate vinyl scheduler fiber");
@@ -455,10 +459,9 @@ vy_scheduler_create(struct vy_scheduler *scheduler, int write_threads,
assert(write_threads > 1);
int dump_threads = MAX(1, write_threads / 4);
int compaction_threads = write_threads - dump_threads;
- vy_worker_pool_create(&scheduler->dump_pool,
- "dump", dump_threads);
- vy_worker_pool_create(&scheduler->compaction_pool,
- "compaction", compaction_threads);
+ vy_worker_pool_create(&scheduler->dump_pool, "dump", dump_threads);
+ vy_worker_pool_create(&scheduler->compaction_pool, "compaction",
+ compaction_threads);
stailq_create(&scheduler->processed_tasks);
@@ -515,8 +518,8 @@ vy_scheduler_on_delete_lsm(struct trigger *trigger, void *event)
{
struct vy_lsm *lsm = event;
struct vy_scheduler *scheduler = trigger->data;
- assert(! heap_node_is_stray(&lsm->in_dump));
- assert(! heap_node_is_stray(&lsm->in_compaction));
+ assert(!heap_node_is_stray(&lsm->in_dump));
+ assert(!heap_node_is_stray(&lsm->in_compaction));
vy_dump_heap_delete(&scheduler->dump_heap, lsm);
vy_compaction_heap_delete(&scheduler->compaction_heap, lsm);
trigger_clear(trigger);
@@ -552,8 +555,8 @@ vy_scheduler_add_lsm(struct vy_scheduler *scheduler, struct vy_lsm *lsm)
static void
vy_scheduler_update_lsm(struct vy_scheduler *scheduler, struct vy_lsm *lsm)
{
- assert(! heap_node_is_stray(&lsm->in_dump));
- assert(! heap_node_is_stray(&lsm->in_compaction));
+ assert(!heap_node_is_stray(&lsm->in_dump));
+ assert(!heap_node_is_stray(&lsm->in_compaction));
vy_dump_heap_update(&scheduler->dump_heap, lsm);
vy_compaction_heap_update(&scheduler->compaction_heap, lsm);
}
@@ -675,8 +678,8 @@ vy_scheduler_complete_dump(struct vy_scheduler *scheduler)
scheduler->dump_start = now;
scheduler->dump_generation = min_generation;
scheduler->stat.dump_count++;
- scheduler->dump_complete_cb(scheduler,
- min_generation - 1, dump_duration);
+ scheduler->dump_complete_cb(scheduler, min_generation - 1,
+ dump_duration);
fiber_cond_signal(&scheduler->dump_cond);
}
@@ -698,7 +701,8 @@ vy_scheduler_begin_checkpoint(struct vy_scheduler *scheduler, bool is_scheduled)
struct error *e = diag_last_error(&scheduler->diag);
diag_set_error(diag_get(), e);
say_error("cannot checkpoint vinyl, "
- "scheduler is throttled with: %s", e->errmsg);
+ "scheduler is throttled with: %s",
+ e->errmsg);
return -1;
}
say_info("scheduler is unthrottled due to manual checkpoint "
@@ -796,9 +800,11 @@ vy_run_discard(struct vy_run *run)
vy_run_unref(run);
- ERROR_INJECT(ERRINJ_VY_RUN_DISCARD,
- {say_error("error injection: run %lld not discarded",
- (long long)run_id); return;});
+ ERROR_INJECT(ERRINJ_VY_RUN_DISCARD, {
+ say_error("error injection: run %lld not discarded",
+ (long long)run_id);
+ return;
+ });
vy_log_tx_begin();
/*
@@ -865,8 +871,8 @@ vy_deferred_delete_process_one(struct space *deferred_delete_space,
return -1;
struct tuple *unused;
- if (space_execute_dml(deferred_delete_space, txn,
- &request, &unused) != 0) {
+ if (space_execute_dml(deferred_delete_space, txn, &request, &unused) !=
+ 0) {
txn_rollback_stmt(txn);
return -1;
}
@@ -885,8 +891,8 @@ vy_deferred_delete_process_one(struct space *deferred_delete_space,
static void
vy_deferred_delete_batch_process_f(struct cmsg *cmsg)
{
- struct vy_deferred_delete_batch *batch = container_of(cmsg,
- struct vy_deferred_delete_batch, cmsg);
+ struct vy_deferred_delete_batch *batch =
+ container_of(cmsg, struct vy_deferred_delete_batch, cmsg);
struct vy_task *task = batch->task;
struct vy_lsm *pk = task->lsm;
@@ -936,8 +942,8 @@ fail:
static void
vy_deferred_delete_batch_free_f(struct cmsg *cmsg)
{
- struct vy_deferred_delete_batch *batch = container_of(cmsg,
- struct vy_deferred_delete_batch, cmsg);
+ struct vy_deferred_delete_batch *batch =
+ container_of(cmsg, struct vy_deferred_delete_batch, cmsg);
struct vy_task *task = batch->task;
for (int i = 0; i < batch->count; i++) {
struct vy_deferred_delete_stmt *stmt = &batch->stmt[i];
@@ -992,8 +998,8 @@ vy_task_deferred_delete_process(struct vy_deferred_delete_handler *handler,
{
enum { MAX_IN_PROGRESS = 10 };
- struct vy_task *task = container_of(handler, struct vy_task,
- deferred_delete_handler);
+ struct vy_task *task =
+ container_of(handler, struct vy_task, deferred_delete_handler);
struct vy_deferred_delete_batch *batch = task->deferred_delete_batch;
/*
@@ -1036,18 +1042,18 @@ vy_task_deferred_delete_process(struct vy_deferred_delete_handler *handler,
static void
vy_task_deferred_delete_destroy(struct vy_deferred_delete_handler *handler)
{
- struct vy_task *task = container_of(handler, struct vy_task,
- deferred_delete_handler);
+ struct vy_task *task =
+ container_of(handler, struct vy_task, deferred_delete_handler);
vy_task_deferred_delete_flush(task);
while (task->deferred_delete_in_progress > 0)
fiber_sleep(TIMEOUT_INFINITY);
}
static const struct vy_deferred_delete_handler_iface
-vy_task_deferred_delete_iface = {
- .process = vy_task_deferred_delete_process,
- .destroy = vy_task_deferred_delete_destroy,
-};
+ vy_task_deferred_delete_iface = {
+ .process = vy_task_deferred_delete_process,
+ .destroy = vy_task_deferred_delete_destroy,
+ };
static int
vy_task_write_run(struct vy_task *task, bool no_compression)
@@ -1057,17 +1063,17 @@ vy_task_write_run(struct vy_task *task, bool no_compression)
struct vy_lsm *lsm = task->lsm;
struct vy_stmt_stream *wi = task->wi;
- ERROR_INJECT(ERRINJ_VY_RUN_WRITE,
- {diag_set(ClientError, ER_INJECTION,
- "vinyl dump"); return -1;});
+ ERROR_INJECT(ERRINJ_VY_RUN_WRITE, {
+ diag_set(ClientError, ER_INJECTION, "vinyl dump");
+ return -1;
+ });
ERROR_INJECT_SLEEP(ERRINJ_VY_RUN_WRITE_DELAY);
struct vy_run_writer writer;
if (vy_run_writer_create(&writer, task->new_run, lsm->env->path,
- lsm->space_id, lsm->index_id,
- task->cmp_def, task->key_def,
- task->page_size, task->bloom_fpr,
- no_compression) != 0)
+ lsm->space_id, lsm->index_id, task->cmp_def,
+ task->key_def, task->page_size,
+ task->bloom_fpr, no_compression) != 0)
goto fail;
if (wi->iface->start(wi) != 0)
@@ -1076,8 +1082,8 @@ vy_task_write_run(struct vy_task *task, bool no_compression)
int loops = 0;
struct vy_entry entry = vy_entry_none();
while ((rc = wi->iface->next(wi, &entry)) == 0 && entry.stmt != NULL) {
- struct errinj *inj = errinj(ERRINJ_VY_RUN_WRITE_STMT_TIMEOUT,
- ERRINJ_DOUBLE);
+ struct errinj *inj =
+ errinj(ERRINJ_VY_RUN_WRITE_STMT_TIMEOUT, ERRINJ_DOUBLE);
if (inj != NULL && inj->dparam > 0)
thread_sleep(inj->dparam);
@@ -1158,8 +1164,8 @@ vy_task_dump_complete(struct vy_task *task)
* Figure out which ranges intersect the new run.
*/
if (vy_lsm_find_range_intersection(lsm, new_run->info.min_key,
- new_run->info.max_key,
- &begin_range, &end_range) != 0)
+ new_run->info.max_key, &begin_range,
+ &end_range) != 0)
goto fail;
/*
@@ -1173,8 +1179,8 @@ vy_task_dump_complete(struct vy_task *task)
}
for (range = begin_range, i = 0; range != end_range;
range = vy_range_tree_next(&lsm->range_tree, range), i++) {
- slice = vy_slice_new(vy_log_next_id(), new_run,
- range->begin, range->end, lsm->cmp_def);
+ slice = vy_slice_new(vy_log_next_id(), new_run, range->begin,
+ range->end, lsm->cmp_def);
if (slice == NULL)
goto fail_free_slices;
@@ -1473,12 +1479,12 @@ vy_task_compaction_complete(struct vy_task *task)
* as a result of compaction.
*/
RLIST_HEAD(unused_runs);
- for (slice = first_slice; ; slice = rlist_next_entry(slice, in_range)) {
+ for (slice = first_slice;; slice = rlist_next_entry(slice, in_range)) {
slice->run->compacted_slice_count++;
if (slice == last_slice)
break;
}
- for (slice = first_slice; ; slice = rlist_next_entry(slice, in_range)) {
+ for (slice = first_slice;; slice = rlist_next_entry(slice, in_range)) {
run = slice->run;
if (run->compacted_slice_count == run->slice_count)
rlist_add_entry(&unused_runs, run, in_unused);
@@ -1491,7 +1497,7 @@ vy_task_compaction_complete(struct vy_task *task)
* Log change in metadata.
*/
vy_log_tx_begin();
- for (slice = first_slice; ; slice = rlist_next_entry(slice, in_range)) {
+ for (slice = first_slice;; slice = rlist_next_entry(slice, in_range)) {
vy_log_delete_slice(slice->id);
if (slice == last_slice)
break;
@@ -1552,7 +1558,7 @@ vy_task_compaction_complete(struct vy_task *task)
if (new_slice != NULL)
vy_range_add_slice_before(range, new_slice, first_slice);
vy_disk_stmt_counter_reset(&compaction_input);
- for (slice = first_slice; ; slice = next_slice) {
+ for (slice = first_slice;; slice = next_slice) {
next_slice = rlist_next_entry(slice, in_range);
vy_range_remove_slice(range, slice);
rlist_add_entry(&compacted_slices, slice, in_range);
@@ -1564,8 +1570,8 @@ vy_task_compaction_complete(struct vy_task *task)
vy_range_update_compaction_priority(range, &lsm->opts);
vy_range_update_dumps_per_compaction(range);
vy_lsm_acct_range(lsm, range);
- vy_lsm_acct_compaction(lsm, compaction_time,
- &compaction_input, &compaction_output);
+ vy_lsm_acct_compaction(lsm, compaction_time, &compaction_input,
+ &compaction_output);
scheduler->stat.compaction_input += compaction_input.bytes;
scheduler->stat.compaction_output += compaction_output.bytes;
scheduler->stat.compaction_time += compaction_time;
@@ -1575,8 +1581,8 @@ vy_task_compaction_complete(struct vy_task *task)
*/
rlist_foreach_entry(run, &unused_runs, in_unused)
vy_lsm_remove_run(lsm, run);
- rlist_foreach_entry_safe(slice, &compacted_slices,
- in_range, next_slice) {
+ rlist_foreach_entry_safe(slice, &compacted_slices, in_range,
+ next_slice) {
vy_slice_wait_pinned(slice);
vy_slice_delete(slice);
}
@@ -1588,8 +1594,8 @@ vy_task_compaction_complete(struct vy_task *task)
vy_range_heap_insert(&lsm->range_heap, range);
vy_scheduler_update_lsm(scheduler, lsm);
- say_info("%s: completed compacting range %s",
- vy_lsm_name(lsm), vy_range_str(range));
+ say_info("%s: completed compacting range %s", vy_lsm_name(lsm),
+ vy_range_str(range));
return 0;
}
@@ -1605,8 +1611,8 @@ vy_task_compaction_abort(struct vy_task *task)
struct error *e = diag_last_error(&task->diag);
error_log(e);
- say_error("%s: failed to compact range %s",
- vy_lsm_name(lsm), vy_range_str(range));
+ say_error("%s: failed to compact range %s", vy_lsm_name(lsm),
+ vy_range_str(range));
vy_run_discard(task->new_run);
@@ -1635,8 +1641,8 @@ vy_task_compaction_new(struct vy_scheduler *scheduler, struct vy_worker *worker,
return 0;
}
- struct vy_task *task = vy_task_new(scheduler, worker, lsm,
- &compaction_ops);
+ struct vy_task *task =
+ vy_task_new(scheduler, worker, lsm, &compaction_ops);
if (task == NULL)
goto err_task;
@@ -1646,10 +1652,10 @@ vy_task_compaction_new(struct vy_scheduler *scheduler, struct vy_worker *worker,
struct vy_stmt_stream *wi;
bool is_last_level = (range->compaction_priority == range->slice_count);
- wi = vy_write_iterator_new(task->cmp_def, lsm->index_id == 0,
- is_last_level, scheduler->read_views,
- lsm->index_id > 0 ? NULL :
- &task->deferred_delete_handler);
+ wi = vy_write_iterator_new(
+ task->cmp_def, lsm->index_id == 0, is_last_level,
+ scheduler->read_views,
+ lsm->index_id > 0 ? NULL : &task->deferred_delete_handler);
if (wi == NULL)
goto err_wi;
@@ -1657,11 +1663,11 @@ vy_task_compaction_new(struct vy_scheduler *scheduler, struct vy_worker *worker,
int32_t dump_count = 0;
int n = range->compaction_priority;
rlist_foreach_entry(slice, &range->slices, in_range) {
- if (vy_write_iterator_new_slice(wi, slice,
- lsm->disk_format) != 0)
+ if (vy_write_iterator_new_slice(wi, slice, lsm->disk_format) !=
+ 0)
goto err_wi_sub;
- new_run->dump_lsn = MAX(new_run->dump_lsn,
- slice->run->dump_lsn);
+ new_run->dump_lsn =
+ MAX(new_run->dump_lsn, slice->run->dump_lsn);
dump_count += slice->run->dump_count;
/* Remember the slices we are compacting. */
if (task->first_slice == NULL)
@@ -1702,7 +1708,7 @@ vy_task_compaction_new(struct vy_scheduler *scheduler, struct vy_worker *worker,
say_info("%s: started compacting range %s, runs %d/%d",
vy_lsm_name(lsm), vy_range_str(range),
- range->compaction_priority, range->slice_count);
+ range->compaction_priority, range->slice_count);
*p_task = task;
return 0;
@@ -1783,8 +1789,8 @@ static void
vy_task_complete_f(struct cmsg *cmsg)
{
struct vy_task *task = container_of(cmsg, struct vy_task, cmsg);
- stailq_add_tail_entry(&task->scheduler->processed_tasks,
- task, in_processed);
+ stailq_add_tail_entry(&task->scheduler->processed_tasks, task,
+ in_processed);
fiber_cond_signal(&task->scheduler->scheduler_cond);
}
@@ -1892,7 +1898,8 @@ vy_scheduler_peek_compaction(struct vy_scheduler *scheduler,
struct vy_worker *worker = NULL;
retry:
*ptask = NULL;
- struct vy_lsm *lsm = vy_compaction_heap_top(&scheduler->compaction_heap);
+ struct vy_lsm *lsm =
+ vy_compaction_heap_top(&scheduler->compaction_heap);
if (lsm == NULL)
goto no_task; /* nothing to do */
if (vy_lsm_compaction_priority(lsm) <= 1)
@@ -1908,7 +1915,7 @@ retry:
}
if (*ptask == NULL)
goto retry; /* LSM tree dropped or range split/coalesced */
- return 0; /* new task */
+ return 0; /* new task */
no_task:
if (worker != NULL)
vy_worker_pool_put(worker);
@@ -1939,7 +1946,6 @@ fail:
assert(!diag_is_empty(diag_get()));
diag_move(diag_get(), &scheduler->diag);
return -1;
-
}
static int
@@ -1956,12 +1962,11 @@ vy_task_complete(struct vy_task *task)
goto fail; /* ->execute fialed */
}
ERROR_INJECT(ERRINJ_VY_TASK_COMPLETE, {
- diag_set(ClientError, ER_INJECTION,
- "vinyl task completion");
- diag_move(diag_get(), diag);
- goto fail; });
- if (task->ops->complete &&
- task->ops->complete(task) != 0) {
+ diag_set(ClientError, ER_INJECTION, "vinyl task completion");
+ diag_move(diag_get(), diag);
+ goto fail;
+ });
+ if (task->ops->complete && task->ops->complete(task) != 0) {
assert(!diag_is_empty(diag_get()));
diag_move(diag_get(), diag);
goto fail;
@@ -1992,7 +1997,8 @@ vy_scheduler_f(va_list va)
/* Complete and delete all processed tasks. */
stailq_foreach_entry_safe(task, next, &processed_tasks,
- in_processed) {
+ in_processed)
+ {
if (vy_task_complete(task) != 0)
tasks_failed++;
else
@@ -2035,7 +2041,7 @@ vy_scheduler_f(va_list va)
fiber_reschedule();
continue;
-error:
+ error:
/* Abort pending checkpoint. */
fiber_cond_signal(&scheduler->dump_cond);
/*
diff --git a/src/box/vy_scheduler.h b/src/box/vy_scheduler.h
index f487b42..68f35ae 100644
--- a/src/box/vy_scheduler.h
+++ b/src/box/vy_scheduler.h
@@ -53,9 +53,9 @@ struct vy_run_env;
struct vy_worker;
struct vy_scheduler;
-typedef void
-(*vy_scheduler_dump_complete_f)(struct vy_scheduler *scheduler,
- int64_t dump_generation, double dump_duration);
+typedef void (*vy_scheduler_dump_complete_f)(struct vy_scheduler *scheduler,
+ int64_t dump_generation,
+ double dump_duration);
struct vy_worker_pool {
/** Name of the pool. Used for naming threads. */
diff --git a/src/box/vy_stmt.c b/src/box/vy_stmt.c
index 92e0aa1..c2be567 100644
--- a/src/box/vy_stmt.c
+++ b/src/box/vy_stmt.c
@@ -33,7 +33,7 @@
#include <stdlib.h>
#include <string.h>
-#include <sys/uio.h> /* struct iovec */
+#include <sys/uio.h> /* struct iovec */
#include <pmatomic.h> /* for refs */
#include "diag.h"
@@ -171,7 +171,7 @@ vy_stmt_alloc(struct tuple_format *format, uint32_t data_offset, uint32_t bsize)
uint32_t total_size = data_offset + bsize;
if (unlikely(total_size > env->max_tuple_size)) {
diag_set(ClientError, ER_VINYL_MAX_TUPLE_SIZE,
- (unsigned) total_size);
+ (unsigned)total_size);
error_log(diag_last_error(diag_get()));
return NULL;
}
@@ -190,8 +190,9 @@ vy_stmt_alloc(struct tuple_format *format, uint32_t data_offset, uint32_t bsize)
diag_set(OutOfMemory, total_size, "malloc", "struct vy_stmt");
return NULL;
}
- say_debug("vy_stmt_alloc(format = %d data_offset = %u, bsize = %u) = %p",
- format->id, data_offset, bsize, tuple);
+ say_debug(
+ "vy_stmt_alloc(format = %d data_offset = %u, bsize = %u) = %p",
+ format->id, data_offset, bsize, tuple);
tuple->refs = 1;
tuple->format_id = tuple_format_id(format);
if (cord_is_main())
@@ -213,8 +214,8 @@ vy_stmt_dup(struct tuple *stmt)
* tuple field map. This map can be simple memcopied from
* the original tuple.
*/
- struct tuple *res = vy_stmt_alloc(tuple_format(stmt),
- stmt->data_offset, stmt->bsize);
+ struct tuple *res = vy_stmt_alloc(tuple_format(stmt), stmt->data_offset,
+ stmt->bsize);
if (res == NULL)
return NULL;
assert(tuple_size(res) == tuple_size(stmt));
@@ -238,8 +239,8 @@ vy_stmt_dup_lsregion(struct tuple *stmt, struct lsregion *lsregion,
if (type == IPROTO_UPSERT)
alloc_size += align;
- mem_stmt = lsregion_aligned_alloc(lsregion, alloc_size, align,
- alloc_id);
+ mem_stmt =
+ lsregion_aligned_alloc(lsregion, alloc_size, align, alloc_id);
if (mem_stmt == NULL) {
diag_set(OutOfMemory, alloc_size, "lsregion_aligned_alloc",
"mem_stmt");
@@ -279,11 +280,12 @@ vy_key_new(struct tuple_format *format, const char *key, uint32_t part_count)
/* Allocate stmt */
uint32_t key_size = key_end - key;
uint32_t bsize = mp_sizeof_array(part_count) + key_size;
- struct tuple *stmt = vy_stmt_alloc(format, sizeof(struct vy_stmt), bsize);
+ struct tuple *stmt =
+ vy_stmt_alloc(format, sizeof(struct vy_stmt), bsize);
if (stmt == NULL)
return NULL;
/* Copy MsgPack data */
- char *raw = (char *) stmt + sizeof(struct vy_stmt);
+ char *raw = (char *)stmt + sizeof(struct vy_stmt);
char *data = mp_encode_array(raw, part_count);
memcpy(data, key, key_size);
assert(data + key_size == raw + bsize);
@@ -312,8 +314,8 @@ vy_key_dup(const char *key)
*/
static struct tuple *
vy_stmt_new_with_ops(struct tuple_format *format, const char *tuple_begin,
- const char *tuple_end, struct iovec *ops,
- int op_count, enum iproto_type type)
+ const char *tuple_end, struct iovec *ops, int op_count,
+ enum iproto_type type)
{
mp_tuple_assert(tuple_begin, tuple_end);
@@ -350,18 +352,17 @@ vy_stmt_new_with_ops(struct tuple_format *format, const char *tuple_begin,
*/
size_t mpsize = (tuple_end - tuple_begin);
size_t bsize = mpsize + ops_size;
- stmt = vy_stmt_alloc(format, sizeof(struct vy_stmt) +
- field_map_size, bsize);
+ stmt = vy_stmt_alloc(format, sizeof(struct vy_stmt) + field_map_size,
+ bsize);
if (stmt == NULL)
goto end;
/* Copy MsgPack data */
- char *raw = (char *) tuple_data(stmt);
+ char *raw = (char *)tuple_data(stmt);
char *wpos = raw;
field_map_build(&builder, wpos - field_map_size);
memcpy(wpos, tuple_begin, mpsize);
wpos += mpsize;
- for (struct iovec *op = ops, *end = ops + op_count;
- op != end; ++op) {
+ for (struct iovec *op = ops, *end = ops + op_count; op != end; ++op) {
memcpy(wpos, op->iov_base, op->iov_len);
wpos += op->iov_len;
}
@@ -376,32 +377,32 @@ vy_stmt_new_upsert(struct tuple_format *format, const char *tuple_begin,
const char *tuple_end, struct iovec *operations,
uint32_t ops_cnt)
{
- return vy_stmt_new_with_ops(format, tuple_begin, tuple_end,
- operations, ops_cnt, IPROTO_UPSERT);
+ return vy_stmt_new_with_ops(format, tuple_begin, tuple_end, operations,
+ ops_cnt, IPROTO_UPSERT);
}
struct tuple *
vy_stmt_new_replace(struct tuple_format *format, const char *tuple_begin,
const char *tuple_end)
{
- return vy_stmt_new_with_ops(format, tuple_begin, tuple_end,
- NULL, 0, IPROTO_REPLACE);
+ return vy_stmt_new_with_ops(format, tuple_begin, tuple_end, NULL, 0,
+ IPROTO_REPLACE);
}
struct tuple *
vy_stmt_new_insert(struct tuple_format *format, const char *tuple_begin,
const char *tuple_end)
{
- return vy_stmt_new_with_ops(format, tuple_begin, tuple_end,
- NULL, 0, IPROTO_INSERT);
+ return vy_stmt_new_with_ops(format, tuple_begin, tuple_end, NULL, 0,
+ IPROTO_INSERT);
}
struct tuple *
vy_stmt_new_delete(struct tuple_format *format, const char *tuple_begin,
const char *tuple_end)
{
- return vy_stmt_new_with_ops(format, tuple_begin, tuple_end,
- NULL, 0, IPROTO_DELETE);
+ return vy_stmt_new_with_ops(format, tuple_begin, tuple_end, NULL, 0,
+ IPROTO_DELETE);
}
struct tuple *
@@ -415,7 +416,8 @@ vy_stmt_replace_from_upsert(struct tuple *upsert)
/* Copy statement data excluding UPSERT operations */
struct tuple_format *format = tuple_format(upsert);
- struct tuple *replace = vy_stmt_alloc(format, upsert->data_offset, bsize);
+ struct tuple *replace =
+ vy_stmt_alloc(format, upsert->data_offset, bsize);
if (replace == NULL)
return NULL;
/* Copy both data and field_map. */
@@ -453,8 +455,8 @@ vy_stmt_new_surrogate_delete_raw(struct tuple_format *format,
uint32_t field_count;
struct tuple_format_iterator it;
if (tuple_format_iterator_create(&it, format, src_data,
- TUPLE_FORMAT_ITERATOR_KEY_PARTS_ONLY, &field_count,
- region) != 0)
+ TUPLE_FORMAT_ITERATOR_KEY_PARTS_ONLY,
+ &field_count, region) != 0)
goto out;
char *pos = mp_encode_array(data, field_count);
struct tuple_format_iterator_entry entry;
@@ -480,8 +482,9 @@ vy_stmt_new_surrogate_delete_raw(struct tuple_format *format,
uint32_t offset_slot = entry.field->offset_slot;
if (offset_slot != TUPLE_OFFSET_SLOT_NIL &&
field_map_builder_set_slot(&builder, offset_slot,
- pos - data, entry.multikey_idx,
- entry.multikey_count, region) != 0)
+ pos - data, entry.multikey_idx,
+ entry.multikey_count,
+ region) != 0)
goto out;
/* Copy field data. */
if (entry.field->type == FIELD_TYPE_ARRAY) {
@@ -502,7 +505,7 @@ vy_stmt_new_surrogate_delete_raw(struct tuple_format *format,
bsize);
if (stmt == NULL)
goto out;
- char *stmt_data = (char *) tuple_data(stmt);
+ char *stmt_data = (char *)tuple_data(stmt);
char *stmt_field_map_begin = stmt_data - field_map_size;
memcpy(stmt_data, data, bsize);
field_map_build(&builder, stmt_field_map_begin);
@@ -519,8 +522,8 @@ vy_stmt_extract_key(struct tuple *stmt, struct key_def *key_def,
{
struct region *region = &fiber()->gc;
size_t region_svp = region_used(region);
- const char *key_raw = tuple_extract_key(stmt, key_def,
- multikey_idx, NULL);
+ const char *key_raw =
+ tuple_extract_key(stmt, key_def, multikey_idx, NULL);
if (key_raw == NULL)
return NULL;
uint32_t part_count = mp_decode_array(&key_raw);
@@ -551,34 +554,36 @@ vy_stmt_extract_key_raw(const char *data, const char *data_end,
}
int
-vy_bloom_builder_add(struct tuple_bloom_builder *builder,
- struct vy_entry entry, struct key_def *key_def)
+vy_bloom_builder_add(struct tuple_bloom_builder *builder, struct vy_entry entry,
+ struct key_def *key_def)
{
struct tuple *stmt = entry.stmt;
if (vy_stmt_is_key(stmt)) {
const char *data = tuple_data(stmt);
uint32_t part_count = mp_decode_array(&data);
- return tuple_bloom_builder_add_key(builder, data,
- part_count, key_def);
+ return tuple_bloom_builder_add_key(builder, data, part_count,
+ key_def);
} else {
return tuple_bloom_builder_add(builder, stmt, key_def,
- vy_entry_multikey_idx(entry, key_def));
+ vy_entry_multikey_idx(entry,
+ key_def));
}
}
bool
-vy_bloom_maybe_has(const struct tuple_bloom *bloom,
- struct vy_entry entry, struct key_def *key_def)
+vy_bloom_maybe_has(const struct tuple_bloom *bloom, struct vy_entry entry,
+ struct key_def *key_def)
{
struct tuple *stmt = entry.stmt;
if (vy_stmt_is_key(stmt)) {
const char *data = tuple_data(stmt);
uint32_t part_count = mp_decode_array(&data);
- return tuple_bloom_maybe_has_key(bloom, data,
- part_count, key_def);
+ return tuple_bloom_maybe_has_key(bloom, data, part_count,
+ key_def);
} else {
return tuple_bloom_maybe_has(bloom, stmt, key_def,
- vy_entry_multikey_idx(entry, key_def));
+ vy_entry_multikey_idx(entry,
+ key_def));
}
}
@@ -652,9 +657,9 @@ vy_stmt_encode_primary(struct tuple *value, struct key_def *key_def,
switch (type) {
case IPROTO_DELETE:
extracted = vy_stmt_is_key(value) ?
- tuple_data_range(value, &size) :
- tuple_extract_key(value, key_def,
- MULTIKEY_NONE, &size);
+ tuple_data_range(value, &size) :
+ tuple_extract_key(value, key_def,
+ MULTIKEY_NONE, &size);
if (extracted == NULL)
return -1;
request.key = extracted;
@@ -696,10 +701,10 @@ vy_stmt_encode_secondary(struct tuple *value, struct key_def *cmp_def,
memset(&request, 0, sizeof(request));
request.type = type;
uint32_t size;
- const char *extracted = vy_stmt_is_key(value) ?
- tuple_data_range(value, &size) :
- tuple_extract_key(value, cmp_def,
- multikey_idx, &size);
+ const char *extracted =
+ vy_stmt_is_key(value) ?
+ tuple_data_range(value, &size) :
+ tuple_extract_key(value, cmp_def, multikey_idx, &size);
if (extracted == NULL)
return -1;
if (type == IPROTO_REPLACE || type == IPROTO_INSERT) {
@@ -733,15 +738,15 @@ vy_stmt_decode(struct xrow_header *xrow, struct tuple_format *format)
switch (request.type) {
case IPROTO_DELETE:
/* Always use key format for DELETE statements. */
- stmt = vy_stmt_new_with_ops(env->key_format,
- request.key, request.key_end,
- NULL, 0, IPROTO_DELETE);
+ stmt = vy_stmt_new_with_ops(env->key_format, request.key,
+ request.key_end, NULL, 0,
+ IPROTO_DELETE);
break;
case IPROTO_INSERT:
case IPROTO_REPLACE:
stmt = vy_stmt_new_with_ops(format, request.tuple,
- request.tuple_end,
- NULL, 0, request.type);
+ request.tuple_end, NULL, 0,
+ request.type);
break;
case IPROTO_UPSERT:
ops.iov_base = (char *)request.ops;
@@ -781,14 +786,14 @@ vy_stmt_snprint(char *buf, int size, struct tuple *stmt)
}
SNPRINT(total, snprintf, buf, size, "%s(",
iproto_type_name(vy_stmt_type(stmt)));
- SNPRINT(total, mp_snprint, buf, size, tuple_data(stmt));
+ SNPRINT(total, mp_snprint, buf, size, tuple_data(stmt));
if (vy_stmt_type(stmt) == IPROTO_UPSERT) {
SNPRINT(total, snprintf, buf, size, ", ops=");
SNPRINT(total, mp_snprint, buf, size,
vy_stmt_upsert_ops(stmt, &mp_size));
}
SNPRINT(total, snprintf, buf, size, ", lsn=%lld)",
- (long long) vy_stmt_lsn(stmt));
+ (long long)vy_stmt_lsn(stmt));
return total;
}
diff --git a/src/box/vy_stmt.h b/src/box/vy_stmt.h
index 2521923..2638761 100644
--- a/src/box/vy_stmt.h
+++ b/src/box/vy_stmt.h
@@ -117,7 +117,7 @@ enum {
* secondary indexes. It makes the write iterator generate
* DELETE statements for them during compaction.
*/
- VY_STMT_DEFERRED_DELETE = 1 << 0,
+ VY_STMT_DEFERRED_DELETE = 1 << 0,
/**
* Statements that have this flag set are ignored by the
* read iterator.
@@ -127,7 +127,7 @@ enum {
* the older a source, the older statements it stores for a
* particular key.
*/
- VY_STMT_SKIP_READ = 1 << 1,
+ VY_STMT_SKIP_READ = 1 << 1,
/**
* This flag is set for those REPLACE statements that were
* generated by UPDATE operations. It is used by the write
@@ -135,12 +135,12 @@ enum {
* indexes so that they can get annihilated with DELETEs on
* compaction. It is never written to disk.
*/
- VY_STMT_UPDATE = 1 << 2,
+ VY_STMT_UPDATE = 1 << 2,
/**
* Bit mask of all statement flags.
*/
- VY_STMT_FLAGS_ALL = (VY_STMT_DEFERRED_DELETE | VY_STMT_SKIP_READ |
- VY_STMT_UPDATE),
+ VY_STMT_FLAGS_ALL =
+ (VY_STMT_DEFERRED_DELETE | VY_STMT_SKIP_READ | VY_STMT_UPDATE),
};
/**
@@ -172,7 +172,7 @@ enum {
struct vy_stmt {
struct tuple base;
int64_t lsn;
- uint8_t type; /* IPROTO_INSERT/REPLACE/UPSERT/DELETE */
+ uint8_t type; /* IPROTO_INSERT/REPLACE/UPSERT/DELETE */
uint8_t flags;
/**
* Offsets array concatenated with MessagePack fields
@@ -185,28 +185,28 @@ struct vy_stmt {
static inline int64_t
vy_stmt_lsn(struct tuple *stmt)
{
- return ((struct vy_stmt *) stmt)->lsn;
+ return ((struct vy_stmt *)stmt)->lsn;
}
/** Set LSN of the vinyl statement. */
static inline void
vy_stmt_set_lsn(struct tuple *stmt, int64_t lsn)
{
- ((struct vy_stmt *) stmt)->lsn = lsn;
+ ((struct vy_stmt *)stmt)->lsn = lsn;
}
/** Get type of the vinyl statement. */
static inline enum iproto_type
vy_stmt_type(struct tuple *stmt)
{
- return (enum iproto_type)((struct vy_stmt *) stmt)->type;
+ return (enum iproto_type)((struct vy_stmt *)stmt)->type;
}
/** Set type of the vinyl statement. */
static inline void
vy_stmt_set_type(struct tuple *stmt, enum iproto_type type)
{
- ((struct vy_stmt *) stmt)->type = type;
+ ((struct vy_stmt *)stmt)->type = type;
}
/** Get flags of the vinyl statement. */
@@ -383,8 +383,7 @@ vy_stmt_hint(struct tuple *stmt, struct key_def *key_def)
* formats (key or tuple) and using comparison hints.
*/
static inline int
-vy_stmt_compare(struct tuple *a, hint_t a_hint,
- struct tuple *b, hint_t b_hint,
+vy_stmt_compare(struct tuple *a, hint_t a_hint, struct tuple *b, hint_t b_hint,
struct key_def *key_def)
{
bool a_is_tuple = !vy_stmt_is_key(a);
@@ -403,8 +402,8 @@ vy_stmt_compare(struct tuple *a, hint_t a_hint,
a_hint, key_def);
} else {
assert(!a_is_tuple && !b_is_tuple);
- return key_compare(tuple_data(a), a_hint,
- tuple_data(b), b_hint, key_def);
+ return key_compare(tuple_data(a), a_hint, tuple_data(b), b_hint,
+ key_def);
}
}
@@ -419,9 +418,8 @@ vy_stmt_compare_with_raw_key(struct tuple *stmt, hint_t stmt_hint,
{
if (!vy_stmt_is_key(stmt)) {
uint32_t part_count = mp_decode_array(&key);
- return tuple_compare_with_key(stmt, stmt_hint, key,
- part_count, key_hint,
- key_def);
+ return tuple_compare_with_key(stmt, stmt_hint, key, part_count,
+ key_hint, key_def);
}
return key_compare(tuple_data(stmt), stmt_hint, key, key_hint, key_def);
}
@@ -465,8 +463,8 @@ vy_key_dup(const char *key);
* @retval NULL Memory or fields format error.
*/
struct tuple *
-vy_stmt_new_surrogate_delete_raw(struct tuple_format *format,
- const char *data, const char *data_end);
+vy_stmt_new_surrogate_delete_raw(struct tuple_format *format, const char *data,
+ const char *data_end);
/** @copydoc vy_stmt_new_surrogate_delete_raw. */
static inline struct tuple *
@@ -489,7 +487,7 @@ vy_stmt_new_surrogate_delete(struct tuple_format *format, struct tuple *tuple)
*/
struct tuple *
vy_stmt_new_replace(struct tuple_format *format, const char *tuple,
- const char *tuple_end);
+ const char *tuple_end);
/**
* Create the INSERT statement from raw MessagePack data.
@@ -519,7 +517,7 @@ struct tuple *
vy_stmt_new_delete(struct tuple_format *format, const char *tuple_begin,
const char *tuple_end);
- /**
+/**
* Create the UPSERT statement from raw MessagePack data.
* @param tuple_begin MessagePack data that contain an array of fields WITH the
* array header.
@@ -533,9 +531,9 @@ vy_stmt_new_delete(struct tuple_format *format, const char *tuple_begin,
* @retval not NULL Success.
*/
struct tuple *
-vy_stmt_new_upsert(struct tuple_format *format,
- const char *tuple_begin, const char *tuple_end,
- struct iovec *operations, uint32_t ops_cnt);
+vy_stmt_new_upsert(struct tuple_format *format, const char *tuple_begin,
+ const char *tuple_end, struct iovec *operations,
+ uint32_t ops_cnt);
/**
* Create REPLACE statement from UPSERT statement.
@@ -623,16 +621,16 @@ vy_stmt_extract_key_raw(const char *data, const char *data_end,
* See tuple_bloom_builder_add() for more details.
*/
int
-vy_bloom_builder_add(struct tuple_bloom_builder *builder,
- struct vy_entry entry, struct key_def *key_def);
+vy_bloom_builder_add(struct tuple_bloom_builder *builder, struct vy_entry entry,
+ struct key_def *key_def);
/**
* Check if a statement hash is present in a bloom filter.
* See tuple_bloom_maybe_has() for more details.
*/
bool
-vy_bloom_maybe_has(const struct tuple_bloom *bloom,
- struct vy_entry entry, struct key_def *key_def);
+vy_bloom_maybe_has(const struct tuple_bloom *bloom, struct vy_entry entry,
+ struct key_def *key_def);
/**
* Encode vy_stmt for a primary key as xrow_header
@@ -742,12 +740,11 @@ vy_entry_compare(struct vy_entry a, struct vy_entry b, struct key_def *key_def)
* (msgpack array).
*/
static inline int
-vy_entry_compare_with_raw_key(struct vy_entry entry,
- const char *key, hint_t key_hint,
- struct key_def *key_def)
+vy_entry_compare_with_raw_key(struct vy_entry entry, const char *key,
+ hint_t key_hint, struct key_def *key_def)
{
- return vy_stmt_compare_with_raw_key(entry.stmt, entry.hint,
- key, key_hint, key_def);
+ return vy_stmt_compare_with_raw_key(entry.stmt, entry.hint, key,
+ key_hint, key_def);
}
/**
@@ -764,15 +761,18 @@ vy_entry_compare_with_raw_key(struct vy_entry entry,
*
* entry.stmt is set to src_stmt on each iteration.
*/
-#define vy_stmt_foreach_entry(entry, src_stmt, key_def) \
- for (uint32_t multikey_idx = 0, \
- multikey_count = !(key_def)->is_multikey ? 1 : \
- tuple_multikey_count((src_stmt), (key_def)); \
- multikey_idx < multikey_count && \
- (((entry).stmt = (src_stmt)), \
- ((entry).hint = !(key_def)->is_multikey ? \
- vy_stmt_hint((src_stmt), (key_def)) : \
- multikey_idx), true); \
+#define vy_stmt_foreach_entry(entry, src_stmt, key_def) \
+ for (uint32_t multikey_idx = 0, \
+ multikey_count = !(key_def)->is_multikey ? \
+ 1 : \
+ tuple_multikey_count( \
+ (src_stmt), (key_def)); \
+ multikey_idx < multikey_count && \
+ (((entry).stmt = (src_stmt)), \
+ ((entry).hint = !(key_def)->is_multikey ? \
+ vy_stmt_hint((src_stmt), (key_def)) : \
+ multikey_idx), \
+ true); \
++multikey_idx)
#if defined(__cplusplus)
diff --git a/src/box/vy_stmt_stream.h b/src/box/vy_stmt_stream.h
index 08e4d5f..c1e0589 100644
--- a/src/box/vy_stmt_stream.h
+++ b/src/box/vy_stmt_stream.h
@@ -48,20 +48,18 @@ struct vy_stmt_stream;
/**
* Start streaming
*/
-typedef NODISCARD int
-(*vy_stream_start_f)(struct vy_stmt_stream *virt_stream);
+typedef NODISCARD int (*vy_stream_start_f)(struct vy_stmt_stream *virt_stream);
/**
* Get next tuple from a stream.
*/
-typedef NODISCARD int
-(*vy_stream_next_f)(struct vy_stmt_stream *virt_stream, struct vy_entry *ret);
+typedef NODISCARD int (*vy_stream_next_f)(struct vy_stmt_stream *virt_stream,
+ struct vy_entry *ret);
/**
* Close the stream.
*/
-typedef void
-(*vy_stream_close_f)(struct vy_stmt_stream *virt_stream);
+typedef void (*vy_stream_close_f)(struct vy_stmt_stream *virt_stream);
/**
* The interface description for streams over run and mem.
diff --git a/src/box/vy_tx.c b/src/box/vy_tx.c
index ff63cd7..0fb9eed 100644
--- a/src/box/vy_tx.c
+++ b/src/box/vy_tx.c
@@ -103,8 +103,8 @@ vy_tx_manager_new(void)
{
struct vy_tx_manager *xm = calloc(1, sizeof(*xm));
if (xm == NULL) {
- diag_set(OutOfMemory, sizeof(*xm),
- "malloc", "struct vy_tx_manager");
+ diag_set(OutOfMemory, sizeof(*xm), "malloc",
+ "struct vy_tx_manager");
return NULL;
}
@@ -113,8 +113,8 @@ vy_tx_manager_new(void)
vy_global_read_view_create((struct vy_read_view *)&xm->global_read_view,
INT64_MAX);
xm->p_global_read_view = &xm->global_read_view;
- vy_global_read_view_create((struct vy_read_view *)&xm->committed_read_view,
- MAX_LSN - 1);
+ vy_global_read_view_create(
+ (struct vy_read_view *)&xm->committed_read_view, MAX_LSN - 1);
xm->p_committed_read_view = &xm->committed_read_view;
struct slab_cache *slab_cache = cord_slab_cache();
@@ -171,15 +171,13 @@ vy_tx_manager_read_view(struct vy_tx_manager *xm)
if ((xm->last_prepared_tx == NULL && rv->vlsn == xm->lsn) ||
(xm->last_prepared_tx != NULL &&
rv->vlsn == MAX_LSN + xm->last_prepared_tx->psn)) {
-
rv->refs++;
- return rv;
+ return rv;
}
}
rv = mempool_alloc(&xm->read_view_mempool);
if (rv == NULL) {
- diag_set(OutOfMemory, sizeof(*rv),
- "mempool", "read view");
+ diag_set(OutOfMemory, sizeof(*rv), "mempool", "read view");
return NULL;
}
if (xm->last_prepared_tx != NULL) {
@@ -196,7 +194,7 @@ vy_tx_manager_read_view(struct vy_tx_manager *xm)
void
vy_tx_manager_destroy_read_view(struct vy_tx_manager *xm,
- struct vy_read_view *rv)
+ struct vy_read_view *rv)
{
if (rv == xm->p_global_read_view)
return;
@@ -268,16 +266,16 @@ vy_read_interval_unacct(struct vy_read_interval *interval)
}
static struct vy_read_interval *
-vy_read_interval_new(struct vy_tx *tx, struct vy_lsm *lsm,
- struct vy_entry left, bool left_belongs,
- struct vy_entry right, bool right_belongs)
+vy_read_interval_new(struct vy_tx *tx, struct vy_lsm *lsm, struct vy_entry left,
+ bool left_belongs, struct vy_entry right,
+ bool right_belongs)
{
struct vy_tx_manager *xm = tx->xm;
struct vy_read_interval *interval;
interval = mempool_alloc(&xm->read_interval_mempool);
if (interval == NULL) {
- diag_set(OutOfMemory, sizeof(*interval),
- "mempool", "struct vy_read_interval");
+ diag_set(OutOfMemory, sizeof(*interval), "mempool",
+ "struct vy_read_interval");
return NULL;
}
interval->tx = tx;
@@ -343,8 +341,7 @@ vy_tx_destroy(struct vy_tx *tx)
vy_tx_manager_destroy_read_view(tx->xm, tx->read_view);
struct txv *v, *tmp;
- stailq_foreach_entry_safe(v, tmp, &tx->log, next_in_log)
- txv_delete(v);
+ stailq_foreach_entry_safe(v, tmp, &tx->log, next_in_log) txv_delete(v);
vy_tx_read_set_iter(&tx->read_set, NULL, vy_tx_read_set_free_cb, NULL);
rlist_del_entry(tx, in_writers);
@@ -484,8 +481,8 @@ vy_tx_write_prepare(struct txv *v)
* @retval -1 Memory error.
*/
static int
-vy_tx_write(struct vy_lsm *lsm, struct vy_mem *mem,
- struct vy_entry entry, struct tuple **region_stmt)
+vy_tx_write(struct vy_lsm *lsm, struct vy_mem *mem, struct vy_entry entry,
+ struct tuple **region_stmt)
{
assert(vy_stmt_is_refable(entry.stmt));
assert(*region_stmt == NULL || !vy_stmt_is_refable(*region_stmt));
@@ -511,7 +508,7 @@ vy_tx_write(struct vy_lsm *lsm, struct vy_mem *mem,
vy_stmt_type(applied.stmt);
assert(applied_type == IPROTO_REPLACE ||
applied_type == IPROTO_INSERT);
- (void) applied_type;
+ (void)applied_type;
int rc = vy_lsm_set(lsm, mem, applied,
region_stmt);
tuple_unref(applied.stmt);
@@ -569,8 +566,8 @@ vy_tx_handle_deferred_delete(struct vy_tx *tx, struct txv *v)
/* Look up the tuple overwritten by this statement. */
struct vy_entry overwritten;
- if (vy_point_lookup_mem(pk, &tx->xm->p_global_read_view,
- v->entry, &overwritten) != 0)
+ if (vy_point_lookup_mem(pk, &tx->xm->p_global_read_view, v->entry,
+ &overwritten) != 0)
return -1;
if (overwritten.stmt == NULL) {
@@ -596,8 +593,8 @@ vy_tx_handle_deferred_delete(struct vy_tx *tx, struct txv *v)
}
struct tuple *delete_stmt;
- delete_stmt = vy_stmt_new_surrogate_delete(pk->mem_format,
- overwritten.stmt);
+ delete_stmt =
+ vy_stmt_new_surrogate_delete(pk->mem_format, overwritten.stmt);
tuple_unref(overwritten.stmt);
if (delete_stmt == NULL)
return -1;
@@ -641,7 +638,7 @@ vy_tx_handle_deferred_delete(struct vy_tx *tx, struct txv *v)
*/
assert(vy_stmt_type(stmt) == IPROTO_REPLACE);
assert(vy_stmt_type(other->entry.stmt) ==
- IPROTO_REPLACE);
+ IPROTO_REPLACE);
other->is_nop = true;
continue;
}
@@ -704,7 +701,8 @@ vy_tx_prepare(struct vy_tx *tx)
/* repsert - REPLACE/UPSERT */
struct tuple *delete = NULL, *repsert = NULL;
MAYBE_UNUSED uint32_t current_space_id = 0;
- stailq_foreach_entry(v, &tx->log, next_in_log) {
+ stailq_foreach_entry(v, &tx->log, next_in_log)
+ {
struct vy_lsm *lsm = v->lsm;
if (lsm->index_id == 0) {
/* The beginning of the new txn_stmt is met. */
@@ -755,8 +753,9 @@ vy_tx_prepare(struct vy_tx *tx)
*/
uint8_t flags = vy_stmt_flags(v->entry.stmt);
if (flags & VY_STMT_DEFERRED_DELETE) {
- vy_stmt_set_flags(v->entry.stmt, flags &
- ~VY_STMT_DEFERRED_DELETE);
+ vy_stmt_set_flags(
+ v->entry.stmt,
+ flags & ~VY_STMT_DEFERRED_DELETE);
}
}
@@ -780,8 +779,8 @@ vy_tx_prepare(struct vy_tx *tx)
/* In secondary indexes only REPLACE/DELETE can be written. */
vy_stmt_set_lsn(v->entry.stmt, MAX_LSN + tx->psn);
- struct tuple **region_stmt =
- (type == IPROTO_DELETE) ? &delete : &repsert;
+ struct tuple **region_stmt = (type == IPROTO_DELETE) ? &delete :
+ &repsert;
if (vy_tx_write(lsm, v->mem, v->entry, region_stmt) != 0)
return -1;
v->region_stmt = *region_stmt;
@@ -809,7 +808,8 @@ vy_tx_commit(struct vy_tx *tx, int64_t lsn)
/* Fix LSNs of the records and commit changes. */
struct txv *v;
- stailq_foreach_entry(v, &tx->log, next_in_log) {
+ stailq_foreach_entry(v, &tx->log, next_in_log)
+ {
if (v->region_stmt != NULL) {
struct vy_entry entry;
entry.stmt = v->region_stmt;
@@ -858,7 +858,8 @@ vy_tx_rollback_after_prepare(struct vy_tx *tx)
xm->last_prepared_tx = NULL;
struct txv *v;
- stailq_foreach_entry(v, &tx->log, next_in_log) {
+ stailq_foreach_entry(v, &tx->log, next_in_log)
+ {
if (v->region_stmt != NULL) {
struct vy_entry entry;
entry.stmt = v->region_stmt;
@@ -908,8 +909,7 @@ vy_tx_begin_statement(struct vy_tx *tx, struct space *space, void **savepoint)
void
vy_tx_rollback_statement(struct vy_tx *tx, void *svp)
{
- if (tx->state == VINYL_TX_ABORT ||
- tx->state == VINYL_TX_COMMIT)
+ if (tx->state == VINYL_TX_ABORT || tx->state == VINYL_TX_COMMIT)
return;
assert(tx->state == VINYL_TX_READY);
@@ -919,7 +919,8 @@ vy_tx_rollback_statement(struct vy_tx *tx, void *svp)
/* Rollback statements in LIFO order. */
stailq_reverse(&tail);
struct txv *v, *tmp;
- stailq_foreach_entry_safe(v, tmp, &tail, next_in_log) {
+ stailq_foreach_entry_safe(v, tmp, &tail, next_in_log)
+ {
write_set_remove(&tx->write_set, v);
if (v->overwritten != NULL) {
/* Restore overwritten statement. */
@@ -935,9 +936,8 @@ vy_tx_rollback_statement(struct vy_tx *tx, void *svp)
}
int
-vy_tx_track(struct vy_tx *tx, struct vy_lsm *lsm,
- struct vy_entry left, bool left_belongs,
- struct vy_entry right, bool right_belongs)
+vy_tx_track(struct vy_tx *tx, struct vy_lsm *lsm, struct vy_entry left,
+ bool left_belongs, struct vy_entry right, bool right_belongs)
{
if (vy_tx_is_in_read_view(tx)) {
/* No point in tracking reads. */
@@ -945,8 +945,8 @@ vy_tx_track(struct vy_tx *tx, struct vy_lsm *lsm,
}
struct vy_read_interval *new_interval;
- new_interval = vy_read_interval_new(tx, lsm, left, left_belongs,
- right, right_belongs);
+ new_interval = vy_read_interval_new(tx, lsm, left, left_belongs, right,
+ right_belongs);
if (new_interval == NULL)
return -1;
@@ -1006,7 +1006,8 @@ vy_tx_track(struct vy_tx *tx, struct vy_lsm *lsm,
}
struct vy_read_interval *next_interval;
stailq_foreach_entry_safe(interval, next_interval, &merge,
- in_merge) {
+ in_merge)
+ {
vy_tx_read_set_remove(&tx->read_set, interval);
vy_lsm_read_set_remove(&lsm->read_set, interval);
vy_read_interval_delete(interval);
@@ -1059,14 +1060,12 @@ vy_tx_set_entry(struct vy_tx *tx, struct vy_lsm *lsm, struct vy_entry entry)
if (old != NULL && vy_stmt_type(entry.stmt) == IPROTO_UPSERT) {
assert(lsm->index_id == 0);
uint8_t old_type = vy_stmt_type(old->entry.stmt);
- assert(old_type == IPROTO_UPSERT ||
- old_type == IPROTO_INSERT ||
- old_type == IPROTO_REPLACE ||
- old_type == IPROTO_DELETE);
- (void) old_type;
-
- applied = vy_entry_apply_upsert(entry, old->entry,
- lsm->cmp_def, true);
+ assert(old_type == IPROTO_UPSERT || old_type == IPROTO_INSERT ||
+ old_type == IPROTO_REPLACE || old_type == IPROTO_DELETE);
+ (void)old_type;
+
+ applied = vy_entry_apply_upsert(entry, old->entry, lsm->cmp_def,
+ true);
lsm->stat.upsert.applied++;
if (applied.stmt == NULL)
return -1;
@@ -1096,8 +1095,8 @@ vy_tx_set_entry(struct vy_tx *tx, struct vy_lsm *lsm, struct vy_entry entry)
*/
if (vy_stmt_flags(old->entry.stmt) & VY_STMT_DEFERRED_DELETE) {
uint8_t flags = vy_stmt_flags(entry.stmt);
- vy_stmt_set_flags(entry.stmt, flags |
- VY_STMT_DEFERRED_DELETE);
+ vy_stmt_set_flags(entry.stmt,
+ flags | VY_STMT_DEFERRED_DELETE);
}
}
@@ -1179,9 +1178,9 @@ vy_tx_manager_abort_writers_for_ro(struct vy_tx_manager *xm)
void
vy_txw_iterator_open(struct vy_txw_iterator *itr,
- struct vy_txw_iterator_stat *stat,
- struct vy_tx *tx, struct vy_lsm *lsm,
- enum iterator_type iterator_type, struct vy_entry key)
+ struct vy_txw_iterator_stat *stat, struct vy_tx *tx,
+ struct vy_lsm *lsm, enum iterator_type iterator_type,
+ struct vy_entry key)
{
itr->stat = stat;
itr->tx = tx;
@@ -1210,7 +1209,8 @@ vy_txw_iterator_seek(struct vy_txw_iterator *itr, struct vy_entry last)
if (last.stmt != NULL) {
key = last;
iterator_type = iterator_direction(iterator_type) > 0 ?
- ITER_GT : ITER_LT;
+ ITER_GT :
+ ITER_LT;
}
struct vy_lsm *lsm = itr->lsm;
@@ -1230,9 +1230,11 @@ vy_txw_iterator_seek(struct vy_txw_iterator *itr, struct vy_entry last)
struct txv *next;
if (iterator_type == ITER_LE ||
iterator_type == ITER_GT)
- next = write_set_next(&itr->tx->write_set, txv);
+ next = write_set_next(
+ &itr->tx->write_set, txv);
else
- next = write_set_prev(&itr->tx->write_set, txv);
+ next = write_set_prev(
+ &itr->tx->write_set, txv);
if (next == NULL || next->lsm != lsm)
break;
if (vy_entry_compare(key, next->entry,
@@ -1260,8 +1262,7 @@ vy_txw_iterator_seek(struct vy_txw_iterator *itr, struct vy_entry last)
}
NODISCARD int
-vy_txw_iterator_next(struct vy_txw_iterator *itr,
- struct vy_history *history)
+vy_txw_iterator_next(struct vy_txw_iterator *itr, struct vy_history *history)
{
vy_history_cleanup(history);
if (!itr->search_started) {
@@ -1273,9 +1274,11 @@ vy_txw_iterator_next(struct vy_txw_iterator *itr,
if (itr->curr_txv == NULL)
return 0;
if (itr->iterator_type == ITER_LE || itr->iterator_type == ITER_LT)
- itr->curr_txv = write_set_prev(&itr->tx->write_set, itr->curr_txv);
+ itr->curr_txv =
+ write_set_prev(&itr->tx->write_set, itr->curr_txv);
else
- itr->curr_txv = write_set_next(&itr->tx->write_set, itr->curr_txv);
+ itr->curr_txv =
+ write_set_next(&itr->tx->write_set, itr->curr_txv);
if (itr->curr_txv != NULL && itr->curr_txv->lsm != itr->lsm)
itr->curr_txv = NULL;
if (itr->curr_txv != NULL && itr->iterator_type == ITER_EQ &&
@@ -1305,8 +1308,9 @@ vy_txw_iterator_skip(struct vy_txw_iterator *itr, struct vy_entry last,
if (itr->search_started &&
(itr->curr_txv == NULL || last.stmt == NULL ||
iterator_direction(itr->iterator_type) *
- vy_entry_compare(itr->curr_txv->entry, last,
- itr->lsm->cmp_def) > 0))
+ vy_entry_compare(itr->curr_txv->entry, last,
+ itr->lsm->cmp_def) >
+ 0))
return 0;
vy_history_cleanup(history);
diff --git a/src/box/vy_tx.h b/src/box/vy_tx.h
index 4fac5f6..ba9be3d 100644
--- a/src/box/vy_tx.h
+++ b/src/box/vy_tx.h
@@ -128,7 +128,8 @@ write_set_key_cmp(struct write_set_key *a, struct txv *b);
typedef rb_tree(struct txv) write_set_t;
rb_gen_ext_key(MAYBE_UNUSED static inline, write_set_, write_set_t, struct txv,
- in_set, write_set_cmp, struct write_set_key *, write_set_key_cmp);
+ in_set, write_set_cmp, struct write_set_key *,
+ write_set_key_cmp);
static inline struct txv *
write_set_search_key(write_set_t *tree, struct vy_lsm *lsm,
@@ -296,7 +297,7 @@ vy_tx_manager_read_view(struct vy_tx_manager *xm);
/** Dereference and possibly destroy a read view. */
void
vy_tx_manager_destroy_read_view(struct vy_tx_manager *xm,
- struct vy_read_view *rv);
+ struct vy_read_view *rv);
/**
* Abort all rw transactions that affect the given space
@@ -309,7 +310,7 @@ vy_tx_manager_destroy_read_view(struct vy_tx_manager *xm,
*/
void
vy_tx_manager_abort_writers_for_ddl(struct vy_tx_manager *xm,
- struct space *space, bool *need_wal_sync);
+ struct space *space, bool *need_wal_sync);
/**
* Abort all local rw transactions that haven't reached WAL yet.
@@ -386,9 +387,8 @@ vy_tx_rollback_statement(struct vy_tx *tx, void *svp);
* @retval -1 Memory error.
*/
int
-vy_tx_track(struct vy_tx *tx, struct vy_lsm *lsm,
- struct vy_entry left, bool left_belongs,
- struct vy_entry right, bool right_belongs);
+vy_tx_track(struct vy_tx *tx, struct vy_lsm *lsm, struct vy_entry left,
+ bool left_belongs, struct vy_entry right, bool right_belongs);
/**
* Remember a point read in the conflict manager index.
@@ -453,9 +453,9 @@ struct vy_txw_iterator {
*/
void
vy_txw_iterator_open(struct vy_txw_iterator *itr,
- struct vy_txw_iterator_stat *stat,
- struct vy_tx *tx, struct vy_lsm *lsm,
- enum iterator_type iterator_type, struct vy_entry key);
+ struct vy_txw_iterator_stat *stat, struct vy_tx *tx,
+ struct vy_lsm *lsm, enum iterator_type iterator_type,
+ struct vy_entry key);
/**
* Advance a txw iterator to the next key.
@@ -463,8 +463,7 @@ vy_txw_iterator_open(struct vy_txw_iterator *itr,
* Returns 0 on success, -1 on memory allocation error.
*/
NODISCARD int
-vy_txw_iterator_next(struct vy_txw_iterator *itr,
- struct vy_history *history);
+vy_txw_iterator_next(struct vy_txw_iterator *itr, struct vy_history *history);
/**
* Advance a txw iterator to the key following @last.
diff --git a/src/box/vy_upsert.c b/src/box/vy_upsert.c
index 797492c..4ddbe26e 100644
--- a/src/box/vy_upsert.c
+++ b/src/box/vy_upsert.c
@@ -47,19 +47,17 @@
* @retval -1 - memory error
*/
static int
-vy_upsert_try_to_squash(struct tuple_format *format,
- const char *key_mp, const char *key_mp_end,
- const char *old_serie, const char *old_serie_end,
- const char *new_serie, const char *new_serie_end,
- struct tuple **result_stmt)
+vy_upsert_try_to_squash(struct tuple_format *format, const char *key_mp,
+ const char *key_mp_end, const char *old_serie,
+ const char *old_serie_end, const char *new_serie,
+ const char *new_serie_end, struct tuple **result_stmt)
{
*result_stmt = NULL;
size_t squashed_size;
- const char *squashed =
- xrow_upsert_squash(old_serie, old_serie_end,
- new_serie, new_serie_end, format,
- &squashed_size, 0);
+ const char *squashed = xrow_upsert_squash(old_serie, old_serie_end,
+ new_serie, new_serie_end,
+ format, &squashed_size, 0);
if (squashed == NULL)
return 0;
/* Successful squash! */
@@ -67,8 +65,8 @@ vy_upsert_try_to_squash(struct tuple_format *format,
operations[0].iov_base = (void *)squashed;
operations[0].iov_len = squashed_size;
- *result_stmt = vy_stmt_new_upsert(format, key_mp, key_mp_end,
- operations, 1);
+ *result_stmt =
+ vy_stmt_new_upsert(format, key_mp, key_mp_end, operations, 1);
if (*result_stmt == NULL)
return -1;
return 0;
@@ -119,21 +117,20 @@ vy_apply_upsert(struct tuple *new_stmt, struct tuple *old_stmt,
uint8_t old_type = vy_stmt_type(old_stmt);
uint64_t column_mask = COLUMN_MASK_FULL;
result_mp = xrow_upsert_execute(new_ops, new_ops_end, result_mp,
- result_mp_end, format, &mp_size,
- 0, suppress_error, &column_mask);
+ result_mp_end, format, &mp_size, 0,
+ suppress_error, &column_mask);
if (result_mp == NULL) {
region_truncate(region, region_svp);
return NULL;
}
result_mp_end = result_mp + mp_size;
if (old_type != IPROTO_UPSERT) {
- assert(old_type == IPROTO_INSERT ||
- old_type == IPROTO_REPLACE);
+ assert(old_type == IPROTO_INSERT || old_type == IPROTO_REPLACE);
/*
* UPDATE case: return the updated old stmt.
*/
- result_stmt = vy_stmt_new_replace(format, result_mp,
- result_mp_end);
+ result_stmt =
+ vy_stmt_new_replace(format, result_mp, result_mp_end);
region_truncate(region, region_svp);
if (result_stmt == NULL)
return NULL; /* OOM */
@@ -154,8 +151,8 @@ vy_apply_upsert(struct tuple *new_stmt, struct tuple *old_stmt,
* UPSERT + UPSERT case: combine operations
*/
assert(old_ops_end - old_ops > 0);
- if (vy_upsert_try_to_squash(format, result_mp, result_mp_end,
- old_ops, old_ops_end, new_ops, new_ops_end,
+ if (vy_upsert_try_to_squash(format, result_mp, result_mp_end, old_ops,
+ old_ops_end, new_ops, new_ops_end,
&result_stmt) != 0) {
region_truncate(region, region_svp);
return NULL;
@@ -195,8 +192,8 @@ check_key:
* Check that key hasn't been changed after applying operations.
*/
if (!key_update_can_be_skipped(cmp_def->column_mask, column_mask) &&
- vy_stmt_compare(old_stmt, HINT_NONE, result_stmt,
- HINT_NONE, cmp_def) != 0) {
+ vy_stmt_compare(old_stmt, HINT_NONE, result_stmt, HINT_NONE,
+ cmp_def) != 0) {
/*
* Key has been changed: ignore this UPSERT and
* @retval the old stmt.
diff --git a/src/box/vy_upsert.h b/src/box/vy_upsert.h
index 9b585e0..6a0daa7 100644
--- a/src/box/vy_upsert.h
+++ b/src/box/vy_upsert.h
@@ -73,8 +73,8 @@ vy_entry_apply_upsert(struct vy_entry new_entry, struct vy_entry old_entry,
{
struct vy_entry result;
result.hint = old_entry.stmt != NULL ? old_entry.hint : new_entry.hint;
- result.stmt = vy_apply_upsert(new_entry.stmt, old_entry.stmt,
- cmp_def, suppress_error);
+ result.stmt = vy_apply_upsert(new_entry.stmt, old_entry.stmt, cmp_def,
+ suppress_error);
return result.stmt != NULL ? result : vy_entry_none();
}
diff --git a/src/box/vy_write_iterator.c b/src/box/vy_write_iterator.c
index 78a52ae..4f99960 100644
--- a/src/box/vy_write_iterator.c
+++ b/src/box/vy_write_iterator.c
@@ -109,8 +109,9 @@ vy_write_history_new(struct vy_entry entry, struct vy_write_history *next)
return NULL;
}
h->entry = entry;
- assert(next == NULL || (next->entry.stmt != NULL &&
- vy_stmt_lsn(next->entry.stmt) > vy_stmt_lsn(entry.stmt)));
+ assert(next == NULL ||
+ (next->entry.stmt != NULL &&
+ vy_stmt_lsn(next->entry.stmt) > vy_stmt_lsn(entry.stmt)));
h->next = next;
vy_stmt_ref_if_possible(entry.stmt);
return h;
@@ -237,8 +238,8 @@ heap_less(heap_t *heap, struct vy_write_src *src1, struct vy_write_src *src2)
* Virtual sources use 0 for LSN, so they are ordered
* last automatically.
*/
- int64_t lsn1 = src1->is_end_of_key ? 0 : vy_stmt_lsn(src1->entry.stmt);
- int64_t lsn2 = src2->is_end_of_key ? 0 : vy_stmt_lsn(src2->entry.stmt);
+ int64_t lsn1 = src1->is_end_of_key ? 0 : vy_stmt_lsn(src1->entry.stmt);
+ int64_t lsn2 = src2->is_end_of_key ? 0 : vy_stmt_lsn(src2->entry.stmt);
if (lsn1 != lsn2)
return lsn1 > lsn2;
@@ -251,7 +252,6 @@ heap_less(heap_t *heap, struct vy_write_src *src1, struct vy_write_src *src2)
*/
return (vy_stmt_type(src1->entry.stmt) == IPROTO_DELETE ? 1 : 0) <
(vy_stmt_type(src2->entry.stmt) == IPROTO_DELETE ? 1 : 0);
-
}
/**
@@ -262,10 +262,10 @@ heap_less(heap_t *heap, struct vy_write_src *src1, struct vy_write_src *src2)
static struct vy_write_src *
vy_write_iterator_new_src(struct vy_write_iterator *stream)
{
- struct vy_write_src *res = (struct vy_write_src *) malloc(sizeof(*res));
+ struct vy_write_src *res = (struct vy_write_src *)malloc(sizeof(*res));
if (res == NULL) {
- diag_set(OutOfMemory, sizeof(*res),
- "malloc", "vinyl write stream");
+ diag_set(OutOfMemory, sizeof(*res), "malloc",
+ "vinyl write stream");
return NULL;
}
heap_node_create(&res->heap_node);
@@ -275,7 +275,6 @@ vy_write_iterator_new_src(struct vy_write_iterator *stream)
return res;
}
-
/** Close a stream, remove it from the write iterator and delete. */
static void
vy_write_iterator_delete_src(struct vy_write_iterator *stream,
@@ -310,8 +309,8 @@ vy_write_iterator_add_src(struct vy_write_iterator *stream,
rc = vy_source_heap_insert(&stream->src_heap, src);
if (rc != 0) {
- diag_set(OutOfMemory, sizeof(void *),
- "malloc", "vinyl write stream heap");
+ diag_set(OutOfMemory, sizeof(void *), "malloc",
+ "vinyl write stream heap");
goto stop;
}
return 0;
@@ -326,7 +325,7 @@ stop:
*/
static void
vy_write_iterator_remove_src(struct vy_write_iterator *stream,
- struct vy_write_src *src)
+ struct vy_write_src *src)
{
if (heap_node_is_stray(&src->heap_node))
return; /* already removed */
@@ -362,7 +361,7 @@ vy_write_iterator_new(struct key_def *cmp_def, bool is_primary,
size_t size = sizeof(struct vy_write_iterator) +
count * sizeof(struct vy_read_view_stmt);
struct vy_write_iterator *stream =
- (struct vy_write_iterator *) calloc(1, size);
+ (struct vy_write_iterator *)calloc(1, size);
if (stream == NULL) {
diag_set(OutOfMemory, size, "malloc", "write stream");
return NULL;
@@ -409,8 +408,8 @@ vy_write_iterator_start(struct vy_stmt_stream *vstream)
if (vy_write_iterator_add_src(stream, src) != 0)
goto fail;
#ifndef NDEBUG
- struct errinj *inj =
- errinj(ERRINJ_VY_WRITE_ITERATOR_START_FAIL, ERRINJ_BOOL);
+ struct errinj *inj = errinj(ERRINJ_VY_WRITE_ITERATOR_START_FAIL,
+ ERRINJ_BOOL);
if (inj != NULL && inj->bparam) {
inj->bparam = false;
diag_set(OutOfMemory, 666, "malloc", "struct vy_stmt");
@@ -447,7 +446,7 @@ vy_write_iterator_stop(struct vy_stmt_stream *vstream)
stream->deferred_delete = vy_entry_none();
}
struct vy_deferred_delete_handler *handler =
- stream->deferred_delete_handler;
+ stream->deferred_delete_handler;
if (handler != NULL) {
handler->iface->destroy(handler);
stream->deferred_delete_handler = NULL;
@@ -556,8 +555,7 @@ vy_write_iterator_push_rv(struct vy_write_iterator *stream,
assert(current_rv_i < stream->rv_count);
struct vy_read_view_stmt *rv = &stream->read_views[current_rv_i];
assert(rv->vlsn >= vy_stmt_lsn(entry.stmt));
- struct vy_write_history *h =
- vy_write_history_new(entry, rv->history);
+ struct vy_write_history *h = vy_write_history_new(entry, rv->history);
if (h == NULL)
return -1;
rv->history = h;
@@ -627,7 +625,7 @@ vy_write_iterator_deferred_delete(struct vy_write_iterator *stream,
*/
if (stream->deferred_delete.stmt != NULL) {
struct vy_deferred_delete_handler *handler =
- stream->deferred_delete_handler;
+ stream->deferred_delete_handler;
if (handler != NULL && vy_stmt_type(stmt) != IPROTO_DELETE &&
handler->iface->process(handler, stmt,
stream->deferred_delete.stmt) != 0)
@@ -669,8 +667,8 @@ vy_write_iterator_deferred_delete(struct vy_write_iterator *stream,
* @retval -1 Memory error.
*/
static NODISCARD int
-vy_write_iterator_build_history(struct vy_write_iterator *stream,
- int *count, bool *is_first_insert)
+vy_write_iterator_build_history(struct vy_write_iterator *stream, int *count,
+ bool *is_first_insert)
{
*count = 0;
*is_first_insert = false;
@@ -695,8 +693,8 @@ vy_write_iterator_build_history(struct vy_write_iterator *stream,
end_of_key_src.entry = src->entry;
int rc = vy_source_heap_insert(&stream->src_heap, &end_of_key_src);
if (rc) {
- diag_set(OutOfMemory, sizeof(void *),
- "malloc", "vinyl write stream heap");
+ diag_set(OutOfMemory, sizeof(void *), "malloc",
+ "vinyl write stream heap");
return rc;
}
vy_stmt_ref_if_possible(src->entry.stmt);
@@ -710,7 +708,8 @@ vy_write_iterator_build_history(struct vy_write_iterator *stream,
int64_t merge_until_lsn = vy_write_iterator_get_vlsn(stream, 1);
while (true) {
- *is_first_insert = vy_stmt_type(src->entry.stmt) == IPROTO_INSERT;
+ *is_first_insert = vy_stmt_type(src->entry.stmt) ==
+ IPROTO_INSERT;
if (!stream->is_primary &&
(vy_stmt_flags(src->entry.stmt) & VY_STMT_UPDATE) != 0) {
@@ -756,9 +755,8 @@ vy_write_iterator_build_history(struct vy_write_iterator *stream,
*/
current_rv_i++;
current_rv_lsn = merge_until_lsn;
- merge_until_lsn =
- vy_write_iterator_get_vlsn(stream,
- current_rv_i + 1);
+ merge_until_lsn = vy_write_iterator_get_vlsn(
+ stream, current_rv_i + 1);
}
/*
@@ -787,11 +785,10 @@ vy_write_iterator_build_history(struct vy_write_iterator *stream,
vy_stmt_type(src->entry.stmt) == IPROTO_DELETE) {
current_rv_i++;
current_rv_lsn = merge_until_lsn;
- merge_until_lsn =
- vy_write_iterator_get_vlsn(stream,
- current_rv_i + 1);
+ merge_until_lsn = vy_write_iterator_get_vlsn(
+ stream, current_rv_i + 1);
}
-next_lsn:
+ next_lsn:
rc = vy_write_iterator_merge_step(stream);
if (rc != 0)
break;
@@ -845,8 +842,7 @@ vy_read_view_merge(struct vy_write_iterator *stream, struct vy_entry prev,
* by a read view if it is preceded by another DELETE for
* the same key.
*/
- if (prev.stmt != NULL &&
- vy_stmt_type(prev.stmt) == IPROTO_DELETE &&
+ if (prev.stmt != NULL && vy_stmt_type(prev.stmt) == IPROTO_DELETE &&
vy_stmt_type(h->entry.stmt) == IPROTO_DELETE) {
vy_write_history_destroy(h);
rv->history = NULL;
@@ -871,13 +867,13 @@ vy_read_view_merge(struct vy_write_iterator *stream, struct vy_entry prev,
* it, whether is_last_level is true or not.
*/
if (vy_stmt_type(h->entry.stmt) == IPROTO_UPSERT &&
- (stream->is_last_level || (prev.stmt != NULL &&
- vy_stmt_type(prev.stmt) != IPROTO_UPSERT))) {
+ (stream->is_last_level ||
+ (prev.stmt != NULL && vy_stmt_type(prev.stmt) != IPROTO_UPSERT))) {
assert(!stream->is_last_level || prev.stmt == NULL ||
vy_stmt_type(prev.stmt) != IPROTO_UPSERT);
struct vy_entry applied;
- applied = vy_entry_apply_upsert(h->entry, prev,
- stream->cmp_def, false);
+ applied = vy_entry_apply_upsert(h->entry, prev, stream->cmp_def,
+ false);
if (applied.stmt == NULL)
return -1;
vy_stmt_unref_if_possible(h->entry.stmt);
@@ -1034,7 +1030,8 @@ vy_write_iterator_build_read_views(struct vy_write_iterator *stream, int *count)
for (; rv >= &stream->read_views[0]; --rv) {
if (rv->history == NULL)
continue;
- if (vy_read_view_merge(stream, prev, rv, is_first_insert) != 0) {
+ if (vy_read_view_merge(stream, prev, rv, is_first_insert) !=
+ 0) {
rc = -1;
goto cleanup;
}
@@ -1123,4 +1120,3 @@ static const struct vy_stmt_stream_iface vy_slice_stream_iface = {
.stop = vy_write_iterator_stop,
.close = vy_write_iterator_close
};
-
diff --git a/src/box/vy_write_iterator.h b/src/box/vy_write_iterator.h
index e217160..41884b0 100644
--- a/src/box/vy_write_iterator.h
+++ b/src/box/vy_write_iterator.h
@@ -215,16 +215,16 @@ struct vy_slice;
*
* @sa VY_STMT_DEFERRED_DELETE.
*/
-typedef int
-(*vy_deferred_delete_process_f)(struct vy_deferred_delete_handler *handler,
- struct tuple *old_stmt, struct tuple *new_stmt);
+typedef int (*vy_deferred_delete_process_f)(
+ struct vy_deferred_delete_handler *handler, struct tuple *old_stmt,
+ struct tuple *new_stmt);
/**
* Callack invoked by the write iterator to destroy a deferred
* DELETE handler when the iteration is stopped.
*/
-typedef void
-(*vy_deferred_delete_destroy_f)(struct vy_deferred_delete_handler *handler);
+typedef void (*vy_deferred_delete_destroy_f)(
+ struct vy_deferred_delete_handler *handler);
struct vy_deferred_delete_handler_iface {
vy_deferred_delete_process_f process;
@@ -269,4 +269,3 @@ vy_write_iterator_new_slice(struct vy_stmt_stream *stream,
struct tuple_format *disk_format);
#endif /* INCLUDES_TARANTOOL_BOX_VY_WRITE_STREAM_H */
-
diff --git a/src/box/wal.c b/src/box/wal.c
index 84abaa7..7210995 100644
--- a/src/box/wal.c
+++ b/src/box/wal.c
@@ -80,8 +80,7 @@ wal_write_none(struct journal *, struct journal_entry *);
* members used mainly in tx thread go first, wal thread members
* following.
*/
-struct wal_writer
-{
+struct wal_writer {
struct journal base;
/* ----------------- tx ------------------- */
wal_on_garbage_collection_f on_garbage_collection;
@@ -214,8 +213,8 @@ static void
tx_complete_batch(struct cmsg *msg);
static struct cmsg_hop wal_request_route[] = {
- {wal_write_to_disk, &wal_writer_singleton.tx_prio_pipe},
- {tx_complete_batch, NULL},
+ { wal_write_to_disk, &wal_writer_singleton.tx_prio_pipe },
+ { tx_complete_batch, NULL },
};
static void
@@ -231,7 +230,7 @@ wal_msg_create(struct wal_msg *batch)
static struct wal_msg *
wal_msg(struct cmsg *msg)
{
- return msg->route == wal_request_route ? (struct wal_msg *) msg : NULL;
+ return msg->route == wal_request_route ? (struct wal_msg *)msg : NULL;
}
/** Write a request to a log in a single transaction. */
@@ -249,7 +248,7 @@ xlog_write_entry(struct xlog *l, struct journal_entry *entry)
if (inj != NULL && inj->iparam == (*row)->lsn) {
(*row)->lsn = inj->iparam - 1;
say_warn("injected broken lsn: %lld",
- (long long) (*row)->lsn);
+ (long long)(*row)->lsn);
}
if (xlog_write_row(l, *row) < 0) {
/*
@@ -314,7 +313,7 @@ wal_begin_rollback(void)
static void
wal_complete_rollback(struct cmsg *base)
{
- (void) base;
+ (void)base;
/* WAL-thread can try writing transactions again. */
wal_writer_singleton.is_in_rollback = false;
}
@@ -329,16 +328,14 @@ tx_complete_rollback(void)
* transactions to rollback are collected, the last entry
* will be exactly, well, the last entry.
*/
- if (stailq_last_entry(&writer->rollback, struct journal_entry,
- fifo) != writer->last_entry)
+ if (stailq_last_entry(&writer->rollback, struct journal_entry, fifo) !=
+ writer->last_entry)
return;
stailq_reverse(&writer->rollback);
tx_schedule_queue(&writer->rollback);
/* TX-thread can try sending transactions to WAL again. */
stailq_create(&writer->rollback);
- static struct cmsg_hop route[] = {
- {wal_complete_rollback, NULL}
- };
+ static struct cmsg_hop route[] = { { wal_complete_rollback, NULL } };
static struct cmsg msg;
cmsg_init(&msg, route);
cpipe_push(&writer->wal_pipe, &msg);
@@ -356,20 +353,21 @@ static void
tx_complete_batch(struct cmsg *msg)
{
struct wal_writer *writer = &wal_writer_singleton;
- struct wal_msg *batch = (struct wal_msg *) msg;
+ struct wal_msg *batch = (struct wal_msg *)msg;
/*
* Move the rollback list to the writer first, since
* wal_msg memory disappears after the first
* iteration of tx_schedule_queue loop.
*/
- if (! stailq_empty(&batch->rollback)) {
+ if (!stailq_empty(&batch->rollback)) {
stailq_concat(&writer->rollback, &batch->rollback);
tx_complete_rollback();
}
/* Update the tx vclock to the latest written by wal. */
vclock_copy(&replicaset.vclock, &batch->vclock);
tx_schedule_queue(&batch->commit);
- mempool_free(&writer->msg_pool, container_of(msg, struct wal_msg, base));
+ mempool_free(&writer->msg_pool,
+ container_of(msg, struct wal_msg, base));
}
/**
@@ -417,10 +415,9 @@ wal_writer_create(struct wal_writer *writer, enum wal_mode wal_mode,
writer->wal_max_size = wal_max_size;
journal_create(&writer->base,
- wal_mode == WAL_NONE ?
- wal_write_none_async : wal_write_async,
- wal_mode == WAL_NONE ?
- wal_write_none : wal_write);
+ wal_mode == WAL_NONE ? wal_write_none_async :
+ wal_write_async,
+ wal_mode == WAL_NONE ? wal_write_none : wal_write);
struct xlog_opts opts = xlog_opts_default;
opts.sync_is_async = true;
@@ -463,8 +460,8 @@ wal_open_f(struct cbus_call_msg *msg)
{
(void)msg;
struct wal_writer *writer = &wal_writer_singleton;
- const char *path = xdir_format_filename(&writer->wal_dir,
- vclock_sum(&writer->vclock), NONE);
+ const char *path = xdir_format_filename(
+ &writer->wal_dir, vclock_sum(&writer->vclock), NONE);
assert(!xlog_is_open(&writer->current_wal));
return xlog_open(&writer->current_wal, path, &writer->wal_dir.opts);
}
@@ -475,8 +472,8 @@ wal_open_f(struct cbus_call_msg *msg)
static int
wal_open(struct wal_writer *writer)
{
- const char *path = xdir_format_filename(&writer->wal_dir,
- vclock_sum(&writer->vclock), NONE);
+ const char *path = xdir_format_filename(
+ &writer->wal_dir, vclock_sum(&writer->vclock), NONE);
if (access(path, F_OK) != 0) {
if (errno == ENOENT) {
/* No WAL, nothing to do. */
@@ -528,8 +525,8 @@ wal_open(struct wal_writer *writer)
}
int
-wal_init(enum wal_mode wal_mode, const char *wal_dirname,
- int64_t wal_max_size, const struct tt_uuid *instance_uuid,
+wal_init(enum wal_mode wal_mode, const char *wal_dirname, int64_t wal_max_size,
+ const struct tt_uuid *instance_uuid,
wal_on_garbage_collection_f on_garbage_collection,
wal_on_checkpoint_threshold_f on_checkpoint_threshold)
{
@@ -590,14 +587,14 @@ wal_free(void)
}
struct wal_vclock_msg {
- struct cbus_call_msg base;
- struct vclock vclock;
+ struct cbus_call_msg base;
+ struct vclock vclock;
};
static int
wal_sync_f(struct cbus_call_msg *data)
{
- struct wal_vclock_msg *msg = (struct wal_vclock_msg *) data;
+ struct wal_vclock_msg *msg = (struct wal_vclock_msg *)data;
struct wal_writer *writer = &wal_writer_singleton;
if (writer->is_in_rollback) {
/* We're rolling back a failed write. */
@@ -629,8 +626,8 @@ wal_sync(struct vclock *vclock)
}
bool cancellable = fiber_set_cancellable(false);
struct wal_vclock_msg msg;
- int rc = cbus_call(&writer->wal_pipe, &writer->tx_prio_pipe,
- &msg.base, wal_sync_f, NULL, TIMEOUT_INFINITY);
+ int rc = cbus_call(&writer->wal_pipe, &writer->tx_prio_pipe, &msg.base,
+ wal_sync_f, NULL, TIMEOUT_INFINITY);
fiber_set_cancellable(cancellable);
if (vclock != NULL)
vclock_copy(vclock, &msg.vclock);
@@ -640,7 +637,7 @@ wal_sync(struct vclock *vclock)
static int
wal_begin_checkpoint_f(struct cbus_call_msg *data)
{
- struct wal_checkpoint *msg = (struct wal_checkpoint *) data;
+ struct wal_checkpoint *msg = (struct wal_checkpoint *)data;
struct wal_writer *writer = &wal_writer_singleton;
if (writer->is_in_rollback) {
/*
@@ -656,8 +653,7 @@ wal_begin_checkpoint_f(struct cbus_call_msg *data)
*/
if (xlog_is_open(&writer->current_wal) &&
vclock_sum(&writer->current_wal.meta.vclock) !=
- vclock_sum(&writer->vclock)) {
-
+ vclock_sum(&writer->vclock)) {
xlog_close(&writer->current_wal, false);
/*
* The next WAL will be created on the first write.
@@ -702,7 +698,7 @@ wal_begin_checkpoint(struct wal_checkpoint *checkpoint)
static int
wal_commit_checkpoint_f(struct cbus_call_msg *data)
{
- struct wal_checkpoint *msg = (struct wal_checkpoint *) data;
+ struct wal_checkpoint *msg = (struct wal_checkpoint *)data;
struct wal_writer *writer = &wal_writer_singleton;
/*
* Now, once checkpoint has been created, we can update
@@ -730,9 +726,8 @@ wal_commit_checkpoint(struct wal_checkpoint *checkpoint)
return;
}
bool cancellable = fiber_set_cancellable(false);
- cbus_call(&writer->wal_pipe, &writer->tx_prio_pipe,
- &checkpoint->base, wal_commit_checkpoint_f, NULL,
- TIMEOUT_INFINITY);
+ cbus_call(&writer->wal_pipe, &writer->tx_prio_pipe, &checkpoint->base,
+ wal_commit_checkpoint_f, NULL, TIMEOUT_INFINITY);
fiber_set_cancellable(cancellable);
}
@@ -760,14 +755,12 @@ wal_set_checkpoint_threshold(int64_t threshold)
struct wal_set_checkpoint_threshold_msg msg;
msg.checkpoint_threshold = threshold;
bool cancellable = fiber_set_cancellable(false);
- cbus_call(&writer->wal_pipe, &writer->tx_prio_pipe,
- &msg.base, wal_set_checkpoint_threshold_f, NULL,
- TIMEOUT_INFINITY);
+ cbus_call(&writer->wal_pipe, &writer->tx_prio_pipe, &msg.base,
+ wal_set_checkpoint_threshold_f, NULL, TIMEOUT_INFINITY);
fiber_set_cancellable(cancellable);
}
-struct wal_gc_msg
-{
+struct wal_gc_msg {
struct cbus_call_msg base;
const struct vclock *vclock;
};
@@ -938,8 +931,8 @@ out:
};
struct tx_notify_gc_msg *msg = malloc(sizeof(*msg));
if (msg != NULL) {
- if (xdir_first_vclock(&writer->wal_dir,
- &msg->vclock) < 0)
+ if (xdir_first_vclock(&writer->wal_dir, &msg->vclock) <
+ 0)
vclock_copy(&msg->vclock, &writer->vclock);
cmsg_init(&msg->base, route);
cpipe_push(&writer->tx_prio_pipe, &msg->base);
@@ -955,14 +948,13 @@ out:
*/
static void
wal_assign_lsn(struct vclock *vclock_diff, struct vclock *base,
- struct xrow_header **row,
- struct xrow_header **end)
+ struct xrow_header **row, struct xrow_header **end)
{
int64_t tsn = 0;
struct xrow_header **start = row;
struct xrow_header **first_glob_row = row;
/** Assign LSN to all local rows. */
- for ( ; row < end; row++) {
+ for (; row < end; row++) {
if ((*row)->replica_id == 0) {
/*
* All rows representing local space data
@@ -975,8 +967,9 @@ wal_assign_lsn(struct vclock *vclock_diff, struct vclock *base,
if ((*row)->group_id != GROUP_LOCAL)
(*row)->replica_id = instance_id;
- (*row)->lsn = vclock_inc(vclock_diff, (*row)->replica_id) +
- vclock_get(base, (*row)->replica_id);
+ (*row)->lsn =
+ vclock_inc(vclock_diff, (*row)->replica_id) +
+ vclock_get(base, (*row)->replica_id);
/*
* Use lsn of the first global row as
* transaction id.
@@ -991,19 +984,22 @@ wal_assign_lsn(struct vclock *vclock_diff, struct vclock *base,
(*row)->tsn = tsn == 0 ? (*start)->lsn : tsn;
(*row)->is_commit = row == end - 1;
} else {
- int64_t diff = (*row)->lsn - vclock_get(base, (*row)->replica_id);
- if (diff <= vclock_get(vclock_diff,
- (*row)->replica_id)) {
+ int64_t diff = (*row)->lsn -
+ vclock_get(base, (*row)->replica_id);
+ if (diff <=
+ vclock_get(vclock_diff, (*row)->replica_id)) {
say_crit("Attempt to write a broken LSN to WAL:"
" replica id: %d, confirmed lsn: %d,"
- " new lsn %d", (*row)->replica_id,
+ " new lsn %d",
+ (*row)->replica_id,
vclock_get(base, (*row)->replica_id) +
- vclock_get(vclock_diff,
- (*row)->replica_id),
- (*row)->lsn);
+ vclock_get(vclock_diff,
+ (*row)->replica_id),
+ (*row)->lsn);
assert(false);
} else {
- vclock_follow(vclock_diff, (*row)->replica_id, diff);
+ vclock_follow(vclock_diff, (*row)->replica_id,
+ diff);
}
}
}
@@ -1021,7 +1017,7 @@ static void
wal_write_to_disk(struct cmsg *msg)
{
struct wal_writer *writer = &wal_writer_singleton;
- struct wal_msg *wal_msg = (struct wal_msg *) msg;
+ struct wal_msg *wal_msg = (struct wal_msg *)msg;
struct error *error;
/*
@@ -1090,11 +1086,12 @@ wal_write_to_disk(struct cmsg *msg)
int rc;
struct journal_entry *entry;
struct stailq_entry *last_committed = NULL;
- stailq_foreach_entry(entry, &wal_msg->commit, fifo) {
- wal_assign_lsn(&vclock_diff, &writer->vclock,
- entry->rows, entry->rows + entry->n_rows);
- entry->res = vclock_sum(&vclock_diff) +
- vclock_sum(&writer->vclock);
+ stailq_foreach_entry(entry, &wal_msg->commit, fifo)
+ {
+ wal_assign_lsn(&vclock_diff, &writer->vclock, entry->rows,
+ entry->rows + entry->n_rows);
+ entry->res =
+ vclock_sum(&vclock_diff) + vclock_sum(&writer->vclock);
rc = xlog_write_entry(l, entry);
if (rc < 0)
goto done;
@@ -1160,8 +1157,7 @@ done:
if (!stailq_empty(&rollback)) {
/* Update status of the successfully committed requests. */
- stailq_foreach_entry(entry, &rollback, fifo)
- entry->res = -1;
+ stailq_foreach_entry(entry, &rollback, fifo) entry->res = -1;
/* Rollback unprocessed requests */
stailq_concat(&wal_msg->rollback, &rollback);
wal_begin_rollback();
@@ -1175,7 +1171,7 @@ done:
static int
wal_writer_f(va_list ap)
{
- (void) ap;
+ (void)ap;
struct wal_writer *writer = &wal_writer_singleton;
/** Initialize eio in this thread */
@@ -1199,11 +1195,11 @@ wal_writer_f(va_list ap)
*/
if (writer->wal_mode != WAL_NONE &&
(!xlog_is_open(&writer->current_wal) ||
- vclock_compare(&writer->vclock,
- &writer->current_wal.meta.vclock) > 0)) {
+ vclock_compare(&writer->vclock, &writer->current_wal.meta.vclock) >
+ 0)) {
struct xlog l;
- if (xdir_create_xlog(&writer->wal_dir, &l,
- &writer->vclock) == 0)
+ if (xdir_create_xlog(&writer->wal_dir, &l, &writer->vclock) ==
+ 0)
xlog_close(&l, false);
else
diag_log();
@@ -1226,13 +1222,11 @@ wal_writer_f(va_list ap)
static int
wal_write_async(struct journal *journal, struct journal_entry *entry)
{
- struct wal_writer *writer = (struct wal_writer *) journal;
+ struct wal_writer *writer = (struct wal_writer *)journal;
- ERROR_INJECT(ERRINJ_WAL_IO, {
- goto fail;
- });
+ ERROR_INJECT(ERRINJ_WAL_IO, { goto fail; });
- if (! stailq_empty(&writer->rollback)) {
+ if (!stailq_empty(&writer->rollback)) {
/*
* The writer rollback queue is not empty,
* roll back this transaction immediately.
@@ -1250,13 +1244,12 @@ wal_write_async(struct journal *journal, struct journal_entry *entry)
if (!stailq_empty(&writer->wal_pipe.input) &&
(batch = wal_msg(stailq_first_entry(&writer->wal_pipe.input,
struct cmsg, fifo)))) {
-
stailq_add_tail_entry(&batch->commit, entry, fifo);
} else {
batch = (struct wal_msg *)mempool_alloc(&writer->msg_pool);
if (batch == NULL) {
- diag_set(OutOfMemory, sizeof(struct wal_msg),
- "region", "struct wal_msg");
+ diag_set(OutOfMemory, sizeof(struct wal_msg), "region",
+ "struct wal_msg");
goto fail;
}
wal_msg_create(batch);
@@ -1305,10 +1298,9 @@ wal_write(struct journal *journal, struct journal_entry *entry)
}
static int
-wal_write_none_async(struct journal *journal,
- struct journal_entry *entry)
+wal_write_none_async(struct journal *journal, struct journal_entry *entry)
{
- struct wal_writer *writer = (struct wal_writer *) journal;
+ struct wal_writer *writer = (struct wal_writer *)journal;
struct vclock vclock_diff;
vclock_create(&vclock_diff);
@@ -1333,8 +1325,7 @@ wal_init_vy_log(void)
xlog_clear(&vy_log_writer.xlog);
}
-struct wal_write_vy_log_msg
-{
+struct wal_write_vy_log_msg {
struct cbus_call_msg base;
struct journal_entry *entry;
};
@@ -1345,7 +1336,7 @@ wal_write_vy_log_f(struct cbus_call_msg *msg)
struct journal_entry *entry =
((struct wal_write_vy_log_msg *)msg)->entry;
- if (! xlog_is_open(&vy_log_writer.xlog)) {
+ if (!xlog_is_open(&vy_log_writer.xlog)) {
if (vy_log_open(&vy_log_writer.xlog) < 0)
return -1;
}
@@ -1364,11 +1355,10 @@ wal_write_vy_log(struct journal_entry *entry)
{
struct wal_writer *writer = &wal_writer_singleton;
struct wal_write_vy_log_msg msg;
- msg.entry= entry;
+ msg.entry = entry;
bool cancellable = fiber_set_cancellable(false);
- int rc = cbus_call(&writer->wal_pipe, &writer->tx_prio_pipe,
- &msg.base, wal_write_vy_log_f, NULL,
- TIMEOUT_INFINITY);
+ int rc = cbus_call(&writer->wal_pipe, &writer->tx_prio_pipe, &msg.base,
+ wal_write_vy_log_f, NULL, TIMEOUT_INFINITY);
fiber_set_cancellable(cancellable);
return rc;
}
@@ -1376,7 +1366,7 @@ wal_write_vy_log(struct journal_entry *entry)
static int
wal_rotate_vy_log_f(struct cbus_call_msg *msg)
{
- (void) msg;
+ (void)msg;
if (xlog_is_open(&vy_log_writer.xlog))
xlog_close(&vy_log_writer.xlog, false);
return 0;
@@ -1419,7 +1409,7 @@ wal_watcher_notify(struct wal_watcher *watcher, unsigned events)
static void
wal_watcher_notify_perform(struct cmsg *cmsg)
{
- struct wal_watcher_msg *msg = (struct wal_watcher_msg *) cmsg;
+ struct wal_watcher_msg *msg = (struct wal_watcher_msg *)cmsg;
struct wal_watcher *watcher = msg->watcher;
unsigned events = msg->events;
@@ -1429,7 +1419,7 @@ wal_watcher_notify_perform(struct cmsg *cmsg)
static void
wal_watcher_notify_complete(struct cmsg *cmsg)
{
- struct wal_watcher_msg *msg = (struct wal_watcher_msg *) cmsg;
+ struct wal_watcher_msg *msg = (struct wal_watcher_msg *)cmsg;
struct wal_watcher *watcher = msg->watcher;
cmsg->route = NULL;
@@ -1452,7 +1442,7 @@ wal_watcher_notify_complete(struct cmsg *cmsg)
static void
wal_watcher_attach(void *arg)
{
- struct wal_watcher *watcher = (struct wal_watcher *) arg;
+ struct wal_watcher *watcher = (struct wal_watcher *)arg;
struct wal_writer *writer = &wal_writer_singleton;
assert(rlist_empty(&watcher->next));
@@ -1468,7 +1458,7 @@ wal_watcher_attach(void *arg)
static void
wal_watcher_detach(void *arg)
{
- struct wal_watcher *watcher = (struct wal_watcher *) arg;
+ struct wal_watcher *watcher = (struct wal_watcher *)arg;
assert(!rlist_empty(&watcher->next));
rlist_del_entry(watcher, next);
@@ -1489,10 +1479,10 @@ wal_set_watcher(struct wal_watcher *watcher, const char *name,
watcher->pending_events = 0;
assert(lengthof(watcher->route) == 2);
- watcher->route[0] = (struct cmsg_hop)
- { wal_watcher_notify_perform, &watcher->wal_pipe };
- watcher->route[1] = (struct cmsg_hop)
- { wal_watcher_notify_complete, NULL };
+ watcher->route[0] = (struct cmsg_hop){ wal_watcher_notify_perform,
+ &watcher->wal_pipe };
+ watcher->route[1] =
+ (struct cmsg_hop){ wal_watcher_notify_complete, NULL };
cbus_pair("wal", name, &watcher->wal_pipe, &watcher->watcher_pipe,
wal_watcher_attach, watcher, process_cb);
}
@@ -1515,7 +1505,6 @@ wal_notify_watchers(struct wal_writer *writer, unsigned events)
wal_watcher_notify(watcher, events);
}
-
/**
* After fork, the WAL writer thread disappears.
* Make sure that atexit() handlers in the child do
diff --git a/src/box/wal.h b/src/box/wal.h
index 581306f..70db02f 100644
--- a/src/box/wal.h
+++ b/src/box/wal.h
@@ -81,8 +81,8 @@ typedef void (*wal_on_checkpoint_threshold_f)(void);
* Start WAL thread and initialize WAL writer.
*/
int
-wal_init(enum wal_mode wal_mode, const char *wal_dirname,
- int64_t wal_max_size, const struct tt_uuid *instance_uuid,
+wal_init(enum wal_mode wal_mode, const char *wal_dirname, int64_t wal_max_size,
+ const struct tt_uuid *instance_uuid,
wal_on_garbage_collection_f on_garbage_collection,
wal_on_checkpoint_threshold_f on_checkpoint_threshold);
@@ -113,9 +113,9 @@ struct wal_watcher_msg {
enum wal_event {
/** A row is written to the current WAL. */
- WAL_EVENT_WRITE = (1 << 0),
+ WAL_EVENT_WRITE = (1 << 0),
/** A new WAL is created. */
- WAL_EVENT_ROTATE = (1 << 1),
+ WAL_EVENT_ROTATE = (1 << 1),
};
struct wal_watcher {
diff --git a/src/box/xlog.c b/src/box/xlog.c
index 974f460..4c60382 100644
--- a/src/box/xlog.c
+++ b/src/box/xlog.c
@@ -54,9 +54,9 @@
* for a while. Define it manually if necessary.
*/
#ifdef HAVE_FALLOCATE
-# ifndef FALLOC_FL_KEEP_SIZE
-# define FALLOC_FL_KEEP_SIZE 0x01
-# endif
+#ifndef FALLOC_FL_KEEP_SIZE
+#define FALLOC_FL_KEEP_SIZE 0x01
+#endif
#endif /* HAVE_FALLOCATE */
/*
@@ -67,9 +67,12 @@
*/
typedef uint32_t log_magic_t;
-static const log_magic_t row_marker = mp_bswap_u32(0xd5ba0bab); /* host byte order */
-static const log_magic_t zrow_marker = mp_bswap_u32(0xd5ba0bba); /* host byte order */
-static const log_magic_t eof_marker = mp_bswap_u32(0xd510aded); /* host byte order */
+static const log_magic_t row_marker =
+ mp_bswap_u32(0xd5ba0bab); /* host byte order */
+static const log_magic_t zrow_marker =
+ mp_bswap_u32(0xd5ba0bba); /* host byte order */
+static const log_magic_t eof_marker =
+ mp_bswap_u32(0xd510aded); /* host byte order */
enum {
/**
@@ -121,8 +124,7 @@ static const char v12[] = "0.12";
void
xlog_meta_create(struct xlog_meta *meta, const char *filetype,
const struct tt_uuid *instance_uuid,
- const struct vclock *vclock,
- const struct vclock *prev_vclock)
+ const struct vclock *vclock, const struct vclock *prev_vclock)
{
snprintf(meta->filetype, sizeof(meta->filetype), "%s", filetype);
meta->instance_uuid = *instance_uuid;
@@ -154,9 +156,7 @@ xlog_meta_format(const struct xlog_meta *meta, char *buf, int size)
int total = 0;
SNPRINT(total, snprintf, buf, size,
"%s\n"
- "%s\n"
- VERSION_KEY ": %s\n"
- INSTANCE_UUID_KEY ": %s\n",
+ "%s\n" VERSION_KEY ": %s\n" INSTANCE_UUID_KEY ": %s\n",
meta->filetype, v13, PACKAGE_VERSION,
tt_uuid_str(&meta->instance_uuid));
if (vclock_is_set(&meta->vclock)) {
@@ -188,8 +188,10 @@ parse_vclock(const char *val, const char *val_end, struct vclock *vclock)
size_t off = vclock_from_string(vclock, str);
ERROR_INJECT(ERRINJ_XLOG_META, { off = 1; });
if (off != 0) {
- diag_set(XlogError, "invalid vclock at "
- "offset %zd", off);
+ diag_set(XlogError,
+ "invalid vclock at "
+ "offset %zd",
+ off);
return -1;
}
return 0;
@@ -211,12 +213,11 @@ xlog_meta_key_equal(const char *key, const char *key_end, const char *str)
* @retval 1 if buffer hasn't enough data
*/
static ssize_t
-xlog_meta_parse(struct xlog_meta *meta, const char **data,
- const char *data_end)
+xlog_meta_parse(struct xlog_meta *meta, const char **data, const char *data_end)
{
memset(meta, 0, sizeof(*meta));
- const char *end = (const char *)memmem(*data, data_end - *data,
- "\n\n", 2);
+ const char *end =
+ (const char *)memmem(*data, data_end - *data, "\n\n", 2);
if (end == NULL)
return 1;
++end; /* include the trailing \n to simplify the checks */
@@ -226,7 +227,7 @@ xlog_meta_parse(struct xlog_meta *meta, const char **data,
* Parse filetype, i.e "SNAP" or "XLOG"
*/
const char *eol = (const char *)memchr(pos, '\n', end - pos);
- if (eol == end || (eol - pos) >= (ptrdiff_t) sizeof(meta->filetype)) {
+ if (eol == end || (eol - pos) >= (ptrdiff_t)sizeof(meta->filetype)) {
diag_set(XlogError, "failed to parse xlog type string");
return -1;
}
@@ -240,7 +241,7 @@ xlog_meta_parse(struct xlog_meta *meta, const char **data,
*/
char version[10];
eol = (const char *)memchr(pos, '\n', end - pos);
- if (eol == end || (eol - pos) >= (ptrdiff_t) sizeof(version)) {
+ if (eol == end || (eol - pos) >= (ptrdiff_t)sizeof(version)) {
diag_set(XlogError, "failed to parse xlog version string");
return -1;
}
@@ -250,9 +251,8 @@ xlog_meta_parse(struct xlog_meta *meta, const char **data,
assert(pos <= end);
if (strncmp(version, v12, sizeof(v12)) != 0 &&
strncmp(version, v13, sizeof(v13)) != 0) {
- diag_set(XlogError,
- "unsupported file format version %s",
- version);
+ diag_set(XlogError, "unsupported file format version %s",
+ version);
return -1;
}
@@ -266,8 +266,7 @@ xlog_meta_parse(struct xlog_meta *meta, const char **data,
eol = (const char *)memchr(pos, '\n', end - pos);
assert(eol <= end);
const char *key = pos;
- const char *key_end = (const char *)
- memchr(key, ':', eol - key);
+ const char *key_end = (const char *)memchr(key, ':', eol - key);
if (key_end == NULL) {
diag_set(XlogError, "can't extract meta value");
return -1;
@@ -286,14 +285,17 @@ xlog_meta_parse(struct xlog_meta *meta, const char **data,
* Instance: <uuid>
*/
if (val_end - val != UUID_STR_LEN) {
- diag_set(XlogError, "can't parse instance UUID");
+ diag_set(XlogError,
+ "can't parse instance UUID");
return -1;
}
char uuid[UUID_STR_LEN + 1];
memcpy(uuid, val, UUID_STR_LEN);
uuid[UUID_STR_LEN] = '\0';
- if (tt_uuid_from_string(uuid, &meta->instance_uuid) != 0) {
- diag_set(XlogError, "can't parse instance UUID");
+ if (tt_uuid_from_string(uuid, &meta->instance_uuid) !=
+ 0) {
+ diag_set(XlogError,
+ "can't parse instance UUID");
return -1;
}
} else if (xlog_meta_key_equal(key, key_end, VCLOCK_KEY)) {
@@ -422,7 +424,7 @@ xdir_index_file(struct xdir *dir, int64_t signature)
* Append the clock describing the file to the
* directory index.
*/
- struct vclock *vclock = (struct vclock *) malloc(sizeof(*vclock));
+ struct vclock *vclock = (struct vclock *)malloc(sizeof(*vclock));
if (vclock == NULL) {
diag_set(OutOfMemory, sizeof(*vclock), "malloc", "vclock");
xlog_cursor_close(&cursor, false);
@@ -452,8 +454,8 @@ xdir_open_cursor(struct xdir *dir, int64_t signature,
struct xlog_meta *meta = &cursor->meta;
if (strcmp(meta->filetype, dir->filetype) != 0) {
xlog_cursor_close(cursor, false);
- diag_set(ClientError, ER_INVALID_XLOG_TYPE,
- dir->filetype, meta->filetype);
+ diag_set(ClientError, ER_INVALID_XLOG_TYPE, dir->filetype,
+ meta->filetype);
return -1;
}
if (!tt_uuid_is_nil(dir->instance_uuid) &&
@@ -521,15 +523,15 @@ xdir_open_cursor(struct xdir *dir, int64_t signature,
int
xdir_scan(struct xdir *dir, bool is_dir_required)
{
- DIR *dh = opendir(dir->dirname); /* log dir */
- int64_t *signatures = NULL; /* log file names */
+ DIR *dh = opendir(dir->dirname); /* log dir */
+ int64_t *signatures = NULL; /* log file names */
size_t s_count = 0, s_capacity = 0;
if (dh == NULL) {
if (!is_dir_required && errno == ENOENT)
return 0;
diag_set(SystemError, "error reading directory '%s'",
- dir->dirname);
+ dir->dirname);
return -1;
}
@@ -569,8 +571,8 @@ xdir_scan(struct xdir *dir, bool is_dir_required)
char *dot;
long long signature = strtoll(dent->d_name, &dot, 10);
- if (ext != dot ||
- signature == LLONG_MAX || signature == LLONG_MIN) {
+ if (ext != dot || signature == LLONG_MAX ||
+ signature == LLONG_MIN) {
say_warn("can't parse `%s', skipping", dent->d_name);
continue;
}
@@ -578,10 +580,10 @@ xdir_scan(struct xdir *dir, bool is_dir_required)
if (s_count == s_capacity) {
s_capacity = s_capacity > 0 ? 2 * s_capacity : 16;
size_t size = sizeof(*signatures) * s_capacity;
- signatures = (int64_t *) realloc(signatures, size);
+ signatures = (int64_t *)realloc(signatures, size);
if (signatures == NULL) {
- diag_set(OutOfMemory,
- size, "realloc", "signatures array");
+ diag_set(OutOfMemory, size, "realloc",
+ "signatures array");
goto exit;
}
}
@@ -612,7 +614,8 @@ xdir_scan(struct xdir *dir, bool is_dir_required)
/*
* force_recovery must not affect OOM
*/
- struct error *e = diag_last_error(&fiber()->diag);
+ struct error *e =
+ diag_last_error(&fiber()->diag);
if (!dir->force_recovery ||
type_assignable(&type_OutOfMemory, e->type))
goto exit;
@@ -621,8 +624,7 @@ xdir_scan(struct xdir *dir, bool is_dir_required)
}
i++;
} else {
- assert(s_old == s_new && i < s_count &&
- vclock != NULL);
+ assert(s_old == s_new && i < s_count && vclock != NULL);
vclock = vclockset_next(&dir->index, vclock);
i++;
}
@@ -638,10 +640,10 @@ exit:
int
xdir_check(struct xdir *dir)
{
- DIR *dh = opendir(dir->dirname); /* log dir */
+ DIR *dh = opendir(dir->dirname); /* log dir */
if (dh == NULL) {
diag_set(SystemError, "error reading directory '%s'",
- dir->dirname);
+ dir->dirname);
return -1;
}
closedir(dh);
@@ -650,12 +652,11 @@ xdir_check(struct xdir *dir)
const char *
xdir_format_filename(struct xdir *dir, int64_t signature,
- enum log_suffix suffix)
+ enum log_suffix suffix)
{
- return tt_snprintf(PATH_MAX, "%s/%020lld%s%s",
- dir->dirname, (long long) signature,
- dir->filename_ext, suffix == INPROGRESS ?
- inprogress_suffix : "");
+ return tt_snprintf(PATH_MAX, "%s/%020lld%s%s", dir->dirname,
+ (long long)signature, dir->filename_ext,
+ suffix == INPROGRESS ? inprogress_suffix : "");
}
static void
@@ -734,7 +735,6 @@ xdir_add_vclock(struct xdir *xdir, const struct vclock *vclock)
/* }}} */
-
/* {{{ struct xlog */
int
@@ -754,8 +754,7 @@ xlog_rename(struct xlog *l)
if (rename(filename, new_filename) != 0) {
say_syserror("can't rename %s to %s", filename, new_filename);
- diag_set(SystemError, "failed to rename '%s' file",
- filename);
+ diag_set(SystemError, "failed to rename '%s' file", filename);
return -1;
}
l->is_inprogress = false;
@@ -824,7 +823,8 @@ xlog_create(struct xlog *xlog, const char *name, int flags,
xlog->meta = *meta;
xlog->is_inprogress = true;
- snprintf(xlog->filename, sizeof(xlog->filename), "%s%s", name, inprogress_suffix);
+ snprintf(xlog->filename, sizeof(xlog->filename), "%s%s", name,
+ inprogress_suffix);
/* Make directory if needed (gh-5090). */
if (mkdirpath(xlog->filename) != 0) {
@@ -927,7 +927,7 @@ xlog_open(struct xlog *xlog, const char *name, const struct xlog_opts *opts)
goto err_read;
}
if (rc != sizeof(magic) || load_u32(magic) != eof_marker) {
-no_eof:
+ no_eof:
xlog->offset = fio_lseek(xlog->fd, 0, SEEK_END);
if (xlog->offset < 0) {
diag_set(SystemError, "failed to seek file '%s'",
@@ -991,12 +991,12 @@ xdir_create_xlog(struct xdir *dir, struct xlog *xlog,
prev_vclock = vclockset_last(&dir->index);
struct xlog_meta meta;
- xlog_meta_create(&meta, dir->filetype, dir->instance_uuid,
- vclock, prev_vclock);
+ xlog_meta_create(&meta, dir->filetype, dir->instance_uuid, vclock,
+ prev_vclock);
const char *filename = xdir_format_filename(dir, signature, NONE);
- if (xlog_create(xlog, filename, dir->open_wflags, &meta,
- &dir->opts) != 0)
+ if (xlog_create(xlog, filename, dir->open_wflags, &meta, &dir->opts) !=
+ 0)
return -1;
/* Rename xlog file */
@@ -1071,8 +1071,7 @@ xlog_tx_write_plain(struct xlog *log)
struct iovec *iov;
size_t offset = XLOG_FIXHEADER_SIZE;
for (iov = log->obuf.iov; iov->iov_len; ++iov) {
- crc32c = crc32_calc(crc32c,
- (char *)iov->iov_base + offset,
+ crc32c = crc32_calc(crc32c, (char *)iov->iov_base + offset,
iov->iov_len - offset);
offset = 0;
}
@@ -1095,7 +1094,8 @@ xlog_tx_write_plain(struct xlog *log)
return -1;
});
- ssize_t written = fio_writevn(log->fd, log->obuf.iov, log->obuf.pos + 1);
+ ssize_t written =
+ fio_writevn(log->fd, log->obuf.iov, log->obuf.pos + 1);
if (written < 0) {
diag_set(SystemError, "failed to write to '%s' file",
log->filename);
@@ -1112,8 +1112,7 @@ xlog_tx_write_plain(struct xlog *log)
static off_t
xlog_tx_write_zstd(struct xlog *log)
{
- char *fixheader = (char *)obuf_alloc(&log->zbuf,
- XLOG_FIXHEADER_SIZE);
+ char *fixheader = (char *)obuf_alloc(&log->zbuf, XLOG_FIXHEADER_SIZE);
uint32_t crc32c = 0;
struct iovec *iov;
@@ -1127,11 +1126,11 @@ xlog_tx_write_zstd(struct xlog *log)
void *zdst = obuf_reserve(&log->zbuf, zmax_size);
if (!zdst) {
diag_set(OutOfMemory, zmax_size, "runtime arena",
- "compression buffer");
+ "compression buffer");
goto error;
}
- size_t (*fcompress)(ZSTD_CCtx *, void *, size_t,
- const void *, size_t);
+ size_t (*fcompress)(ZSTD_CCtx *, void *, size_t, const void *,
+ size_t);
/*
* If it's the last iov or the last
* log has 0 bytes, end the stream.
@@ -1185,8 +1184,7 @@ xlog_tx_write_zstd(struct xlog *log)
});
ssize_t written;
- written = fio_writevn(log->fd, log->zbuf.iov,
- log->zbuf.pos + 1);
+ written = fio_writevn(log->fd, log->zbuf.iov, log->zbuf.pos + 1);
if (written < 0) {
diag_set(SystemError, "failed to write to '%s' file",
log->filename);
@@ -1200,9 +1198,9 @@ error:
}
/* file syncing and posix_fadvise() should be rounded by a page boundary */
-#define SYNC_MASK (4096 - 1)
-#define SYNC_ROUND_DOWN(size) ((size) & ~(4096 - 1))
-#define SYNC_ROUND_UP(size) (SYNC_ROUND_DOWN(size + SYNC_MASK))
+#define SYNC_MASK (4096 - 1)
+#define SYNC_ROUND_DOWN(size) ((size) & ~(4096 - 1))
+#define SYNC_ROUND_UP(size) (SYNC_ROUND_DOWN(size + SYNC_MASK))
/**
* Writes xlog batch to file
@@ -1234,7 +1232,8 @@ xlog_tx_write(struct xlog *log)
if (written < 0) {
if (lseek(log->fd, log->offset, SEEK_SET) < 0 ||
ftruncate(log->fd, log->offset) != 0)
- panic_syserror("failed to truncate xlog after write error");
+ panic_syserror(
+ "failed to truncate xlog after write error");
log->allocated = 0;
return -1;
}
@@ -1245,17 +1244,18 @@ xlog_tx_write(struct xlog *log)
log->offset += written;
log->rows += log->tx_rows;
log->tx_rows = 0;
- if ((log->opts.sync_interval && log->offset >=
- (off_t)(log->synced_size + log->opts.sync_interval)) ||
- (log->opts.rate_limit && log->offset >=
- (off_t)(log->synced_size + log->opts.rate_limit))) {
+ if ((log->opts.sync_interval &&
+ log->offset >=
+ (off_t)(log->synced_size + log->opts.sync_interval)) ||
+ (log->opts.rate_limit &&
+ log->offset >= (off_t)(log->synced_size + log->opts.rate_limit))) {
off_t sync_from = SYNC_ROUND_DOWN(log->synced_size);
- size_t sync_len = SYNC_ROUND_UP(log->offset) -
- sync_from;
+ size_t sync_len = SYNC_ROUND_UP(log->offset) - sync_from;
if (log->opts.rate_limit > 0) {
double throttle_time;
- throttle_time = (double)sync_len / log->opts.rate_limit -
- (ev_monotonic_time() - log->sync_time);
+ throttle_time =
+ (double)sync_len / log->opts.rate_limit -
+ (ev_monotonic_time() - log->sync_time);
if (throttle_time > 0)
ev_sleep(throttle_time);
}
@@ -1263,8 +1263,8 @@ xlog_tx_write(struct xlog *log)
#ifdef HAVE_SYNC_FILE_RANGE
sync_file_range(log->fd, sync_from, sync_len,
SYNC_FILE_RANGE_WAIT_BEFORE |
- SYNC_FILE_RANGE_WRITE |
- SYNC_FILE_RANGE_WAIT_AFTER);
+ SYNC_FILE_RANGE_WRITE |
+ SYNC_FILE_RANGE_WAIT_AFTER);
#else
fdatasync(log->fd);
#endif /* HAVE_SYNC_FILE_RANGE */
@@ -1277,8 +1277,8 @@ xlog_tx_write(struct xlog *log)
say_syserror("posix_fadvise, fd=%i", log->fd);
}
#else
- (void) sync_from;
- (void) sync_len;
+ (void)sync_from;
+ (void)sync_len;
#endif /* HAVE_POSIX_FADVISE */
}
log->synced_size = log->offset;
@@ -1303,7 +1303,7 @@ xlog_write_row(struct xlog *log, const struct xrow_header *packet)
if (obuf_size(&log->obuf) == 0) {
if (!obuf_alloc(&log->obuf, XLOG_FIXHEADER_SIZE)) {
diag_set(OutOfMemory, XLOG_FIXHEADER_SIZE,
- "runtime arena", "xlog tx output buffer");
+ "runtime arena", "xlog tx output buffer");
return -1;
}
}
@@ -1319,8 +1319,8 @@ xlog_write_row(struct xlog *log, const struct xrow_header *packet)
return -1;
}
for (int i = 0; i < iovcnt; ++i) {
- struct errinj *inj = errinj(ERRINJ_WAL_WRITE_PARTIAL,
- ERRINJ_INT);
+ struct errinj *inj =
+ errinj(ERRINJ_WAL_WRITE_PARTIAL, ERRINJ_INT);
if (inj != NULL && inj->iparam >= 0 &&
obuf_size(&log->obuf) > (size_t)inj->iparam) {
diag_set(ClientError, ER_INJECTION,
@@ -1331,7 +1331,7 @@ xlog_write_row(struct xlog *log, const struct xrow_header *packet)
if (obuf_dup(&log->obuf, iov[i].iov_base, iov[i].iov_len) <
iov[i].iov_len) {
diag_set(OutOfMemory, XLOG_FIXHEADER_SIZE,
- "runtime arena", "xlog tx output buffer");
+ "runtime arena", "xlog tx output buffer");
obuf_rollback_to_svp(&log->obuf, &svp);
return -1;
}
@@ -1404,11 +1404,10 @@ xlog_flush(struct xlog *log)
static int
sync_cb(eio_req *req)
{
- int fd = (intptr_t) req->data;
+ int fd = (intptr_t)req->data;
if (req->result) {
errno = req->errorno;
- say_syserror("%s: fsync() failed",
- fio_filename(fd));
+ say_syserror("%s: fsync() failed", fio_filename(fd));
errno = 0;
}
close(fd);
@@ -1424,7 +1423,7 @@ xlog_sync(struct xlog *l)
say_syserror("%s: dup() failed", l->filename);
return -1;
}
- eio_fsync(fd, 0, sync_cb, (void *) (intptr_t) fd);
+ eio_fsync(fd, 0, sync_cb, (void *)(intptr_t)fd);
} else if (fsync(l->fd) < 0) {
say_syserror("%s: fsync failed", l->filename);
return -1;
@@ -1503,7 +1502,7 @@ xlog_atfork(struct xlog *xlog)
/* {{{ struct xlog_cursor */
-#define XLOG_READ_AHEAD (1 << 14)
+#define XLOG_READ_AHEAD (1 << 14)
/**
* Ensure that at least count bytes are in read buffer
@@ -1531,8 +1530,7 @@ xlog_cursor_ensure(struct xlog_cursor *cursor, size_t count)
return -1;
}
ssize_t readen;
- readen = fio_pread(cursor->fd, dst, to_load,
- cursor->read_offset);
+ readen = fio_pread(cursor->fd, dst, to_load, cursor->read_offset);
struct errinj *inj = errinj(ERRINJ_XLOG_READ, ERRINJ_INT);
if (inj != NULL && inj->iparam >= 0 &&
inj->iparam < cursor->read_offset) {
@@ -1540,15 +1538,14 @@ xlog_cursor_ensure(struct xlog_cursor *cursor, size_t count)
errno = EIO;
};
if (readen < 0) {
- diag_set(SystemError, "failed to read '%s' file",
- cursor->name);
+ diag_set(SystemError, "failed to read '%s' file", cursor->name);
return -1;
}
/* ibuf_reserve() has been called above, ibuf_alloc() must not fail */
assert((size_t)readen <= to_load);
ibuf_alloc(&cursor->rbuf, readen);
cursor->read_offset += readen;
- return ibuf_used(&cursor->rbuf) >= count ? 0: 1;
+ return ibuf_used(&cursor->rbuf) >= count ? 0 : 1;
}
/**
@@ -1562,8 +1559,8 @@ static int
xlog_cursor_decompress(char **rows, char *rows_end, const char **data,
const char *data_end, ZSTD_DStream *zdctx)
{
- ZSTD_inBuffer input = {*data, (size_t)(data_end - *data), 0};
- ZSTD_outBuffer output = {*rows, (size_t)(rows_end - *rows), 0};
+ ZSTD_inBuffer input = { *data, (size_t)(data_end - *data), 0 };
+ ZSTD_outBuffer output = { *rows, (size_t)(rows_end - *rows), 0 };
while (input.pos < input.size && output.pos < output.size) {
size_t rc = ZSTD_decompressStream(zdctx, &output, &input);
@@ -1576,7 +1573,7 @@ xlog_cursor_decompress(char **rows, char *rows_end, const char **data,
*rows = (char *)output.dst + output.pos;
*data = (char *)input.src + input.pos;
}
- return input.pos == input.size ? 0: 1;
+ return input.pos == input.size ? 0 : 1;
}
/**
@@ -1610,8 +1607,8 @@ struct xlog_fixheader {
* @retval count of bytes left to parse header
*/
static ssize_t
-xlog_fixheader_decode(struct xlog_fixheader *fixheader,
- const char **data, const char *data_end)
+xlog_fixheader_decode(struct xlog_fixheader *fixheader, const char **data,
+ const char *data_end)
{
if (data_end - *data < (ptrdiff_t)XLOG_FIXHEADER_SIZE)
return XLOG_FIXHEADER_SIZE - (data_end - *data);
@@ -1620,8 +1617,7 @@ xlog_fixheader_decode(struct xlog_fixheader *fixheader,
/* Decode magic */
fixheader->magic = load_u32(pos);
- if (fixheader->magic != row_marker &&
- fixheader->magic != zrow_marker) {
+ if (fixheader->magic != row_marker && fixheader->magic != zrow_marker) {
diag_set(XlogError, "invalid magic: 0x%x", fixheader->magic);
return -1;
}
@@ -1671,8 +1667,8 @@ xlog_fixheader_decode(struct xlog_fixheader *fixheader,
}
int
-xlog_tx_decode(const char *data, const char *data_end,
- char *rows, char *rows_end, ZSTD_DStream *zdctx)
+xlog_tx_decode(const char *data, const char *data_end, char *rows,
+ char *rows_end, ZSTD_DStream *zdctx)
{
/* Decode fixheader */
struct xlog_fixheader fixheader;
@@ -1681,14 +1677,16 @@ xlog_tx_decode(const char *data, const char *data_end,
/* Check that buffer has enough bytes */
if (data + fixheader.len != data_end) {
- diag_set(XlogError, "invalid compressed length: "
- "expected %zd, got %u",
- data_end - data, fixheader.len);
+ diag_set(XlogError,
+ "invalid compressed length: "
+ "expected %zd, got %u",
+ data_end - data, fixheader.len);
return -1;
}
ERROR_INJECT(ERRINJ_XLOG_GARBAGE, {
- *((char *)data + fixheader.len / 2) = ~*((char *)data + fixheader.len / 2);
+ *((char *)data + fixheader.len / 2) =
+ ~*((char *)data + fixheader.len / 2);
});
/* Validate checksum */
@@ -1700,9 +1698,10 @@ xlog_tx_decode(const char *data, const char *data_end,
/* Copy uncompressed rows */
if (fixheader.magic == row_marker) {
if (rows_end - rows != (ptrdiff_t)fixheader.len) {
- diag_set(XlogError, "invalid unpacked length: "
- "expected %zd, got %u",
- rows_end - data, fixheader.len);
+ diag_set(XlogError,
+ "invalid unpacked length: "
+ "expected %zd, got %u",
+ rows_end - data, fixheader.len);
return -1;
}
memcpy(rows, data, fixheader.len);
@@ -1712,14 +1711,16 @@ xlog_tx_decode(const char *data, const char *data_end,
/* Decompress zstd rows */
assert(fixheader.magic == zrow_marker);
ZSTD_initDStream(zdctx);
- int rc = xlog_cursor_decompress(&rows, rows_end, &data, data_end,
- zdctx);
+ int rc =
+ xlog_cursor_decompress(&rows, rows_end, &data, data_end, zdctx);
if (rc < 0) {
return -1;
} else if (rc > 0) {
- diag_set(XlogError, "invalid decompressed length: "
- "expected %zd, got %zd", rows_end - data,
- rows_end - data + XLOG_TX_AUTOCOMMIT_THRESHOLD);
+ diag_set(XlogError,
+ "invalid decompressed length: "
+ "expected %zd, got %zd",
+ rows_end - data,
+ rows_end - data + XLOG_TX_AUTOCOMMIT_THRESHOLD);
return -1;
}
@@ -1733,9 +1734,8 @@ xlog_tx_decode(const char *data, const char *data_end,
* @retval >0 how many bytes we will have for continue
*/
ssize_t
-xlog_tx_cursor_create(struct xlog_tx_cursor *tx_cursor,
- const char **data, const char *data_end,
- ZSTD_DStream *zdctx)
+xlog_tx_cursor_create(struct xlog_tx_cursor *tx_cursor, const char **data,
+ const char *data_end, ZSTD_DStream *zdctx)
{
const char *rpos = *data;
struct xlog_fixheader fixheader;
@@ -1749,7 +1749,8 @@ xlog_tx_cursor_create(struct xlog_tx_cursor *tx_cursor,
return fixheader.len - (data_end - rpos);
ERROR_INJECT(ERRINJ_XLOG_GARBAGE, {
- *((char *)rpos + fixheader.len / 2) = ~*((char *)rpos + fixheader.len / 2);
+ *((char *)rpos + fixheader.len / 2) =
+ ~*((char *)rpos + fixheader.len / 2);
});
/* Validate checksum */
@@ -1764,8 +1765,8 @@ xlog_tx_cursor_create(struct xlog_tx_cursor *tx_cursor,
if (fixheader.magic == row_marker) {
void *dst = ibuf_alloc(&tx_cursor->rows, fixheader.len);
if (dst == NULL) {
- diag_set(OutOfMemory, fixheader.len,
- "runtime", "xlog rows buffer");
+ diag_set(OutOfMemory, fixheader.len, "runtime",
+ "xlog rows buffer");
ibuf_destroy(&tx_cursor->rows);
return -1;
}
@@ -1783,7 +1784,7 @@ xlog_tx_cursor_create(struct xlog_tx_cursor *tx_cursor,
if (ibuf_reserve(&tx_cursor->rows,
XLOG_TX_AUTOCOMMIT_THRESHOLD) == NULL) {
diag_set(OutOfMemory, XLOG_TX_AUTOCOMMIT_THRESHOLD,
- "runtime", "xlog output buffer");
+ "runtime", "xlog output buffer");
ibuf_destroy(&tx_cursor->rows);
return -1;
}
@@ -1801,13 +1802,12 @@ xlog_tx_cursor_create(struct xlog_tx_cursor *tx_cursor,
int
xlog_tx_cursor_next_row(struct xlog_tx_cursor *tx_cursor,
- struct xrow_header *xrow)
+ struct xrow_header *xrow)
{
if (ibuf_used(&tx_cursor->rows) == 0)
return 1;
/* Return row from xlog tx buffer */
- int rc = xrow_header_decode(xrow,
- (const char **)&tx_cursor->rows.rpos,
+ int rc = xrow_header_decode(xrow, (const char **)&tx_cursor->rows.rpos,
(const char *)tx_cursor->rows.wpos, false);
if (rc != 0) {
diag_set(XlogError, "can't parse row");
@@ -1897,9 +1897,10 @@ eof_found:
if (rc < 0)
return -1;
if (rc == 0) {
- diag_set(XlogError, "%s: has some data after "
- "eof marker at %lld", i->name,
- xlog_cursor_pos(i));
+ diag_set(XlogError,
+ "%s: has some data after "
+ "eof marker at %lld",
+ i->name, xlog_cursor_pos(i));
return -1;
}
i->state = XLOG_CURSOR_EOF;
@@ -1921,8 +1922,8 @@ xlog_cursor_next_row(struct xlog_cursor *cursor, struct xrow_header *xrow)
}
int
-xlog_cursor_next(struct xlog_cursor *cursor,
- struct xrow_header *xrow, bool force_recovery)
+xlog_cursor_next(struct xlog_cursor *cursor, struct xrow_header *xrow,
+ bool force_recovery)
{
assert(xlog_cursor_is_open(cursor));
while (true) {
@@ -1932,15 +1933,13 @@ xlog_cursor_next(struct xlog_cursor *cursor,
break;
if (rc < 0) {
struct error *e = diag_last_error(diag_get());
- if (!force_recovery ||
- e->type != &type_XlogError)
+ if (!force_recovery || e->type != &type_XlogError)
return -1;
say_error("can't decode row: %s", e->errmsg);
}
while ((rc = xlog_cursor_next_tx(cursor)) < 0) {
struct error *e = diag_last_error(diag_get());
- if (!force_recovery ||
- e->type != &type_XlogError)
+ if (!force_recovery || e->type != &type_XlogError)
return -1;
say_error("can't open tx: %s", e->errmsg);
if ((rc = xlog_cursor_find_tx_magic(cursor)) < 0)
@@ -1970,13 +1969,14 @@ xlog_cursor_openfd(struct xlog_cursor *i, int fd, const char *name)
rc = xlog_cursor_ensure(i, XLOG_META_LEN_MAX);
if (rc == -1)
goto error;
- rc = xlog_meta_parse(&i->meta,
- (const char **)&i->rbuf.rpos,
+ rc = xlog_meta_parse(&i->meta, (const char **)&i->rbuf.rpos,
(const char *)i->rbuf.wpos);
if (rc == -1)
goto error;
if (rc > 0) {
- diag_set(XlogError, "Unexpected end of file, run with 'force_recovery = true'");
+ diag_set(
+ XlogError,
+ "Unexpected end of file, run with 'force_recovery = true'");
goto error;
}
snprintf(i->name, sizeof(i->name), "%s", name);
@@ -2021,14 +2021,13 @@ xlog_cursor_openmem(struct xlog_cursor *i, const char *data, size_t size,
void *dst = ibuf_alloc(&i->rbuf, size);
if (dst == NULL) {
diag_set(OutOfMemory, size, "runtime",
- "xlog cursor read buffer");
+ "xlog cursor read buffer");
goto error;
}
memcpy(dst, data, size);
i->read_offset = size;
int rc;
- rc = xlog_meta_parse(&i->meta,
- (const char **)&i->rbuf.rpos,
+ rc = xlog_meta_parse(&i->meta, (const char **)&i->rbuf.rpos,
(const char *)i->rbuf.wpos);
if (rc < 0)
goto error;
@@ -2061,8 +2060,8 @@ xlog_cursor_close(struct xlog_cursor *i, bool reuse_fd)
if (i->state == XLOG_CURSOR_TX)
xlog_tx_cursor_destroy(&i->tx_cursor);
ZSTD_freeDStream(i->zdctx);
- i->state = (i->state == XLOG_CURSOR_EOF ?
- XLOG_CURSOR_EOF_CLOSED : XLOG_CURSOR_CLOSED);
+ i->state = (i->state == XLOG_CURSOR_EOF ? XLOG_CURSOR_EOF_CLOSED :
+ XLOG_CURSOR_CLOSED);
/*
* Do not trash the cursor object since the caller might
* still want to access its state and/or meta information.
diff --git a/src/box/xlog.h b/src/box/xlog.h
index 5b1f42c..cb7c2d3 100644
--- a/src/box/xlog.h
+++ b/src/box/xlog.h
@@ -94,9 +94,9 @@ extern const struct xlog_opts xlog_opts_default;
* but an xlog object sees only those files which match its type.
*/
enum xdir_type {
- SNAP, /* memtx snapshot */
- XLOG, /* write ahead log */
- VYLOG, /* vinyl metadata log */
+ SNAP, /* memtx snapshot */
+ XLOG, /* write ahead log */
+ VYLOG, /* vinyl metadata log */
};
/**
@@ -323,8 +323,7 @@ struct xlog_meta {
void
xlog_meta_create(struct xlog_meta *meta, const char *filetype,
const struct tt_uuid *instance_uuid,
- const struct vclock *vclock,
- const struct vclock *prev_vclock);
+ const struct vclock *vclock, const struct vclock *prev_vclock);
/* }}} */
@@ -455,7 +454,6 @@ xlog_create(struct xlog *xlog, const char *name, int flags,
int
xlog_open(struct xlog *xlog, const char *name, const struct xlog_opts *opts);
-
/**
* Reset an xlog object without opening it.
* The object is in limbo state: it doesn't hold
@@ -465,7 +463,6 @@ xlog_open(struct xlog *xlog, const char *name, const struct xlog_opts *opts);
void
xlog_clear(struct xlog *xlog);
-
/** Returns true if the xlog file is open. */
static inline bool
xlog_is_open(struct xlog *l)
@@ -531,7 +528,6 @@ xlog_tx_rollback(struct xlog *log);
ssize_t
xlog_flush(struct xlog *log);
-
/**
* Sync a log file. The exact action is defined
* by xdir flags.
@@ -563,8 +559,7 @@ xlog_atfork(struct xlog *xlog);
/**
* xlog tx iterator
*/
-struct xlog_tx_cursor
-{
+struct xlog_tx_cursor {
/** rows buffer */
struct ibuf rows;
/** tx size */
@@ -580,9 +575,8 @@ struct xlog_tx_cursor
* @retval >0 how many additional bytes should be read to parse tx
*/
ssize_t
-xlog_tx_cursor_create(struct xlog_tx_cursor *cursor,
- const char **data, const char *data_end,
- ZSTD_DStream *zdctx);
+xlog_tx_cursor_create(struct xlog_tx_cursor *cursor, const char **data,
+ const char *data_end, ZSTD_DStream *zdctx);
/**
* Destroy xlog tx cursor and free all associated memory
@@ -598,7 +592,8 @@ xlog_tx_cursor_destroy(struct xlog_tx_cursor *tx_cursor);
* @retval -1 for error
*/
int
-xlog_tx_cursor_next_row(struct xlog_tx_cursor *tx_cursor, struct xrow_header *xrow);
+xlog_tx_cursor_next_row(struct xlog_tx_cursor *tx_cursor,
+ struct xrow_header *xrow);
/**
* Return current tx cursor position
@@ -624,9 +619,8 @@ xlog_tx_cursor_pos(struct xlog_tx_cursor *tx_cursor)
* @retval -1 error, check diag
*/
int
-xlog_tx_decode(const char *data, const char *data_end,
- char *rows, char *rows_end,
- ZSTD_DStream *zdctx);
+xlog_tx_decode(const char *data, const char *data_end, char *rows,
+ char *rows_end, ZSTD_DStream *zdctx);
/* }}} */
@@ -762,8 +756,8 @@ xlog_cursor_next_row(struct xlog_cursor *cursor, struct xrow_header *xrow);
* @retval -1 for error
*/
int
-xlog_cursor_next(struct xlog_cursor *cursor,
- struct xrow_header *xrow, bool force_recovery);
+xlog_cursor_next(struct xlog_cursor *cursor, struct xrow_header *xrow,
+ bool force_recovery);
/**
* Move to the next xlog tx
@@ -875,8 +869,8 @@ xlog_cursor_open_xc(struct xlog_cursor *cursor, const char *name)
* @copydoc xlog_cursor_next
*/
static inline int
-xlog_cursor_next_xc(struct xlog_cursor *cursor,
- struct xrow_header *xrow, bool force_recovery)
+xlog_cursor_next_xc(struct xlog_cursor *cursor, struct xrow_header *xrow,
+ bool force_recovery)
{
int rc = xlog_cursor_next(cursor, xrow, force_recovery);
if (rc == -1)
diff --git a/src/box/xrow.c b/src/box/xrow.c
index da5c6ff..4237bb5 100644
--- a/src/box/xrow.c
+++ b/src/box/xrow.c
@@ -46,15 +46,16 @@
#include "mpstream/mpstream.h"
static_assert(IPROTO_DATA < 0x7f && IPROTO_METADATA < 0x7f &&
- IPROTO_SQL_INFO < 0x7f, "encoded IPROTO_BODY keys must fit into "\
+ IPROTO_SQL_INFO < 0x7f,
+ "encoded IPROTO_BODY keys must fit into "
"one byte");
static inline uint32_t
mp_sizeof_vclock_ignore0(const struct vclock *vclock)
{
uint32_t size = vclock_size_ignore0(vclock);
- return mp_sizeof_map(size) + size * (mp_sizeof_uint(UINT32_MAX) +
- mp_sizeof_uint(UINT64_MAX));
+ return mp_sizeof_map(size) +
+ size * (mp_sizeof_uint(UINT32_MAX) + mp_sizeof_uint(UINT64_MAX));
}
static inline char *
@@ -67,7 +68,7 @@ mp_encode_vclock_ignore0(char *data, const struct vclock *vclock)
replica = vclock_iterator_next(&it);
if (replica.id == 0)
replica = vclock_iterator_next(&it);
- for ( ; replica.id < VCLOCK_MAX; replica = vclock_iterator_next(&it)) {
+ for (; replica.id < VCLOCK_MAX; replica = vclock_iterator_next(&it)) {
data = mp_encode_uint(data, replica.id);
data = mp_encode_uint(data, replica.lsn);
}
@@ -104,7 +105,9 @@ mp_decode_vclock_ignore0(const char **data, struct vclock *vclock)
*
* The format is similar to the xxd utility.
*/
-void dump_row_hex(const char *start, const char *end) {
+void
+dump_row_hex(const char *start, const char *end)
+{
if (!say_log_level_is_enabled(S_VERBOSE))
return;
@@ -116,7 +119,8 @@ void dump_row_hex(const char *start, const char *end) {
char *pos = buf;
pos += snprintf(pos, buf_end - pos, "%08lX: ", cur - start);
for (size_t i = 0; i < 16; ++i) {
- pos += snprintf(pos, buf_end - pos, "%02X ", (unsigned char)*cur++);
+ pos += snprintf(pos, buf_end - pos, "%02X ",
+ (unsigned char)*cur++);
if (cur >= end || pos == buf_end)
break;
}
@@ -125,10 +129,11 @@ void dump_row_hex(const char *start, const char *end) {
}
}
-#define xrow_on_decode_err(start, end, what, desc_str) do {\
- diag_set(ClientError, what, desc_str);\
- dump_row_hex(start, end);\
-} while (0);
+#define xrow_on_decode_err(start, end, what, desc_str) \
+ do { \
+ diag_set(ClientError, what, desc_str); \
+ dump_row_hex(start, end); \
+ } while (0);
int
xrow_header_decode(struct xrow_header *header, const char **pos,
@@ -136,10 +141,11 @@ xrow_header_decode(struct xrow_header *header, const char **pos,
{
memset(header, 0, sizeof(struct xrow_header));
const char *tmp = *pos;
- const char * const start = *pos;
+ const char *const start = *pos;
if (mp_check(&tmp, end) != 0) {
-error:
- xrow_on_decode_err(start, end, ER_INVALID_MSGPACK, "packet header");
+ error:
+ xrow_on_decode_err(start, end, ER_INVALID_MSGPACK,
+ "packet header");
return -1;
}
@@ -206,15 +212,17 @@ error:
if (*pos < end && header->type != IPROTO_NOP) {
const char *body = *pos;
if (mp_check(pos, end)) {
- xrow_on_decode_err(start, end, ER_INVALID_MSGPACK, "packet body");
+ xrow_on_decode_err(start, end, ER_INVALID_MSGPACK,
+ "packet body");
return -1;
}
header->bodycnt = 1;
- header->body[0].iov_base = (void *) body;
+ header->body[0].iov_base = (void *)body;
header->body[0].iov_len = *pos - body;
}
if (end_is_exact && *pos < end) {
- xrow_on_decode_err(start,end, ER_INVALID_MSGPACK, "packet body");
+ xrow_on_decode_err(start, end, ER_INVALID_MSGPACK,
+ "packet body");
return -1;
}
return 0;
@@ -240,14 +248,14 @@ xrow_header_encode(const struct xrow_header *header, uint64_t sync,
struct iovec *out, size_t fixheader_len)
{
/* allocate memory for sign + header */
- out->iov_base = region_alloc(&fiber()->gc, XROW_HEADER_LEN_MAX +
- fixheader_len);
+ out->iov_base =
+ region_alloc(&fiber()->gc, XROW_HEADER_LEN_MAX + fixheader_len);
if (out->iov_base == NULL) {
diag_set(OutOfMemory, XROW_HEADER_LEN_MAX + fixheader_len,
"gc arena", "xrow header encode");
return -1;
}
- char *data = (char *) out->iov_base + fixheader_len;
+ char *data = (char *)out->iov_base + fixheader_len;
/* Header */
char *d = data + 1; /* Skip 1 byte for MP_MAP */
@@ -323,7 +331,7 @@ xrow_header_encode(const struct xrow_header *header, uint64_t sync,
}
assert(d <= data + XROW_HEADER_LEN_MAX);
mp_encode_map(data, map_size);
- out->iov_len = d - (char *) out->iov_base;
+ out->iov_len = d - (char *)out->iov_base;
out++;
memcpy(out, header->body, sizeof(*out) * header->bodycnt);
@@ -339,18 +347,18 @@ xrow_encode_uuid(char *pos, const struct tt_uuid *in)
/* m_ - msgpack meta, k_ - key, v_ - value */
struct PACKED iproto_header_bin {
- uint8_t m_len; /* MP_UINT32 */
- uint32_t v_len; /* length */
- uint8_t m_header; /* MP_MAP */
- uint8_t k_code; /* IPROTO_REQUEST_TYPE */
- uint8_t m_code; /* MP_UINT32 */
- uint32_t v_code; /* response status */
- uint8_t k_sync; /* IPROTO_SYNC */
- uint8_t m_sync; /* MP_UINT64 */
- uint64_t v_sync; /* sync */
- uint8_t k_schema_version; /* IPROTO_SCHEMA_VERSION */
- uint8_t m_schema_version; /* MP_UINT32 */
- uint32_t v_schema_version; /* schema_version */
+ uint8_t m_len; /* MP_UINT32 */
+ uint32_t v_len; /* length */
+ uint8_t m_header; /* MP_MAP */
+ uint8_t k_code; /* IPROTO_REQUEST_TYPE */
+ uint8_t m_code; /* MP_UINT32 */
+ uint32_t v_code; /* response status */
+ uint8_t k_sync; /* IPROTO_SYNC */
+ uint8_t m_sync; /* MP_UINT64 */
+ uint64_t v_sync; /* sync */
+ uint8_t k_schema_version; /* IPROTO_SCHEMA_VERSION */
+ uint8_t m_schema_version; /* MP_UINT32 */
+ uint32_t v_schema_version; /* schema_version */
};
static_assert(sizeof(struct iproto_header_bin) == IPROTO_HEADER_LEN,
@@ -378,18 +386,18 @@ iproto_header_encode(char *out, uint32_t type, uint64_t sync,
}
struct PACKED iproto_body_bin {
- uint8_t m_body; /* MP_MAP */
- uint8_t k_data; /* IPROTO_DATA or errors */
- uint8_t m_data; /* MP_STR or MP_ARRAY */
- uint32_t v_data_len; /* string length of array size */
+ uint8_t m_body; /* MP_MAP */
+ uint8_t k_data; /* IPROTO_DATA or errors */
+ uint8_t m_data; /* MP_STR or MP_ARRAY */
+ uint32_t v_data_len; /* string length of array size */
};
static_assert(sizeof(struct iproto_body_bin) + IPROTO_HEADER_LEN ==
- IPROTO_SELECT_HEADER_LEN, "size of the prepared select");
+ IPROTO_SELECT_HEADER_LEN,
+ "size of the prepared select");
-static const struct iproto_body_bin iproto_body_bin = {
- 0x81, IPROTO_DATA, 0xdd, 0
-};
+static const struct iproto_body_bin iproto_body_bin = { 0x81, IPROTO_DATA, 0xdd,
+ 0 };
/** Return a 4-byte numeric error code, with status flags. */
static inline uint32_t
@@ -417,12 +425,12 @@ iproto_reply_vclock(struct obuf *out, const struct vclock *vclock,
uint64_t sync, uint32_t schema_version)
{
size_t max_size = IPROTO_HEADER_LEN + mp_sizeof_map(1) +
- mp_sizeof_uint(UINT32_MAX) + mp_sizeof_vclock_ignore0(vclock);
+ mp_sizeof_uint(UINT32_MAX) +
+ mp_sizeof_vclock_ignore0(vclock);
char *buf = obuf_reserve(out, max_size);
if (buf == NULL) {
- diag_set(OutOfMemory, max_size,
- "obuf_alloc", "buf");
+ diag_set(OutOfMemory, max_size, "obuf_alloc", "buf");
return -1;
}
@@ -437,30 +445,30 @@ iproto_reply_vclock(struct obuf *out, const struct vclock *vclock,
size - IPROTO_HEADER_LEN);
char *ptr = obuf_alloc(out, size);
- (void) ptr;
+ (void)ptr;
assert(ptr == buf);
return 0;
}
int
-iproto_reply_vote(struct obuf *out, const struct ballot *ballot,
- uint64_t sync, uint32_t schema_version)
+iproto_reply_vote(struct obuf *out, const struct ballot *ballot, uint64_t sync,
+ uint32_t schema_version)
{
- size_t max_size = IPROTO_HEADER_LEN + mp_sizeof_map(1) +
+ size_t max_size =
+ IPROTO_HEADER_LEN + mp_sizeof_map(1) +
mp_sizeof_uint(UINT32_MAX) + mp_sizeof_map(5) +
mp_sizeof_uint(UINT32_MAX) + mp_sizeof_bool(ballot->is_ro) +
- mp_sizeof_uint(UINT32_MAX) + mp_sizeof_bool(ballot->is_loading) +
- mp_sizeof_uint(IPROTO_BALLOT_IS_ANON) +
- mp_sizeof_bool(ballot->is_anon) +
mp_sizeof_uint(UINT32_MAX) +
+ mp_sizeof_bool(ballot->is_loading) +
+ mp_sizeof_uint(IPROTO_BALLOT_IS_ANON) +
+ mp_sizeof_bool(ballot->is_anon) + mp_sizeof_uint(UINT32_MAX) +
mp_sizeof_vclock_ignore0(&ballot->vclock) +
mp_sizeof_uint(UINT32_MAX) +
mp_sizeof_vclock_ignore0(&ballot->gc_vclock);
char *buf = obuf_reserve(out, max_size);
if (buf == NULL) {
- diag_set(OutOfMemory, max_size,
- "obuf_alloc", "buf");
+ diag_set(OutOfMemory, max_size, "obuf_alloc", "buf");
return -1;
}
@@ -485,7 +493,7 @@ iproto_reply_vote(struct obuf *out, const struct ballot *ballot,
size - IPROTO_HEADER_LEN);
char *ptr = obuf_alloc(out, size);
- (void) ptr;
+ (void)ptr;
assert(ptr == buf);
return 0;
}
@@ -560,7 +568,7 @@ iproto_write_error(int fd, const struct error *e, uint32_t schema_version,
ssize_t unused;
unused = write(fd, header, sizeof(header));
unused = write(fd, payload, payload_size);
- (void) unused;
+ (void)unused;
cleanup:
region_truncate(region, region_svp);
}
@@ -580,7 +588,7 @@ iproto_prepare_header(struct obuf *buf, struct obuf_svp *svp, size_t size)
}
*svp = obuf_create_svp(buf);
ptr = obuf_alloc(buf, size);
- assert(ptr != NULL);
+ assert(ptr != NULL);
return 0;
}
@@ -588,10 +596,9 @@ void
iproto_reply_select(struct obuf *buf, struct obuf_svp *svp, uint64_t sync,
uint32_t schema_version, uint32_t count)
{
- char *pos = (char *) obuf_svp_to_ptr(buf, svp);
+ char *pos = (char *)obuf_svp_to_ptr(buf, svp);
iproto_header_encode(pos, IPROTO_OK, sync, schema_version,
- obuf_size(buf) - svp->used -
- IPROTO_HEADER_LEN);
+ obuf_size(buf) - svp->used - IPROTO_HEADER_LEN);
struct iproto_body_bin body = iproto_body_bin;
body.v_data_len = mp_bswap_u32(count);
@@ -603,18 +610,19 @@ int
xrow_decode_sql(const struct xrow_header *row, struct sql_request *request)
{
if (row->bodycnt == 0) {
- diag_set(ClientError, ER_INVALID_MSGPACK, "missing request body");
+ diag_set(ClientError, ER_INVALID_MSGPACK,
+ "missing request body");
return 1;
}
assert(row->bodycnt == 1);
- const char *data = (const char *) row->body[0].iov_base;
+ const char *data = (const char *)row->body[0].iov_base;
const char *end = data + row->body[0].iov_len;
assert((end - data) > 0);
if (mp_typeof(*data) != MP_MAP || mp_check_map(data, end) > 0) {
-error:
- xrow_on_decode_err(row->body[0].iov_base, end, ER_INVALID_MSGPACK,
- "packet body");
+ error:
+ xrow_on_decode_err(row->body[0].iov_base, end,
+ ER_INVALID_MSGPACK, "packet body");
return -1;
}
@@ -626,12 +634,12 @@ error:
uint8_t key = *data;
if (key != IPROTO_SQL_BIND && key != IPROTO_SQL_TEXT &&
key != IPROTO_STMT_ID) {
- mp_check(&data, end); /* skip the key */
- mp_check(&data, end); /* skip the value */
+ mp_check(&data, end); /* skip the key */
+ mp_check(&data, end); /* skip the value */
continue;
}
- const char *value = ++data; /* skip the key */
- if (mp_check(&data, end) != 0) /* check the value */
+ const char *value = ++data; /* skip the key */
+ if (mp_check(&data, end) != 0) /* check the value */
goto error;
if (key == IPROTO_SQL_BIND)
request->bind = value;
@@ -641,17 +649,17 @@ error:
request->stmt_id = value;
}
if (request->sql_text != NULL && request->stmt_id != NULL) {
- xrow_on_decode_err(row->body[0].iov_base, end, ER_INVALID_MSGPACK,
- "SQL text and statement id are incompatible "\
+ xrow_on_decode_err(row->body[0].iov_base, end,
+ ER_INVALID_MSGPACK,
+ "SQL text and statement id are incompatible "
"options in one request: choose one");
return -1;
}
if (request->sql_text == NULL && request->stmt_id == NULL) {
- xrow_on_decode_err(row->body[0].iov_base, end,
- ER_MISSING_REQUEST_FIELD,
- tt_sprintf("%s or %s",
- iproto_key_name(IPROTO_SQL_TEXT),
- iproto_key_name(IPROTO_STMT_ID)));
+ xrow_on_decode_err(
+ row->body[0].iov_base, end, ER_MISSING_REQUEST_FIELD,
+ tt_sprintf("%s or %s", iproto_key_name(IPROTO_SQL_TEXT),
+ iproto_key_name(IPROTO_STMT_ID)));
return -1;
}
if (data != end)
@@ -663,7 +671,7 @@ void
iproto_reply_sql(struct obuf *buf, struct obuf_svp *svp, uint64_t sync,
uint32_t schema_version)
{
- char *pos = (char *) obuf_svp_to_ptr(buf, svp);
+ char *pos = (char *)obuf_svp_to_ptr(buf, svp);
iproto_header_encode(pos, IPROTO_OK, sync, schema_version,
obuf_size(buf) - svp->used - IPROTO_HEADER_LEN);
}
@@ -672,7 +680,7 @@ void
iproto_reply_chunk(struct obuf *buf, struct obuf_svp *svp, uint64_t sync,
uint32_t schema_version)
{
- char *pos = (char *) obuf_svp_to_ptr(buf, svp);
+ char *pos = (char *)obuf_svp_to_ptr(buf, svp);
iproto_header_encode(pos, IPROTO_CHUNK, sync, schema_version,
obuf_size(buf) - svp->used - IPROTO_HEADER_LEN);
struct iproto_body_bin body = iproto_body_bin;
@@ -695,20 +703,20 @@ xrow_decode_dml(struct xrow_header *row, struct request *request,
goto done;
assert(row->bodycnt == 1);
- const char *data = start = (const char *) row->body[0].iov_base;
+ const char *data = start = (const char *)row->body[0].iov_base;
end = data + row->body[0].iov_len;
assert((end - data) > 0);
if (mp_typeof(*data) != MP_MAP || mp_check_map(data, end) > 0) {
-error:
- xrow_on_decode_err(row->body[0].iov_base, end, ER_INVALID_MSGPACK,
- "packet body");
+ error:
+ xrow_on_decode_err(row->body[0].iov_base, end,
+ ER_INVALID_MSGPACK, "packet body");
return -1;
}
uint32_t size = mp_decode_map(&data);
for (uint32_t i = 0; i < size; i++) {
- if (! iproto_dml_body_has_key(data, end)) {
+ if (!iproto_dml_body_has_key(data, end)) {
if (mp_check(&data, end) != 0 ||
mp_check(&data, end) != 0)
goto error;
@@ -716,8 +724,7 @@ error:
}
uint64_t key = mp_decode_uint(&data);
const char *value = data;
- if (mp_check(&data, end) ||
- key >= IPROTO_KEY_MAX ||
+ if (mp_check(&data, end) || key >= IPROTO_KEY_MAX ||
iproto_key_type[key] != mp_typeof(*value))
goto error;
key_map &= ~iproto_key_bit(key);
@@ -761,13 +768,13 @@ error:
}
}
if (data != end) {
- xrow_on_decode_err(row->body[0].iov_base, end, ER_INVALID_MSGPACK,
- "packet end");
+ xrow_on_decode_err(row->body[0].iov_base, end,
+ ER_INVALID_MSGPACK, "packet end");
return -1;
}
done:
if (key_map) {
- enum iproto_key key = (enum iproto_key) bit_ctz_u64(key_map);
+ enum iproto_key key = (enum iproto_key)bit_ctz_u64(key_map);
xrow_on_decode_err(start, end, ER_MISSING_REQUEST_FIELD,
iproto_key_name(key));
return -1;
@@ -779,14 +786,14 @@ static int
request_snprint(char *buf, int size, const struct request *request)
{
int total = 0;
- SNPRINT(total, snprintf, buf, size, "{type: '%s', "
- "replica_id: %u, lsn: %lld, "
- "space_id: %u, index_id: %u",
- iproto_type_name(request->type),
- (unsigned) request->header->replica_id,
- (long long) request->header->lsn,
- (unsigned) request->space_id,
- (unsigned) request->index_id);
+ SNPRINT(total, snprintf, buf, size,
+ "{type: '%s', "
+ "replica_id: %u, lsn: %lld, "
+ "space_id: %u, index_id: %u",
+ iproto_type_name(request->type),
+ (unsigned)request->header->replica_id,
+ (long long)request->header->lsn, (unsigned)request->space_id,
+ (unsigned)request->index_id);
if (request->key != NULL) {
SNPRINT(total, snprintf, buf, size, ", key: ");
SNPRINT(total, mp_snprint, buf, size, request->key);
@@ -822,14 +829,14 @@ xrow_encode_dml(const struct request *request, struct region *region,
uint32_t ops_len = request->ops_end - request->ops;
uint32_t tuple_meta_len = request->tuple_meta_end - request->tuple_meta;
uint32_t tuple_len = request->tuple_end - request->tuple;
- uint32_t len = MAP_LEN_MAX + key_len + ops_len + tuple_meta_len +
- tuple_len;
- char *begin = (char *) region_alloc(region, len);
+ uint32_t len =
+ MAP_LEN_MAX + key_len + ops_len + tuple_meta_len + tuple_len;
+ char *begin = (char *)region_alloc(region, len);
if (begin == NULL) {
diag_set(OutOfMemory, len, "region_alloc", "begin");
return -1;
}
- char *pos = begin + 1; /* skip 1 byte for MP_MAP */
+ char *pos = begin + 1; /* skip 1 byte for MP_MAP */
int map_size = 0;
if (request->space_id) {
pos = mp_encode_uint(pos, IPROTO_SPACE_ID);
@@ -883,8 +890,7 @@ xrow_encode_dml(const struct request *request, struct region *region,
}
void
-xrow_encode_synchro(struct xrow_header *row,
- struct synchro_body_bin *body,
+xrow_encode_synchro(struct xrow_header *row, struct synchro_body_bin *body,
const struct synchro_request *req)
{
/*
@@ -918,8 +924,8 @@ xrow_decode_synchro(const struct xrow_header *row, struct synchro_request *req)
assert(row->bodycnt == 1);
- const char * const data = (const char *)row->body[0].iov_base;
- const char * const end = data + row->body[0].iov_len;
+ const char *const data = (const char *)row->body[0].iov_base;
+ const char *const end = data + row->body[0].iov_len;
const char *d = data;
if (mp_check(&d, end) != 0 || mp_typeof(*data) != MP_MAP) {
xrow_on_decode_err(data, end, ER_INVALID_MSGPACK,
@@ -967,8 +973,8 @@ xrow_encode_raft(struct xrow_header *row, struct region *region,
* the term is too old.
*/
int map_size = 1;
- size_t size = mp_sizeof_uint(IPROTO_RAFT_TERM) +
- mp_sizeof_uint(r->term);
+ size_t size =
+ mp_sizeof_uint(IPROTO_RAFT_TERM) + mp_sizeof_uint(r->term);
if (r->vote != 0) {
++map_size;
size += mp_sizeof_uint(IPROTO_RAFT_VOTE) +
@@ -1033,8 +1039,7 @@ xrow_decode_raft(const struct xrow_header *row, struct raft_request *r,
const char *end = begin + row->body[0].iov_len;
const char *pos = begin;
uint32_t map_size = mp_decode_map(&pos);
- for (uint32_t i = 0; i < map_size; ++i)
- {
+ for (uint32_t i = 0; i < map_size; ++i) {
if (mp_typeof(*pos) != MP_UINT)
goto bad_msgpack;
uint64_t key = mp_decode_uint(&pos);
@@ -1085,7 +1090,7 @@ xrow_to_iovec(const struct xrow_header *row, struct iovec *out)
len += out[i].iov_len;
/* Encode length */
- char *data = (char *) out[0].iov_base;
+ char *data = (char *)out[0].iov_base;
*(data++) = 0xce; /* MP_UINT32 */
store_u32(data, mp_bswap_u32(len));
@@ -1103,14 +1108,14 @@ xrow_decode_call(const struct xrow_header *row, struct call_request *request)
}
assert(row->bodycnt == 1);
- const char *data = (const char *) row->body[0].iov_base;
+ const char *data = (const char *)row->body[0].iov_base;
const char *end = data + row->body[0].iov_len;
assert((end - data) > 0);
if (mp_typeof(*data) != MP_MAP || mp_check_map(data, end) > 0) {
-error:
- xrow_on_decode_err(row->body[0].iov_base, end, ER_INVALID_MSGPACK,
- "packet body");
+ error:
+ xrow_on_decode_err(row->body[0].iov_base, end,
+ ER_INVALID_MSGPACK, "packet body");
return -1;
}
@@ -1149,20 +1154,21 @@ error:
}
}
if (data != end) {
- xrow_on_decode_err(row->body[0].iov_base, end, ER_INVALID_MSGPACK,
- "packet end");
+ xrow_on_decode_err(row->body[0].iov_base, end,
+ ER_INVALID_MSGPACK, "packet end");
return -1;
}
if (row->type == IPROTO_EVAL) {
if (request->expr == NULL) {
- xrow_on_decode_err(row->body[0].iov_base, end, ER_MISSING_REQUEST_FIELD,
+ xrow_on_decode_err(row->body[0].iov_base, end,
+ ER_MISSING_REQUEST_FIELD,
iproto_key_name(IPROTO_EXPR));
return -1;
}
} else if (request->name == NULL) {
- assert(row->type == IPROTO_CALL_16 ||
- row->type == IPROTO_CALL);
- xrow_on_decode_err(row->body[0].iov_base, end, ER_MISSING_REQUEST_FIELD,
+ assert(row->type == IPROTO_CALL_16 || row->type == IPROTO_CALL);
+ xrow_on_decode_err(row->body[0].iov_base, end,
+ ER_MISSING_REQUEST_FIELD,
iproto_key_name(IPROTO_FUNCTION_NAME));
return -1;
}
@@ -1184,14 +1190,14 @@ xrow_decode_auth(const struct xrow_header *row, struct auth_request *request)
}
assert(row->bodycnt == 1);
- const char *data = (const char *) row->body[0].iov_base;
+ const char *data = (const char *)row->body[0].iov_base;
const char *end = data + row->body[0].iov_len;
assert((end - data) > 0);
if (mp_typeof(*data) != MP_MAP || mp_check_map(data, end) > 0) {
-error:
- xrow_on_decode_err(row->body[0].iov_base, end, ER_INVALID_MSGPACK,
- "packet body");
+ error:
+ xrow_on_decode_err(row->body[0].iov_base, end,
+ ER_INVALID_MSGPACK, "packet body");
return -1;
}
@@ -1223,17 +1229,19 @@ error:
}
}
if (data != end) {
- xrow_on_decode_err(row->body[0].iov_base, end, ER_INVALID_MSGPACK,
- "packet end");
+ xrow_on_decode_err(row->body[0].iov_base, end,
+ ER_INVALID_MSGPACK, "packet end");
return -1;
}
if (request->user_name == NULL) {
- xrow_on_decode_err(row->body[0].iov_base, end, ER_MISSING_REQUEST_FIELD,
+ xrow_on_decode_err(row->body[0].iov_base, end,
+ ER_MISSING_REQUEST_FIELD,
iproto_key_name(IPROTO_USER_NAME));
return -1;
}
if (request->scramble == NULL) {
- xrow_on_decode_err(row->body[0].iov_base, end, ER_MISSING_REQUEST_FIELD,
+ xrow_on_decode_err(row->body[0].iov_base, end,
+ ER_MISSING_REQUEST_FIELD,
iproto_key_name(IPROTO_TUPLE));
return -1;
}
@@ -1242,14 +1250,14 @@ error:
int
xrow_encode_auth(struct xrow_header *packet, const char *salt, size_t salt_len,
- const char *login, size_t login_len,
- const char *password, size_t password_len)
+ const char *login, size_t login_len, const char *password,
+ size_t password_len)
{
assert(login != NULL);
memset(packet, 0, sizeof(*packet));
size_t buf_size = XROW_BODY_LEN_MAX + login_len + SCRAMBLE_SIZE;
- char *buf = (char *) region_alloc(&fiber()->gc, buf_size);
+ char *buf = (char *)region_alloc(&fiber()->gc, buf_size);
if (buf == NULL) {
diag_set(OutOfMemory, buf_size, "region_alloc", "buf");
return -1;
@@ -1259,9 +1267,9 @@ xrow_encode_auth(struct xrow_header *packet, const char *salt, size_t salt_len,
d = mp_encode_map(d, password != NULL ? 2 : 1);
d = mp_encode_uint(d, IPROTO_USER_NAME);
d = mp_encode_str(d, login, login_len);
- if (password != NULL) { /* password can be omitted */
+ if (password != NULL) { /* password can be omitted */
assert(salt_len >= SCRAMBLE_SIZE); /* greetingbuf_decode */
- (void) salt_len;
+ (void)salt_len;
char scramble[SCRAMBLE_SIZE];
scramble_prepare(scramble, salt, password, password_len);
d = mp_encode_uint(d, IPROTO_TUPLE);
@@ -1289,11 +1297,11 @@ xrow_decode_error(struct xrow_header *row)
if (row->bodycnt == 0)
goto error;
- pos = (char *) row->body[0].iov_base;
+ pos = (char *)row->body[0].iov_base;
if (mp_check(&pos, pos + row->body[0].iov_len))
goto error;
- pos = (char *) row->body[0].iov_base;
+ pos = (char *)row->body[0].iov_base;
if (mp_typeof(*pos) != MP_MAP)
goto error;
map_size = mp_decode_map(&pos);
@@ -1314,7 +1322,8 @@ xrow_decode_error(struct xrow_header *row)
uint32_t len;
const char *str = mp_decode_str(&pos, &len);
if (!is_stack_parsed) {
- snprintf(error, sizeof(error), "%.*s", len, str);
+ snprintf(error, sizeof(error), "%.*s", len,
+ str);
box_error_set(__FILE__, __LINE__, code, error);
}
} else if (key == IPROTO_ERROR) {
@@ -1356,7 +1365,7 @@ xrow_decode_ballot(struct xrow_header *row, struct ballot *ballot)
goto err;
assert(row->bodycnt == 1);
- const char *data = start = (const char *) row->body[0].iov_base;
+ const char *data = start = (const char *)row->body[0].iov_base;
end = data + row->body[0].iov_len;
const char *tmp = data;
if (mp_check(&tmp, end) != 0 || mp_typeof(*data) != MP_MAP)
@@ -1402,8 +1411,8 @@ xrow_decode_ballot(struct xrow_header *row, struct ballot *ballot)
ballot->is_anon = mp_decode_bool(&data);
break;
case IPROTO_BALLOT_VCLOCK:
- if (mp_decode_vclock_ignore0(&data,
- &ballot->vclock) != 0)
+ if (mp_decode_vclock_ignore0(&data, &ballot->vclock) !=
+ 0)
goto err;
break;
case IPROTO_BALLOT_GC_VCLOCK:
@@ -1427,12 +1436,11 @@ xrow_encode_register(struct xrow_header *row,
const struct vclock *vclock)
{
memset(row, 0, sizeof(*row));
- size_t size = mp_sizeof_map(2) +
- mp_sizeof_uint(IPROTO_INSTANCE_UUID) +
+ size_t size = mp_sizeof_map(2) + mp_sizeof_uint(IPROTO_INSTANCE_UUID) +
mp_sizeof_str(UUID_STR_LEN) +
mp_sizeof_uint(IPROTO_VCLOCK) +
mp_sizeof_vclock_ignore0(vclock);
- char *buf = (char *) region_alloc(&fiber()->gc, size);
+ char *buf = (char *)region_alloc(&fiber()->gc, size);
if (buf == NULL) {
diag_set(OutOfMemory, size, "region_alloc", "buf");
return -1;
@@ -1459,9 +1467,8 @@ xrow_encode_subscribe(struct xrow_header *row,
uint32_t id_filter)
{
memset(row, 0, sizeof(*row));
- size_t size = XROW_BODY_LEN_MAX +
- mp_sizeof_vclock_ignore0(vclock);
- char *buf = (char *) region_alloc(&fiber()->gc, size);
+ size_t size = XROW_BODY_LEN_MAX + mp_sizeof_vclock_ignore0(vclock);
+ char *buf = (char *)region_alloc(&fiber()->gc, size);
if (buf == NULL) {
diag_set(OutOfMemory, size, "region_alloc", "buf");
return -1;
@@ -1483,8 +1490,7 @@ xrow_encode_subscribe(struct xrow_header *row,
data = mp_encode_uint(data, IPROTO_ID_FILTER);
data = mp_encode_array(data, filter_size);
struct bit_iterator it;
- bit_iterator_init(&it, &id_filter, sizeof(id_filter),
- true);
+ bit_iterator_init(&it, &id_filter, sizeof(id_filter), true);
for (size_t id = bit_iterator_next(&it); id < VCLOCK_MAX;
id = bit_iterator_next(&it)) {
data = mp_encode_uint(data, id);
@@ -1501,15 +1507,14 @@ xrow_encode_subscribe(struct xrow_header *row,
int
xrow_decode_subscribe(struct xrow_header *row, struct tt_uuid *replicaset_uuid,
struct tt_uuid *instance_uuid, struct vclock *vclock,
- uint32_t *version_id, bool *anon,
- uint32_t *id_filter)
+ uint32_t *version_id, bool *anon, uint32_t *id_filter)
{
if (row->bodycnt == 0) {
diag_set(ClientError, ER_INVALID_MSGPACK, "request body");
return -1;
}
assert(row->bodycnt == 1);
- const char * const data = (const char *) row->body[0].iov_base;
+ const char *const data = (const char *)row->body[0].iov_base;
const char *end = data + row->body[0].iov_len;
const char *d = data;
if (mp_check(&d, end) != 0 || mp_typeof(*data) != MP_MAP) {
@@ -1536,8 +1541,8 @@ xrow_decode_subscribe(struct xrow_header *row, struct tt_uuid *replicaset_uuid,
if (replicaset_uuid == NULL)
goto skip;
if (xrow_decode_uuid(&d, replicaset_uuid) != 0) {
- xrow_on_decode_err(data, end, ER_INVALID_MSGPACK,
- "UUID");
+ xrow_on_decode_err(data, end,
+ ER_INVALID_MSGPACK, "UUID");
return -1;
}
break;
@@ -1545,8 +1550,8 @@ xrow_decode_subscribe(struct xrow_header *row, struct tt_uuid *replicaset_uuid,
if (instance_uuid == NULL)
goto skip;
if (xrow_decode_uuid(&d, instance_uuid) != 0) {
- xrow_on_decode_err(data, end, ER_INVALID_MSGPACK,
- "UUID");
+ xrow_on_decode_err(data, end,
+ ER_INVALID_MSGPACK, "UUID");
return -1;
}
break;
@@ -1554,7 +1559,8 @@ xrow_decode_subscribe(struct xrow_header *row, struct tt_uuid *replicaset_uuid,
if (vclock == NULL)
goto skip;
if (mp_decode_vclock_ignore0(&d, vclock) != 0) {
- xrow_on_decode_err(data, end, ER_INVALID_MSGPACK,
+ xrow_on_decode_err(data, end,
+ ER_INVALID_MSGPACK,
"invalid VCLOCK");
return -1;
}
@@ -1563,7 +1569,8 @@ xrow_decode_subscribe(struct xrow_header *row, struct tt_uuid *replicaset_uuid,
if (version_id == NULL)
goto skip;
if (mp_typeof(*d) != MP_UINT) {
- xrow_on_decode_err(data, end, ER_INVALID_MSGPACK,
+ xrow_on_decode_err(data, end,
+ ER_INVALID_MSGPACK,
"invalid VERSION");
return -1;
}
@@ -1573,7 +1580,8 @@ xrow_decode_subscribe(struct xrow_header *row, struct tt_uuid *replicaset_uuid,
if (anon == NULL)
goto skip;
if (mp_typeof(*d) != MP_BOOL) {
- xrow_on_decode_err(data, end, ER_INVALID_MSGPACK,
+ xrow_on_decode_err(data, end,
+ ER_INVALID_MSGPACK,
"invalid REPLICA_ANON flag");
return -1;
}
@@ -1583,7 +1591,9 @@ xrow_decode_subscribe(struct xrow_header *row, struct tt_uuid *replicaset_uuid,
if (id_filter == NULL)
goto skip;
if (mp_typeof(*d) != MP_ARRAY) {
-id_filter_decode_err: xrow_on_decode_err(data, end, ER_INVALID_MSGPACK,
+ id_filter_decode_err:
+ xrow_on_decode_err(data, end,
+ ER_INVALID_MSGPACK,
"invalid ID_FILTER");
return -1;
}
@@ -1597,7 +1607,8 @@ id_filter_decode_err: xrow_on_decode_err(data, end, ER_INVALID_MSGPACK,
*id_filter |= 1 << val;
}
break;
- default: skip:
+ default:
+ skip:
mp_next(&d); /* value */
}
}
@@ -1610,7 +1621,7 @@ xrow_encode_join(struct xrow_header *row, const struct tt_uuid *instance_uuid)
memset(row, 0, sizeof(*row));
size_t size = 64;
- char *buf = (char *) region_alloc(&fiber()->gc, size);
+ char *buf = (char *)region_alloc(&fiber()->gc, size);
if (buf == NULL) {
diag_set(OutOfMemory, size, "region_alloc", "buf");
return -1;
@@ -1636,7 +1647,7 @@ xrow_encode_vclock(struct xrow_header *row, const struct vclock *vclock)
/* Add vclock to response body */
size_t size = 8 + mp_sizeof_vclock_ignore0(vclock);
- char *buf = (char *) region_alloc(&fiber()->gc, size);
+ char *buf = (char *)region_alloc(&fiber()->gc, size);
if (buf == NULL) {
diag_set(OutOfMemory, size, "region_alloc", "buf");
return -1;
@@ -1659,12 +1670,11 @@ xrow_encode_subscribe_response(struct xrow_header *row,
const struct vclock *vclock)
{
memset(row, 0, sizeof(*row));
- size_t size = mp_sizeof_map(2) +
- mp_sizeof_uint(IPROTO_VCLOCK) +
+ size_t size = mp_sizeof_map(2) + mp_sizeof_uint(IPROTO_VCLOCK) +
mp_sizeof_vclock_ignore0(vclock) +
mp_sizeof_uint(IPROTO_CLUSTER_UUID) +
mp_sizeof_str(UUID_STR_LEN);
- char *buf = (char *) region_alloc(&fiber()->gc, size);
+ char *buf = (char *)region_alloc(&fiber()->gc, size);
if (buf == NULL) {
diag_set(OutOfMemory, size, "region_alloc", "buf");
return -1;
@@ -1698,8 +1708,9 @@ greeting_encode(char *greetingbuf, uint32_t version_id,
{
int h = IPROTO_GREETING_SIZE / 2;
int r = snprintf(greetingbuf, h + 1, "Tarantool %u.%u.%u (Binary) ",
- version_id_major(version_id), version_id_minor(version_id),
- version_id_patch(version_id));
+ version_id_major(version_id),
+ version_id_minor(version_id),
+ version_id_patch(version_id));
assert(r + UUID_STR_LEN < h);
tt_uuid_to_string(uuid, greetingbuf + r);
@@ -1726,17 +1737,19 @@ greeting_decode(const char *greetingbuf, struct greeting *greeting)
int h = IPROTO_GREETING_SIZE / 2;
const char *pos = greetingbuf + strlen("Tarantool ");
const char *end = greetingbuf + h;
- for (; pos < end && *pos == ' '; ++pos); /* skip spaces */
+ for (; pos < end && *pos == ' '; ++pos)
+ ; /* skip spaces */
/* Extract a version string - a string until ' ' */
char version[20];
- const char *vend = (const char *) memchr(pos, ' ', end - pos);
+ const char *vend = (const char *)memchr(pos, ' ', end - pos);
if (vend == NULL || (size_t)(vend - pos) >= sizeof(version))
return -1;
memcpy(version, pos, vend - pos);
version[vend - pos] = '\0';
pos = vend + 1;
- for (; pos < end && *pos == ' '; ++pos); /* skip spaces */
+ for (; pos < end && *pos == ' '; ++pos)
+ ; /* skip spaces */
/* Parse a version string - 1.6.6-83-gc6b2129 or 1.6.7 */
unsigned major, minor, patch;
@@ -1746,7 +1759,7 @@ greeting_decode(const char *greetingbuf, struct greeting *greeting)
if (*pos == '(') {
/* Extract protocol name - a string between (parentheses) */
- vend = (const char *) memchr(pos + 1, ')', end - pos);
+ vend = (const char *)memchr(pos + 1, ')', end - pos);
if (!vend || (vend - pos - 1) > GREETING_PROTOCOL_LEN_MAX)
return -1;
memcpy(greeting->protocol, pos + 1, vend - pos - 1);
@@ -1759,10 +1772,12 @@ greeting_decode(const char *greetingbuf, struct greeting *greeting)
if (greeting->version_id >= version_id(1, 6, 7)) {
if (*(pos++) != ' ')
return -1;
- for (; pos < end && *pos == ' '; ++pos); /* spaces */
+ for (; pos < end && *pos == ' '; ++pos)
+ ; /* spaces */
if (end - pos < UUID_STR_LEN)
return -1;
- if (tt_uuid_from_strl(pos, UUID_STR_LEN, &greeting->uuid))
+ if (tt_uuid_from_strl(pos, UUID_STR_LEN,
+ &greeting->uuid))
return -1;
}
} else if (greeting->version_id < version_id(1, 6, 7)) {
@@ -1773,10 +1788,10 @@ greeting_decode(const char *greetingbuf, struct greeting *greeting)
}
/* Decode salt for binary protocol */
- greeting->salt_len = base64_decode(greetingbuf + h, h - 1,
- greeting->salt,
- sizeof(greeting->salt));
- if (greeting->salt_len < SCRAMBLE_SIZE || greeting->salt_len >= (uint32_t)h)
+ greeting->salt_len = base64_decode(
+ greetingbuf + h, h - 1, greeting->salt, sizeof(greeting->salt));
+ if (greeting->salt_len < SCRAMBLE_SIZE ||
+ greeting->salt_len >= (uint32_t)h)
return -1;
return 0;
diff --git a/src/box/xrow.h b/src/box/xrow.h
index 25985ad..7fcc672 100644
--- a/src/box/xrow.h
+++ b/src/box/xrow.h
@@ -250,8 +250,7 @@ struct PACKED synchro_body_bin {
* @param req Request parameters.
*/
void
-xrow_encode_synchro(struct xrow_header *row,
- struct synchro_body_bin *body,
+xrow_encode_synchro(struct xrow_header *row, struct synchro_body_bin *body,
const struct synchro_request *req);
/**
@@ -429,8 +428,7 @@ xrow_encode_subscribe(struct xrow_header *row,
int
xrow_decode_subscribe(struct xrow_header *row, struct tt_uuid *replicaset_uuid,
struct tt_uuid *instance_uuid, struct vclock *vclock,
- uint32_t *version_id, bool *anon,
- uint32_t *id_filter);
+ uint32_t *version_id, bool *anon, uint32_t *id_filter);
/**
* Encode JOIN command.
@@ -510,8 +508,8 @@ xrow_decode_vclock(struct xrow_header *row, struct vclock *vclock)
*/
int
xrow_encode_subscribe_response(struct xrow_header *row,
- const struct tt_uuid *replicaset_uuid,
- const struct vclock *vclock);
+ const struct tt_uuid *replicaset_uuid,
+ const struct vclock *vclock);
/**
* Decode a response to subscribe request.
@@ -632,8 +630,8 @@ iproto_reply_vclock(struct obuf *out, const struct vclock *vclock,
* @retval -1 Memory error.
*/
int
-iproto_reply_vote(struct obuf *out, const struct ballot *ballot,
- uint64_t sync, uint32_t schema_version);
+iproto_reply_vote(struct obuf *out, const struct ballot *ballot, uint64_t sync,
+ uint32_t schema_version);
/**
* Write an error packet int output buffer. Doesn't throw if out
@@ -781,7 +779,7 @@ xrow_decode_error(struct xrow_header *row);
* @return Previous LSN value.
*/
static inline int64_t
-vclock_follow_xrow(struct vclock* vclock, const struct xrow_header *row)
+vclock_follow_xrow(struct vclock *vclock, const struct xrow_header *row)
{
assert(row);
assert(row->replica_id < VCLOCK_MAX);
@@ -793,10 +791,9 @@ vclock_follow_xrow(struct vclock* vclock, const struct xrow_header *row)
/* Never confirm LSN out of order. */
panic("LSN for %u is used twice or COMMIT order is broken: "
"confirmed: %lld, new: %lld, req: %s",
- (unsigned) row->replica_id,
- (long long) vclock_get(vclock, row->replica_id),
- (long long) row->lsn,
- req_str);
+ (unsigned)row->replica_id,
+ (long long)vclock_get(vclock, row->replica_id),
+ (long long)row->lsn, req_str);
}
return vclock_follow(vclock, row->replica_id, row->lsn);
}
@@ -853,8 +850,7 @@ xrow_encode_dml_xc(const struct request *request, struct region *region,
/** @copydoc xrow_decode_call. */
static inline void
-xrow_decode_call_xc(const struct xrow_header *row,
- struct call_request *request)
+xrow_decode_call_xc(const struct xrow_header *row, struct call_request *request)
{
if (xrow_decode_call(row, request) != 0)
diag_raise();
@@ -862,8 +858,7 @@ xrow_decode_call_xc(const struct xrow_header *row,
/** @copydoc xrow_decode_auth. */
static inline void
-xrow_decode_auth_xc(const struct xrow_header *row,
- struct auth_request *request)
+xrow_decode_auth_xc(const struct xrow_header *row, struct auth_request *request)
{
if (xrow_decode_auth(row, request) != 0)
diag_raise();
@@ -891,8 +886,8 @@ xrow_decode_ballot_xc(struct xrow_header *row, struct ballot *ballot)
/** @copydoc xrow_encode_register. */
static inline void
xrow_encode_register_xc(struct xrow_header *row,
- const struct tt_uuid *instance_uuid,
- const struct vclock *vclock)
+ const struct tt_uuid *instance_uuid,
+ const struct vclock *vclock)
{
if (xrow_encode_register(row, instance_uuid, vclock) != 0)
diag_raise();
@@ -906,8 +901,8 @@ xrow_encode_subscribe_xc(struct xrow_header *row,
const struct vclock *vclock, bool anon,
uint32_t id_filter)
{
- if (xrow_encode_subscribe(row, replicaset_uuid, instance_uuid,
- vclock, anon, id_filter) != 0)
+ if (xrow_encode_subscribe(row, replicaset_uuid, instance_uuid, vclock,
+ anon, id_filter) != 0)
diag_raise();
}
@@ -919,9 +914,8 @@ xrow_decode_subscribe_xc(struct xrow_header *row,
uint32_t *replica_version_id, bool *anon,
uint32_t *id_filter)
{
- if (xrow_decode_subscribe(row, replicaset_uuid, instance_uuid,
- vclock, replica_version_id, anon,
- id_filter) != 0)
+ if (xrow_decode_subscribe(row, replicaset_uuid, instance_uuid, vclock,
+ replica_version_id, anon, id_filter) != 0)
diag_raise();
}
@@ -1007,7 +1001,7 @@ iproto_reply_vclock_xc(struct obuf *out, const struct vclock *vclock,
/** @copydoc iproto_reply_vote. */
static inline void
iproto_reply_vote_xc(struct obuf *out, const struct ballot *ballot,
- uint64_t sync, uint32_t schema_version)
+ uint64_t sync, uint32_t schema_version)
{
if (iproto_reply_vote(out, ballot, sync, schema_version) != 0)
diag_raise();
diff --git a/src/box/xrow_io.cc b/src/box/xrow_io.cc
index 4870798..f746b52 100644
--- a/src/box/xrow_io.cc
+++ b/src/box/xrow_io.cc
@@ -44,21 +44,20 @@ coio_read_xrow(struct ev_io *coio, struct ibuf *in, struct xrow_header *row)
/* Read length */
if (mp_typeof(*in->rpos) != MP_UINT) {
- tnt_raise(ClientError, ER_INVALID_MSGPACK,
- "packet length");
+ tnt_raise(ClientError, ER_INVALID_MSGPACK, "packet length");
}
ssize_t to_read = mp_check_uint(in->rpos, in->wpos);
if (to_read > 0)
coio_breadn(coio, in, to_read);
- uint32_t len = mp_decode_uint((const char **) &in->rpos);
+ uint32_t len = mp_decode_uint((const char **)&in->rpos);
/* Read header and body */
to_read = len - ibuf_used(in);
if (to_read > 0)
coio_breadn(coio, in, to_read);
- xrow_header_decode_xc(row, (const char **) &in->rpos, in->rpos + len,
+ xrow_header_decode_xc(row, (const char **)&in->rpos, in->rpos + len,
true);
}
@@ -75,26 +74,24 @@ coio_read_xrow_timeout_xc(struct ev_io *coio, struct ibuf *in,
/* Read length */
if (mp_typeof(*in->rpos) != MP_UINT) {
- tnt_raise(ClientError, ER_INVALID_MSGPACK,
- "packet length");
+ tnt_raise(ClientError, ER_INVALID_MSGPACK, "packet length");
}
ssize_t to_read = mp_check_uint(in->rpos, in->wpos);
if (to_read > 0)
coio_breadn_timeout(coio, in, to_read, delay);
coio_timeout_update(&start, &delay);
- uint32_t len = mp_decode_uint((const char **) &in->rpos);
+ uint32_t len = mp_decode_uint((const char **)&in->rpos);
/* Read header and body */
to_read = len - ibuf_used(in);
if (to_read > 0)
coio_breadn_timeout(coio, in, to_read, delay);
- xrow_header_decode_xc(row, (const char **) &in->rpos, in->rpos + len,
+ xrow_header_decode_xc(row, (const char **)&in->rpos, in->rpos + len,
true);
}
-
void
coio_write_xrow(struct ev_io *coio, const struct xrow_header *row)
{
@@ -102,4 +99,3 @@ coio_write_xrow(struct ev_io *coio, const struct xrow_header *row)
int iovcnt = xrow_to_iovec_xc(row, iov);
coio_writev(coio, iov, iovcnt, 0);
}
-
diff --git a/src/box/xrow_io.h b/src/box/xrow_io.h
index 0eb7a8a..eedad3d 100644
--- a/src/box/xrow_io.h
+++ b/src/box/xrow_io.h
@@ -48,7 +48,6 @@ coio_read_xrow_timeout_xc(struct ev_io *coio, struct ibuf *in,
void
coio_write_xrow(struct ev_io *coio, const struct xrow_header *row);
-
#if defined(__cplusplus)
} /* extern "C" */
#endif
diff --git a/src/box/xrow_update.c b/src/box/xrow_update.c
index 0493c0d..68e1395 100644
--- a/src/box/xrow_update.c
+++ b/src/box/xrow_update.c
@@ -102,8 +102,7 @@
*/
/** Update internal state */
-struct xrow_update
-{
+struct xrow_update {
/** Operations array. */
struct xrow_update_op *ops;
/** Length of ops. */
@@ -168,9 +167,8 @@ xrow_update_read_ops(struct xrow_update *update, const char *expr,
}
int size = update->op_count * sizeof(update->ops[0]);
- update->ops = (struct xrow_update_op *)
- region_aligned_alloc(&fiber()->gc, size,
- alignof(struct xrow_update_op));
+ update->ops = (struct xrow_update_op *)region_aligned_alloc(
+ &fiber()->gc, size, alignof(struct xrow_update_op));
if (update->ops == NULL) {
diag_set(OutOfMemory, size, "region_aligned_alloc",
"update->ops");
@@ -253,7 +251,7 @@ xrow_update_read_ops(struct xrow_update *update, const char *expr,
if (opcode == '!')
++field_count_hint;
else if (opcode == '#')
- field_count_hint -= (int32_t) op->arg.del.count;
+ field_count_hint -= (int32_t)op->arg.del.count;
if (opcode == '!' || opcode == '#')
/*
@@ -349,7 +347,7 @@ xrow_update_finish(struct xrow_update *update, struct tuple_format *format,
uint32_t *p_tuple_len)
{
uint32_t tuple_len = xrow_update_array_sizeof(&update->root);
- char *buffer = (char *) region_alloc(&fiber()->gc, tuple_len);
+ char *buffer = (char *)region_alloc(&fiber()->gc, tuple_len);
if (buffer == NULL) {
diag_set(OutOfMemory, tuple_len, "region_alloc", "buffer");
return NULL;
@@ -371,7 +369,7 @@ xrow_update_check_ops(const char *expr, const char *expr_end,
}
const char *
-xrow_update_execute(const char *expr,const char *expr_end,
+xrow_update_execute(const char *expr, const char *expr_end,
const char *old_data, const char *old_data_end,
struct tuple_format *format, uint32_t *p_tuple_len,
int index_base, uint64_t *column_mask)
@@ -394,7 +392,7 @@ xrow_update_execute(const char *expr,const char *expr_end,
}
const char *
-xrow_upsert_execute(const char *expr,const char *expr_end,
+xrow_upsert_execute(const char *expr, const char *expr_end,
const char *old_data, const char *old_data_end,
struct tuple_format *format, uint32_t *p_tuple_len,
int index_base, bool suppress_error, uint64_t *column_mask)
@@ -417,19 +415,18 @@ xrow_upsert_execute(const char *expr,const char *expr_end,
}
const char *
-xrow_upsert_squash(const char *expr1, const char *expr1_end,
- const char *expr2, const char *expr2_end,
- struct tuple_format *format, size_t *result_size,
- int index_base)
+xrow_upsert_squash(const char *expr1, const char *expr1_end, const char *expr2,
+ const char *expr2_end, struct tuple_format *format,
+ size_t *result_size, int index_base)
{
- const char *expr[2] = {expr1, expr2};
- const char *expr_end[2] = {expr1_end, expr2_end};
+ const char *expr[2] = { expr1, expr2 };
+ const char *expr_end[2] = { expr1_end, expr2_end };
struct xrow_update update[2];
struct tuple_dictionary *dict = format->dict;
for (int j = 0; j < 2; j++) {
xrow_update_init(&update[j], index_base);
- if (xrow_update_read_ops(&update[j], expr[j], expr_end[j],
- dict, 0) != 0)
+ if (xrow_update_read_ops(&update[j], expr[j], expr_end[j], dict,
+ 0) != 0)
return NULL;
mp_decode_array(&expr[j]);
int32_t prev_field_no = index_base - 1;
@@ -454,8 +451,8 @@ xrow_upsert_squash(const char *expr1, const char *expr1_end,
}
size_t possible_size = expr1_end - expr1 + expr2_end - expr2;
const uint32_t space_for_arr_tag = 5;
- char *buf = (char *) region_alloc(&fiber()->gc,
- possible_size + space_for_arr_tag);
+ char *buf = (char *)region_alloc(&fiber()->gc,
+ possible_size + space_for_arr_tag);
if (buf == NULL) {
diag_set(OutOfMemory, possible_size + space_for_arr_tag,
"region_alloc", "buf");
@@ -465,16 +462,16 @@ xrow_upsert_squash(const char *expr1, const char *expr1_end,
char *res_ops = buf + space_for_arr_tag;
uint32_t res_count = 0; /* number of resulting operations */
- uint32_t op_count[2] = {update[0].op_count, update[1].op_count};
- uint32_t op_no[2] = {0, 0};
+ uint32_t op_count[2] = { update[0].op_count, update[1].op_count };
+ uint32_t op_no[2] = { 0, 0 };
struct json_tree *format_tree = &format->fields;
struct json_token *root = &format_tree->root;
struct json_token token;
token.type = JSON_TOKEN_NUM;
while (op_no[0] < op_count[0] || op_no[1] < op_count[1]) {
res_count++;
- struct xrow_update_op *op[2] = {update[0].ops + op_no[0],
- update[1].ops + op_no[1]};
+ struct xrow_update_op *op[2] = { update[0].ops + op_no[0],
+ update[1].ops + op_no[1] };
/*
* from:
* 0 - take op from first update,
@@ -482,11 +479,13 @@ xrow_upsert_squash(const char *expr1, const char *expr1_end,
* 2 - merge both ops
*/
uint32_t from;
- uint32_t has[2] = {op_no[0] < op_count[0], op_no[1] < op_count[1]};
+ uint32_t has[2] = { op_no[0] < op_count[0],
+ op_no[1] < op_count[1] };
assert(has[0] || has[1]);
if (has[0] && has[1]) {
from = op[0]->field_no < op[1]->field_no ? 0 :
- op[0]->field_no > op[1]->field_no ? 1 : 2;
+ op[0]->field_no > op[1]->field_no ? 1 :
+ 2;
} else {
assert(has[0] != has[1]);
from = has[1];
@@ -527,21 +526,20 @@ xrow_upsert_squash(const char *expr1, const char *expr1_end,
*/
if (op[0]->opcode == '=') {
if (xrow_mp_read_arg_arith(op[0], &op[0]->arg.set.value,
- &arith) != 0)
+ &arith) != 0)
return NULL;
} else {
arith = op[0]->arg.arith;
}
struct xrow_update_op res;
- if (xrow_update_arith_make(op[1], arith,
- &res.arg.arith) != 0)
+ if (xrow_update_arith_make(op[1], arith, &res.arg.arith) != 0)
return NULL;
res_ops = mp_encode_array(res_ops, 3);
- res_ops = mp_encode_str(res_ops,
- (const char *)&op[0]->opcode, 1);
+ res_ops =
+ mp_encode_str(res_ops, (const char *)&op[0]->opcode, 1);
token.num = op[0]->field_no;
- res_ops = mp_encode_uint(res_ops, token.num +
- update[0].index_base);
+ res_ops = mp_encode_uint(res_ops,
+ token.num + update[0].index_base);
struct json_token *this_node =
json_tree_lookup(format_tree, root, &token);
xrow_update_op_store_arith(&res, format_tree, this_node, NULL,
@@ -554,8 +552,7 @@ xrow_upsert_squash(const char *expr1, const char *expr1_end,
}
assert(op_no[0] == op_count[0] && op_no[1] == op_count[1]);
assert(expr[0] == expr_end[0] && expr[1] == expr_end[1]);
- char *arr_start = buf + space_for_arr_tag -
- mp_sizeof_array(res_count);
+ char *arr_start = buf + space_for_arr_tag - mp_sizeof_array(res_count);
mp_encode_array(arr_start, res_count);
*result_size = res_ops - arr_start;
return arr_start;
diff --git a/src/box/xrow_update.h b/src/box/xrow_update.h
index d48c379..281c2fb 100644
--- a/src/box/xrow_update.h
+++ b/src/box/xrow_update.h
@@ -51,7 +51,7 @@ xrow_update_check_ops(const char *expr, const char *expr_end,
struct tuple_format *format, int index_base);
const char *
-xrow_update_execute(const char *expr,const char *expr_end,
+xrow_update_execute(const char *expr, const char *expr_end,
const char *old_data, const char *old_data_end,
struct tuple_format *format, uint32_t *p_new_size,
int index_base, uint64_t *column_mask);
@@ -60,8 +60,7 @@ const char *
xrow_upsert_execute(const char *expr, const char *expr_end,
const char *old_data, const char *old_data_end,
struct tuple_format *format, uint32_t *p_new_size,
- int index_base, bool suppress_error,
- uint64_t *column_mask);
+ int index_base, bool suppress_error, uint64_t *column_mask);
/**
* Try to merge two update/upsert expressions to an equivalent one.
@@ -74,10 +73,9 @@ xrow_upsert_execute(const char *expr, const char *expr_end,
* If it isn't possible to merge expressions NULL is returned.
*/
const char *
-xrow_upsert_squash(const char *expr1, const char *expr1_end,
- const char *expr2, const char *expr2_end,
- struct tuple_format *format, size_t *result_size,
- int index_base);
+xrow_upsert_squash(const char *expr1, const char *expr1_end, const char *expr2,
+ const char *expr2_end, struct tuple_format *format,
+ size_t *result_size, int index_base);
#if defined(__cplusplus)
} /* extern "C" */
diff --git a/src/box/xrow_update_array.c b/src/box/xrow_update_array.c
index 717466b..f90e539 100644
--- a/src/box/xrow_update_array.c
+++ b/src/box/xrow_update_array.c
@@ -47,8 +47,8 @@ xrow_update_op_prepare_num_token(struct xrow_update_op *op)
if (op->is_token_consumed && xrow_update_op_next_token(op) != 0)
return -1;
if (op->token_type != JSON_TOKEN_NUM) {
- return xrow_update_err(op, "can't update an array by a "\
- "non-numeric index");
+ return xrow_update_err(op, "can't update an array by a "
+ "non-numeric index");
}
return 0;
}
@@ -117,9 +117,10 @@ xrow_update_array_item_split(struct region *region,
struct xrow_update_array_item *prev, size_t size,
size_t offset)
{
- (void) size;
- struct xrow_update_array_item *next = (struct xrow_update_array_item *)
- xrow_update_alloc(region, sizeof(*next));
+ (void)size;
+ struct xrow_update_array_item *next =
+ (struct xrow_update_array_item *)xrow_update_alloc(
+ region, sizeof(*next));
if (next == NULL)
return NULL;
assert(offset > 0 && prev->tail_size > 0);
@@ -175,8 +176,9 @@ xrow_update_array_create(struct xrow_update_field *field, const char *header,
field->array.rope = xrow_update_rope_new(region);
if (field->array.rope == NULL)
return -1;
- struct xrow_update_array_item *item = (struct xrow_update_array_item *)
- xrow_update_alloc(region, sizeof(*item));
+ struct xrow_update_array_item *item =
+ (struct xrow_update_array_item *)xrow_update_alloc(
+ region, sizeof(*item));
if (item == NULL)
return -1;
if (data == data_end)
@@ -206,8 +208,9 @@ xrow_update_array_create_with_child(struct xrow_update_field *field,
struct xrow_update_rope *rope = xrow_update_rope_new(region);
if (rope == NULL)
return -1;
- struct xrow_update_array_item *item = (struct xrow_update_array_item *)
- xrow_update_alloc(region, sizeof(*item));
+ struct xrow_update_array_item *item =
+ (struct xrow_update_array_item *)xrow_update_alloc(
+ region, sizeof(*item));
if (item == NULL)
return -1;
const char *end = first_field_end;
@@ -219,8 +222,8 @@ xrow_update_array_create_with_child(struct xrow_update_field *field,
end - first_field_end);
if (xrow_update_rope_append(rope, item, field_no) != 0)
return -1;
- item = (struct xrow_update_array_item *)
- xrow_update_alloc(region, sizeof(*item));
+ item = (struct xrow_update_array_item *)xrow_update_alloc(
+ region, sizeof(*item));
if (item == NULL)
return -1;
first_field = end;
@@ -292,10 +295,12 @@ xrow_update_array_store(struct xrow_update_field *field,
for (; node != NULL; node = xrow_update_rope_iter_next(&it)) {
struct xrow_update_array_item *item =
xrow_update_rope_leaf_data(node);
- next_node = json_tree_lookup(format_tree, this_node, &token);
+ next_node = json_tree_lookup(format_tree, this_node,
+ &token);
uint32_t field_count = xrow_update_rope_leaf_size(node);
- out += xrow_update_field_store(&item->field, format_tree,
- next_node, out, out_end);
+ out += xrow_update_field_store(&item->field,
+ format_tree, next_node,
+ out, out_end);
assert(item->tail_size == 0 || field_count > 1);
memcpy(out, item->field.data + item->field.size,
item->tail_size);
@@ -304,7 +309,7 @@ xrow_update_array_store(struct xrow_update_field *field,
total_field_count += field_count;
}
}
- (void) total_field_count;
+ (void)total_field_count;
assert(xrow_update_rope_size(field->array.rope) == total_field_count);
assert(out <= out_end);
return out - out_begin;
@@ -332,8 +337,8 @@ xrow_update_op_do_array_insert(struct xrow_update_op *op,
if (xrow_update_op_adjust_field_no(op, size + 1) != 0)
return -1;
- item = (struct xrow_update_array_item *)
- xrow_update_alloc(rope->ctx, sizeof(*item));
+ item = (struct xrow_update_array_item *)xrow_update_alloc(
+ rope->ctx, sizeof(*item));
if (item == NULL)
return -1;
xrow_update_array_item_create(item, XUPDATE_NOP, op->arg.set.value,
@@ -351,7 +356,7 @@ xrow_update_op_do_array_set(struct xrow_update_op *op,
return -1;
/* Interpret '=' for n + 1 field as insert. */
- if (op->field_no == (int32_t) xrow_update_rope_size(rope))
+ if (op->field_no == (int32_t)xrow_update_rope_size(rope))
return xrow_update_op_do_array_insert(op, field);
struct xrow_update_array_item *item =
@@ -396,7 +401,7 @@ xrow_update_op_do_array_delete(struct xrow_update_op *op,
if (xrow_update_op_adjust_field_no(op, size) != 0)
return -1;
uint32_t delete_count = op->arg.del.count;
- if ((uint64_t) op->field_no + delete_count > size)
+ if ((uint64_t)op->field_no + delete_count > size)
delete_count = size - op->field_no;
assert(delete_count > 0);
for (uint32_t u = delete_count; u != 0; --u)
@@ -404,29 +409,29 @@ xrow_update_op_do_array_delete(struct xrow_update_op *op,
return 0;
}
-#define DO_SCALAR_OP_GENERIC(op_type) \
-int \
-xrow_update_op_do_array_##op_type(struct xrow_update_op *op, \
- struct xrow_update_field *field) \
-{ \
- if (xrow_update_op_prepare_num_token(op) != 0) \
- return -1; \
- struct xrow_update_array_item *item = \
- xrow_update_array_extract_item(field, op); \
- if (item == NULL) \
- return -1; \
- if (!xrow_update_op_is_term(op)) { \
- op->is_token_consumed = true; \
- return xrow_update_op_do_field_##op_type(op, &item->field); \
- } \
- if (item->field.type != XUPDATE_NOP) \
- return xrow_update_err_double(op); \
- if (xrow_update_op_do_##op_type(op, item->field.data) != 0) \
- return -1; \
- item->field.type = XUPDATE_SCALAR; \
- item->field.scalar.op = op; \
- return 0; \
-}
+#define DO_SCALAR_OP_GENERIC(op_type) \
+ int xrow_update_op_do_array_##op_type(struct xrow_update_op *op, \
+ struct xrow_update_field *field) \
+ { \
+ if (xrow_update_op_prepare_num_token(op) != 0) \
+ return -1; \
+ struct xrow_update_array_item *item = \
+ xrow_update_array_extract_item(field, op); \
+ if (item == NULL) \
+ return -1; \
+ if (!xrow_update_op_is_term(op)) { \
+ op->is_token_consumed = true; \
+ return xrow_update_op_do_field_##op_type( \
+ op, &item->field); \
+ } \
+ if (item->field.type != XUPDATE_NOP) \
+ return xrow_update_err_double(op); \
+ if (xrow_update_op_do_##op_type(op, item->field.data) != 0) \
+ return -1; \
+ item->field.type = XUPDATE_SCALAR; \
+ item->field.scalar.op = op; \
+ return 0; \
+ }
DO_SCALAR_OP_GENERIC(arith)
diff --git a/src/box/xrow_update_bar.c b/src/box/xrow_update_bar.c
index 796f340..28ff572 100644
--- a/src/box/xrow_update_bar.c
+++ b/src/box/xrow_update_bar.c
@@ -66,8 +66,7 @@ xrow_update_bar_finish(struct xrow_update_field *field)
*/
static inline int
xrow_update_bar_locate(struct xrow_update_op *op,
- struct xrow_update_field *field,
- int *key_len_or_index)
+ struct xrow_update_field *field, int *key_len_or_index)
{
/*
* Bar update is not flat by definition. It always has a
@@ -88,7 +87,6 @@ xrow_update_bar_locate(struct xrow_update_op *op,
struct json_token token;
while ((rc = json_lexer_next_token(&op->lexer, &token)) == 0 &&
token.type != JSON_TOKEN_END) {
-
switch (token.type) {
case JSON_TOKEN_NUM:
field->bar.parent = pos;
@@ -186,11 +184,11 @@ xrow_update_bar_locate_opt(struct xrow_update_op *op,
if (token.type == JSON_TOKEN_NUM) {
const char *tmp = field->bar.parent;
if (mp_typeof(*tmp) != MP_ARRAY) {
- return xrow_update_err(op, "can not access by index a "\
- "non-array field");
+ return xrow_update_err(op, "can not access by index a "
+ "non-array field");
}
uint32_t size = mp_decode_array(&tmp);
- if ((uint32_t) token.num > size)
+ if ((uint32_t)token.num > size)
return xrow_update_err_no_such_field(op);
/*
* The updated point is in an array, its position
@@ -199,7 +197,7 @@ xrow_update_bar_locate_opt(struct xrow_update_op *op,
* to append a new array element. The following
* code tries to find the array's end.
*/
- assert((uint32_t) token.num == size);
+ assert((uint32_t)token.num == size);
if (field->bar.parent == field->data) {
/*
* Optimization for the case when the path
@@ -220,8 +218,8 @@ xrow_update_bar_locate_opt(struct xrow_update_op *op,
field->bar.new_key = token.str;
field->bar.new_key_len = token.len;
if (mp_typeof(*field->bar.parent) != MP_MAP) {
- return xrow_update_err(op, "can not access by key a "\
- "non-map field");
+ return xrow_update_err(op, "can not access by key a "
+ "non-map field");
}
}
return 0;
@@ -306,19 +304,18 @@ xrow_update_op_do_nop_delete(struct xrow_update_op *op,
return xrow_update_bar_finish(field);
}
-#define DO_NOP_OP_GENERIC(op_type) \
-int \
-xrow_update_op_do_nop_##op_type(struct xrow_update_op *op, \
- struct xrow_update_field *field) \
-{ \
- assert(field->type == XUPDATE_NOP); \
- int key_len_or_index; \
- if (xrow_update_bar_locate(op, field, &key_len_or_index) != 0) \
- return -1; \
- if (xrow_update_op_do_##op_type(op, field->bar.point) != 0) \
- return -1; \
- return xrow_update_bar_finish(field); \
-}
+#define DO_NOP_OP_GENERIC(op_type) \
+ int xrow_update_op_do_nop_##op_type(struct xrow_update_op *op, \
+ struct xrow_update_field *field) \
+ { \
+ assert(field->type == XUPDATE_NOP); \
+ int key_len_or_index; \
+ if (xrow_update_bar_locate(op, field, &key_len_or_index) != 0) \
+ return -1; \
+ if (xrow_update_op_do_##op_type(op, field->bar.point) != 0) \
+ return -1; \
+ return xrow_update_bar_finish(field); \
+ }
DO_NOP_OP_GENERIC(arith)
@@ -328,17 +325,16 @@ DO_NOP_OP_GENERIC(splice)
#undef DO_NOP_OP_GENERIC
-#define DO_BAR_OP_GENERIC(op_type) \
-int \
-xrow_update_op_do_bar_##op_type(struct xrow_update_op *op, \
- struct xrow_update_field *field) \
-{ \
- assert(field->type == XUPDATE_BAR); \
- field = xrow_update_route_branch(field, op); \
- if (field == NULL) \
- return -1; \
- return xrow_update_op_do_field_##op_type(op, field); \
-}
+#define DO_BAR_OP_GENERIC(op_type) \
+ int xrow_update_op_do_bar_##op_type(struct xrow_update_op *op, \
+ struct xrow_update_field *field) \
+ { \
+ assert(field->type == XUPDATE_BAR); \
+ field = xrow_update_route_branch(field, op); \
+ if (field == NULL) \
+ return -1; \
+ return xrow_update_op_do_field_##op_type(op, field); \
+ }
DO_BAR_OP_GENERIC(insert)
@@ -358,7 +354,7 @@ uint32_t
xrow_update_bar_sizeof(struct xrow_update_field *field)
{
assert(field->type == XUPDATE_BAR);
- switch(field->bar.op->opcode) {
+ switch (field->bar.op->opcode) {
case '!': {
const char *parent = field->bar.parent;
uint32_t size = field->size + field->bar.op->new_field_len;
@@ -401,10 +397,10 @@ xrow_update_bar_store(struct xrow_update_field *field,
struct json_token *this_node, char *out, char *out_end)
{
assert(field->type == XUPDATE_BAR);
- (void) out_end;
+ (void)out_end;
struct xrow_update_op *op = field->bar.op;
char *out_saved = out;
- switch(op->opcode) {
+ switch (op->opcode) {
case '!': {
const char *pos = field->bar.parent;
uint32_t before_parent = pos - field->data;
diff --git a/src/box/xrow_update_field.c b/src/box/xrow_update_field.c
index 1095ece..6f9118b 100644
--- a/src/box/xrow_update_field.c
+++ b/src/box/xrow_update_field.c
@@ -85,8 +85,9 @@ int
xrow_update_err_no_such_field(const struct xrow_update_op *op)
{
if (op->lexer.src == NULL) {
- diag_set(ClientError, ER_NO_SUCH_FIELD_NO, op->field_no +
- (op->field_no >= 0 ? TUPLE_INDEX_BASE : 0));
+ diag_set(ClientError, ER_NO_SUCH_FIELD_NO,
+ op->field_no +
+ (op->field_no >= 0 ? TUPLE_INDEX_BASE : 0));
return -1;
}
diag_set(ClientError, ER_NO_SUCH_FIELD_NAME,
@@ -132,7 +133,7 @@ xrow_update_field_store(struct xrow_update_field *field,
struct json_token *this_node, char *out, char *out_end)
{
struct xrow_update_op *op;
- switch(field->type) {
+ switch (field->type) {
case XUPDATE_NOP:
assert(out_end - out >= field->size);
memcpy(out, field->data, field->size);
@@ -188,7 +189,7 @@ xrow_mp_read_arg_arith(struct xrow_update_op *op, const char **expr,
{
int8_t ext_type;
uint32_t len;
- switch(mp_typeof(**expr)) {
+ switch (mp_typeof(**expr)) {
case MP_UINT:
ret->type = XUPDATE_TYPE_INT;
int96_set_unsigned(&ret->int96, mp_decode_uint(expr));
@@ -237,10 +238,10 @@ static int
xrow_update_read_arg_set(struct xrow_update_op *op, const char **expr,
int index_base)
{
- (void) index_base;
+ (void)index_base;
op->arg.set.value = *expr;
mp_next(expr);
- op->arg.set.length = (uint32_t) (*expr - op->arg.set.value);
+ op->arg.set.length = (uint32_t)(*expr - op->arg.set.value);
return 0;
}
@@ -248,9 +249,9 @@ static int
xrow_update_read_arg_delete(struct xrow_update_op *op, const char **expr,
int index_base)
{
- (void) index_base;
+ (void)index_base;
if (mp_typeof(**expr) == MP_UINT) {
- op->arg.del.count = (uint32_t) mp_decode_uint(expr);
+ op->arg.del.count = (uint32_t)mp_decode_uint(expr);
if (op->arg.del.count != 0)
return 0;
return xrow_update_err(op, "cannot delete 0 fields");
@@ -262,7 +263,7 @@ static int
xrow_update_read_arg_arith(struct xrow_update_op *op, const char **expr,
int index_base)
{
- (void) index_base;
+ (void)index_base;
return xrow_mp_read_arg_arith(op, expr, &op->arg.arith);
}
@@ -270,7 +271,7 @@ static int
xrow_update_read_arg_bit(struct xrow_update_op *op, const char **expr,
int index_base)
{
- (void) index_base;
+ (void)index_base;
return xrow_update_mp_read_uint(op, expr, &op->arg.bit.val);
}
@@ -373,7 +374,7 @@ xrow_update_arith_make(struct xrow_update_op *op,
lowest_type = arg2.type;
if (lowest_type == XUPDATE_TYPE_INT) {
- switch(opcode) {
+ switch (opcode) {
case '+':
int96_add(&arg1.int96, &arg2.int96);
break;
@@ -393,7 +394,7 @@ xrow_update_arith_make(struct xrow_update_op *op,
double a = xrow_update_arg_arith_to_double(arg1);
double b = xrow_update_arg_arith_to_double(arg2);
double c;
- switch(opcode) {
+ switch (opcode) {
case '+':
c = a + b;
break;
@@ -417,13 +418,13 @@ xrow_update_arith_make(struct xrow_update_op *op,
return 0;
} else {
decimal_t a, b, c;
- if (! xrow_update_arg_arith_to_decimal(arg1, &a) ||
- ! xrow_update_arg_arith_to_decimal(arg2, &b)) {
- return xrow_update_err_arg_type(op, "a number "\
- "convertible to "\
- "decimal");
+ if (!xrow_update_arg_arith_to_decimal(arg1, &a) ||
+ !xrow_update_arg_arith_to_decimal(arg2, &b)) {
+ return xrow_update_err_arg_type(op, "a number "
+ "convertible to "
+ "decimal");
}
- switch(opcode) {
+ switch (opcode) {
case '+':
if (decimal_add(&c, &a, &b) == NULL)
return xrow_update_err_decimal_overflow(op);
@@ -482,7 +483,7 @@ xrow_update_op_do_splice(struct xrow_update_op *op, const char *old)
{
struct xrow_update_arg_splice *arg = &op->arg.splice;
int32_t str_len = 0;
- if (xrow_update_mp_read_str(op, &old, (uint32_t *) &str_len, &old) != 0)
+ if (xrow_update_mp_read_str(op, &old, (uint32_t *)&str_len, &old) != 0)
return -1;
if (arg->offset < 0) {
@@ -520,9 +521,9 @@ xrow_update_op_store_set(struct xrow_update_op *op,
struct json_token *this_node, const char *in,
char *out)
{
- (void) format_tree;
- (void) this_node;
- (void) in;
+ (void)format_tree;
+ (void)this_node;
+ (void)in;
memcpy(out, op->arg.set.value, op->arg.set.length);
return op->arg.set.length;
}
@@ -533,19 +534,19 @@ xrow_update_op_store_arith(struct xrow_update_op *op,
struct json_token *this_node, const char *in,
char *out)
{
- (void) format_tree;
- (void) in;
+ (void)format_tree;
+ (void)in;
char *begin = out;
struct xrow_update_arg_arith *arg = &op->arg.arith;
switch (arg->type) {
case XUPDATE_TYPE_INT:
if (int96_is_uint64(&arg->int96)) {
- out = mp_encode_uint(
- out, int96_extract_uint64(&arg->int96));
+ out = mp_encode_uint(out,
+ int96_extract_uint64(&arg->int96));
} else {
assert(int96_is_neg_int64(&arg->int96));
out = mp_encode_int(
- out, int96_extract_neg_int64( &arg->int96));
+ out, int96_extract_neg_int64(&arg->int96));
}
break;
case XUPDATE_TYPE_DOUBLE:
@@ -555,7 +556,8 @@ xrow_update_op_store_arith(struct xrow_update_op *op,
if (this_node != NULL) {
enum field_type type =
json_tree_entry(this_node, struct tuple_field,
- token)->type;
+ token)
+ ->type;
if (type == FIELD_TYPE_DOUBLE) {
out = mp_encode_double(out, arg->flt);
break;
@@ -577,9 +579,9 @@ xrow_update_op_store_bit(struct xrow_update_op *op,
struct json_token *this_node, const char *in,
char *out)
{
- (void) format_tree;
- (void) this_node;
- (void) in;
+ (void)format_tree;
+ (void)this_node;
+ (void)in;
char *end = mp_encode_uint(out, op->arg.bit.val);
return end - out;
}
@@ -590,13 +592,13 @@ xrow_update_op_store_splice(struct xrow_update_op *op,
struct json_token *this_node, const char *in,
char *out)
{
- (void) format_tree;
- (void) this_node;
+ (void)format_tree;
+ (void)this_node;
struct xrow_update_arg_splice *arg = &op->arg.splice;
- uint32_t new_str_len = arg->offset + arg->paste_length +
- arg->tail_length;
+ uint32_t new_str_len =
+ arg->offset + arg->paste_length + arg->tail_length;
char *begin = out;
- (void) mp_decode_strl(&in);
+ (void)mp_decode_strl(&in);
out = mp_encode_strl(out, new_str_len);
/* Copy field head. */
memcpy(out, in, arg->offset);
@@ -614,27 +616,27 @@ xrow_update_op_store_splice(struct xrow_update_op *op,
static const struct xrow_update_op_meta op_set = {
xrow_update_read_arg_set, xrow_update_op_do_field_set,
- (xrow_update_op_store_f) xrow_update_op_store_set, 3
+ (xrow_update_op_store_f)xrow_update_op_store_set, 3
};
static const struct xrow_update_op_meta op_insert = {
xrow_update_read_arg_set, xrow_update_op_do_field_insert,
- (xrow_update_op_store_f) xrow_update_op_store_set, 3
+ (xrow_update_op_store_f)xrow_update_op_store_set, 3
};
static const struct xrow_update_op_meta op_arith = {
xrow_update_read_arg_arith, xrow_update_op_do_field_arith,
- (xrow_update_op_store_f) xrow_update_op_store_arith, 3
+ (xrow_update_op_store_f)xrow_update_op_store_arith, 3
};
static const struct xrow_update_op_meta op_bit = {
xrow_update_read_arg_bit, xrow_update_op_do_field_bit,
- (xrow_update_op_store_f) xrow_update_op_store_bit, 3
+ (xrow_update_op_store_f)xrow_update_op_store_bit, 3
};
static const struct xrow_update_op_meta op_splice = {
xrow_update_read_arg_splice, xrow_update_op_do_field_splice,
- (xrow_update_op_store_f) xrow_update_op_store_splice, 5
+ (xrow_update_op_store_f)xrow_update_op_store_splice, 5
};
static const struct xrow_update_op_meta op_delete = {
xrow_update_read_arg_delete, xrow_update_op_do_field_delete,
- (xrow_update_op_store_f) NULL, 3
+ (xrow_update_op_store_f)NULL, 3
};
static inline const struct xrow_update_op_meta *
@@ -689,13 +691,15 @@ xrow_update_op_decode(struct xrow_update_op *op, int op_num, int index_base,
struct tuple_dictionary *dict, const char **expr)
{
if (mp_typeof(**expr) != MP_ARRAY) {
- diag_set(ClientError, ER_ILLEGAL_PARAMS, "update operation "
+ diag_set(ClientError, ER_ILLEGAL_PARAMS,
+ "update operation "
"must be an array {op,..}");
return -1;
}
uint32_t len, arg_count = mp_decode_array(expr);
if (arg_count < 1) {
- diag_set(ClientError, ER_ILLEGAL_PARAMS, "update operation "\
+ diag_set(ClientError, ER_ILLEGAL_PARAMS,
+ "update operation "
"must be an array {op,..}, got empty array");
return -1;
}
@@ -710,7 +714,7 @@ xrow_update_op_decode(struct xrow_update_op *op, int op_num, int index_base,
return -1;
op->opcode = *opcode;
if (arg_count != op->meta->arg_count) {
- const char *str = tt_sprintf("wrong number of arguments, "\
+ const char *str = tt_sprintf("wrong number of arguments, "
"expected %u, got %u",
op->meta->arg_count, arg_count);
diag_set(ClientError, ER_UNKNOWN_UPDATE_OP, op_num, str);
@@ -724,7 +728,7 @@ xrow_update_op_decode(struct xrow_update_op *op, int op_num, int index_base,
op->token_type = JSON_TOKEN_NUM;
op->is_token_consumed = false;
int32_t field_no = 0;
- switch(mp_typeof(**expr)) {
+ switch (mp_typeof(**expr)) {
case MP_INT:
case MP_UINT: {
json_lexer_create(&op->lexer, NULL, 0, 0);
@@ -744,9 +748,9 @@ xrow_update_op_decode(struct xrow_update_op *op, int op_num, int index_base,
const char *path = mp_decode_str(expr, &len);
uint32_t field_no, hash = field_name_hash(path, len);
json_lexer_create(&op->lexer, path, len, TUPLE_INDEX_BASE);
- if (tuple_fieldno_by_name(dict, path, len, hash,
- &field_no) == 0) {
- op->field_no = (int32_t) field_no;
+ if (tuple_fieldno_by_name(dict, path, len, hash, &field_no) ==
+ 0) {
+ op->field_no = (int32_t)field_no;
op->lexer.offset = len;
break;
}
@@ -762,7 +766,7 @@ xrow_update_op_decode(struct xrow_update_op *op, int op_num, int index_base,
hash = field_name_hash(token.str, token.len);
if (tuple_fieldno_by_name(dict, token.str, token.len,
hash, &field_no) == 0) {
- op->field_no = (int32_t) field_no;
+ op->field_no = (int32_t)field_no;
break;
}
FALLTHROUGH;
diff --git a/src/box/xrow_update_field.h b/src/box/xrow_update_field.h
index 193df58..3d5c39b 100644
--- a/src/box/xrow_update_field.h
+++ b/src/box/xrow_update_field.h
@@ -76,9 +76,9 @@ struct xrow_update_arg_del {
*/
enum xrow_update_arith_type {
XUPDATE_TYPE_DECIMAL = 0, /* MP_EXT + MP_DECIMAL */
- XUPDATE_TYPE_DOUBLE = 1, /* MP_DOUBLE */
- XUPDATE_TYPE_FLOAT = 2, /* MP_FLOAT */
- XUPDATE_TYPE_INT = 3 /* MP_INT/MP_UINT */
+ XUPDATE_TYPE_DOUBLE = 1, /* MP_DOUBLE */
+ XUPDATE_TYPE_FLOAT = 2, /* MP_FLOAT */
+ XUPDATE_TYPE_INT = 3 /* MP_INT/MP_UINT */
};
/**
@@ -143,19 +143,16 @@ union xrow_update_arg {
struct xrow_update_arg_splice splice;
};
-typedef int
-(*xrow_update_op_read_arg_f)(struct xrow_update_op *op, const char **expr,
- int index_base);
+typedef int (*xrow_update_op_read_arg_f)(struct xrow_update_op *op,
+ const char **expr, int index_base);
-typedef int
-(*xrow_update_op_do_f)(struct xrow_update_op *op,
- struct xrow_update_field *field);
+typedef int (*xrow_update_op_do_f)(struct xrow_update_op *op,
+ struct xrow_update_field *field);
-typedef uint32_t
-(*xrow_update_op_store_f)(struct xrow_update_op *op,
- struct json_tree *format_tree,
- struct json_token *this_node, const char *in,
- char *out);
+typedef uint32_t (*xrow_update_op_store_f)(struct xrow_update_op *op,
+ struct json_tree *format_tree,
+ struct json_token *this_node,
+ const char *in, char *out);
/**
* A set of functions and properties to initialize, do and store
@@ -482,39 +479,31 @@ xrow_update_field_store(struct xrow_update_field *field,
* etc. Each complex type has basic operations of the same
* signature: insert, set, delete, arith, bit, splice.
*/
-#define OP_DECL_GENERIC(type) \
-int \
-xrow_update_op_do_##type##_insert(struct xrow_update_op *op, \
- struct xrow_update_field *field); \
- \
-int \
-xrow_update_op_do_##type##_set(struct xrow_update_op *op, \
- struct xrow_update_field *field); \
- \
-int \
-xrow_update_op_do_##type##_delete(struct xrow_update_op *op, \
- struct xrow_update_field *field); \
- \
-int \
-xrow_update_op_do_##type##_arith(struct xrow_update_op *op, \
- struct xrow_update_field *field); \
- \
-int \
-xrow_update_op_do_##type##_bit(struct xrow_update_op *op, \
- struct xrow_update_field *field); \
- \
-int \
-xrow_update_op_do_##type##_splice(struct xrow_update_op *op, \
- struct xrow_update_field *field); \
- \
-uint32_t \
-xrow_update_##type##_sizeof(struct xrow_update_field *field); \
- \
-uint32_t \
-xrow_update_##type##_store(struct xrow_update_field *field, \
- struct json_tree *format_tree, \
- struct json_token *this_node, char *out, \
- char *out_end);
+#define OP_DECL_GENERIC(type) \
+ int xrow_update_op_do_##type##_insert( \
+ struct xrow_update_op *op, struct xrow_update_field *field); \
+ \
+ int xrow_update_op_do_##type##_set(struct xrow_update_op *op, \
+ struct xrow_update_field *field); \
+ \
+ int xrow_update_op_do_##type##_delete( \
+ struct xrow_update_op *op, struct xrow_update_field *field); \
+ \
+ int xrow_update_op_do_##type##_arith(struct xrow_update_op *op, \
+ struct xrow_update_field *field); \
+ \
+ int xrow_update_op_do_##type##_bit(struct xrow_update_op *op, \
+ struct xrow_update_field *field); \
+ \
+ int xrow_update_op_do_##type##_splice( \
+ struct xrow_update_op *op, struct xrow_update_field *field); \
+ \
+ uint32_t xrow_update_##type##_sizeof(struct xrow_update_field *field); \
+ \
+ uint32_t xrow_update_##type##_store(struct xrow_update_field *field, \
+ struct json_tree *format_tree, \
+ struct json_token *this_node, \
+ char *out, char *out_end);
/* }}} xrow_update_field */
@@ -666,27 +655,26 @@ OP_DECL_GENERIC(route)
* fit ~10k update tree depth - incredible number, even though the
* real limit is 4k due to limited number of operations.
*/
-#define OP_DECL_GENERIC(op_type) \
-static inline int \
-xrow_update_op_do_field_##op_type(struct xrow_update_op *op, \
- struct xrow_update_field *field) \
-{ \
- switch (field->type) { \
- case XUPDATE_ARRAY: \
- return xrow_update_op_do_array_##op_type(op, field); \
- case XUPDATE_NOP: \
- return xrow_update_op_do_nop_##op_type(op, field); \
- case XUPDATE_BAR: \
- return xrow_update_op_do_bar_##op_type(op, field); \
- case XUPDATE_ROUTE: \
- return xrow_update_op_do_route_##op_type(op, field); \
- case XUPDATE_MAP: \
- return xrow_update_op_do_map_##op_type(op, field); \
- default: \
- unreachable(); \
- } \
- return 0; \
-}
+#define OP_DECL_GENERIC(op_type) \
+ static inline int xrow_update_op_do_field_##op_type( \
+ struct xrow_update_op *op, struct xrow_update_field *field) \
+ { \
+ switch (field->type) { \
+ case XUPDATE_ARRAY: \
+ return xrow_update_op_do_array_##op_type(op, field); \
+ case XUPDATE_NOP: \
+ return xrow_update_op_do_nop_##op_type(op, field); \
+ case XUPDATE_BAR: \
+ return xrow_update_op_do_bar_##op_type(op, field); \
+ case XUPDATE_ROUTE: \
+ return xrow_update_op_do_route_##op_type(op, field); \
+ case XUPDATE_MAP: \
+ return xrow_update_op_do_map_##op_type(op, field); \
+ default: \
+ unreachable(); \
+ } \
+ return 0; \
+ }
OP_DECL_GENERIC(insert)
@@ -758,15 +746,15 @@ xrow_update_err_double(const struct xrow_update_op *op)
static inline int
xrow_update_err_bad_json(const struct xrow_update_op *op, int pos)
{
- return xrow_update_err(op, tt_sprintf("invalid JSON in position %d",
- pos));
+ return xrow_update_err(op,
+ tt_sprintf("invalid JSON in position %d", pos));
}
static inline int
xrow_update_err_delete1(const struct xrow_update_op *op)
{
- return xrow_update_err(op, "can delete only 1 field from a map in a "\
- "row");
+ return xrow_update_err(op, "can delete only 1 field from a map in a "
+ "row");
}
static inline int
diff --git a/src/box/xrow_update_map.c b/src/box/xrow_update_map.c
index 57fb27f..4bd7c94 100644
--- a/src/box/xrow_update_map.c
+++ b/src/box/xrow_update_map.c
@@ -83,8 +83,7 @@ xrow_update_map_create_item(struct xrow_update_field *field,
item->key_len = key_len;
item->field.type = type;
item->field.data = data;
- item->field.size = data_size,
- item->tail_size = tail_size;
+ item->field.size = data_size, item->tail_size = tail_size;
/*
* Each time a new item it created it is stored in the
* head of update map item list. It helps in case the
@@ -123,8 +122,8 @@ xrow_update_map_extract_opt_item(struct xrow_update_field *field,
if (xrow_update_op_next_token(op) != 0)
return -1;
if (op->token_type != JSON_TOKEN_STR) {
- return xrow_update_err(op, "can't update a map by not "\
- "a string key");
+ return xrow_update_err(op, "can't update a map by not "
+ "a string key");
}
}
struct stailq *items = &field->map.items;
@@ -136,7 +135,8 @@ xrow_update_map_extract_opt_item(struct xrow_update_field *field,
* passing this key, so it should be here for all except
* first updates.
*/
- stailq_foreach_entry(i, items, in_items) {
+ stailq_foreach_entry(i, items, in_items)
+ {
if (i->key != NULL && i->key_len == op->key_len &&
memcmp(i->key, op->key, i->key_len) == 0) {
*res = i;
@@ -149,12 +149,13 @@ xrow_update_map_extract_opt_item(struct xrow_update_field *field,
*/
uint32_t key_len, i_tail_size;
const char *pos, *key, *end, *tmp, *begin;
- stailq_foreach_entry(i, items, in_items) {
+ stailq_foreach_entry(i, items, in_items)
+ {
begin = i->field.data + i->field.size;
pos = begin;
end = begin + i->tail_size;
i_tail_size = 0;
- while(pos < end) {
+ while (pos < end) {
if (mp_typeof(*pos) != MP_STR) {
mp_next(&pos);
mp_next(&pos);
@@ -309,28 +310,28 @@ xrow_update_op_do_map_delete(struct xrow_update_op *op,
return 0;
}
-#define DO_SCALAR_OP_GENERIC(op_type) \
-int \
-xrow_update_op_do_map_##op_type(struct xrow_update_op *op, \
- struct xrow_update_field *field) \
-{ \
- assert(field->type == XUPDATE_MAP); \
- struct xrow_update_map_item *item = \
- xrow_update_map_extract_item(field, op); \
- if (item == NULL) \
- return -1; \
- if (!xrow_update_op_is_term(op)) { \
- op->is_token_consumed = true; \
- return xrow_update_op_do_field_##op_type(op, &item->field); \
- } \
- if (item->field.type != XUPDATE_NOP) \
- return xrow_update_err_double(op); \
- if (xrow_update_op_do_##op_type(op, item->field.data) != 0) \
- return -1; \
- item->field.type = XUPDATE_SCALAR; \
- item->field.scalar.op = op; \
- return 0; \
-}
+#define DO_SCALAR_OP_GENERIC(op_type) \
+ int xrow_update_op_do_map_##op_type(struct xrow_update_op *op, \
+ struct xrow_update_field *field) \
+ { \
+ assert(field->type == XUPDATE_MAP); \
+ struct xrow_update_map_item *item = \
+ xrow_update_map_extract_item(field, op); \
+ if (item == NULL) \
+ return -1; \
+ if (!xrow_update_op_is_term(op)) { \
+ op->is_token_consumed = true; \
+ return xrow_update_op_do_field_##op_type( \
+ op, &item->field); \
+ } \
+ if (item->field.type != XUPDATE_NOP) \
+ return xrow_update_err_double(op); \
+ if (xrow_update_op_do_##op_type(op, item->field.data) != 0) \
+ return -1; \
+ item->field.type = XUPDATE_SCALAR; \
+ item->field.scalar.op = op; \
+ return 0; \
+ }
DO_SCALAR_OP_GENERIC(arith)
@@ -349,9 +350,8 @@ xrow_update_map_create(struct xrow_update_field *field, const char *header,
stailq_create(&field->map.items);
if (field_count == 0)
return 0;
- struct xrow_update_map_item *first =
- xrow_update_map_new_item(field, XUPDATE_NOP, NULL, 0, data, 0,
- data_end - data);
+ struct xrow_update_map_item *first = xrow_update_map_new_item(
+ field, XUPDATE_NOP, NULL, 0, data, 0, data_end - data);
return first != NULL ? 0 : -1;
}
@@ -418,7 +418,8 @@ xrow_update_map_sizeof(struct xrow_update_field *field)
assert(field->type == XUPDATE_MAP);
uint32_t res = mp_sizeof_map(field->map.size);
struct xrow_update_map_item *i;
- stailq_foreach_entry(i, &field->map.items, in_items) {
+ stailq_foreach_entry(i, &field->map.items, in_items)
+ {
res += i->tail_size;
if (i->key != NULL) {
res += mp_sizeof_str(i->key_len) +
@@ -442,25 +443,25 @@ xrow_update_map_store(struct xrow_update_field *field,
* others. The first cycle doesn't save unchanged tails.
*/
if (this_node == NULL) {
- stailq_foreach_entry(i, &field->map.items, in_items) {
+ stailq_foreach_entry(i, &field->map.items, in_items)
+ {
if (i->key != NULL) {
out = mp_encode_str(out, i->key, i->key_len);
- out += xrow_update_field_store(&i->field, NULL,
- NULL, out,
- out_end);
+ out += xrow_update_field_store(
+ &i->field, NULL, NULL, out, out_end);
}
}
} else {
struct json_token token;
token.type = JSON_TOKEN_STR;
struct json_token *next_node;
- stailq_foreach_entry(i, &field->map.items, in_items) {
+ stailq_foreach_entry(i, &field->map.items, in_items)
+ {
if (i->key != NULL) {
token.str = i->key;
token.len = i->key_len;
next_node = json_tree_lookup(format_tree,
- this_node,
- &token);
+ this_node, &token);
out = mp_encode_str(out, i->key, i->key_len);
out += xrow_update_field_store(&i->field,
format_tree,
@@ -469,7 +470,8 @@ xrow_update_map_store(struct xrow_update_field *field,
}
}
}
- stailq_foreach_entry(i, &field->map.items, in_items) {
+ stailq_foreach_entry(i, &field->map.items, in_items)
+ {
memcpy(out, i->field.data + i->field.size, i->tail_size);
out += i->tail_size;
}
diff --git a/src/box/xrow_update_route.c b/src/box/xrow_update_route.c
index 0352aec..ee23bbe 100644
--- a/src/box/xrow_update_route.c
+++ b/src/box/xrow_update_route.c
@@ -151,8 +151,8 @@ xrow_update_route_branch_map(struct xrow_update_field *next_hop,
mp_next(&end);
mp_next(&end);
}
- if (xrow_update_map_create(next_hop, parent, data, end,
- field_count) != 0)
+ if (xrow_update_map_create(next_hop, parent, data, end, field_count) !=
+ 0)
return -1;
return op->meta->do_op(op, next_hop);
}
@@ -209,7 +209,7 @@ xrow_update_route_branch(struct xrow_update_field *field,
}
if (json_token_cmp(&old_token, &new_token) != 0)
break;
- switch(new_token.type) {
+ switch (new_token.type) {
case JSON_TOKEN_NUM:
rc = tuple_field_go_to_index(&parent, new_token.num);
break;
@@ -281,8 +281,8 @@ xrow_update_route_branch(struct xrow_update_field *field,
if (type == MP_ARRAY) {
if (new_token.type != JSON_TOKEN_NUM) {
- xrow_update_err(new_op, "can not update array by "\
- "non-integer index");
+ xrow_update_err(new_op, "can not update array by "
+ "non-integer index");
return NULL;
}
new_op->is_token_consumed = false;
@@ -293,8 +293,8 @@ xrow_update_route_branch(struct xrow_update_field *field,
return NULL;
} else if (type == MP_MAP) {
if (new_token.type != JSON_TOKEN_STR) {
- xrow_update_err(new_op, "can not update map by "\
- "non-string key");
+ xrow_update_err(new_op, "can not update map by "
+ "non-string key");
return NULL;
}
new_op->is_token_consumed = false;
@@ -327,7 +327,8 @@ xrow_update_route_branch(struct xrow_update_field *field,
* the route is just followed, via a lexer offset increase.
*/
static struct xrow_update_field *
-xrow_update_route_next(struct xrow_update_field *field, struct xrow_update_op *op)
+xrow_update_route_next(struct xrow_update_field *field,
+ struct xrow_update_op *op)
{
assert(field->type == XUPDATE_ROUTE);
assert(!xrow_update_op_is_term(op));
@@ -346,17 +347,17 @@ xrow_update_route_next(struct xrow_update_field *field, struct xrow_update_op *o
return xrow_update_route_branch(field, op);
}
-#define DO_SCALAR_OP_GENERIC(op_type) \
-int \
-xrow_update_op_do_route_##op_type(struct xrow_update_op *op, \
- struct xrow_update_field *field) \
-{ \
- assert(field->type == XUPDATE_ROUTE); \
- struct xrow_update_field *next_hop = xrow_update_route_next(field, op); \
- if (next_hop == NULL) \
- return -1; \
- return xrow_update_op_do_field_##op_type(op, next_hop); \
-}
+#define DO_SCALAR_OP_GENERIC(op_type) \
+ int xrow_update_op_do_route_##op_type(struct xrow_update_op *op, \
+ struct xrow_update_field *field) \
+ { \
+ assert(field->type == XUPDATE_ROUTE); \
+ struct xrow_update_field *next_hop = \
+ xrow_update_route_next(field, op); \
+ if (next_hop == NULL) \
+ return -1; \
+ return xrow_update_op_do_field_##op_type(op, next_hop); \
+ }
DO_SCALAR_OP_GENERIC(set)
@@ -383,9 +384,9 @@ xrow_update_route_store(struct xrow_update_field *field,
struct json_token *this_node, char *out, char *out_end)
{
if (this_node != NULL) {
- this_node = json_tree_lookup_path(
- format_tree, this_node, field->route.path,
- field->route.path_len, 0);
+ this_node = json_tree_lookup_path(format_tree, this_node,
+ field->route.path,
+ field->route.path_len, 0);
}
char *saved_out = out;
int before_hop = field->route.next_hop->data - field->data;
--
1.8.3.1
prev parent reply other threads:[~2020-10-07 14:11 UTC|newest]
Thread overview: 4+ messages / expand[flat|nested] mbox.gz Atom feed top
2020-10-07 13:24 [Tarantool-patches] [PATCH 0/3] Add clang format Kirill Yukhin
2020-10-07 13:24 ` [Tarantool-patches] [PATCH 1/3] clang-format: guard various declarations Kirill Yukhin
2020-10-07 13:24 ` [Tarantool-patches] [PATCH 2/3] Add .clang-format for src/box/ Kirill Yukhin
2020-10-07 14:11 ` Kirill Yukhin [this message]
Reply instructions:
You may reply publicly to this message via plain-text email
using any one of the following methods:
* Save the following mbox file, import it into your mail client,
and reply-to-all from there: mbox
Avoid top-posting and favor interleaved quoting:
https://en.wikipedia.org/wiki/Posting_style#Interleaved_style
* Reply using the --to, --cc, and --in-reply-to
switches of git-send-email(1):
git send-email \
--in-reply-to=96ac986f4ecc27e2bdff835e9fe2fef5a5a89a7a.1602076356.git.kyukhin@tarantool.org \
--to=kyukhin@tarantool.org \
--cc=tarantool-patches@dev.tarantool.org \
--subject='Re: [Tarantool-patches] [PATCH 3/3] Apply clang-format' \
/path/to/YOUR_REPLY
https://kernel.org/pub/software/scm/git/docs/git-send-email.html
* If your mail client supports setting the In-Reply-To header
via mailto: links, try the mailto: link
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox