mirror of
https://github.com/netdata/netdata.git
synced 2025-04-05 22:15:32 +00:00

* rrdset - in progress * rrdset optimal constructor; rrdset conflict * rrdset final touches * re-organization of rrdset object members * prevent use-after-free * dictionary dfe supports also counting of iterations * rrddim managed by dictionary * rrd.h cleanup * DICTIONARY_ITEM now is referencing actual dictionary items in the code * removed rrdset linked list * Revert "removed rrdset linked list" This reverts commit 690d6a588b4b99619c2c5e10f84e8f868ae6def5. * removed rrdset linked list * added comments * Switch chart uuid to static allocation in rrdset Remove unused functions * rrdset_archive() and friends... * always create rrdfamily * enable ml_free_dimension * rrddim_foreach done with dfe * most custom rrddim loops replaced with rrddim_foreach * removed accesses to rrddim->dimensions * removed locks that are no longer needed * rrdsetvar is now managed by the dictionary * set rrdset is rrdsetvar, fixes https://github.com/netdata/netdata/pull/13646#issuecomment-1242574853 * conflict callback of rrdsetvar now properly checks if it has to reset the variable * dictionary registered callbacks accept as first parameter the DICTIONARY_ITEM * dictionary dfe now uses internal counter to report; avoided excess variables defined with dfe * dictionary walkthrough callbacks get dictionary acquired items * dictionary reference counters that can be dupped from zero * added advanced functions for get and del * rrdvar managed by dictionaries * thread safety for rrdsetvar * faster rrdvar initialization * rrdvar string lengths should match in all add, del, get functions * rrdvar internals hidden from the rest of the world * rrdvar is now acquired throughout netdata * hide the internal structures of rrdsetvar * rrdsetvar is now acquired through out netdata * rrddimvar managed by dictionary; rrddimvar linked list removed; rrddimvar structures hidden from the rest of netdata * better error handling * dont create variables if not initialized for health * dont create variables if not initialized for health again * rrdfamily is now managed by dictionaries; references of it are acquired dictionary items * type checking on acquired objects * rrdcalc renaming of functions * type checking for rrdfamily_acquired * rrdcalc managed by dictionaries * rrdcalc double free fix * host rrdvars is always needed * attempt to fix deadlock 1 * attempt to fix deadlock 2 * Remove unused variable * attempt to fix deadlock 3 * snprintfz * rrdcalc index in rrdset fix * Stop storing active charts and computing chart hashes * Remove store active chart function * Remove compute chart hash function * Remove sql_store_chart_hash function * Remove store_active_dimension function * dictionary delayed destruction * formatting and cleanup * zero dictionary base on rrdsetvar * added internal error to log delayed destructions of dictionaries * typo in rrddimvar * added debugging info to dictionary * debug info * fix for rrdcalc keys being empty * remove forgotten unlock * remove deadlock * Switch to metadata version 5 and drop chart_hash chart_hash_map chart_active dimension_active v_chart_hash * SQL cosmetic changes * do not busy wait while destroying a referenced dictionary * remove deadlock * code cleanup; re-organization; * fast cleanup and flushing of dictionaries * number formatting fixes * do not delete configured alerts when archiving a chart * rrddim obsolete linked list management outside dictionaries * removed duplicate contexts call * fix crash when rrdfamily is not initialized * dont keep rrddimvar referenced * properly cleanup rrdvar * removed some locks * Do not attempt to cleanup chart_hash / chart_hash_map * rrdcalctemplate managed by dictionary * register callbacks on the right dictionary * removed some more locks * rrdcalc secondary index replaced with linked-list; rrdcalc labels updates are now executed by health thread * when looking up for an alarm look using both chart id and chart name * host initialization a bit more modular * init rrdlabels on host update * preparation for dictionary views * improved comment * unused variables without internal checks * service threads isolation and worker info * more worker info in service thread * thread cancelability debugging with internal checks * strings data races addressed; fixes https://github.com/netdata/netdata/issues/13647 * dictionary modularization * Remove unused SQL statement definition * unit-tested thread safety of dictionaries; removed data race conditions on dictionaries and strings; dictionaries now can detect if the caller is holds a write lock and automatically all the calls become their unsafe versions; all direct calls to unsafe version is eliminated * remove worker_is_idle() from the exit of service functions, because we lose the lock time between loops * rewritten dictionary to have 2 separate locks, one for indexing and another for traversal * Update collectors/cgroups.plugin/sys_fs_cgroup.c Co-authored-by: Vladimir Kobal <vlad@prokk.net> * Update collectors/cgroups.plugin/sys_fs_cgroup.c Co-authored-by: Vladimir Kobal <vlad@prokk.net> * Update collectors/proc.plugin/proc_net_dev.c Co-authored-by: Vladimir Kobal <vlad@prokk.net> * fix memory leak in rrdset cache_dir * minor dictionary changes * dont use index locks in single threaded * obsolete dict option * rrddim options and flags separation; rrdset_done() optimization to keep array of reference pointers to rrddim; * fix jump on uninitialized value in dictionary; remove double free of cache_dir * addressed codacy findings * removed debugging code * use the private refcount on dictionaries * make dictionary item desctructors work on dictionary destruction; strictier control on dictionary API; proper cleanup sequence on rrddim; * more dictionary statistics * global statistics about dictionary operations, memory, items, callbacks * dictionary support for views - missing the public API * removed warning about unused parameter * chart and context name for cloud * chart and context name for cloud, again * dictionary statistics fixed; first implementation of dictionary views - not currently used * only the master can globally delete an item * context needs netdata prefix * fix context and chart it of spins * fix for host variables when health is not enabled * run garbage collector on item insert too * Fix info message; remove extra "using" * update dict unittest for new placement of garbage collector * we need RRDHOST->rrdvars for maintaining custom host variables * Health initialization needs the host->host_uuid * split STRING to its own files; no code changes other than that * initialize health unconditionally * unit tests do not pollute the global scope with their variables * Skip initialization when creating archived hosts on startup. When a child connects it will initialize properly Co-authored-by: Stelios Fragkakis <52996999+stelfrag@users.noreply.github.com> Co-authored-by: Vladimir Kobal <vlad@prokk.net>
1091 lines
41 KiB
C
Executable file
1091 lines
41 KiB
C
Executable file
// SPDX-License-Identifier: GPL-3.0-or-later
|
|
#include "rrdengine.h"
|
|
|
|
/* Default global database instance */
|
|
struct rrdengine_instance multidb_ctx_storage_tier0;
|
|
struct rrdengine_instance multidb_ctx_storage_tier1;
|
|
struct rrdengine_instance multidb_ctx_storage_tier2;
|
|
struct rrdengine_instance multidb_ctx_storage_tier3;
|
|
struct rrdengine_instance multidb_ctx_storage_tier4;
|
|
#if RRD_STORAGE_TIERS != 5
|
|
#error RRD_STORAGE_TIERS is not 5 - you need to add allocations here
|
|
#endif
|
|
struct rrdengine_instance *multidb_ctx[RRD_STORAGE_TIERS];
|
|
uint8_t tier_page_type[RRD_STORAGE_TIERS] = {PAGE_METRICS, PAGE_TIER, PAGE_TIER, PAGE_TIER, PAGE_TIER};
|
|
|
|
#if PAGE_TYPE_MAX != 1
|
|
#error PAGE_TYPE_MAX is not 1 - you need to add allocations here
|
|
#endif
|
|
size_t page_type_size[256] = {sizeof(storage_number), sizeof(storage_number_tier1_t)};
|
|
|
|
__attribute__((constructor)) void initialize_multidb_ctx(void) {
|
|
multidb_ctx[0] = &multidb_ctx_storage_tier0;
|
|
multidb_ctx[1] = &multidb_ctx_storage_tier1;
|
|
multidb_ctx[2] = &multidb_ctx_storage_tier2;
|
|
multidb_ctx[3] = &multidb_ctx_storage_tier3;
|
|
multidb_ctx[4] = &multidb_ctx_storage_tier4;
|
|
}
|
|
|
|
int db_engine_use_malloc = 0;
|
|
int default_rrdeng_page_fetch_timeout = 3;
|
|
int default_rrdeng_page_fetch_retries = 3;
|
|
int default_rrdeng_page_cache_mb = 32;
|
|
int default_rrdeng_disk_quota_mb = 256;
|
|
int default_multidb_disk_quota_mb = 256;
|
|
/* Default behaviour is to unblock data collection if the page cache is full of dirty pages by dropping metrics */
|
|
uint8_t rrdeng_drop_metrics_under_page_cache_pressure = 1;
|
|
|
|
static inline struct rrdengine_instance *get_rrdeng_ctx_from_host(RRDHOST *host, int tier) {
|
|
if(tier < 0 || tier >= RRD_STORAGE_TIERS) tier = 0;
|
|
if(!host->storage_instance[tier]) tier = 0;
|
|
return (struct rrdengine_instance *)host->storage_instance[tier];
|
|
}
|
|
|
|
/* This UUID is not unique across hosts */
|
|
void rrdeng_generate_legacy_uuid(const char *dim_id, char *chart_id, uuid_t *ret_uuid)
|
|
{
|
|
EVP_MD_CTX *evpctx;
|
|
unsigned char hash_value[EVP_MAX_MD_SIZE];
|
|
unsigned int hash_len;
|
|
|
|
evpctx = EVP_MD_CTX_create();
|
|
EVP_DigestInit_ex(evpctx, EVP_sha256(), NULL);
|
|
EVP_DigestUpdate(evpctx, dim_id, strlen(dim_id));
|
|
EVP_DigestUpdate(evpctx, chart_id, strlen(chart_id));
|
|
EVP_DigestFinal_ex(evpctx, hash_value, &hash_len);
|
|
EVP_MD_CTX_destroy(evpctx);
|
|
fatal_assert(hash_len > sizeof(uuid_t));
|
|
memcpy(ret_uuid, hash_value, sizeof(uuid_t));
|
|
}
|
|
|
|
/* Transform legacy UUID to be unique across hosts deterministically */
|
|
void rrdeng_convert_legacy_uuid_to_multihost(char machine_guid[GUID_LEN + 1], uuid_t *legacy_uuid, uuid_t *ret_uuid)
|
|
{
|
|
EVP_MD_CTX *evpctx;
|
|
unsigned char hash_value[EVP_MAX_MD_SIZE];
|
|
unsigned int hash_len;
|
|
|
|
evpctx = EVP_MD_CTX_create();
|
|
EVP_DigestInit_ex(evpctx, EVP_sha256(), NULL);
|
|
EVP_DigestUpdate(evpctx, machine_guid, GUID_LEN);
|
|
EVP_DigestUpdate(evpctx, *legacy_uuid, sizeof(uuid_t));
|
|
EVP_DigestFinal_ex(evpctx, hash_value, &hash_len);
|
|
EVP_MD_CTX_destroy(evpctx);
|
|
fatal_assert(hash_len > sizeof(uuid_t));
|
|
memcpy(ret_uuid, hash_value, sizeof(uuid_t));
|
|
}
|
|
|
|
struct rrdeng_metric_handle {
|
|
RRDDIM *rd;
|
|
struct rrdengine_instance *ctx;
|
|
uuid_t *rrdeng_uuid; // database engine metric UUID
|
|
struct pg_cache_page_index *page_index;
|
|
};
|
|
|
|
void rrdeng_metric_free(STORAGE_METRIC_HANDLE *db_metric_handle) {
|
|
freez(db_metric_handle);
|
|
}
|
|
|
|
STORAGE_METRIC_HANDLE *rrdeng_metric_init(RRDDIM *rd, STORAGE_INSTANCE *db_instance) {
|
|
struct rrdengine_instance *ctx = (struct rrdengine_instance *)db_instance;
|
|
struct page_cache *pg_cache;
|
|
uuid_t legacy_uuid;
|
|
uuid_t multihost_legacy_uuid;
|
|
Pvoid_t *PValue;
|
|
struct pg_cache_page_index *page_index = NULL;
|
|
int is_multihost_child = 0;
|
|
RRDHOST *host = rd->rrdset->rrdhost;
|
|
|
|
pg_cache = &ctx->pg_cache;
|
|
|
|
rrdeng_generate_legacy_uuid(rrddim_id(rd), (char *)rrdset_id(rd->rrdset), &legacy_uuid);
|
|
if (host != localhost && is_storage_engine_shared((STORAGE_INSTANCE *)ctx))
|
|
is_multihost_child = 1;
|
|
|
|
uv_rwlock_rdlock(&pg_cache->metrics_index.lock);
|
|
PValue = JudyHSGet(pg_cache->metrics_index.JudyHS_array, &legacy_uuid, sizeof(uuid_t));
|
|
if (likely(NULL != PValue)) {
|
|
page_index = *PValue;
|
|
}
|
|
uv_rwlock_rdunlock(&pg_cache->metrics_index.lock);
|
|
if (is_multihost_child || NULL == PValue) {
|
|
/* First time we see the legacy UUID or metric belongs to child host in multi-host DB.
|
|
* Drop legacy support, normal path */
|
|
|
|
uv_rwlock_rdlock(&pg_cache->metrics_index.lock);
|
|
PValue = JudyHSGet(pg_cache->metrics_index.JudyHS_array, &rd->metric_uuid, sizeof(uuid_t));
|
|
if (likely(NULL != PValue)) {
|
|
page_index = *PValue;
|
|
}
|
|
uv_rwlock_rdunlock(&pg_cache->metrics_index.lock);
|
|
if (NULL == PValue) {
|
|
uv_rwlock_wrlock(&pg_cache->metrics_index.lock);
|
|
PValue = JudyHSIns(&pg_cache->metrics_index.JudyHS_array, &rd->metric_uuid, sizeof(uuid_t), PJE0);
|
|
fatal_assert(NULL == *PValue); /* TODO: figure out concurrency model */
|
|
*PValue = page_index = create_page_index(&rd->metric_uuid);
|
|
page_index->prev = pg_cache->metrics_index.last_page_index;
|
|
pg_cache->metrics_index.last_page_index = page_index;
|
|
uv_rwlock_wrunlock(&pg_cache->metrics_index.lock);
|
|
}
|
|
} else {
|
|
/* There are legacy UUIDs in the database, implement backward compatibility */
|
|
|
|
rrdeng_convert_legacy_uuid_to_multihost(rd->rrdset->rrdhost->machine_guid, &legacy_uuid,
|
|
&multihost_legacy_uuid);
|
|
|
|
int need_to_store = uuid_compare(rd->metric_uuid, multihost_legacy_uuid);
|
|
|
|
uuid_copy(rd->metric_uuid, multihost_legacy_uuid);
|
|
|
|
if (unlikely(need_to_store && !ctx->tier))
|
|
(void)sql_store_dimension(&rd->metric_uuid, &rd->rrdset->chart_uuid, rrddim_id(rd), rrddim_name(rd), rd->multiplier, rd->divisor, rd->algorithm);
|
|
}
|
|
|
|
struct rrdeng_metric_handle *mh = mallocz(sizeof(struct rrdeng_metric_handle));
|
|
mh->rd = rd;
|
|
mh->ctx = ctx;
|
|
mh->rrdeng_uuid = &page_index->id;
|
|
mh->page_index = page_index;
|
|
return (STORAGE_METRIC_HANDLE *)mh;
|
|
}
|
|
|
|
/*
|
|
* Gets a handle for storing metrics to the database.
|
|
* The handle must be released with rrdeng_store_metric_final().
|
|
*/
|
|
STORAGE_COLLECT_HANDLE *rrdeng_store_metric_init(STORAGE_METRIC_HANDLE *db_metric_handle) {
|
|
struct rrdeng_metric_handle *metric_handle = (struct rrdeng_metric_handle *)db_metric_handle;
|
|
|
|
struct rrdeng_collect_handle *handle;
|
|
struct pg_cache_page_index *page_index;
|
|
|
|
handle = callocz(1, sizeof(struct rrdeng_collect_handle));
|
|
handle->metric_handle = metric_handle;
|
|
handle->ctx = metric_handle->ctx;
|
|
handle->descr = NULL;
|
|
handle->unaligned_page = 0;
|
|
|
|
page_index = metric_handle->page_index;
|
|
uv_rwlock_wrlock(&page_index->lock);
|
|
++page_index->writers;
|
|
uv_rwlock_wrunlock(&page_index->lock);
|
|
|
|
return (STORAGE_COLLECT_HANDLE *)handle;
|
|
}
|
|
|
|
/* The page must be populated and referenced */
|
|
static int page_has_only_empty_metrics(struct rrdeng_page_descr *descr)
|
|
{
|
|
switch(descr->type) {
|
|
case PAGE_METRICS: {
|
|
size_t slots = descr->page_length / PAGE_POINT_SIZE_BYTES(descr);
|
|
storage_number *array = (storage_number *)descr->pg_cache_descr->page;
|
|
for (size_t i = 0 ; i < slots; ++i) {
|
|
if(does_storage_number_exist(array[i]))
|
|
return 0;
|
|
}
|
|
}
|
|
break;
|
|
|
|
case PAGE_TIER: {
|
|
size_t slots = descr->page_length / PAGE_POINT_SIZE_BYTES(descr);
|
|
storage_number_tier1_t *array = (storage_number_tier1_t *)descr->pg_cache_descr->page;
|
|
for (size_t i = 0 ; i < slots; ++i) {
|
|
if(fpclassify(array[i].sum_value) != FP_NAN)
|
|
return 0;
|
|
}
|
|
}
|
|
break;
|
|
|
|
default: {
|
|
static bool logged = false;
|
|
if(!logged) {
|
|
error("DBENGINE: cannot check page for nulls on unknown page type id %d", descr->type);
|
|
logged = true;
|
|
}
|
|
return 0;
|
|
}
|
|
}
|
|
|
|
return 1;
|
|
}
|
|
|
|
void rrdeng_store_metric_flush_current_page(STORAGE_COLLECT_HANDLE *collection_handle) {
|
|
struct rrdeng_collect_handle *handle = (struct rrdeng_collect_handle *)collection_handle;
|
|
// struct rrdeng_metric_handle *metric_handle = (struct rrdeng_metric_handle *)handle->metric_handle;
|
|
struct rrdengine_instance *ctx = handle->ctx;
|
|
struct rrdeng_page_descr *descr = handle->descr;
|
|
|
|
if (unlikely(!ctx)) return;
|
|
if (unlikely(!descr)) return;
|
|
|
|
if (likely(descr->page_length)) {
|
|
int page_is_empty;
|
|
|
|
rrd_stat_atomic_add(&ctx->stats.metric_API_producers, -1);
|
|
|
|
page_is_empty = page_has_only_empty_metrics(descr);
|
|
if (page_is_empty) {
|
|
debug(D_RRDENGINE, "Page has empty metrics only, deleting:");
|
|
if (unlikely(debug_flags & D_RRDENGINE))
|
|
print_page_cache_descr(descr);
|
|
pg_cache_put(ctx, descr);
|
|
pg_cache_punch_hole(ctx, descr, 1, 0, NULL);
|
|
} else
|
|
rrdeng_commit_page(ctx, descr, handle->page_correlation_id);
|
|
} else {
|
|
dbengine_page_free(descr->pg_cache_descr->page);
|
|
rrdeng_destroy_pg_cache_descr(ctx, descr->pg_cache_descr);
|
|
rrdeng_page_descr_freez(descr);
|
|
}
|
|
handle->descr = NULL;
|
|
}
|
|
|
|
void rrdeng_store_metric_next(STORAGE_COLLECT_HANDLE *collection_handle,
|
|
usec_t point_in_time,
|
|
NETDATA_DOUBLE n,
|
|
NETDATA_DOUBLE min_value,
|
|
NETDATA_DOUBLE max_value,
|
|
uint16_t count,
|
|
uint16_t anomaly_count,
|
|
SN_FLAGS flags)
|
|
{
|
|
struct rrdeng_collect_handle *handle = (struct rrdeng_collect_handle *)collection_handle;
|
|
struct rrdeng_metric_handle *metric_handle = (struct rrdeng_metric_handle *)handle->metric_handle;
|
|
struct rrdengine_instance *ctx = handle->ctx;
|
|
struct page_cache *pg_cache = &ctx->pg_cache;
|
|
struct rrdeng_page_descr *descr = handle->descr;
|
|
RRDDIM *rd = metric_handle->rd;
|
|
|
|
void *page;
|
|
uint8_t must_flush_unaligned_page = 0, perfect_page_alignment = 0;
|
|
|
|
if (descr) {
|
|
/* Make alignment decisions */
|
|
|
|
if (descr->page_length == rd->rrdset->rrddim_page_alignment) {
|
|
/* this is the leading dimension that defines chart alignment */
|
|
perfect_page_alignment = 1;
|
|
}
|
|
/* is the metric far enough out of alignment with the others? */
|
|
if (unlikely(descr->page_length + PAGE_POINT_SIZE_BYTES(descr) < rd->rrdset->rrddim_page_alignment)) {
|
|
handle->unaligned_page = 1;
|
|
debug(D_RRDENGINE, "Metric page is not aligned with chart:");
|
|
if (unlikely(debug_flags & D_RRDENGINE))
|
|
print_page_cache_descr(descr);
|
|
}
|
|
if (unlikely(handle->unaligned_page &&
|
|
/* did the other metrics change page? */
|
|
rd->rrdset->rrddim_page_alignment <= PAGE_POINT_SIZE_BYTES(descr))) {
|
|
debug(D_RRDENGINE, "Flushing unaligned metric page.");
|
|
must_flush_unaligned_page = 1;
|
|
handle->unaligned_page = 0;
|
|
}
|
|
}
|
|
if (unlikely(NULL == descr ||
|
|
descr->page_length + PAGE_POINT_SIZE_BYTES(descr) > RRDENG_BLOCK_SIZE ||
|
|
must_flush_unaligned_page)) {
|
|
rrdeng_store_metric_flush_current_page(collection_handle);
|
|
|
|
page = rrdeng_create_page(ctx, &metric_handle->page_index->id, &descr);
|
|
fatal_assert(page);
|
|
|
|
handle->descr = descr;
|
|
|
|
handle->page_correlation_id = rrd_atomic_fetch_add(&pg_cache->committed_page_index.latest_corr_id, 1);
|
|
|
|
if (0 == rd->rrdset->rrddim_page_alignment) {
|
|
/* this is the leading dimension that defines chart alignment */
|
|
perfect_page_alignment = 1;
|
|
}
|
|
}
|
|
|
|
page = descr->pg_cache_descr->page;
|
|
|
|
switch (descr->type) {
|
|
case PAGE_METRICS: {
|
|
((storage_number *)page)[descr->page_length / PAGE_POINT_SIZE_BYTES(descr)] = pack_storage_number(n, flags);
|
|
}
|
|
break;
|
|
|
|
case PAGE_TIER: {
|
|
storage_number_tier1_t number_tier1;
|
|
number_tier1.sum_value = (float)n;
|
|
number_tier1.min_value = (float)min_value;
|
|
number_tier1.max_value = (float)max_value;
|
|
number_tier1.anomaly_count = anomaly_count;
|
|
number_tier1.count = count;
|
|
((storage_number_tier1_t *)page)[descr->page_length / PAGE_POINT_SIZE_BYTES(descr)] = number_tier1;
|
|
}
|
|
break;
|
|
|
|
default: {
|
|
static bool logged = false;
|
|
if(!logged) {
|
|
error("DBENGINE: cannot store metric on unknown page type id %d", descr->type);
|
|
logged = true;
|
|
}
|
|
}
|
|
break;
|
|
}
|
|
|
|
pg_cache_atomic_set_pg_info(descr, point_in_time, descr->page_length + PAGE_POINT_SIZE_BYTES(descr));
|
|
|
|
if (perfect_page_alignment)
|
|
rd->rrdset->rrddim_page_alignment = descr->page_length;
|
|
if (unlikely(INVALID_TIME == descr->start_time)) {
|
|
unsigned long new_metric_API_producers, old_metric_API_max_producers, ret_metric_API_max_producers;
|
|
descr->start_time = point_in_time;
|
|
|
|
new_metric_API_producers = rrd_atomic_add_fetch(&ctx->stats.metric_API_producers, 1);
|
|
while (unlikely(new_metric_API_producers > (old_metric_API_max_producers = ctx->metric_API_max_producers))) {
|
|
/* Increase ctx->metric_API_max_producers */
|
|
ret_metric_API_max_producers = ulong_compare_and_swap(&ctx->metric_API_max_producers,
|
|
old_metric_API_max_producers,
|
|
new_metric_API_producers);
|
|
if (old_metric_API_max_producers == ret_metric_API_max_producers) {
|
|
/* success */
|
|
break;
|
|
}
|
|
}
|
|
|
|
pg_cache_insert(ctx, metric_handle->page_index, descr);
|
|
} else {
|
|
pg_cache_add_new_metric_time(metric_handle->page_index, descr);
|
|
}
|
|
}
|
|
|
|
/*
|
|
* Releases the database reference from the handle for storing metrics.
|
|
* Returns 1 if it's safe to delete the dimension.
|
|
*/
|
|
int rrdeng_store_metric_finalize(STORAGE_COLLECT_HANDLE *collection_handle) {
|
|
struct rrdeng_collect_handle *handle = (struct rrdeng_collect_handle *)collection_handle;
|
|
struct rrdeng_metric_handle *metric_handle = (struct rrdeng_metric_handle *)handle->metric_handle;
|
|
struct pg_cache_page_index *page_index = metric_handle->page_index;
|
|
|
|
uint8_t can_delete_metric = 0;
|
|
|
|
rrdeng_store_metric_flush_current_page(collection_handle);
|
|
uv_rwlock_wrlock(&page_index->lock);
|
|
if (!--page_index->writers && !page_index->page_count) {
|
|
can_delete_metric = 1;
|
|
}
|
|
uv_rwlock_wrunlock(&page_index->lock);
|
|
freez(handle);
|
|
|
|
return can_delete_metric;
|
|
}
|
|
|
|
//static inline uint32_t *pginfo_to_dt(struct rrdeng_page_info *page_info)
|
|
//{
|
|
// return (uint32_t *)&page_info->scratch[0];
|
|
//}
|
|
//
|
|
//static inline uint32_t *pginfo_to_points(struct rrdeng_page_info *page_info)
|
|
//{
|
|
// return (uint32_t *)&page_info->scratch[sizeof(uint32_t)];
|
|
//}
|
|
//
|
|
/*
|
|
* Gets a handle for loading metrics from the database.
|
|
* The handle must be released with rrdeng_load_metric_final().
|
|
*/
|
|
void rrdeng_load_metric_init(STORAGE_METRIC_HANDLE *db_metric_handle, struct rrddim_query_handle *rrdimm_handle, time_t start_time, time_t end_time, TIER_QUERY_FETCH tier_query_fetch_type)
|
|
{
|
|
struct rrdeng_metric_handle *metric_handle = (struct rrdeng_metric_handle *)db_metric_handle;
|
|
struct rrdengine_instance *ctx = metric_handle->ctx;
|
|
RRDDIM *rd = metric_handle->rd;
|
|
|
|
// fprintf(stderr, "%s: %s/%s start time %ld, end time %ld\n", __FUNCTION__ , rd->rrdset->name, rd->name, start_time, end_time);
|
|
|
|
struct rrdeng_query_handle *handle;
|
|
unsigned pages_nr;
|
|
|
|
rrdimm_handle->start_time = start_time;
|
|
rrdimm_handle->end_time = end_time;
|
|
|
|
handle = callocz(1, sizeof(struct rrdeng_query_handle));
|
|
handle->next_page_time = start_time;
|
|
handle->now = start_time;
|
|
handle->tier_query_fetch_type = tier_query_fetch_type;
|
|
// TODO we should store the dt of each page in each page
|
|
// this will produce wrong values for dt in case the user changes
|
|
// the update every of the charts or the tier grouping iterations
|
|
handle->dt_sec = get_tier_grouping(ctx->tier) * (time_t)rd->update_every;
|
|
handle->dt = handle->dt_sec * USEC_PER_SEC;
|
|
handle->position = 0;
|
|
handle->ctx = ctx;
|
|
handle->metric_handle = metric_handle;
|
|
handle->descr = NULL;
|
|
rrdimm_handle->handle = (STORAGE_QUERY_HANDLE *)handle;
|
|
pages_nr = pg_cache_preload(ctx, metric_handle->rrdeng_uuid, start_time * USEC_PER_SEC, end_time * USEC_PER_SEC,
|
|
NULL, &handle->page_index);
|
|
if (unlikely(NULL == handle->page_index || 0 == pages_nr))
|
|
// there are no metrics to load
|
|
handle->next_page_time = INVALID_TIME;
|
|
}
|
|
|
|
static int rrdeng_load_page_next(struct rrddim_query_handle *rrdimm_handle) {
|
|
struct rrdeng_query_handle *handle = (struct rrdeng_query_handle *)rrdimm_handle->handle;
|
|
|
|
struct rrdengine_instance *ctx = handle->ctx;
|
|
struct rrdeng_page_descr *descr = handle->descr;
|
|
|
|
uint32_t page_length;
|
|
usec_t page_end_time;
|
|
unsigned position;
|
|
|
|
if (likely(descr)) {
|
|
// Drop old page's reference
|
|
|
|
#ifdef NETDATA_INTERNAL_CHECKS
|
|
rrd_stat_atomic_add(&ctx->stats.metric_API_consumers, -1);
|
|
#endif
|
|
|
|
pg_cache_put(ctx, descr);
|
|
handle->descr = NULL;
|
|
handle->next_page_time = (handle->page_end_time / USEC_PER_SEC) + 1;
|
|
|
|
if (unlikely(handle->next_page_time > rrdimm_handle->end_time))
|
|
return 1;
|
|
}
|
|
|
|
usec_t next_page_time = handle->next_page_time * USEC_PER_SEC;
|
|
descr = pg_cache_lookup_next(ctx, handle->page_index, &handle->page_index->id, next_page_time, rrdimm_handle->end_time * USEC_PER_SEC);
|
|
if (NULL == descr)
|
|
return 1;
|
|
|
|
#ifdef NETDATA_INTERNAL_CHECKS
|
|
rrd_stat_atomic_add(&ctx->stats.metric_API_consumers, 1);
|
|
#endif
|
|
|
|
handle->descr = descr;
|
|
pg_cache_atomic_get_pg_info(descr, &page_end_time, &page_length);
|
|
if (unlikely(INVALID_TIME == descr->start_time || INVALID_TIME == page_end_time))
|
|
return 1;
|
|
|
|
if (unlikely(descr->start_time != page_end_time && next_page_time > descr->start_time)) {
|
|
// we're in the middle of the page somewhere
|
|
unsigned entries = page_length / PAGE_POINT_SIZE_BYTES(descr);
|
|
position = ((uint64_t)(next_page_time - descr->start_time)) * (entries - 1) /
|
|
(page_end_time - descr->start_time);
|
|
}
|
|
else
|
|
position = 0;
|
|
|
|
handle->page_end_time = page_end_time;
|
|
handle->page_length = page_length;
|
|
handle->page = descr->pg_cache_descr->page;
|
|
usec_t entries = handle->entries = page_length / PAGE_POINT_SIZE_BYTES(descr);
|
|
if (likely(entries > 1))
|
|
handle->dt = (page_end_time - descr->start_time) / (entries - 1);
|
|
else {
|
|
// TODO we should store the dt of each page in each page
|
|
// now we keep the dt of whatever was before
|
|
;
|
|
}
|
|
|
|
handle->dt_sec = (time_t)(handle->dt / USEC_PER_SEC);
|
|
handle->position = position;
|
|
|
|
return 0;
|
|
}
|
|
|
|
// Returns the metric and sets its timestamp into current_time
|
|
// IT IS REQUIRED TO **ALWAYS** SET ALL RETURN VALUES (current_time, end_time, flags)
|
|
// IT IS REQUIRED TO **ALWAYS** KEEP TRACK OF TIME, EVEN OUTSIDE THE DATABASE BOUNDARIES
|
|
STORAGE_POINT rrdeng_load_metric_next(struct rrddim_query_handle *rrdimm_handle) {
|
|
struct rrdeng_query_handle *handle = (struct rrdeng_query_handle *)rrdimm_handle->handle;
|
|
// struct rrdeng_metric_handle *metric_handle = handle->metric_handle;
|
|
|
|
STORAGE_POINT sp;
|
|
struct rrdeng_page_descr *descr = handle->descr;
|
|
unsigned position = handle->position + 1;
|
|
time_t now = handle->now + handle->dt_sec;
|
|
storage_number_tier1_t tier1_value;
|
|
|
|
if (unlikely(INVALID_TIME == handle->next_page_time)) {
|
|
handle->next_page_time = INVALID_TIME;
|
|
handle->now = now;
|
|
storage_point_empty(sp, now - handle->dt_sec, now);
|
|
return sp;
|
|
}
|
|
|
|
if (unlikely(!descr || position >= handle->entries)) {
|
|
// We need to get a new page
|
|
if(rrdeng_load_page_next(rrdimm_handle)) {
|
|
// next calls will not load any more metrics
|
|
handle->next_page_time = INVALID_TIME;
|
|
handle->now = now;
|
|
storage_point_empty(sp, now - handle->dt_sec, now);
|
|
return sp;
|
|
}
|
|
|
|
descr = handle->descr;
|
|
position = handle->position;
|
|
now = (time_t)((descr->start_time + position * handle->dt) / USEC_PER_SEC);
|
|
}
|
|
|
|
sp.start_time = now - handle->dt_sec;
|
|
sp.end_time = now;
|
|
|
|
handle->position = position;
|
|
handle->now = now;
|
|
|
|
switch(descr->type) {
|
|
case PAGE_METRICS: {
|
|
storage_number n = handle->page[position];
|
|
sp.min = sp.max = sp.sum = unpack_storage_number(n);
|
|
sp.flags = n & SN_USER_FLAGS;
|
|
sp.count = 1;
|
|
sp.anomaly_count = is_storage_number_anomalous(n) ? 1 : 0;
|
|
}
|
|
break;
|
|
|
|
case PAGE_TIER: {
|
|
tier1_value = ((storage_number_tier1_t *)handle->page)[position];
|
|
sp.flags = tier1_value.anomaly_count ? SN_FLAG_NONE : SN_FLAG_NOT_ANOMALOUS;
|
|
sp.count = tier1_value.count;
|
|
sp.anomaly_count = tier1_value.anomaly_count;
|
|
sp.min = tier1_value.min_value;
|
|
sp.max = tier1_value.max_value;
|
|
sp.sum = tier1_value.sum_value;
|
|
}
|
|
break;
|
|
|
|
// we don't know this page type
|
|
default: {
|
|
static bool logged = false;
|
|
if(!logged) {
|
|
error("DBENGINE: unknown page type %d found. Cannot decode it. Ignoring its metrics.", descr->type);
|
|
logged = true;
|
|
}
|
|
storage_point_empty(sp, sp.start_time, sp.end_time);
|
|
}
|
|
break;
|
|
}
|
|
|
|
if (unlikely(now >= rrdimm_handle->end_time)) {
|
|
// next calls will not load any more metrics
|
|
handle->next_page_time = INVALID_TIME;
|
|
}
|
|
|
|
return sp;
|
|
}
|
|
|
|
int rrdeng_load_metric_is_finished(struct rrddim_query_handle *rrdimm_handle)
|
|
{
|
|
struct rrdeng_query_handle *handle = (struct rrdeng_query_handle *)rrdimm_handle->handle;
|
|
return (INVALID_TIME == handle->next_page_time);
|
|
}
|
|
|
|
/*
|
|
* Releases the database reference from the handle for loading metrics.
|
|
*/
|
|
void rrdeng_load_metric_finalize(struct rrddim_query_handle *rrdimm_handle)
|
|
{
|
|
struct rrdeng_query_handle *handle = (struct rrdeng_query_handle *)rrdimm_handle->handle;
|
|
struct rrdengine_instance *ctx = handle->ctx;
|
|
struct rrdeng_page_descr *descr = handle->descr;
|
|
|
|
if (descr) {
|
|
#ifdef NETDATA_INTERNAL_CHECKS
|
|
rrd_stat_atomic_add(&ctx->stats.metric_API_consumers, -1);
|
|
#endif
|
|
pg_cache_put(ctx, descr);
|
|
}
|
|
|
|
// whatever is allocated at rrdeng_load_metric_init() should be freed here
|
|
freez(handle);
|
|
rrdimm_handle->handle = NULL;
|
|
}
|
|
|
|
time_t rrdeng_metric_latest_time(STORAGE_METRIC_HANDLE *db_metric_handle) {
|
|
struct rrdeng_metric_handle *metric_handle = (struct rrdeng_metric_handle *)db_metric_handle;
|
|
|
|
struct pg_cache_page_index *page_index = metric_handle->page_index;
|
|
return page_index->latest_time / USEC_PER_SEC;
|
|
}
|
|
time_t rrdeng_metric_oldest_time(STORAGE_METRIC_HANDLE *db_metric_handle) {
|
|
struct rrdeng_metric_handle *metric_handle = (struct rrdeng_metric_handle *)db_metric_handle;
|
|
|
|
struct pg_cache_page_index *page_index = metric_handle->page_index;
|
|
return page_index->oldest_time / USEC_PER_SEC;
|
|
}
|
|
|
|
int rrdeng_metric_latest_time_by_uuid(uuid_t *dim_uuid, time_t *first_entry_t, time_t *last_entry_t, int tier)
|
|
{
|
|
struct page_cache *pg_cache;
|
|
struct rrdengine_instance *ctx;
|
|
Pvoid_t *PValue;
|
|
struct pg_cache_page_index *page_index = NULL;
|
|
|
|
ctx = get_rrdeng_ctx_from_host(localhost, tier);
|
|
if (unlikely(!ctx)) {
|
|
error("Failed to fetch multidb context");
|
|
return 1;
|
|
}
|
|
pg_cache = &ctx->pg_cache;
|
|
|
|
uv_rwlock_rdlock(&pg_cache->metrics_index.lock);
|
|
PValue = JudyHSGet(pg_cache->metrics_index.JudyHS_array, dim_uuid, sizeof(uuid_t));
|
|
if (likely(NULL != PValue)) {
|
|
page_index = *PValue;
|
|
}
|
|
uv_rwlock_rdunlock(&pg_cache->metrics_index.lock);
|
|
|
|
if (likely(page_index)) {
|
|
*first_entry_t = page_index->oldest_time / USEC_PER_SEC;
|
|
*last_entry_t = page_index->latest_time / USEC_PER_SEC;
|
|
return 0;
|
|
}
|
|
|
|
return 1;
|
|
}
|
|
|
|
int rrdeng_metric_retention_by_uuid(STORAGE_INSTANCE *si, uuid_t *dim_uuid, time_t *first_entry_t, time_t *last_entry_t)
|
|
{
|
|
struct page_cache *pg_cache;
|
|
struct rrdengine_instance *ctx;
|
|
Pvoid_t *PValue;
|
|
struct pg_cache_page_index *page_index = NULL;
|
|
|
|
ctx = (struct rrdengine_instance *)si;
|
|
if (unlikely(!ctx)) {
|
|
error("DBENGINE: invalid STORAGE INSTANCE to %s()", __FUNCTION__);
|
|
return 1;
|
|
}
|
|
pg_cache = &ctx->pg_cache;
|
|
|
|
uv_rwlock_rdlock(&pg_cache->metrics_index.lock);
|
|
PValue = JudyHSGet(pg_cache->metrics_index.JudyHS_array, dim_uuid, sizeof(uuid_t));
|
|
if (likely(NULL != PValue)) {
|
|
page_index = *PValue;
|
|
}
|
|
uv_rwlock_rdunlock(&pg_cache->metrics_index.lock);
|
|
|
|
if (likely(page_index)) {
|
|
*first_entry_t = page_index->oldest_time / USEC_PER_SEC;
|
|
*last_entry_t = page_index->latest_time / USEC_PER_SEC;
|
|
return 0;
|
|
}
|
|
|
|
return 1;
|
|
}
|
|
|
|
/* Also gets a reference for the page */
|
|
void *rrdeng_create_page(struct rrdengine_instance *ctx, uuid_t *id, struct rrdeng_page_descr **ret_descr)
|
|
{
|
|
struct rrdeng_page_descr *descr;
|
|
struct page_cache_descr *pg_cache_descr;
|
|
void *page;
|
|
/* TODO: check maximum number of pages in page cache limit */
|
|
|
|
descr = pg_cache_create_descr();
|
|
descr->id = id; /* TODO: add page type: metric, log, something? */
|
|
descr->type = ctx->page_type;
|
|
page = dbengine_page_alloc(); /*TODO: add page size */
|
|
rrdeng_page_descr_mutex_lock(ctx, descr);
|
|
pg_cache_descr = descr->pg_cache_descr;
|
|
pg_cache_descr->page = page;
|
|
pg_cache_descr->flags = RRD_PAGE_DIRTY /*| RRD_PAGE_LOCKED */ | RRD_PAGE_POPULATED /* | BEING_COLLECTED */;
|
|
pg_cache_descr->refcnt = 1;
|
|
|
|
debug(D_RRDENGINE, "Created new page:");
|
|
if (unlikely(debug_flags & D_RRDENGINE))
|
|
print_page_cache_descr(descr);
|
|
rrdeng_page_descr_mutex_unlock(ctx, descr);
|
|
*ret_descr = descr;
|
|
return page;
|
|
}
|
|
|
|
/* The page must not be empty */
|
|
void rrdeng_commit_page(struct rrdengine_instance *ctx, struct rrdeng_page_descr *descr,
|
|
Word_t page_correlation_id)
|
|
{
|
|
struct page_cache *pg_cache = &ctx->pg_cache;
|
|
Pvoid_t *PValue;
|
|
unsigned nr_committed_pages;
|
|
|
|
if (unlikely(NULL == descr)) {
|
|
debug(D_RRDENGINE, "%s: page descriptor is NULL, page has already been force-committed.", __func__);
|
|
return;
|
|
}
|
|
fatal_assert(descr->page_length);
|
|
|
|
uv_rwlock_wrlock(&pg_cache->committed_page_index.lock);
|
|
PValue = JudyLIns(&pg_cache->committed_page_index.JudyL_array, page_correlation_id, PJE0);
|
|
*PValue = descr;
|
|
nr_committed_pages = ++pg_cache->committed_page_index.nr_committed_pages;
|
|
uv_rwlock_wrunlock(&pg_cache->committed_page_index.lock);
|
|
|
|
if (nr_committed_pages >= pg_cache_hard_limit(ctx) / 2) {
|
|
/* over 50% of pages have not been committed yet */
|
|
|
|
if (ctx->drop_metrics_under_page_cache_pressure &&
|
|
nr_committed_pages >= pg_cache_committed_hard_limit(ctx)) {
|
|
/* 100% of pages are dirty */
|
|
struct rrdeng_cmd cmd;
|
|
|
|
cmd.opcode = RRDENG_INVALIDATE_OLDEST_MEMORY_PAGE;
|
|
rrdeng_enq_cmd(&ctx->worker_config, &cmd);
|
|
} else {
|
|
if (0 == (unsigned long) ctx->stats.pg_cache_over_half_dirty_events) {
|
|
/* only print the first time */
|
|
errno = 0;
|
|
error("Failed to flush dirty buffers quickly enough in dbengine instance \"%s\". "
|
|
"Metric data at risk of not being stored in the database, "
|
|
"please reduce disk load or use a faster disk.", ctx->dbfiles_path);
|
|
}
|
|
rrd_stat_atomic_add(&ctx->stats.pg_cache_over_half_dirty_events, 1);
|
|
rrd_stat_atomic_add(&global_pg_cache_over_half_dirty_events, 1);
|
|
}
|
|
}
|
|
|
|
pg_cache_put(ctx, descr);
|
|
}
|
|
|
|
/* Gets a reference for the page */
|
|
void *rrdeng_get_latest_page(struct rrdengine_instance *ctx, uuid_t *id, void **handle)
|
|
{
|
|
struct rrdeng_page_descr *descr;
|
|
struct page_cache_descr *pg_cache_descr;
|
|
|
|
debug(D_RRDENGINE, "Reading existing page:");
|
|
descr = pg_cache_lookup(ctx, NULL, id, INVALID_TIME);
|
|
if (NULL == descr) {
|
|
*handle = NULL;
|
|
|
|
return NULL;
|
|
}
|
|
*handle = descr;
|
|
pg_cache_descr = descr->pg_cache_descr;
|
|
|
|
return pg_cache_descr->page;
|
|
}
|
|
|
|
/* Gets a reference for the page */
|
|
void *rrdeng_get_page(struct rrdengine_instance *ctx, uuid_t *id, usec_t point_in_time, void **handle)
|
|
{
|
|
struct rrdeng_page_descr *descr;
|
|
struct page_cache_descr *pg_cache_descr;
|
|
|
|
debug(D_RRDENGINE, "Reading existing page:");
|
|
descr = pg_cache_lookup(ctx, NULL, id, point_in_time);
|
|
if (NULL == descr) {
|
|
*handle = NULL;
|
|
|
|
return NULL;
|
|
}
|
|
*handle = descr;
|
|
pg_cache_descr = descr->pg_cache_descr;
|
|
|
|
return pg_cache_descr->page;
|
|
}
|
|
|
|
/*
|
|
* Gathers Database Engine statistics.
|
|
* Careful when modifying this function.
|
|
* You must not change the indices of the statistics or user code will break.
|
|
* You must not exceed RRDENG_NR_STATS or it will crash.
|
|
*/
|
|
void rrdeng_get_37_statistics(struct rrdengine_instance *ctx, unsigned long long *array)
|
|
{
|
|
if (ctx == NULL)
|
|
return;
|
|
|
|
struct page_cache *pg_cache = &ctx->pg_cache;
|
|
|
|
array[0] = (uint64_t)ctx->stats.metric_API_producers;
|
|
array[1] = (uint64_t)ctx->stats.metric_API_consumers;
|
|
array[2] = (uint64_t)pg_cache->page_descriptors;
|
|
array[3] = (uint64_t)pg_cache->populated_pages;
|
|
array[4] = (uint64_t)pg_cache->committed_page_index.nr_committed_pages;
|
|
array[5] = (uint64_t)ctx->stats.pg_cache_insertions;
|
|
array[6] = (uint64_t)ctx->stats.pg_cache_deletions;
|
|
array[7] = (uint64_t)ctx->stats.pg_cache_hits;
|
|
array[8] = (uint64_t)ctx->stats.pg_cache_misses;
|
|
array[9] = (uint64_t)ctx->stats.pg_cache_backfills;
|
|
array[10] = (uint64_t)ctx->stats.pg_cache_evictions;
|
|
array[11] = (uint64_t)ctx->stats.before_compress_bytes;
|
|
array[12] = (uint64_t)ctx->stats.after_compress_bytes;
|
|
array[13] = (uint64_t)ctx->stats.before_decompress_bytes;
|
|
array[14] = (uint64_t)ctx->stats.after_decompress_bytes;
|
|
array[15] = (uint64_t)ctx->stats.io_write_bytes;
|
|
array[16] = (uint64_t)ctx->stats.io_write_requests;
|
|
array[17] = (uint64_t)ctx->stats.io_read_bytes;
|
|
array[18] = (uint64_t)ctx->stats.io_read_requests;
|
|
array[19] = (uint64_t)ctx->stats.io_write_extent_bytes;
|
|
array[20] = (uint64_t)ctx->stats.io_write_extents;
|
|
array[21] = (uint64_t)ctx->stats.io_read_extent_bytes;
|
|
array[22] = (uint64_t)ctx->stats.io_read_extents;
|
|
array[23] = (uint64_t)ctx->stats.datafile_creations;
|
|
array[24] = (uint64_t)ctx->stats.datafile_deletions;
|
|
array[25] = (uint64_t)ctx->stats.journalfile_creations;
|
|
array[26] = (uint64_t)ctx->stats.journalfile_deletions;
|
|
array[27] = (uint64_t)ctx->stats.page_cache_descriptors;
|
|
array[28] = (uint64_t)ctx->stats.io_errors;
|
|
array[29] = (uint64_t)ctx->stats.fs_errors;
|
|
array[30] = (uint64_t)global_io_errors;
|
|
array[31] = (uint64_t)global_fs_errors;
|
|
array[32] = (uint64_t)rrdeng_reserved_file_descriptors;
|
|
array[33] = (uint64_t)ctx->stats.pg_cache_over_half_dirty_events;
|
|
array[34] = (uint64_t)global_pg_cache_over_half_dirty_events;
|
|
array[35] = (uint64_t)ctx->stats.flushing_pressure_page_deletions;
|
|
array[36] = (uint64_t)global_flushing_pressure_page_deletions;
|
|
fatal_assert(RRDENG_NR_STATS == 37);
|
|
}
|
|
|
|
/* Releases reference to page */
|
|
void rrdeng_put_page(struct rrdengine_instance *ctx, void *handle)
|
|
{
|
|
(void)ctx;
|
|
pg_cache_put(ctx, (struct rrdeng_page_descr *)handle);
|
|
}
|
|
|
|
/*
|
|
* Returns 0 on success, negative on error
|
|
*/
|
|
int rrdeng_init(RRDHOST *host, struct rrdengine_instance **ctxp, char *dbfiles_path, unsigned page_cache_mb,
|
|
unsigned disk_space_mb, int tier) {
|
|
struct rrdengine_instance *ctx;
|
|
int error;
|
|
uint32_t max_open_files;
|
|
|
|
max_open_files = rlimit_nofile.rlim_cur / 4;
|
|
|
|
/* reserve RRDENG_FD_BUDGET_PER_INSTANCE file descriptors for this instance */
|
|
rrd_stat_atomic_add(&rrdeng_reserved_file_descriptors, RRDENG_FD_BUDGET_PER_INSTANCE);
|
|
if (rrdeng_reserved_file_descriptors > max_open_files) {
|
|
error(
|
|
"Exceeded the budget of available file descriptors (%u/%u), cannot create new dbengine instance.",
|
|
(unsigned)rrdeng_reserved_file_descriptors,
|
|
(unsigned)max_open_files);
|
|
|
|
rrd_stat_atomic_add(&global_fs_errors, 1);
|
|
rrd_stat_atomic_add(&rrdeng_reserved_file_descriptors, -RRDENG_FD_BUDGET_PER_INSTANCE);
|
|
return UV_EMFILE;
|
|
}
|
|
|
|
if(NULL == ctxp) {
|
|
ctx = multidb_ctx[tier];
|
|
memset(ctx, 0, sizeof(*ctx));
|
|
}
|
|
else {
|
|
*ctxp = ctx = callocz(1, sizeof(*ctx));
|
|
}
|
|
ctx->tier = tier;
|
|
ctx->page_type = tier_page_type[tier];
|
|
ctx->global_compress_alg = RRD_LZ4;
|
|
if (page_cache_mb < RRDENG_MIN_PAGE_CACHE_SIZE_MB)
|
|
page_cache_mb = RRDENG_MIN_PAGE_CACHE_SIZE_MB;
|
|
ctx->max_cache_pages = page_cache_mb * (1048576LU / RRDENG_BLOCK_SIZE);
|
|
/* try to keep 5% of the page cache free */
|
|
ctx->cache_pages_low_watermark = (ctx->max_cache_pages * 95LLU) / 100;
|
|
if (disk_space_mb < RRDENG_MIN_DISK_SPACE_MB)
|
|
disk_space_mb = RRDENG_MIN_DISK_SPACE_MB;
|
|
ctx->max_disk_space = disk_space_mb * 1048576LLU;
|
|
strncpyz(ctx->dbfiles_path, dbfiles_path, sizeof(ctx->dbfiles_path) - 1);
|
|
ctx->dbfiles_path[sizeof(ctx->dbfiles_path) - 1] = '\0';
|
|
if (NULL == host)
|
|
strncpyz(ctx->machine_guid, registry_get_this_machine_guid(), GUID_LEN);
|
|
else
|
|
strncpyz(ctx->machine_guid, host->machine_guid, GUID_LEN);
|
|
|
|
ctx->drop_metrics_under_page_cache_pressure = rrdeng_drop_metrics_under_page_cache_pressure;
|
|
ctx->metric_API_max_producers = 0;
|
|
ctx->quiesce = NO_QUIESCE;
|
|
ctx->metalog_ctx = NULL; /* only set this after the metadata log has finished initializing */
|
|
ctx->host = host;
|
|
|
|
memset(&ctx->worker_config, 0, sizeof(ctx->worker_config));
|
|
ctx->worker_config.ctx = ctx;
|
|
init_page_cache(ctx);
|
|
init_commit_log(ctx);
|
|
error = init_rrd_files(ctx);
|
|
if (error) {
|
|
goto error_after_init_rrd_files;
|
|
}
|
|
|
|
completion_init(&ctx->rrdengine_completion);
|
|
fatal_assert(0 == uv_thread_create(&ctx->worker_config.thread, rrdeng_worker, &ctx->worker_config));
|
|
/* wait for worker thread to initialize */
|
|
completion_wait_for(&ctx->rrdengine_completion);
|
|
completion_destroy(&ctx->rrdengine_completion);
|
|
uv_thread_set_name_np(ctx->worker_config.thread, "LIBUV_WORKER");
|
|
if (ctx->worker_config.error) {
|
|
goto error_after_rrdeng_worker;
|
|
}
|
|
error = metalog_init(ctx);
|
|
if (error) {
|
|
error("Failed to initialize metadata log file event loop.");
|
|
goto error_after_rrdeng_worker;
|
|
}
|
|
|
|
return 0;
|
|
|
|
error_after_rrdeng_worker:
|
|
finalize_rrd_files(ctx);
|
|
error_after_init_rrd_files:
|
|
free_page_cache(ctx);
|
|
if (!is_storage_engine_shared((STORAGE_INSTANCE *)ctx)) {
|
|
freez(ctx);
|
|
if (ctxp)
|
|
*ctxp = NULL;
|
|
}
|
|
rrd_stat_atomic_add(&rrdeng_reserved_file_descriptors, -RRDENG_FD_BUDGET_PER_INSTANCE);
|
|
return UV_EIO;
|
|
}
|
|
|
|
/*
|
|
* Returns 0 on success, 1 on error
|
|
*/
|
|
int rrdeng_exit(struct rrdengine_instance *ctx)
|
|
{
|
|
struct rrdeng_cmd cmd;
|
|
|
|
if (NULL == ctx) {
|
|
return 1;
|
|
}
|
|
|
|
/* TODO: add page to page cache */
|
|
cmd.opcode = RRDENG_SHUTDOWN;
|
|
rrdeng_enq_cmd(&ctx->worker_config, &cmd);
|
|
|
|
fatal_assert(0 == uv_thread_join(&ctx->worker_config.thread));
|
|
|
|
finalize_rrd_files(ctx);
|
|
//metalog_exit(ctx->metalog_ctx);
|
|
free_page_cache(ctx);
|
|
|
|
if(!is_storage_engine_shared((STORAGE_INSTANCE *)ctx))
|
|
freez(ctx);
|
|
|
|
rrd_stat_atomic_add(&rrdeng_reserved_file_descriptors, -RRDENG_FD_BUDGET_PER_INSTANCE);
|
|
return 0;
|
|
}
|
|
|
|
void rrdeng_prepare_exit(struct rrdengine_instance *ctx)
|
|
{
|
|
struct rrdeng_cmd cmd;
|
|
|
|
if (NULL == ctx) {
|
|
return;
|
|
}
|
|
|
|
completion_init(&ctx->rrdengine_completion);
|
|
cmd.opcode = RRDENG_QUIESCE;
|
|
rrdeng_enq_cmd(&ctx->worker_config, &cmd);
|
|
|
|
/* wait for dbengine to quiesce */
|
|
completion_wait_for(&ctx->rrdengine_completion);
|
|
completion_destroy(&ctx->rrdengine_completion);
|
|
|
|
//metalog_prepare_exit(ctx->metalog_ctx);
|
|
}
|
|
|
|
RRDENG_SIZE_STATS rrdeng_size_statistics(struct rrdengine_instance *ctx) {
|
|
RRDENG_SIZE_STATS stats = { 0 };
|
|
|
|
for(struct pg_cache_page_index *page_index = ctx->pg_cache.metrics_index.last_page_index;
|
|
page_index != NULL ;page_index = page_index->prev) {
|
|
stats.metrics++;
|
|
stats.metrics_pages += page_index->page_count;
|
|
}
|
|
|
|
for(struct rrdengine_datafile *df = ctx->datafiles.first; df ;df = df->next) {
|
|
stats.datafiles++;
|
|
|
|
for(struct extent_info *ei = df->extents.first; ei ; ei = ei->next) {
|
|
stats.extents++;
|
|
stats.extents_compressed_bytes += ei->size;
|
|
|
|
for(int p = 0; p < ei->number_of_pages ;p++) {
|
|
struct rrdeng_page_descr *descr = ei->pages[p];
|
|
|
|
usec_t update_every_usec;
|
|
|
|
size_t points = descr->page_length / PAGE_POINT_SIZE_BYTES(descr);
|
|
|
|
if(likely(points > 1))
|
|
update_every_usec = (descr->end_time - descr->start_time) / (points - 1);
|
|
else {
|
|
update_every_usec = default_rrd_update_every * get_tier_grouping(ctx->tier) * USEC_PER_SEC;
|
|
stats.single_point_pages++;
|
|
}
|
|
|
|
time_t duration_secs = (time_t)((descr->end_time - descr->start_time + update_every_usec)/USEC_PER_SEC);
|
|
|
|
stats.extents_pages++;
|
|
stats.pages_uncompressed_bytes += descr->page_length;
|
|
stats.pages_duration_secs += duration_secs;
|
|
stats.points += points;
|
|
|
|
stats.page_types[descr->type].pages++;
|
|
stats.page_types[descr->type].pages_uncompressed_bytes += descr->page_length;
|
|
stats.page_types[descr->type].pages_duration_secs += duration_secs;
|
|
stats.page_types[descr->type].points += points;
|
|
|
|
if(!stats.first_t || (descr->start_time - update_every_usec) < stats.first_t)
|
|
stats.first_t = (descr->start_time - update_every_usec) / USEC_PER_SEC;
|
|
|
|
if(!stats.last_t || descr->end_time > stats.last_t)
|
|
stats.last_t = descr->end_time / USEC_PER_SEC;
|
|
}
|
|
}
|
|
}
|
|
|
|
|
|
stats.currently_collected_metrics = ctx->stats.metric_API_producers;
|
|
stats.max_concurrently_collected_metrics = ctx->metric_API_max_producers;
|
|
|
|
internal_error(stats.metrics_pages != stats.extents_pages + stats.currently_collected_metrics,
|
|
"DBENGINE: metrics pages is %zu, but extents pages is %zu and API consumers is %zu",
|
|
stats.metrics_pages, stats.extents_pages, stats.currently_collected_metrics);
|
|
|
|
stats.disk_space = ctx->disk_space;
|
|
stats.max_disk_space = ctx->max_disk_space;
|
|
|
|
stats.database_retention_secs = (time_t)(stats.last_t - stats.first_t);
|
|
|
|
if(stats.extents_pages)
|
|
stats.average_page_size_bytes = (double)stats.pages_uncompressed_bytes / (double)stats.extents_pages;
|
|
|
|
if(stats.pages_uncompressed_bytes > 0)
|
|
stats.average_compression_savings = 100.0 - ((double)stats.extents_compressed_bytes * 100.0 / (double)stats.pages_uncompressed_bytes);
|
|
|
|
if(stats.points)
|
|
stats.average_point_duration_secs = (double)stats.pages_duration_secs / (double)stats.points;
|
|
|
|
if(stats.metrics) {
|
|
stats.average_metric_retention_secs = (double)stats.pages_duration_secs / (double)stats.metrics;
|
|
|
|
if(stats.database_retention_secs) {
|
|
double metric_coverage = stats.average_metric_retention_secs / (double)stats.database_retention_secs;
|
|
double db_retention_days = (double)stats.database_retention_secs / 86400.0;
|
|
|
|
stats.estimated_concurrently_collected_metrics = stats.metrics * metric_coverage;
|
|
|
|
stats.ephemeral_metrics_per_day_percent = ((double)stats.metrics * 100.0 / (double)stats.estimated_concurrently_collected_metrics - 100.0) / (double)db_retention_days;
|
|
}
|
|
}
|
|
|
|
stats.sizeof_metric = struct_natural_alignment(sizeof(struct pg_cache_page_index));
|
|
stats.sizeof_page = struct_natural_alignment(sizeof(struct rrdeng_page_descr));
|
|
stats.sizeof_datafile = struct_natural_alignment(sizeof(struct rrdengine_datafile)) + struct_natural_alignment(sizeof(struct rrdengine_journalfile));
|
|
stats.sizeof_page_in_cache = struct_natural_alignment(sizeof(struct page_cache_descr));
|
|
stats.sizeof_point_data = page_type_size[ctx->page_type];
|
|
stats.sizeof_page_data = RRDENG_BLOCK_SIZE;
|
|
stats.pages_per_extent = rrdeng_pages_per_extent;
|
|
|
|
stats.sizeof_extent = sizeof(struct extent_info);
|
|
stats.sizeof_page_in_extent = sizeof(struct rrdeng_page_descr *);
|
|
|
|
stats.sizeof_metric_in_index = 40;
|
|
stats.sizeof_page_in_index = 24;
|
|
|
|
stats.default_granularity_secs = (size_t)default_rrd_update_every * get_tier_grouping(ctx->tier);
|
|
|
|
return stats;
|
|
}
|