0
0
Fork 0
mirror of https://github.com/netdata/netdata.git synced 2025-04-14 01:29:11 +00:00
netdata_netdata/src/database/engine/metric.c
Costa Tsaousis e9b9fb75c4
Free all memory on exit ()
* destroy caches

* cleanup the mrg and all hosts

* cleanup uuidmap

* cleanup inline functions and replication

* Add destroy_aclk_config function to free aclk_config in RRDHOST

* flush all hot pages after the collection has been finalized

* increase daemon status file version to 13

---------

Co-authored-by: Stelios Fragkakis <52996999+stelfrag@users.noreply.github.com>
2025-03-11 10:41:25 +02:00

997 lines
36 KiB
C

// SPDX-License-Identifier: GPL-3.0-or-later
#include "metric.h"
#include "cache.h"
#include "libnetdata/locks/locks.h"
#include "rrddiskprotocol.h"
struct metric {
Word_t section; // never changes
UUIDMAP_ID uuid; // never changes
REFCOUNT refcount;
uint8_t partition;
uint32_t latest_update_every_s; // the latest data collection frequency
time_t first_time_s; // the timestamp of the oldest point in the database
time_t latest_time_s_clean; // the timestamp of the newest point in the database
time_t latest_time_s_hot; // the timestamp of the latest point that has been collected (not yet stored)
#ifdef NETDATA_INTERNAL_CHECKS
pid_t writer;
#endif
// THIS IS allocated with malloc()
// YOU HAVE TO INITIALIZE IT YOURSELF !
};
#define set_metric_field_with_condition(field, value, condition) ({ \
typeof(field) _current = __atomic_load_n(&(field), __ATOMIC_RELAXED); \
typeof(field) _wanted = value; \
bool did_it = true; \
\
do { \
if((condition) && (_current != _wanted)) { \
; \
} \
else { \
did_it = false; \
break; \
} \
} while(!__atomic_compare_exchange_n(&(field), &_current, _wanted, \
false, __ATOMIC_RELAXED, __ATOMIC_RELAXED)); \
\
did_it; \
})
static struct aral_statistics mrg_aral_statistics;
struct mrg {
struct mrg_partition {
ARAL *aral; // not protected by our spinlock - it has its own
RW_SPINLOCK rw_spinlock;
Pvoid_t uuid_judy; // JudyL: each UUID has a JudyL of sections (tiers)
struct mrg_statistics stats;
} index[UUIDMAP_PARTITIONS];
};
static inline void MRG_STATS_DUPLICATE_ADD(MRG *mrg, size_t partition) {
mrg->index[partition].stats.additions_duplicate++;
}
static inline void MRG_STATS_ADDED_METRIC(MRG *mrg, size_t partition) {
mrg->index[partition].stats.entries++;
mrg->index[partition].stats.additions++;
mrg->index[partition].stats.size += sizeof(METRIC);
}
static inline void MRG_STATS_DELETED_METRIC(MRG *mrg, size_t partition) {
mrg->index[partition].stats.entries--;
mrg->index[partition].stats.size -= sizeof(METRIC);
mrg->index[partition].stats.deletions++;
}
static inline void MRG_STATS_SEARCH_HIT(MRG *mrg, size_t partition) {
__atomic_add_fetch(&mrg->index[partition].stats.search_hits, 1, __ATOMIC_RELAXED);
}
static inline void MRG_STATS_SEARCH_MISS(MRG *mrg, size_t partition) {
__atomic_add_fetch(&mrg->index[partition].stats.search_misses, 1, __ATOMIC_RELAXED);
}
static inline void MRG_STATS_DELETE_MISS(MRG *mrg, size_t partition) {
mrg->index[partition].stats.delete_misses++;
}
#define mrg_index_read_lock(mrg, partition) rw_spinlock_read_lock(&(mrg)->index[partition].rw_spinlock)
#define mrg_index_read_unlock(mrg, partition) rw_spinlock_read_unlock(&(mrg)->index[partition].rw_spinlock)
#define mrg_index_write_lock(mrg, partition) rw_spinlock_write_lock(&(mrg)->index[partition].rw_spinlock)
#define mrg_index_write_unlock(mrg, partition) rw_spinlock_write_unlock(&(mrg)->index[partition].rw_spinlock)
static inline void mrg_stats_judy_mem(MRG *mrg, size_t partition, int64_t judy_mem) {
__atomic_add_fetch(&mrg->index[partition].stats.size, judy_mem, __ATOMIC_RELAXED);
}
static ALWAYS_INLINE time_t mrg_metric_get_first_time_s_smart(MRG *mrg __maybe_unused, METRIC *metric) {
time_t first_time_s = __atomic_load_n(&metric->first_time_s, __ATOMIC_RELAXED);
if(first_time_s <= 0) {
first_time_s = __atomic_load_n(&metric->latest_time_s_clean, __ATOMIC_RELAXED);
if(first_time_s <= 0)
first_time_s = __atomic_load_n(&metric->latest_time_s_hot, __ATOMIC_RELAXED);
if(first_time_s <= 0)
first_time_s = 0;
else
__atomic_store_n(&metric->first_time_s, first_time_s, __ATOMIC_RELAXED);
}
return first_time_s;
}
static inline void metric_log(MRG *mrg __maybe_unused, METRIC *metric, const char *msg) {
struct rrdengine_instance *ctx = (struct rrdengine_instance *)metric->section;
nd_uuid_t uuid;
uuidmap_uuid(metric->uuid, uuid);
char uuid_txt[UUID_STR_LEN];
uuid_unparse_lower(uuid, uuid_txt);
nd_log(NDLS_DAEMON, NDLP_ERR,
"METRIC: %s on %s at tier %d, refcount %d, partition %u, "
"retention [%ld - %ld (hot), %ld (clean)], update every %"PRIu32
#ifdef NETDATA_INTERNAL_CHECKS
", writer pid %d "
#endif
" --- PLEASE OPEN A GITHUB ISSUE TO REPORT THIS LOG LINE TO NETDATA --- ",
msg,
uuid_txt,
ctx->config.tier,
metric->refcount,
metric->partition,
metric->first_time_s,
metric->latest_time_s_hot,
metric->latest_time_s_clean,
metric->latest_update_every_s
#ifdef NETDATA_INTERNAL_CHECKS
, (int)metric->writer
#endif
);
}
static ALWAYS_INLINE bool acquired_metric_has_retention(MRG *mrg, METRIC *metric) {
time_t first, last;
mrg_metric_get_retention(mrg, metric, &first, &last, NULL);
bool rc = (first != 0 && last != 0 && first <= last);
if(!rc && __atomic_load_n(&mrg->index[metric->partition].stats.writers, __ATOMIC_RELAXED) > 0)
rc = true;
return rc;
}
static ALWAYS_INLINE void acquired_for_deletion_metric_delete(MRG *mrg, METRIC *metric) {
JudyAllocThreadPulseReset();
size_t partition = metric->partition;
mrg_index_write_lock(mrg, partition);
Pvoid_t *sections_judy_pptr = JudyLGet(mrg->index[partition].uuid_judy, metric->uuid, PJE0);
if(unlikely(sections_judy_pptr == PJERR))
fatal("METRIC: corrupted JudyL");
if(unlikely(!sections_judy_pptr || !*sections_judy_pptr)) {
MRG_STATS_DELETE_MISS(mrg, partition);
mrg_index_write_unlock(mrg, partition);
return;
}
int rc = JudyLDel(sections_judy_pptr, metric->section, PJE0);
if(unlikely(!rc)) {
MRG_STATS_DELETE_MISS(mrg, partition);
mrg_index_write_unlock(mrg, partition);
mrg_stats_judy_mem(mrg, partition, JudyAllocThreadPulseGetAndReset());
return;
}
if(!*sections_judy_pptr) {
rc = JudyLDel(&mrg->index[partition].uuid_judy, metric->uuid, PJE0);
if(unlikely(!rc))
fatal("DBENGINE METRIC: cannot delete UUID from JudyL");
}
MRG_STATS_DELETED_METRIC(mrg, partition);
mrg_index_write_unlock(mrg, partition);
uuidmap_free(metric->uuid);
aral_freez(mrg->index[partition].aral, metric);
mrg_stats_judy_mem(mrg, partition, JudyAllocThreadPulseGetAndReset());
}
static ALWAYS_INLINE bool metric_acquire(MRG *mrg, METRIC *metric) {
REFCOUNT rc = refcount_acquire_advanced(&metric->refcount);
if(!REFCOUNT_ACQUIRED(rc))
return false;
size_t partition = metric->partition;
if(rc == 1)
__atomic_add_fetch(&mrg->index[partition].stats.entries_acquired, 1, __ATOMIC_RELAXED);
__atomic_add_fetch(&mrg->index[partition].stats.current_references, 1, __ATOMIC_RELAXED);
return true;
}
static ALWAYS_INLINE bool metric_release(MRG *mrg, METRIC *metric) {
size_t partition = metric->partition;
REFCOUNT refcount = refcount_release(&metric->refcount);
if(!refcount && !acquired_metric_has_retention(mrg, metric) && refcount_acquire_for_deletion(&metric->refcount))
refcount = REFCOUNT_DELETED;
if(refcount == 0 || refcount == REFCOUNT_DELETED) {
__atomic_sub_fetch(&mrg->index[partition].stats.entries_acquired, 1, __ATOMIC_RELAXED);
if(refcount == REFCOUNT_DELETED)
acquired_for_deletion_metric_delete(mrg, metric);
}
__atomic_sub_fetch(&mrg->index[partition].stats.current_references, 1, __ATOMIC_RELAXED);
return refcount == REFCOUNT_DELETED;
}
static ALWAYS_INLINE METRIC *metric_add_and_acquire(MRG *mrg, MRG_ENTRY *entry, bool *ret) {
JudyAllocThreadPulseReset();
UUIDMAP_ID id = uuidmap_create(*entry->uuid);
size_t partition = uuid_to_uuidmap_partition(*entry->uuid);
METRIC *allocation = aral_mallocz(mrg->index[partition].aral);
Pvoid_t *PValue;
while(1) {
mrg_index_write_lock(mrg, partition);
Pvoid_t *sections_judy_pptr = JudyLIns(&mrg->index[partition].uuid_judy, id, PJE0);
if (unlikely(!sections_judy_pptr || sections_judy_pptr == PJERR))
fatal("DBENGINE METRIC: corrupted UUIDs JudyL array");
PValue = JudyLIns(sections_judy_pptr, entry->section, PJE0);
if (unlikely(!PValue || PValue == PJERR))
fatal("DBENGINE METRIC: corrupted section JudyL array");
if (unlikely(*PValue != NULL)) {
METRIC *metric = *PValue;
if(!metric_acquire(mrg, metric)) {
mrg_index_write_unlock(mrg, partition);
continue;
}
MRG_STATS_DUPLICATE_ADD(mrg, partition);
mrg_index_write_unlock(mrg, partition);
if (ret)
*ret = false;
uuidmap_free(id);
aral_freez(mrg->index[partition].aral, allocation);
mrg_stats_judy_mem(mrg, partition, JudyAllocThreadPulseGetAndReset());
return metric;
}
break;
}
METRIC *metric = allocation;
metric->uuid = id;
metric->section = entry->section;
metric->first_time_s = MAX(0, entry->first_time_s);
metric->latest_time_s_clean = MAX(0, entry->last_time_s);
metric->latest_time_s_hot = 0;
metric->latest_update_every_s = entry->latest_update_every_s;
#ifdef NETDATA_INTERNAL_CHECKS
metric->writer = 0;
#endif
metric->refcount = 1;
metric->partition = partition;
*PValue = metric;
__atomic_add_fetch(&mrg->index[partition].stats.entries_acquired, 1, __ATOMIC_RELAXED);
__atomic_add_fetch(&mrg->index[partition].stats.current_references, 1, __ATOMIC_RELAXED);
MRG_STATS_ADDED_METRIC(mrg, partition);
mrg_index_write_unlock(mrg, partition);
if(ret)
*ret = true;
mrg_stats_judy_mem(mrg, partition, JudyAllocThreadPulseGetAndReset());
return metric;
}
static ALWAYS_INLINE METRIC *metric_get_and_acquire_by_id(MRG *mrg, UUIDMAP_ID id, Word_t section) {
size_t partition = uuidmap_id_to_partition(id);
while(1) {
mrg_index_read_lock(mrg, partition);
Pvoid_t *sections_judy_pptr = JudyLGet(mrg->index[partition].uuid_judy, id, PJE0);
if (unlikely(!sections_judy_pptr)) {
mrg_index_read_unlock(mrg, partition);
MRG_STATS_SEARCH_MISS(mrg, partition);
return NULL;
}
Pvoid_t *PValue = JudyLGet(*sections_judy_pptr, section, PJE0);
if (unlikely(!PValue)) {
mrg_index_read_unlock(mrg, partition);
MRG_STATS_SEARCH_MISS(mrg, partition);
return NULL;
}
METRIC *metric = *PValue;
if(metric && !metric_acquire(mrg, metric))
metric = NULL;
mrg_index_read_unlock(mrg, partition);
if(metric) {
MRG_STATS_SEARCH_HIT(mrg, partition);
return metric;
}
}
}
// ----------------------------------------------------------------------------
// public API
inline MRG *mrg_create(void) {
MRG *mrg = callocz(1, sizeof(MRG));
for(size_t i = 0; i < _countof(mrg->index) ; i++) {
rw_spinlock_init(&mrg->index[i].rw_spinlock);
char buf[ARAL_MAX_NAME + 1];
snprintfz(buf, ARAL_MAX_NAME, "mrg[%zu]", i);
mrg->index[i].aral = aral_create(buf, sizeof(METRIC), 0, 16384, &mrg_aral_statistics, NULL, NULL,
false, false, true);
}
pulse_aral_register_statistics(&mrg_aral_statistics, "mrg");
return mrg;
}
struct aral_statistics *mrg_aral_stats(void) {
return &mrg_aral_statistics;
}
size_t mrg_destroy(MRG *mrg) {
if (unlikely(!mrg))
return 0;
size_t referenced = 0;
// Traverse all partitions
for (size_t partition = 0; partition < UUIDMAP_PARTITIONS; partition++) {
// Lock the partition to prevent new entries while we're cleaning up
mrg_index_write_lock(mrg, partition);
Pvoid_t uuid_judy = mrg->index[partition].uuid_judy;
Word_t uuid_index = 0;
Pvoid_t *uuid_pvalue;
// Traverse all UUIDs in this partition
for (uuid_pvalue = JudyLFirst(uuid_judy, &uuid_index, PJE0);
uuid_pvalue != NULL && uuid_pvalue != PJERR;
uuid_pvalue = JudyLNext(uuid_judy, &uuid_index, PJE0)) {
if (!(*uuid_pvalue))
continue;
// Get the sections judy for this UUID
Pvoid_t sections_judy = *uuid_pvalue;
Word_t section_index = 0;
Pvoid_t *section_pvalue;
// Traverse all sections for this UUID
for (section_pvalue = JudyLFirst(sections_judy, &section_index, PJE0);
section_pvalue != NULL && section_pvalue != PJERR;
section_pvalue = JudyLNext(sections_judy, &section_index, PJE0)) {
if (!(*section_pvalue))
continue;
METRIC *metric = *section_pvalue;
// Try to acquire metric for deletion
if (!refcount_acquire_for_deletion(&metric->refcount))
referenced++;
uuidmap_free(metric->uuid);
aral_freez(mrg->index[partition].aral, metric);
MRG_STATS_DELETED_METRIC(mrg, partition);
}
JudyLFreeArray(&sections_judy, PJE0);
}
JudyLFreeArray(&uuid_judy, PJE0);
// Update the main Judy array reference
mrg->index[partition].uuid_judy = uuid_judy;
// Unlock the partition
mrg_index_write_unlock(mrg, partition);
// Destroy the aral for this partition
aral_destroy(mrg->index[partition].aral);
}
// Unregister the aral statistics
pulse_aral_unregister_statistics(&mrg_aral_statistics);
// Free the MRG structure
freez(mrg);
return referenced;
}
ALWAYS_INLINE METRIC *mrg_metric_add_and_acquire(MRG *mrg, MRG_ENTRY entry, bool *ret) {
// internal_fatal(entry.latest_time_s > max_acceptable_collected_time(),
// "DBENGINE METRIC: metric latest time is in the future");
return metric_add_and_acquire(mrg, &entry, ret);
}
ALWAYS_INLINE METRIC *mrg_metric_get_and_acquire_by_uuid(MRG *mrg, nd_uuid_t *uuid, Word_t section) {
UUIDMAP_ID id = uuidmap_create(*uuid);
METRIC *metric = metric_get_and_acquire_by_id(mrg, id, section);
uuidmap_free(id);
return metric;
}
ALWAYS_INLINE METRIC *mrg_metric_get_and_acquire_by_id(MRG *mrg, UUIDMAP_ID id, Word_t section) {
return metric_get_and_acquire_by_id(mrg, id, section);
}
ALWAYS_INLINE bool mrg_metric_release_and_delete(MRG *mrg, METRIC *metric) {
return metric_release(mrg, metric);
}
ALWAYS_INLINE METRIC *mrg_metric_dup(MRG *mrg, METRIC *metric) {
metric_acquire(mrg, metric);
return metric;
}
ALWAYS_INLINE void mrg_metric_release(MRG *mrg, METRIC *metric) {
metric_release(mrg, metric);
}
ALWAYS_INLINE Word_t mrg_metric_id(MRG *mrg __maybe_unused, METRIC *metric) {
return (Word_t)metric;
}
ALWAYS_INLINE nd_uuid_t *mrg_metric_uuid(MRG *mrg __maybe_unused, METRIC *metric) {
return uuidmap_uuid_ptr(metric->uuid);
}
ALWAYS_INLINE UUIDMAP_ID mrg_metric_uuidmap_id_dup(MRG *mrg __maybe_unused, METRIC *metric) {
return uuidmap_dup(metric->uuid);
}
ALWAYS_INLINE Word_t mrg_metric_section(MRG *mrg __maybe_unused, METRIC *metric) {
return metric->section;
}
ALWAYS_INLINE bool mrg_metric_set_first_time_s(MRG *mrg __maybe_unused, METRIC *metric, time_t first_time_s) {
internal_fatal(first_time_s < 0, "DBENGINE METRIC: timestamp is negative");
if(first_time_s == LONG_MAX)
first_time_s = 0;
if(unlikely(first_time_s < 0))
return false;
__atomic_store_n(&metric->first_time_s, first_time_s, __ATOMIC_RELAXED);
return true;
}
ALWAYS_INLINE void mrg_metric_expand_retention(MRG *mrg __maybe_unused, METRIC *metric, time_t first_time_s, time_t last_time_s, uint32_t update_every_s) {
internal_fatal(first_time_s < 0 || last_time_s < 0,
"DBENGINE METRIC: timestamp is negative");
internal_fatal(first_time_s > max_acceptable_collected_time(),
"DBENGINE METRIC: metric first time is in the future");
internal_fatal(last_time_s > max_acceptable_collected_time(),
"DBENGINE METRIC: metric last time is in the future");
if(first_time_s > 0 && first_time_s != LONG_MAX)
set_metric_field_with_condition(metric->first_time_s, first_time_s, _current <= 0 || (_wanted != 0 && _wanted != LONG_MAX && _wanted < _current));
if(last_time_s > 0) {
if(set_metric_field_with_condition(metric->latest_time_s_clean, last_time_s, _current <= 0 || _wanted > _current) &&
update_every_s > 0)
// set the latest update every too
set_metric_field_with_condition(metric->latest_update_every_s, update_every_s, true);
}
else if(update_every_s > 0)
// set it only if it is invalid
set_metric_field_with_condition(metric->latest_update_every_s, update_every_s, _current <= 0);
}
ALWAYS_INLINE bool mrg_metric_set_first_time_s_if_bigger(MRG *mrg __maybe_unused, METRIC *metric, time_t first_time_s) {
internal_fatal(first_time_s < 0, "DBENGINE METRIC: timestamp is negative");
return set_metric_field_with_condition(metric->first_time_s, first_time_s, _wanted != 0 && _wanted != LONG_MAX && _wanted > _current);
}
ALWAYS_INLINE time_t mrg_metric_get_first_time_s(MRG *mrg __maybe_unused, METRIC *metric) {
return mrg_metric_get_first_time_s_smart(mrg, metric);
}
void mrg_metric_clear_retention(MRG *mrg __maybe_unused, METRIC *metric) {
__atomic_store_n(&metric->first_time_s, 0, __ATOMIC_RELAXED);
__atomic_store_n(&metric->latest_time_s_clean, 0, __ATOMIC_RELAXED);
__atomic_store_n(&metric->latest_time_s_hot, 0, __ATOMIC_RELAXED);
}
ALWAYS_INLINE_HOT void mrg_metric_get_retention(MRG *mrg __maybe_unused, METRIC *metric, time_t *first_time_s, time_t *last_time_s, uint32_t *update_every_s) {
time_t clean = __atomic_load_n(&metric->latest_time_s_clean, __ATOMIC_RELAXED);
time_t hot = __atomic_load_n(&metric->latest_time_s_hot, __ATOMIC_RELAXED);
*last_time_s = MAX(clean, hot);
*first_time_s = mrg_metric_get_first_time_s_smart(mrg, metric);
if (update_every_s)
*update_every_s = __atomic_load_n(&metric->latest_update_every_s, __ATOMIC_RELAXED);
}
ALWAYS_INLINE bool mrg_metric_set_clean_latest_time_s(MRG *mrg __maybe_unused, METRIC *metric, time_t latest_time_s) {
internal_fatal(latest_time_s < 0, "DBENGINE METRIC: timestamp is negative");
// internal_fatal(latest_time_s > max_acceptable_collected_time(),
// "DBENGINE METRIC: metric latest time is in the future");
// internal_fatal(metric->latest_time_s_clean > latest_time_s,
// "DBENGINE METRIC: metric new clean latest time is older than the previous one");
if(latest_time_s > 0) {
if(set_metric_field_with_condition(metric->latest_time_s_clean, latest_time_s, true)) {
set_metric_field_with_condition(metric->first_time_s, latest_time_s, _current <= 0 || _wanted < _current);
return true;
}
}
return false;
}
// returns true when metric still has retention
ALWAYS_INLINE bool mrg_metric_has_zero_disk_retention(MRG *mrg __maybe_unused, METRIC *metric) {
Word_t section = mrg_metric_section(mrg, metric);
bool do_again = false;
size_t countdown = 5;
do {
time_t min_first_time_s = LONG_MAX;
time_t max_end_time_s = 0;
PGC_PAGE *page;
PGC_SEARCH method = PGC_SEARCH_FIRST;
time_t page_first_time_s = 0;
time_t page_end_time_s = 0;
while ((page = pgc_page_get_and_acquire(main_cache, section, (Word_t)metric, page_first_time_s, method))) {
method = PGC_SEARCH_NEXT;
bool is_hot = pgc_is_page_hot(page);
bool is_dirty = pgc_is_page_dirty(page);
page_first_time_s = pgc_page_start_time_s(page);
page_end_time_s = pgc_page_end_time_s(page);
if ((is_hot || is_dirty) && page_first_time_s > 0 && page_first_time_s < min_first_time_s)
min_first_time_s = page_first_time_s;
if (is_dirty && page_end_time_s > max_end_time_s)
max_end_time_s = page_end_time_s;
pgc_page_release(main_cache, page);
}
if (min_first_time_s == LONG_MAX)
min_first_time_s = 0;
if (--countdown && !min_first_time_s && __atomic_load_n(&metric->latest_time_s_hot, __ATOMIC_RELAXED))
do_again = true;
else {
internal_error(!countdown, "METRIC: giving up on updating the retention of metric without disk retention");
do_again = false;
set_metric_field_with_condition(metric->first_time_s, min_first_time_s, true);
set_metric_field_with_condition(metric->latest_time_s_clean, max_end_time_s, true);
}
} while(do_again);
time_t first, last;
mrg_metric_get_retention(mrg, metric, &first, &last, NULL);
return (first && last && first < last);
}
ALWAYS_INLINE bool mrg_metric_set_hot_latest_time_s(MRG *mrg __maybe_unused, METRIC *metric, time_t latest_time_s) {
internal_fatal(latest_time_s < 0, "DBENGINE METRIC: timestamp is negative");
// internal_fatal(latest_time_s > max_acceptable_collected_time(),
// "DBENGINE METRIC: metric latest time is in the future");
if(likely(latest_time_s > 0)) {
__atomic_store_n(&metric->latest_time_s_hot, latest_time_s, __ATOMIC_RELAXED);
return true;
}
return false;
}
ALWAYS_INLINE time_t mrg_metric_get_latest_clean_time_s(MRG *mrg __maybe_unused, METRIC *metric) {
time_t clean = __atomic_load_n(&metric->latest_time_s_clean, __ATOMIC_RELAXED);
return clean;
}
ALWAYS_INLINE time_t mrg_metric_get_latest_time_s(MRG *mrg __maybe_unused, METRIC *metric) {
time_t clean = __atomic_load_n(&metric->latest_time_s_clean, __ATOMIC_RELAXED);
time_t hot = __atomic_load_n(&metric->latest_time_s_hot, __ATOMIC_RELAXED);
return MAX(clean, hot);
}
ALWAYS_INLINE bool mrg_metric_set_update_every(MRG *mrg __maybe_unused, METRIC *metric, uint32_t update_every_s) {
if(likely(update_every_s > 0))
return set_metric_field_with_condition(metric->latest_update_every_s, update_every_s, true);
return false;
}
ALWAYS_INLINE_HOT bool mrg_metric_set_update_every_s_if_zero(MRG *mrg __maybe_unused, METRIC *metric, uint32_t update_every_s) {
if(likely(update_every_s > 0))
return set_metric_field_with_condition(metric->latest_update_every_s, update_every_s, _current <= 0);
return false;
}
ALWAYS_INLINE uint32_t mrg_metric_get_update_every_s(MRG *mrg __maybe_unused, METRIC *metric) {
return __atomic_load_n(&metric->latest_update_every_s, __ATOMIC_RELAXED);
}
#ifdef NETDATA_INTERNAL_CHECKS
ALWAYS_INLINE bool mrg_metric_set_writer(MRG *mrg, METRIC *metric) {
pid_t expected = __atomic_load_n(&metric->writer, __ATOMIC_RELAXED);
pid_t wanted = gettid_cached();
bool done = true;
do {
if(expected != 0) {
done = false;
break;
}
} while(!__atomic_compare_exchange_n(&metric->writer, &expected, wanted, false, __ATOMIC_RELAXED, __ATOMIC_RELAXED));
if(done)
__atomic_add_fetch(&mrg->index[metric->partition].stats.writers, 1, __ATOMIC_RELAXED);
else
__atomic_add_fetch(&mrg->index[metric->partition].stats.writers_conflicts, 1, __ATOMIC_RELAXED);
return done;
}
ALWAYS_INLINE bool mrg_metric_clear_writer(MRG *mrg, METRIC *metric) {
// this function can be called from a different thread than the one than the writer
pid_t expected = __atomic_load_n(&metric->writer, __ATOMIC_RELAXED);
pid_t wanted = 0;
bool done = true;
do {
if(!expected) {
done = false;
break;
}
} while(!__atomic_compare_exchange_n(&metric->writer, &expected, wanted, false, __ATOMIC_RELAXED, __ATOMIC_RELAXED));
if(done)
__atomic_sub_fetch(&mrg->index[metric->partition].stats.writers, 1, __ATOMIC_RELAXED);
return done;
}
#endif
inline void mrg_update_metric_retention_and_granularity_by_uuid(
MRG *mrg, Word_t section, nd_uuid_t *uuid,
time_t first_time_s, time_t last_time_s,
uint32_t update_every_s, time_t now_s)
{
if(unlikely(last_time_s > now_s)) {
nd_log_limit_static_global_var(erl, 1, 0);
nd_log_limit(&erl, NDLS_DAEMON, NDLP_WARNING,
"DBENGINE JV2: wrong last time on-disk (%ld - %ld, now %ld), "
"fixing last time to now",
first_time_s, last_time_s, now_s);
last_time_s = now_s;
}
if (unlikely(first_time_s > last_time_s)) {
nd_log_limit_static_global_var(erl, 1, 0);
nd_log_limit(&erl, NDLS_DAEMON, NDLP_WARNING,
"DBENGINE JV2: wrong first time on-disk (%ld - %ld, now %ld), "
"fixing first time to last time",
first_time_s, last_time_s, now_s);
first_time_s = last_time_s;
}
if (unlikely(first_time_s == 0 || last_time_s == 0)) {
nd_log_limit_static_global_var(erl, 1, 0);
nd_log_limit(&erl, NDLS_DAEMON, NDLP_WARNING,
"DBENGINE JV2: zero on-disk timestamps (%ld - %ld, now %ld), "
"using them as-is",
first_time_s, last_time_s, now_s);
}
bool added = false;
METRIC *metric = mrg_metric_get_and_acquire_by_uuid(mrg, uuid, section);
if (!metric) {
MRG_ENTRY entry = {
.uuid = uuid,
.section = section,
.first_time_s = first_time_s,
.last_time_s = last_time_s,
.latest_update_every_s = update_every_s
};
metric = mrg_metric_add_and_acquire(mrg, entry, &added);
}
struct rrdengine_instance *ctx = (struct rrdengine_instance *) section;
if (likely(!added)) {
uint64_t old_samples = 0;
if (update_every_s && metric->latest_update_every_s && metric->latest_time_s_clean)
old_samples = (metric->latest_time_s_clean - metric->first_time_s) / metric->latest_update_every_s;
mrg_metric_expand_retention(mrg, metric, first_time_s, last_time_s, update_every_s);
uint64_t new_samples = 0;
if (update_every_s && metric->latest_update_every_s && metric->latest_time_s_clean)
new_samples = (metric->latest_time_s_clean - metric->first_time_s) / metric->latest_update_every_s;
__atomic_add_fetch(&ctx->atomic.samples, new_samples - old_samples, __ATOMIC_RELAXED);
}
else {
// Newly added
if (update_every_s) {
uint64_t samples = (last_time_s - first_time_s) / update_every_s;
__atomic_add_fetch(&ctx->atomic.samples, samples, __ATOMIC_RELAXED);
}
__atomic_add_fetch(&ctx->atomic.metrics, 1, __ATOMIC_RELAXED);
}
mrg_metric_release(mrg, metric);
}
inline void mrg_get_statistics(MRG *mrg, struct mrg_statistics *s) {
memset(s, 0, sizeof(struct mrg_statistics));
for(size_t i = 0; i < _countof(mrg->index) ;i++) {
s->entries += __atomic_load_n(&mrg->index[i].stats.entries, __ATOMIC_RELAXED);
s->entries_acquired += __atomic_load_n(&mrg->index[i].stats.entries_acquired, __ATOMIC_RELAXED);
s->size += __atomic_load_n(&mrg->index[i].stats.size, __ATOMIC_RELAXED);
s->current_references += __atomic_load_n(&mrg->index[i].stats.current_references, __ATOMIC_RELAXED);
s->additions += __atomic_load_n(&mrg->index[i].stats.additions, __ATOMIC_RELAXED);
s->additions_duplicate += __atomic_load_n(&mrg->index[i].stats.additions_duplicate, __ATOMIC_RELAXED);
s->deletions += __atomic_load_n(&mrg->index[i].stats.deletions, __ATOMIC_RELAXED);
s->delete_having_retention_or_referenced += __atomic_load_n(&mrg->index[i].stats.delete_having_retention_or_referenced, __ATOMIC_RELAXED);
s->delete_misses += __atomic_load_n(&mrg->index[i].stats.delete_misses, __ATOMIC_RELAXED);
s->search_hits += __atomic_load_n(&mrg->index[i].stats.search_hits, __ATOMIC_RELAXED);
s->search_misses += __atomic_load_n(&mrg->index[i].stats.search_misses, __ATOMIC_RELAXED);
s->writers += __atomic_load_n(&mrg->index[i].stats.writers, __ATOMIC_RELAXED);
s->writers_conflicts += __atomic_load_n(&mrg->index[i].stats.writers_conflicts, __ATOMIC_RELAXED);
}
s->size += sizeof(MRG);
}
// ----------------------------------------------------------------------------
// unit test
struct mrg_stress_entry {
nd_uuid_t uuid;
time_t after;
time_t before;
};
struct mrg_stress {
MRG *mrg;
bool stop;
size_t entries;
struct mrg_stress_entry *array;
size_t updates;
};
static void *mrg_stress(void *ptr) {
struct mrg_stress *t = ptr;
MRG *mrg = t->mrg;
ssize_t start = 0;
ssize_t end = (ssize_t)t->entries;
ssize_t step = 1;
if(gettid_cached() % 2) {
start = (ssize_t)t->entries - 1;
end = -1;
step = -1;
}
while(!__atomic_load_n(&t->stop, __ATOMIC_RELAXED) && !nd_thread_signaled_to_cancel()) {
for (ssize_t i = start; i != end; i += step) {
struct mrg_stress_entry *e = &t->array[i];
time_t after = __atomic_sub_fetch(&e->after, 1, __ATOMIC_RELAXED);
time_t before = __atomic_add_fetch(&e->before, 1, __ATOMIC_RELAXED);
mrg_update_metric_retention_and_granularity_by_uuid(
mrg, 0x01,
&e->uuid,
after,
before,
1,
before);
__atomic_add_fetch(&t->updates, 1, __ATOMIC_RELAXED);
}
}
return ptr;
}
int mrg_unittest(void) {
MRG *mrg = mrg_create();
METRIC *m1_t0, *m2_t0, *m3_t0, *m4_t0;
METRIC *m1_t1, *m2_t1, *m3_t1, *m4_t1;
bool ret;
nd_uuid_t test_uuid;
uuid_generate(test_uuid);
MRG_ENTRY entry = {
.uuid = &test_uuid,
.section = 0,
.first_time_s = 2,
.last_time_s = 3,
.latest_update_every_s = 4,
};
m1_t0 = mrg_metric_add_and_acquire(mrg, entry, &ret);
if(!ret)
fatal("DBENGINE METRIC: failed to add metric");
// add the same metric again
m2_t0 = mrg_metric_add_and_acquire(mrg, entry, &ret);
if(m2_t0 != m1_t0)
fatal("DBENGINE METRIC: adding the same metric twice, does not return the same pointer");
if(ret)
fatal("DBENGINE METRIC: managed to add the same metric twice");
m3_t0 = mrg_metric_get_and_acquire_by_uuid(mrg, entry.uuid, entry.section);
if(m3_t0 != m1_t0)
fatal("DBENGINE METRIC: cannot find the metric added");
// add the same metric again
m4_t0 = mrg_metric_add_and_acquire(mrg, entry, &ret);
if(m4_t0 != m1_t0)
fatal("DBENGINE METRIC: adding the same metric twice, does not return the same pointer");
if(ret)
fatal("DBENGINE METRIC: managed to add the same metric twice");
// add the same metric in another section
entry.section = 1;
m1_t1 = mrg_metric_add_and_acquire(mrg, entry, &ret);
if(!ret)
fatal("DBENGINE METRIC: failed to add metric in section %zu", (size_t)entry.section);
// add the same metric again
m2_t1 = mrg_metric_add_and_acquire(mrg, entry, &ret);
if(m2_t1 != m1_t1)
fatal("DBENGINE METRIC: adding the same metric twice (section %zu), does not return the same pointer", (size_t)entry.section);
if(ret)
fatal("DBENGINE METRIC: managed to add the same metric twice in (section 0)");
m3_t1 = mrg_metric_get_and_acquire_by_uuid(mrg, entry.uuid, entry.section);
if(m3_t1 != m1_t1)
fatal("DBENGINE METRIC: cannot find the metric added (section %zu)", (size_t)entry.section);
// delete the first metric
mrg_metric_release(mrg, m2_t0);
mrg_metric_release(mrg, m3_t0);
mrg_metric_release(mrg, m4_t0);
mrg_metric_set_first_time_s(mrg, m1_t0, 0);
mrg_metric_set_clean_latest_time_s(mrg, m1_t0, 0);
mrg_metric_set_hot_latest_time_s(mrg, m1_t0, 0);
if(!mrg_metric_release_and_delete(mrg, m1_t0))
fatal("DBENGINE METRIC: cannot delete the first metric");
m4_t1 = mrg_metric_get_and_acquire_by_uuid(mrg, entry.uuid, entry.section);
if(m4_t1 != m1_t1)
fatal("DBENGINE METRIC: cannot find the metric added (section %zu), after deleting the first one", (size_t)entry.section);
// delete the second metric
mrg_metric_release(mrg, m2_t1);
mrg_metric_release(mrg, m3_t1);
mrg_metric_release(mrg, m4_t1);
mrg_metric_set_first_time_s(mrg, m1_t1, 0);
mrg_metric_set_clean_latest_time_s(mrg, m1_t1, 0);
mrg_metric_set_hot_latest_time_s(mrg, m1_t1, 0);
if(!mrg_metric_release_and_delete(mrg, m1_t1))
fatal("DBENGINE METRIC: cannot delete the second metric");
struct mrg_statistics s;
mrg_get_statistics(mrg, &s);
if(s.entries != 0)
fatal("DBENGINE METRIC: invalid entries counter");
size_t entries = 1000000;
size_t threads = _countof(mrg->index) / 3 + 1;
size_t tiers = 3;
size_t run_for_secs = 5;
netdata_log_info("preparing stress test of %zu entries...", entries);
struct mrg_stress t = {
.mrg = mrg,
.entries = entries,
.array = callocz(entries, sizeof(struct mrg_stress_entry)),
};
time_t now = max_acceptable_collected_time();
for(size_t i = 0; i < entries ;i++) {
uuid_generate_random(t.array[i].uuid);
t.array[i].after = now / 3;
t.array[i].before = now / 2;
}
netdata_log_info("stress test is populating MRG with 3 tiers...");
for(size_t i = 0; i < entries ;i++) {
struct mrg_stress_entry *e = &t.array[i];
for(size_t tier = 1; tier <= tiers ;tier++) {
mrg_update_metric_retention_and_granularity_by_uuid(
mrg, tier,
&e->uuid,
e->after,
e->before,
1,
e->before);
}
}
netdata_log_info("stress test ready to run...");
usec_t started_ut = now_monotonic_usec();
ND_THREAD *th[threads];
for(size_t i = 0; i < threads ; i++) {
char buf[15 + 1];
snprintfz(buf, sizeof(buf) - 1, "TH[%zu]", i);
th[i] = nd_thread_create(buf, NETDATA_THREAD_OPTION_JOINABLE | NETDATA_THREAD_OPTION_DONT_LOG, mrg_stress, &t);
}
sleep_usec(run_for_secs * USEC_PER_SEC);
__atomic_store_n(&t.stop, true, __ATOMIC_RELAXED);
for(size_t i = 0; i < threads ; i++)
nd_thread_signal_cancel(th[i]);
for(size_t i = 0; i < threads ; i++)
nd_thread_join(th[i]);
usec_t ended_ut = now_monotonic_usec();
struct mrg_statistics stats;
mrg_get_statistics(mrg, &stats);
netdata_log_info("DBENGINE METRIC: did %zu additions, %zu duplicate additions, "
"%zu deletions, %zu wrong deletions, "
"%zu successful searches, %zu wrong searches, "
"in %"PRIu64" usecs",
stats.additions, stats.additions_duplicate,
stats.deletions, stats.delete_misses,
stats.search_hits, stats.search_misses,
ended_ut - started_ut);
netdata_log_info("DBENGINE METRIC: updates performance: %0.2fk/sec total, %0.2fk/sec/thread",
(double)t.updates / (double)((ended_ut - started_ut) / USEC_PER_SEC) / 1000.0,
(double)t.updates / (double)((ended_ut - started_ut) / USEC_PER_SEC) / 1000.0 / threads);
mrg_destroy(mrg);
netdata_log_info("DBENGINE METRIC: all tests passed!");
return 0;
}