0
0
Fork 0
mirror of https://github.com/netdata/netdata.git synced 2025-04-22 04:31:08 +00:00
netdata_netdata/exporting/mongodb/mongodb.c
2023-07-06 15:46:48 +00:00

392 lines
14 KiB
C

// SPDX-License-Identifier: GPL-3.0-or-later
#define EXPORTING_INTERNALS
#include "mongodb.h"
#define CONFIG_FILE_LINE_MAX ((CONFIG_MAX_NAME + CONFIG_MAX_VALUE + 1024) * 2)
/**
* Initialize MongoDB connector specific data, including a ring buffer
*
* @param instance an instance data structure.
* @return Returns 0 on success, 1 on failure.
*/
int mongodb_init(struct instance *instance)
{
struct mongodb_specific_config *connector_specific_config = instance->config.connector_specific_config;
mongoc_uri_t *uri;
bson_error_t bson_error;
if (unlikely(!connector_specific_config->collection || !*connector_specific_config->collection)) {
netdata_log_error("EXPORTING: collection name is a mandatory MongoDB parameter, but it is not configured");
return 1;
}
uri = mongoc_uri_new_with_error(instance->config.destination, &bson_error);
if (unlikely(!uri)) {
netdata_log_error("EXPORTING: failed to parse URI: %s. Error message: %s",
instance->config.destination,
bson_error.message);
return 1;
}
int32_t socket_timeout =
mongoc_uri_get_option_as_int32(uri, MONGOC_URI_SOCKETTIMEOUTMS, instance->config.timeoutms);
if (!mongoc_uri_set_option_as_int32(uri, MONGOC_URI_SOCKETTIMEOUTMS, socket_timeout)) {
netdata_log_error("EXPORTING: failed to set %s to the value %d", MONGOC_URI_SOCKETTIMEOUTMS, socket_timeout);
return 1;
};
struct mongodb_specific_data *connector_specific_data =
(struct mongodb_specific_data *)instance->connector_specific_data;
connector_specific_data->client = mongoc_client_new_from_uri(uri);
if (unlikely(!connector_specific_data->client)) {
netdata_log_error("EXPORTING: failed to create a new client");
return 1;
}
if (!mongoc_client_set_appname(connector_specific_data->client, "netdata")) {
netdata_log_error("EXPORTING: failed to set client appname");
};
connector_specific_data->collection = mongoc_client_get_collection(
connector_specific_data->client, connector_specific_config->database, connector_specific_config->collection);
mongoc_uri_destroy(uri);
// create a ring buffer
struct bson_buffer *first_buffer = NULL;
if (instance->config.buffer_on_failures < 2)
instance->config.buffer_on_failures = 1;
else
instance->config.buffer_on_failures -= 1;
for (int i = 0; i < instance->config.buffer_on_failures; i++) {
struct bson_buffer *current_buffer = callocz(1, sizeof(struct bson_buffer));
if (!connector_specific_data->first_buffer)
first_buffer = current_buffer;
else
current_buffer->next = connector_specific_data->first_buffer;
connector_specific_data->first_buffer = current_buffer;
}
first_buffer->next = connector_specific_data->first_buffer;
connector_specific_data->last_buffer = connector_specific_data->first_buffer;
return 0;
}
/**
* Initialize a MongoDB connector instance
*
* @param instance an instance data structure.
* @return Returns 0 on success, 1 on failure.
*/
int init_mongodb_instance(struct instance *instance)
{
instance->worker = mongodb_connector_worker;
instance->start_batch_formatting = NULL;
instance->start_host_formatting = format_host_labels_json_plaintext;
instance->start_chart_formatting = NULL;
if (EXPORTING_OPTIONS_DATA_SOURCE(instance->config.options) == EXPORTING_SOURCE_DATA_AS_COLLECTED)
instance->metric_formatting = format_dimension_collected_json_plaintext;
else
instance->metric_formatting = format_dimension_stored_json_plaintext;
instance->end_chart_formatting = NULL;
instance->variables_formatting = NULL;
instance->end_host_formatting = flush_host_labels;
instance->end_batch_formatting = format_batch_mongodb;
instance->prepare_header = NULL;
instance->check_response = NULL;
instance->buffer = (void *)buffer_create(0, &netdata_buffers_statistics.buffers_exporters);
if (!instance->buffer) {
netdata_log_error("EXPORTING: cannot create buffer for MongoDB exporting connector instance %s",
instance->config.name);
return 1;
}
if (uv_mutex_init(&instance->mutex))
return 1;
if (uv_cond_init(&instance->cond_var))
return 1;
struct mongodb_specific_data *connector_specific_data = callocz(1, sizeof(struct mongodb_specific_data));
instance->connector_specific_data = (void *)connector_specific_data;
instance->config.timeoutms =
(instance->config.update_every >= 2) ? (instance->engine->config.update_every * MSEC_PER_SEC - 500) : 1000;
if (!instance->engine->mongoc_initialized) {
mongoc_init();
instance->engine->mongoc_initialized = 1;
}
if (unlikely(mongodb_init(instance))) {
netdata_log_error("EXPORTING: cannot initialize MongoDB exporting connector");
return 1;
}
return 0;
}
/**
* Free an array of BSON structures
*
* @param insert an array of documents.
* @param documents_inserted the number of documents inserted.
*/
void free_bson(bson_t **insert, size_t documents_inserted)
{
size_t i;
for (i = 0; i < documents_inserted; i++)
bson_destroy(insert[i]);
freez(insert);
}
/**
* Format a batch for the MongoDB connector
*
* @param instance an instance data structure.
* @return Returns 0 on success, 1 on failure.
*/
int format_batch_mongodb(struct instance *instance)
{
struct mongodb_specific_data *connector_specific_data =
(struct mongodb_specific_data *)instance->connector_specific_data;
struct stats *stats = &instance->stats;
bson_t **insert = connector_specific_data->last_buffer->insert;
if (insert) {
// ring buffer is full, reuse the oldest element
connector_specific_data->first_buffer = connector_specific_data->first_buffer->next;
free_bson(insert, connector_specific_data->last_buffer->documents_inserted);
connector_specific_data->total_documents_inserted -= connector_specific_data->last_buffer->documents_inserted;
stats->buffered_bytes -= connector_specific_data->last_buffer->buffered_bytes;
}
insert = callocz((size_t)stats->buffered_metrics, sizeof(bson_t *));
connector_specific_data->last_buffer->insert = insert;
BUFFER *buffer = (BUFFER *)instance->buffer;
char *start = (char *)buffer_tostring(buffer);
char *end = start;
size_t documents_inserted = 0;
while (*end && documents_inserted <= (size_t)stats->buffered_metrics) {
while (*end && *end != '\n')
end++;
if (likely(*end)) {
*end = '\0';
end++;
} else {
break;
}
bson_error_t bson_error;
insert[documents_inserted] = bson_new_from_json((const uint8_t *)start, -1, &bson_error);
if (unlikely(!insert[documents_inserted])) {
netdata_log_error(
"EXPORTING: Failed creating a BSON document from a JSON string \"%s\" : %s", start, bson_error.message);
free_bson(insert, documents_inserted);
return 1;
}
start = end;
documents_inserted++;
}
stats->buffered_bytes += connector_specific_data->last_buffer->buffered_bytes = buffer_strlen(buffer);
buffer_flush(buffer);
// The stats->buffered_metrics is used in the MongoDB batch formatting as a variable for the number
// of metrics, added in the current iteration, so we are clearing it here. We will use the
// connector_specific_data->total_documents_inserted in the worker to show the statistics.
stats->buffered_metrics = 0;
connector_specific_data->total_documents_inserted += documents_inserted;
connector_specific_data->last_buffer->documents_inserted = documents_inserted;
connector_specific_data->last_buffer = connector_specific_data->last_buffer->next;
return 0;
}
/**
* Clean a MongoDB connector instance up
*
* @param instance an instance data structure.
*/
void mongodb_cleanup(struct instance *instance)
{
netdata_log_info("EXPORTING: cleaning up instance %s ...", instance->config.name);
struct mongodb_specific_data *connector_specific_data =
(struct mongodb_specific_data *)instance->connector_specific_data;
mongoc_collection_destroy(connector_specific_data->collection);
mongoc_client_destroy(connector_specific_data->client);
if (instance->engine->mongoc_initialized) {
mongoc_cleanup();
instance->engine->mongoc_initialized = 0;
}
buffer_free(instance->buffer);
struct bson_buffer *next_buffer = connector_specific_data->first_buffer;
for (int i = 0; i < instance->config.buffer_on_failures; i++) {
struct bson_buffer *current_buffer = next_buffer;
next_buffer = next_buffer->next;
if (current_buffer->insert)
free_bson(current_buffer->insert, current_buffer->documents_inserted);
freez(current_buffer);
}
freez(connector_specific_data);
struct mongodb_specific_config *connector_specific_config =
(struct mongodb_specific_config *)instance->config.connector_specific_config;
freez(connector_specific_config->database);
freez(connector_specific_config->collection);
freez(connector_specific_config);
netdata_log_info("EXPORTING: instance %s exited", instance->config.name);
instance->exited = 1;
return;
}
/**
* MongoDB connector worker
*
* Runs in a separate thread for every instance.
*
* @param instance_p an instance data structure.
*/
void mongodb_connector_worker(void *instance_p)
{
struct instance *instance = (struct instance *)instance_p;
#ifdef NETDATA_INTERNAL_CHECKS
struct mongodb_specific_config *connector_specific_config = instance->config.connector_specific_config;
#endif
struct mongodb_specific_data *connector_specific_data =
(struct mongodb_specific_data *)instance->connector_specific_data;
while (!instance->engine->exit) {
struct stats *stats = &instance->stats;
uv_mutex_lock(&instance->mutex);
if (!connector_specific_data->first_buffer->insert ||
!connector_specific_data->first_buffer->documents_inserted) {
while (!instance->data_is_ready)
uv_cond_wait(&instance->cond_var, &instance->mutex);
instance->data_is_ready = 0;
}
if (unlikely(instance->engine->exit)) {
uv_mutex_unlock(&instance->mutex);
break;
}
// reset the monitoring chart counters
stats->received_bytes =
stats->sent_bytes =
stats->sent_metrics =
stats->lost_metrics =
stats->receptions =
stats->transmission_successes =
stats->transmission_failures =
stats->data_lost_events =
stats->lost_bytes =
stats->reconnects = 0;
bson_t **insert = connector_specific_data->first_buffer->insert;
size_t documents_inserted = connector_specific_data->first_buffer->documents_inserted;
size_t buffered_bytes = connector_specific_data->first_buffer->buffered_bytes;
connector_specific_data->first_buffer->insert = NULL;
connector_specific_data->first_buffer->documents_inserted = 0;
connector_specific_data->first_buffer->buffered_bytes = 0;
connector_specific_data->first_buffer = connector_specific_data->first_buffer->next;
uv_mutex_unlock(&instance->mutex);
size_t data_size = 0;
for (size_t i = 0; i < documents_inserted; i++) {
data_size += insert[i]->len;
}
debug(
D_EXPORTING,
"EXPORTING: mongodb_insert(): destination = %s, database = %s, collection = %s, data size = %zu",
instance->config.destination,
connector_specific_config->database,
connector_specific_config->collection,
data_size);
if (likely(documents_inserted != 0)) {
bson_error_t bson_error;
if (likely(mongoc_collection_insert_many(
connector_specific_data->collection,
(const bson_t **)insert,
documents_inserted,
NULL,
NULL,
&bson_error))) {
stats->sent_metrics = documents_inserted;
stats->sent_bytes += data_size;
stats->transmission_successes++;
stats->receptions++;
} else {
// oops! we couldn't send (all or some of the) data
netdata_log_error("EXPORTING: %s", bson_error.message);
netdata_log_error(
"EXPORTING: failed to write data to the database '%s'. "
"Willing to write %zu bytes, wrote %zu bytes.",
instance->config.destination, data_size, 0UL);
stats->transmission_failures++;
stats->data_lost_events++;
stats->lost_bytes += buffered_bytes;
stats->lost_metrics += documents_inserted;
}
}
free_bson(insert, documents_inserted);
if (unlikely(instance->engine->exit))
break;
uv_mutex_lock(&instance->mutex);
stats->buffered_metrics = connector_specific_data->total_documents_inserted;
send_internal_metrics(instance);
connector_specific_data->total_documents_inserted -= documents_inserted;
stats->buffered_metrics = 0;
stats->buffered_bytes -= buffered_bytes;
uv_mutex_unlock(&instance->mutex);
#ifdef UNIT_TESTING
return;
#endif
}
mongodb_cleanup(instance);
}