0
0
Fork 0
mirror of https://github.com/netdata/netdata.git synced 2025-04-15 10:04:15 +00:00
netdata_netdata/collectors/cups.plugin/cups_plugin.c
Costa Tsaousis 3e508c8f95
New logging layer ()
* cleanup of logging - wip

* first working iteration

* add errno annotator

* replace old logging functions with netdata_logger()

* cleanup

* update error_limit

* fix remanining error_limit references

* work on fatal()

* started working on structured logs

* full cleanup

* default logging to files; fix all plugins initialization

* fix formatting of numbers

* cleanup and reorg

* fix coverity issues

* cleanup obsolete code

* fix formatting of numbers

* fix log rotation

* fix for older systems

* add detection of systemd journal via stderr

* finished on access.log

* remove left-over transport

* do not add empty fields to the logs

* journal get compact uuids; X-Transaction-ID header is added in web responses

* allow compiling on systems without memfd sealing

* added libnetdata/uuid directory

* move datetime formatters to libnetdata

* add missing files

* link the makefiles in libnetdata

* added uuid_parse_flexi() to parse UUIDs with and without hyphens; the web server now read X-Transaction-ID and uses it for functions and web responses

* added stream receiver, sender, proc plugin and pluginsd log stack

* iso8601 advanced usage; line_splitter module in libnetdata; code cleanup

* add message ids to streaming inbound and outbound connections

* cleanup line_splitter between lines to avoid logging garbage; when killing children, kill them with SIGABRT if internal checks is enabled

* send SIGABRT to external plugins only if we are not shutting down

* fix cross cleanup in pluginsd parser

* fatal when there is a stack error in logs

* compile netdata with -fexceptions

* do not kill external plugins with SIGABRT

* metasync info logs to debug level

* added severity to logs

* added json output; added options per log output; added documentation; fixed issues mentioned

* allow memfd only on linux

* moved journal low level functions to journal.c/h

* move health logs to daemon.log with proper priorities

* fixed a couple of bugs; health log in journal

* updated docs

* systemd-cat-native command to push structured logs to journal from the command line

* fix makefiles

* restored NETDATA_LOG_SEVERITY_LEVEL

* fix makefiles

* systemd-cat-native can also work as the logger of Netdata scripts

* do not require a socket to systemd-journal to log-as-netdata

* alarm notify logs in native format

* properly compare log ids

* fatals log alerts; alarm-notify.sh working

* fix overflow warning

* alarm-notify.sh now logs the request (command line)

* anotate external plugins logs with the function cmd they run

* added context, component and type to alarm-notify.sh; shell sanitization removes control character and characters that may be expanded by bash

* reformatted alarm-notify logs

* unify cgroup-network-helper.sh

* added quotes around params

* charts.d.plugin switched logging to journal native

* quotes for logfmt

* unify the status codes of streaming receivers and senders

* alarm-notify: dont log anything, if there is nothing to do

* all external plugins log to stderr when running outside netdata; alarm-notify now shows an error when notifications menthod are needed but are not available

* migrate cgroup-name.sh to new logging

* systemd-cat-native now supports messages with newlines

* socket.c logs use priority

* cleanup log field types

* inherit the systemd set INVOCATION_ID if found

* allow systemd-cat-native to send messages to a systemd-journal-remote URL

* log2journal command that can convert structured logs to journal export format

* various fixes and documentation of log2journal

* updated log2journal docs

* updated log2journal docs

* updated documentation of fields

* allow compiling without libcurl

* do not use socket as format string

* added version information to newly added tools

* updated documentation and help messages

* fix the namespace socket path

* print errno with error

* do not timeout

* updated docs

* updated docs

* updated docs

* log2journal updated docs and params

* when talking to a remote journal, systemd-cat-native batches the messages

* enable lz4 compression for systemd-cat-native when sending messages to a systemd-journal-remote

* Revert "enable lz4 compression for systemd-cat-native when sending messages to a systemd-journal-remote"

This reverts commit b079d53c11.

* note about uncompressed traffic

* log2journal: code reorg and cleanup to make modular

* finished rewriting log2journal

* more comments

* rewriting rules support

* increased limits

* updated docs

* updated docs

* fix old log call

* use journal only when stderr is connected to journal

* update netdata.spec for libcurl, libpcre2 and log2journal

* pcre2-devel

* do not require pcre2 in centos < 8, amazonlinux < 2023, open suse

* log2journal only on systems pcre2 is available

* ignore log2journal in .gitignore

* avoid log2journal on centos 7, amazonlinux 2 and opensuse

* add pcre2-8 to static build

* undo last commit

* Bundle to static

Signed-off-by: Tasos Katsoulas <tasos@netdata.cloud>

* Add build deps for deb packages

Signed-off-by: Tasos Katsoulas <tasos@netdata.cloud>

* Add dependencies; build from source

Signed-off-by: Tasos Katsoulas <tasos@netdata.cloud>

* Test build for amazon linux and centos expect to fail for suse

Signed-off-by: Tasos Katsoulas <tasos@netdata.cloud>

* fix minor oversight

Signed-off-by: Tasos Katsoulas <tasos@netdata.cloud>

* Reorg code

* Add the install from source (deps) as a TODO
* Not enable the build on suse ecosystem

Signed-off-by: Tasos Katsoulas <tasos@netdata.cloud>

---------

Signed-off-by: Tasos Katsoulas <tasos@netdata.cloud>
Co-authored-by: Tasos Katsoulas <tasos@netdata.cloud>
2023-11-22 10:27:25 +02:00

430 lines
14 KiB
C

// SPDX-License-Identifier: GPL-3.0-or-later
/*
* netdata cups.plugin
* (C) Copyright 2017-2018 Simon Nagl <simon.nagl@gmx.de>
* Released under GPL v3+
*/
#include "libnetdata/libnetdata.h"
#include "libnetdata/required_dummies.h"
#include <cups/cups.h>
#include <limits.h>
// Variables
static int debug = 0;
static int netdata_update_every = 1;
static uint32_t netdata_priority = 100004;
http_t *http; // connection to the cups daemon
/*
* Used to aggregate job metrics for a destination (and all destinations).
*/
struct job_metrics {
uint32_t id;
bool is_collected; // flag if this was collected in the current cycle
int num_pending;
int num_processing;
int num_held;
int size_pending; // in kilobyte
int size_processing; // in kilobyte
int size_held; // in kilobyte
};
DICTIONARY *dict_dest_job_metrics = NULL;
struct job_metrics global_job_metrics;
int num_dest_total;
int num_dest_accepting_jobs;
int num_dest_shared;
int num_dest_idle;
int num_dest_printing;
int num_dest_stopped;
void print_help() {
fprintf(stderr,
"\n"
"netdata cups.plugin %s\n"
"\n"
"Copyright (C) 2017-2018 Simon Nagl <simon.nagl@gmx.de>\n"
"Released under GNU General Public License v3+.\n"
"All rights reserved.\n"
"\n"
"This program is a data collector plugin for netdata.\n"
"\n"
"SYNOPSIS: cups.plugin [-d][-h][-v] COLLECTION_FREQUENCY\n"
"\n"
"Options:"
"\n"
" COLLECTION_FREQUENCY data collection frequency in seconds\n"
"\n"
" -d enable verbose output\n"
" default: disabled\n"
"\n"
" -v print version and exit\n"
"\n"
" -h print this message and exit\n"
"\n",
VERSION);
}
void parse_command_line(int argc, char **argv) {
int i;
int freq = 0;
int update_every_found = 0;
for (i = 1; i < argc; i++) {
if (isdigit(*argv[i]) && !update_every_found) {
int n = str2i(argv[i]);
if (n > 0 && n < 86400) {
freq = n;
continue;
}
} else if (strcmp("-v", argv[i]) == 0) {
printf("cups.plugin %s\n", VERSION);
exit(0);
} else if (strcmp("-d", argv[i]) == 0) {
debug = 1;
continue;
} else if (strcmp("-h", argv[i]) == 0) {
print_help();
exit(0);
}
print_help();
exit(1);
}
if (freq >= netdata_update_every) {
netdata_update_every = freq;
} else if (freq) {
netdata_log_error("update frequency %d seconds is too small for CUPS. Using %d.", freq, netdata_update_every);
}
}
/*
* 'cupsGetIntegerOption()' - Get an integer option value.
*
* INT_MIN is returned when the option does not exist, is not an integer, or
* exceeds the range of values for the "int" type.
*
* @since CUPS 2.2.4/macOS 10.13@
*/
int /* O - Option value or @code INT_MIN@ */
getIntegerOption(
const char *name, /* I - Name of option */
int num_options, /* I - Number of options */
cups_option_t *options) /* I - Options */
{
const char *value = cupsGetOption(name, num_options, options);
/* String value of option */
char *ptr; /* Pointer into string value */
long intvalue; /* Integer value */
if (!value || !*value)
return (INT_MIN);
intvalue = strtol(value, &ptr, 10);
if (intvalue < INT_MIN || intvalue > INT_MAX || *ptr)
return (INT_MIN);
return ((int)intvalue);
}
static int reset_job_metrics(const DICTIONARY_ITEM *item __maybe_unused, void *entry, void *data __maybe_unused) {
struct job_metrics *jm = (struct job_metrics *)entry;
jm->is_collected = false;
jm->num_held = 0;
jm->num_pending = 0;
jm->num_processing = 0;
jm->size_held = 0;
jm->size_pending = 0;
jm->size_processing = 0;
return 0;
}
void send_job_charts_definitions_to_netdata(const char *name, uint32_t job_id, bool obsolete) {
printf("CHART cups.job_num_%s '' 'Active jobs of %s' jobs '%s' cups.destination_job_num stacked %u %i %s\n",
name, name, name, netdata_priority + job_id, netdata_update_every, obsolete?"obsolete":"");
printf("DIMENSION pending '' absolute 1 1\n");
printf("DIMENSION held '' absolute 1 1\n");
printf("DIMENSION processing '' absolute 1 1\n");
printf("CHART cups.job_size_%s '' 'Active jobs size of %s' KB '%s' cups.destination_job_size stacked %u %i %s\n",
name, name, name, netdata_priority + 1 + job_id, netdata_update_every, obsolete?"obsolete":"");
printf("DIMENSION pending '' absolute 1 1\n");
printf("DIMENSION held '' absolute 1 1\n");
printf("DIMENSION processing '' absolute 1 1\n");
}
struct job_metrics *get_job_metrics(char *dest) {
struct job_metrics *jm = dictionary_get(dict_dest_job_metrics, dest);
if (unlikely(!jm)) {
static uint32_t job_id = 0;
struct job_metrics new_job_metrics = { .id = ++job_id };
jm = dictionary_set(dict_dest_job_metrics, dest, &new_job_metrics, sizeof(struct job_metrics));
send_job_charts_definitions_to_netdata(dest, jm->id, false);
};
return jm;
}
int send_job_metrics_to_netdata(const DICTIONARY_ITEM *item, void *entry, void *data __maybe_unused) {
const char *name = dictionary_acquired_item_name(item);
struct job_metrics *jm = (struct job_metrics *)entry;
if (jm->is_collected) {
printf(
"BEGIN cups.job_num_%s\n"
"SET pending = %d\n"
"SET held = %d\n"
"SET processing = %d\n"
"END\n",
name, jm->num_pending, jm->num_held, jm->num_processing);
printf(
"BEGIN cups.job_size_%s\n"
"SET pending = %d\n"
"SET held = %d\n"
"SET processing = %d\n"
"END\n",
name, jm->size_pending, jm->size_held, jm->size_processing);
}
else {
// mark it obsolete
send_job_charts_definitions_to_netdata(name, jm->id, true);
// delete it
dictionary_del(dict_dest_job_metrics, name);
}
return 0;
}
void reset_metrics() {
num_dest_total = 0;
num_dest_accepting_jobs = 0;
num_dest_shared = 0;
num_dest_idle = 0;
num_dest_printing = 0;
num_dest_stopped = 0;
reset_job_metrics(NULL, &global_job_metrics, NULL);
dictionary_walkthrough_write(dict_dest_job_metrics, reset_job_metrics, NULL);
}
int main(int argc, char **argv) {
clocks_init();
nd_log_initialize_for_external_plugins("cups.plugin");
parse_command_line(argc, argv);
errno = 0;
dict_dest_job_metrics = dictionary_create(DICT_OPTION_SINGLE_THREADED);
// ------------------------------------------------------------------------
// the main loop
if (debug)
fprintf(stderr, "starting data collection\n");
time_t started_t = now_monotonic_sec();
size_t iteration = 0;
usec_t step = netdata_update_every * USEC_PER_SEC;
heartbeat_t hb;
heartbeat_init(&hb);
for (iteration = 0; 1; iteration++) {
heartbeat_next(&hb, step);
if (unlikely(netdata_exit))
break;
reset_metrics();
cups_dest_t *dests;
num_dest_total = cupsGetDests2(http, &dests);
if(unlikely(num_dest_total == 0)) {
// reconnect to cups to check if the server is down.
httpClose(http);
http = httpConnect2(cupsServer(), ippPort(), NULL, AF_UNSPEC, cupsEncryption(), 0, netdata_update_every * 1000, NULL);
if(http == NULL) {
netdata_log_error("cups daemon is not running. Exiting!");
exit(1);
}
}
cups_dest_t *curr_dest = dests;
int counter = 0;
while (counter < num_dest_total) {
if (counter != 0) {
curr_dest++;
}
counter++;
const char *printer_uri_supported = cupsGetOption("printer-uri-supported", curr_dest->num_options, curr_dest->options);
if (!printer_uri_supported) {
if(debug)
fprintf(stderr, "destination %s discovered, but not yet setup as a local printer", curr_dest->name);
continue;
}
const char *printer_is_accepting_jobs = cupsGetOption("printer-is-accepting-jobs", curr_dest->num_options, curr_dest->options);
if (printer_is_accepting_jobs && !strcmp(printer_is_accepting_jobs, "true")) {
num_dest_accepting_jobs++;
}
const char *printer_is_shared = cupsGetOption("printer-is-shared", curr_dest->num_options, curr_dest->options);
if (printer_is_shared && !strcmp(printer_is_shared, "true")) {
num_dest_shared++;
}
int printer_state = getIntegerOption("printer-state", curr_dest->num_options, curr_dest->options);
switch (printer_state) {
case 3:
num_dest_idle++;
break;
case 4:
num_dest_printing++;
break;
case 5:
num_dest_stopped++;
break;
case INT_MIN:
if(debug)
fprintf(stderr, "printer state is missing for destination %s", curr_dest->name);
break;
default:
netdata_log_error("Unknown printer state (%d) found.", printer_state);
break;
}
/*
* flag job metrics to print values.
* This is needed to report also destinations with zero active jobs.
*/
struct job_metrics *jm = get_job_metrics(curr_dest->name);
jm->is_collected = true;
}
cupsFreeDests(num_dest_total, dests);
if (unlikely(netdata_exit))
break;
cups_job_t *jobs, *curr_job;
int num_jobs = cupsGetJobs2(http, &jobs, NULL, 0, CUPS_WHICHJOBS_ACTIVE);
int i;
for (i = num_jobs, curr_job = jobs; i > 0; i--, curr_job++) {
struct job_metrics *jm = get_job_metrics(curr_job->dest);
jm->is_collected = true;
switch (curr_job->state) {
case IPP_JOB_PENDING:
jm->num_pending++;
jm->size_pending += curr_job->size;
global_job_metrics.num_pending++;
global_job_metrics.size_pending += curr_job->size;
break;
case IPP_JOB_HELD:
jm->num_held++;
jm->size_held += curr_job->size;
global_job_metrics.num_held++;
global_job_metrics.size_held += curr_job->size;
break;
case IPP_JOB_PROCESSING:
jm->num_processing++;
jm->size_processing += curr_job->size;
global_job_metrics.num_processing++;
global_job_metrics.size_processing += curr_job->size;
break;
default:
netdata_log_error("Unsupported job state (%u) found.", curr_job->state);
break;
}
}
cupsFreeJobs(num_jobs, jobs);
dictionary_walkthrough_write(dict_dest_job_metrics, send_job_metrics_to_netdata, NULL);
dictionary_garbage_collect(dict_dest_job_metrics);
static int cups_printer_by_option_created = 0;
if (unlikely(!cups_printer_by_option_created))
{
cups_printer_by_option_created = 1;
printf("CHART cups.dest_state '' 'Destinations by state' dests overview cups.dests_state stacked 100000 %i\n", netdata_update_every);
printf("DIMENSION idle '' absolute 1 1\n");
printf("DIMENSION printing '' absolute 1 1\n");
printf("DIMENSION stopped '' absolute 1 1\n");
printf("CHART cups.dest_option '' 'Destinations by option' dests overview cups.dests_option line 100001 %i\n", netdata_update_every);
printf("DIMENSION total '' absolute 1 1\n");
printf("DIMENSION acceptingjobs '' absolute 1 1\n");
printf("DIMENSION shared '' absolute 1 1\n");
printf("CHART cups.job_num '' 'Active jobs' jobs overview cups.job_num stacked 100002 %i\n", netdata_update_every);
printf("DIMENSION pending '' absolute 1 1\n");
printf("DIMENSION held '' absolute 1 1\n");
printf("DIMENSION processing '' absolute 1 1\n");
printf("CHART cups.job_size '' 'Active jobs size' KB overview cups.job_size stacked 100003 %i\n", netdata_update_every);
printf("DIMENSION pending '' absolute 1 1\n");
printf("DIMENSION held '' absolute 1 1\n");
printf("DIMENSION processing '' absolute 1 1\n");
}
printf(
"BEGIN cups.dest_state\n"
"SET idle = %d\n"
"SET printing = %d\n"
"SET stopped = %d\n"
"END\n",
num_dest_idle, num_dest_printing, num_dest_stopped);
printf(
"BEGIN cups.dest_option\n"
"SET total = %d\n"
"SET acceptingjobs = %d\n"
"SET shared = %d\n"
"END\n",
num_dest_total, num_dest_accepting_jobs, num_dest_shared);
printf(
"BEGIN cups.job_num\n"
"SET pending = %d\n"
"SET held = %d\n"
"SET processing = %d\n"
"END\n",
global_job_metrics.num_pending, global_job_metrics.num_held, global_job_metrics.num_processing);
printf(
"BEGIN cups.job_size\n"
"SET pending = %d\n"
"SET held = %d\n"
"SET processing = %d\n"
"END\n",
global_job_metrics.size_pending, global_job_metrics.size_held, global_job_metrics.size_processing);
fflush(stdout);
if (unlikely(netdata_exit))
break;
// restart check (14400 seconds)
if (!now_monotonic_sec() - started_t > 14400)
break;
}
httpClose(http);
netdata_log_info("CUPS process exiting");
}