0
0
Fork 0
mirror of https://github.com/netdata/netdata.git synced 2025-04-21 20:23:00 +00:00
netdata_netdata/collectors/python.d.plugin/rabbitmq/rabbitmq.chart.py
Ilya Mashchenko 56336f5acf
py stuff: change l2isbad to ilyam8 ()
* change l2isbad to ilyam8

* change l2isbad to ilyam8
2019-03-21 12:35:44 +03:00

185 lines
5 KiB
Python

# -*- coding: utf-8 -*-
# Description: rabbitmq netdata python.d module
# Author: ilyam8
# SPDX-License-Identifier: GPL-3.0-or-later
from json import loads
from bases.FrameworkServices.UrlService import UrlService
API_NODE = 'api/nodes'
API_OVERVIEW = 'api/overview'
NODE_STATS = [
'fd_used',
'mem_used',
'sockets_used',
'proc_used',
'disk_free',
'run_queue'
]
OVERVIEW_STATS = [
'object_totals.channels',
'object_totals.consumers',
'object_totals.connections',
'object_totals.queues',
'object_totals.exchanges',
'queue_totals.messages_ready',
'queue_totals.messages_unacknowledged',
'message_stats.ack',
'message_stats.redeliver',
'message_stats.deliver',
'message_stats.publish'
]
ORDER = [
'queued_messages',
'message_rates',
'global_counts',
'file_descriptors',
'socket_descriptors',
'erlang_processes',
'erlang_run_queue',
'memory',
'disk_space'
]
CHARTS = {
'file_descriptors': {
'options': [None, 'File Descriptors', 'descriptors', 'overview', 'rabbitmq.file_descriptors', 'line'],
'lines': [
['fd_used', 'used', 'absolute']
]
},
'memory': {
'options': [None, 'Memory', 'MiB', 'overview', 'rabbitmq.memory', 'area'],
'lines': [
['mem_used', 'used', 'absolute', 1, 1 << 20]
]
},
'disk_space': {
'options': [None, 'Disk Space', 'GiB', 'overview', 'rabbitmq.disk_space', 'area'],
'lines': [
['disk_free', 'free', 'absolute', 1, 1 << 30]
]
},
'socket_descriptors': {
'options': [None, 'Socket Descriptors', 'descriptors', 'overview', 'rabbitmq.sockets', 'line'],
'lines': [
['sockets_used', 'used', 'absolute']
]
},
'erlang_processes': {
'options': [None, 'Erlang Processes', 'processes', 'overview', 'rabbitmq.processes', 'line'],
'lines': [
['proc_used', 'used', 'absolute']
]
},
'erlang_run_queue': {
'options': [None, 'Erlang Run Queue', 'processes', 'overview', 'rabbitmq.erlang_run_queue', 'line'],
'lines': [
['run_queue', 'length', 'absolute']
]
},
'global_counts': {
'options': [None, 'Global Counts', 'counts', 'overview', 'rabbitmq.global_counts', 'line'],
'lines': [
['object_totals_channels', 'channels', 'absolute'],
['object_totals_consumers', 'consumers', 'absolute'],
['object_totals_connections', 'connections', 'absolute'],
['object_totals_queues', 'queues', 'absolute'],
['object_totals_exchanges', 'exchanges', 'absolute']
]
},
'queued_messages': {
'options': [None, 'Queued Messages', 'messages', 'overview', 'rabbitmq.queued_messages', 'stacked'],
'lines': [
['queue_totals_messages_ready', 'ready', 'absolute'],
['queue_totals_messages_unacknowledged', 'unacknowledged', 'absolute']
]
},
'message_rates': {
'options': [None, 'Message Rates', 'messages/s', 'overview', 'rabbitmq.message_rates', 'line'],
'lines': [
['message_stats_ack', 'ack', 'incremental'],
['message_stats_redeliver', 'redeliver', 'incremental'],
['message_stats_deliver', 'deliver', 'incremental'],
['message_stats_publish', 'publish', 'incremental']
]
}
}
class Service(UrlService):
def __init__(self, configuration=None, name=None):
UrlService.__init__(self, configuration=configuration, name=name)
self.order = ORDER
self.definitions = CHARTS
self.url = '{0}://{1}:{2}'.format(
configuration.get('scheme', 'http'),
configuration.get('host', '127.0.0.1'),
configuration.get('port', 15672),
)
self.node_name = str()
def _get_data(self):
data = dict()
stats = self.get_overview_stats()
if not stats:
return None
data.update(stats)
stats = self.get_nodes_stats()
if not stats:
return None
data.update(stats)
return data or None
def get_overview_stats(self):
url = '{0}/{1}'.format(self.url, API_OVERVIEW)
raw = self._get_raw_data(url)
if not raw:
return None
data = loads(raw)
self.node_name = data['node']
return fetch_data(raw_data=data, metrics=OVERVIEW_STATS)
def get_nodes_stats(self):
url = '{0}/{1}/{2}'.format(self.url, API_NODE, self.node_name)
raw = self._get_raw_data(url)
if not raw:
return None
data = loads(raw)
return fetch_data(raw_data=data, metrics=NODE_STATS)
def fetch_data(raw_data, metrics):
data = dict()
for metric in metrics:
value = raw_data
metrics_list = metric.split('.')
try:
for m in metrics_list:
value = value[m]
except KeyError:
continue
data['_'.join(metrics_list)] = value
return data