mirror of
https://github.com/netdata/netdata.git
synced 2025-04-30 23:50:04 +00:00
Update change-metrics-storage.md (#15896)
* Update change-metrics-storage.md * Update docs/store/change-metrics-storage.md Co-authored-by: Emmanuel Vasilakis <mrzammler@mm.st> --------- Co-authored-by: Emmanuel Vasilakis <mrzammler@mm.st>
This commit is contained in:
parent
76f9cfd332
commit
968642701e
1 changed files with 12 additions and 12 deletions
|
@ -43,8 +43,8 @@ we will have a data point every minute in tier 1 and every minute in tier 2.
|
||||||
Up to 5 tiers are supported. You may add, or remove tiers and/or modify these multipliers, as long as the
|
Up to 5 tiers are supported. You may add, or remove tiers and/or modify these multipliers, as long as the
|
||||||
product of all the "update every iterations" does not exceed 65535 (number of points for each tier0 point).
|
product of all the "update every iterations" does not exceed 65535 (number of points for each tier0 point).
|
||||||
|
|
||||||
e.g. If you simply add a fourth tier by setting `storage tiers = 4` and defining the disk space for the new tier,
|
e.g. If you simply add a fourth tier by setting `storage tiers = 4` and define the disk space for the new tier,
|
||||||
the product of the "update every iterations" will be 60 * 60 * 60 = 216,000, which is > 65535. So you'd need to reduce
|
the product of the "update every iterations" will be 60 \* 60 \* 60 = 216,000, which is > 65535. So you'd need to reduce
|
||||||
the `update every iterations` of the tiers, to stay under the limit.
|
the `update every iterations` of the tiers, to stay under the limit.
|
||||||
|
|
||||||
The exact retention that can be achieved by each tier depends on the number of metrics collected. The more
|
The exact retention that can be achieved by each tier depends on the number of metrics collected. The more
|
||||||
|
@ -163,6 +163,16 @@ Save the file and restart the Agent with `sudo systemctl restart netdata`, or
|
||||||
the [appropriate method](https://github.com/netdata/netdata/blob/master/docs/configure/start-stop-restart.md)
|
the [appropriate method](https://github.com/netdata/netdata/blob/master/docs/configure/start-stop-restart.md)
|
||||||
for your system, to change the database engine's size.
|
for your system, to change the database engine's size.
|
||||||
|
|
||||||
|
## Scaling dedicated parent nodes
|
||||||
|
|
||||||
|
When you use streaming in medium to large infrastructures, you can have potentially millions of metrics per second reaching each parent node.
|
||||||
|
In the lab we have reliably collected 1 million metrics/sec with 16cores and 32GB RAM.
|
||||||
|
|
||||||
|
Our suggestion for scaling parents is to have them running on dedicated VMs, using a maximum of 50% of cpu, and ensuring you have enough RAM
|
||||||
|
for the desired retention. When your infrastructure can lead a parent to exceed these characteristics, split the load to multiple parents that
|
||||||
|
do not communicate with each other. With each child sending data to only one of the parents, you can still have replication, high availability,
|
||||||
|
and infrastructure level observability via the Netdata Cloud UI.
|
||||||
|
|
||||||
## Legacy configuration
|
## Legacy configuration
|
||||||
|
|
||||||
### v1.35.1 and prior
|
### v1.35.1 and prior
|
||||||
|
@ -195,13 +205,3 @@ All new child nodes are automatically transferred to the multihost dbengine inst
|
||||||
space. If you want to migrate a child node from its legacy dbengine instance to the multihost dbengine instance, you
|
space. If you want to migrate a child node from its legacy dbengine instance to the multihost dbengine instance, you
|
||||||
must delete the instance's directory, which is located in `/var/cache/netdata/MACHINE_GUID/dbengine`, after stopping the
|
must delete the instance's directory, which is located in `/var/cache/netdata/MACHINE_GUID/dbengine`, after stopping the
|
||||||
Agent.
|
Agent.
|
||||||
|
|
||||||
## Scaling dedicated parent nodes
|
|
||||||
|
|
||||||
When you use streaming in medium to large infrastructures, you can have potentially millions of metrics per second reaching each parent node.
|
|
||||||
In the lab we have reliably collected 1 million metrics/sec with 16cores and 32GB RAM.
|
|
||||||
|
|
||||||
Our suggestion for scaling parents is to have them running on dedicated VMs, using a maximum of 50% of cpu, and ensuring you have enough RAM
|
|
||||||
for the desired retention. When your infrastructure can lead a parent to exceed these characteristics, split the load to multiple parents that
|
|
||||||
do not communicate with each other. With each child sending data to only one of the parents, you can still have replication, high availability,
|
|
||||||
and infrastructure level observability via the Netdata Cloud UI.
|
|
||||||
|
|
Loading…
Add table
Add a link
Reference in a new issue