From f4193c3b5c013df00b6d05805bf1cc99bebe02bf Mon Sep 17 00:00:00 2001
From: Josh Soref <jsoref@users.noreply.github.com>
Date: Mon, 18 Jan 2021 07:43:43 -0500
Subject: [PATCH] Spelling md (#10508)

* spelling: activity

Signed-off-by: Josh Soref <jsoref@users.noreply.github.com>

* spelling: adding

Signed-off-by: Josh Soref <jsoref@users.noreply.github.com>

* spelling: addresses

Signed-off-by: Josh Soref <jsoref@users.noreply.github.com>

* spelling: administrators

Signed-off-by: Josh Soref <jsoref@users.noreply.github.com>

* spelling: alarm

Signed-off-by: Josh Soref <jsoref@users.noreply.github.com>

* spelling: alignment

Signed-off-by: Josh Soref <jsoref@users.noreply.github.com>

* spelling: analyzing

Signed-off-by: Josh Soref <jsoref@users.noreply.github.com>

* spelling: apcupsd

Signed-off-by: Josh Soref <jsoref@users.noreply.github.com>

* spelling: apply

Signed-off-by: Josh Soref <jsoref@users.noreply.github.com>

* spelling: around

Signed-off-by: Josh Soref <jsoref@users.noreply.github.com>

* spelling: associated

Signed-off-by: Josh Soref <jsoref@users.noreply.github.com>

* spelling: automatically

Signed-off-by: Josh Soref <jsoref@users.noreply.github.com>

* spelling: availability

Signed-off-by: Josh Soref <jsoref@users.noreply.github.com>

* spelling: background

Signed-off-by: Josh Soref <jsoref@users.noreply.github.com>

* spelling: bandwidth

Signed-off-by: Josh Soref <jsoref@users.noreply.github.com>

* spelling: berkeley

Signed-off-by: Josh Soref <jsoref@users.noreply.github.com>

* spelling: between

Signed-off-by: Josh Soref <jsoref@users.noreply.github.com>

* spelling: celsius

Signed-off-by: Josh Soref <jsoref@users.noreply.github.com>

* spelling: centos

Signed-off-by: Josh Soref <jsoref@users.noreply.github.com>

* spelling: certificate

Signed-off-by: Josh Soref <jsoref@users.noreply.github.com>

* spelling: cockroach

Signed-off-by: Josh Soref <jsoref@users.noreply.github.com>

* spelling: collectors

Signed-off-by: Josh Soref <jsoref@users.noreply.github.com>

* spelling: concatenation

Signed-off-by: Josh Soref <jsoref@users.noreply.github.com>

* spelling: configuration

Signed-off-by: Josh Soref <jsoref@users.noreply.github.com>

* spelling: configured

Signed-off-by: Josh Soref <jsoref@users.noreply.github.com>

* spelling: continuous

Signed-off-by: Josh Soref <jsoref@users.noreply.github.com>

* spelling: correctly

Signed-off-by: Josh Soref <jsoref@users.noreply.github.com>

* spelling: corresponding

Signed-off-by: Josh Soref <jsoref@users.noreply.github.com>

* spelling: cyberpower

Signed-off-by: Josh Soref <jsoref@users.noreply.github.com>

* spelling: daemon

Signed-off-by: Josh Soref <jsoref@users.noreply.github.com>

* spelling: dashboard

Signed-off-by: Josh Soref <jsoref@users.noreply.github.com>

* spelling: database

Signed-off-by: Josh Soref <jsoref@users.noreply.github.com>

* spelling: deactivating

Signed-off-by: Josh Soref <jsoref@users.noreply.github.com>

* spelling: dependencies

Signed-off-by: Josh Soref <jsoref@users.noreply.github.com>

* spelling: deployment

Signed-off-by: Josh Soref <jsoref@users.noreply.github.com>

* spelling: determine

Signed-off-by: Josh Soref <jsoref@users.noreply.github.com>

* spelling: downloading

Signed-off-by: Josh Soref <jsoref@users.noreply.github.com>

* spelling: either

Signed-off-by: Josh Soref <jsoref@users.noreply.github.com>

* spelling: electric

Signed-off-by: Josh Soref <jsoref@users.noreply.github.com>

* spelling: entity

Signed-off-by: Josh Soref <jsoref@users.noreply.github.com>

* spelling: entrant

Signed-off-by: Josh Soref <jsoref@users.noreply.github.com>

* spelling: enumerating

Signed-off-by: Josh Soref <jsoref@users.noreply.github.com>

* spelling: environment

Signed-off-by: Josh Soref <jsoref@users.noreply.github.com>

* spelling: equivalent

Signed-off-by: Josh Soref <jsoref@users.noreply.github.com>

* spelling: etsy

Signed-off-by: Josh Soref <jsoref@users.noreply.github.com>

* spelling: everything

Signed-off-by: Josh Soref <jsoref@users.noreply.github.com>

* spelling: examining

Signed-off-by: Josh Soref <jsoref@users.noreply.github.com>

* spelling: expectations

Signed-off-by: Josh Soref <jsoref@users.noreply.github.com>

* spelling: explicit

Signed-off-by: Josh Soref <jsoref@users.noreply.github.com>

* spelling: explicitly

Signed-off-by: Josh Soref <jsoref@users.noreply.github.com>

* spelling: finally

Signed-off-by: Josh Soref <jsoref@users.noreply.github.com>

* spelling: flexible

Signed-off-by: Josh Soref <jsoref@users.noreply.github.com>

* spelling: further

Signed-off-by: Josh Soref <jsoref@users.noreply.github.com>

* spelling: hddtemp

Signed-off-by: Josh Soref <jsoref@users.noreply.github.com>

* spelling: humidity

Signed-off-by: Josh Soref <jsoref@users.noreply.github.com>

* spelling: identify

Signed-off-by: Josh Soref <jsoref@users.noreply.github.com>

* spelling: importance

Signed-off-by: Josh Soref <jsoref@users.noreply.github.com>

* spelling: incoming

Signed-off-by: Josh Soref <jsoref@users.noreply.github.com>

* spelling: individual

Signed-off-by: Josh Soref <jsoref@users.noreply.github.com>

* spelling: initiate

Signed-off-by: Josh Soref <jsoref@users.noreply.github.com>

* spelling: installation

Signed-off-by: Josh Soref <jsoref@users.noreply.github.com>

* spelling: integration

Signed-off-by: Josh Soref <jsoref@users.noreply.github.com>

* spelling: integrity

Signed-off-by: Josh Soref <jsoref@users.noreply.github.com>

* spelling: involuntary

Signed-off-by: Josh Soref <jsoref@users.noreply.github.com>

* spelling: issues

Signed-off-by: Josh Soref <jsoref@users.noreply.github.com>

* spelling: kernel

Signed-off-by: Josh Soref <jsoref@users.noreply.github.com>

* spelling: language

Signed-off-by: Josh Soref <jsoref@users.noreply.github.com>

* spelling: libwebsockets

Signed-off-by: Josh Soref <jsoref@users.noreply.github.com>

* spelling: lighttpd

Signed-off-by: Josh Soref <jsoref@users.noreply.github.com>

* spelling: maintained

Signed-off-by: Josh Soref <jsoref@users.noreply.github.com>

* spelling: meaningful

Signed-off-by: Josh Soref <jsoref@users.noreply.github.com>

* spelling: memory

Signed-off-by: Josh Soref <jsoref@users.noreply.github.com>

* spelling: metrics

Signed-off-by: Josh Soref <jsoref@users.noreply.github.com>

* spelling: miscellaneous

Signed-off-by: Josh Soref <jsoref@users.noreply.github.com>

* spelling: monitoring

Signed-off-by: Josh Soref <jsoref@users.noreply.github.com>

* spelling: monitors

Signed-off-by: Josh Soref <jsoref@users.noreply.github.com>

* spelling: monolithic

Signed-off-by: Josh Soref <jsoref@users.noreply.github.com>

* spelling: multi

Signed-off-by: Josh Soref <jsoref@users.noreply.github.com>

* spelling: multiplier

Signed-off-by: Josh Soref <jsoref@users.noreply.github.com>

* spelling: navigation

Signed-off-by: Josh Soref <jsoref@users.noreply.github.com>

* spelling: noisy

Signed-off-by: Josh Soref <jsoref@users.noreply.github.com>

* spelling: number

Signed-off-by: Josh Soref <jsoref@users.noreply.github.com>

* spelling: observing

Signed-off-by: Josh Soref <jsoref@users.noreply.github.com>

* spelling: omitted

Signed-off-by: Josh Soref <jsoref@users.noreply.github.com>

* spelling: orchestrator

Signed-off-by: Josh Soref <jsoref@users.noreply.github.com>

* spelling: overall

Signed-off-by: Josh Soref <jsoref@users.noreply.github.com>

* spelling: overridden

Signed-off-by: Josh Soref <jsoref@users.noreply.github.com>

* spelling: package

Signed-off-by: Josh Soref <jsoref@users.noreply.github.com>

* spelling: packages

Signed-off-by: Josh Soref <jsoref@users.noreply.github.com>

* spelling: packet

Signed-off-by: Josh Soref <jsoref@users.noreply.github.com>

* spelling: pages

Signed-off-by: Josh Soref <jsoref@users.noreply.github.com>

* spelling: parameter

Signed-off-by: Josh Soref <jsoref@users.noreply.github.com>

* spelling: parsable

Signed-off-by: Josh Soref <jsoref@users.noreply.github.com>

* spelling: percentage

Signed-off-by: Josh Soref <jsoref@users.noreply.github.com>

* spelling: perfect

Signed-off-by: Josh Soref <jsoref@users.noreply.github.com>

* spelling: phpfpm

Signed-off-by: Josh Soref <jsoref@users.noreply.github.com>

* spelling: platform

Signed-off-by: Josh Soref <jsoref@users.noreply.github.com>

* spelling: preferred

Signed-off-by: Josh Soref <jsoref@users.noreply.github.com>

* spelling: prioritize

Signed-off-by: Josh Soref <jsoref@users.noreply.github.com>

* spelling: probabilities

Signed-off-by: Josh Soref <jsoref@users.noreply.github.com>

* spelling: process

Signed-off-by: Josh Soref <jsoref@users.noreply.github.com>

* spelling: processes

Signed-off-by: Josh Soref <jsoref@users.noreply.github.com>

* spelling: program

Signed-off-by: Josh Soref <jsoref@users.noreply.github.com>

* spelling: qos

Signed-off-by: Josh Soref <jsoref@users.noreply.github.com>

* spelling: quick

Signed-off-by: Josh Soref <jsoref@users.noreply.github.com>

* spelling: raspberry

Signed-off-by: Josh Soref <jsoref@users.noreply.github.com>

* spelling: received

Signed-off-by: Josh Soref <jsoref@users.noreply.github.com>

* spelling: recvfile

Signed-off-by: Josh Soref <jsoref@users.noreply.github.com>

* spelling: red hat

Signed-off-by: Josh Soref <jsoref@users.noreply.github.com>

* spelling: relatively

Signed-off-by: Josh Soref <jsoref@users.noreply.github.com>

* spelling: reliability

Signed-off-by: Josh Soref <jsoref@users.noreply.github.com>

* spelling: repository

Signed-off-by: Josh Soref <jsoref@users.noreply.github.com>

* spelling: requested

Signed-off-by: Josh Soref <jsoref@users.noreply.github.com>

* spelling: requests

Signed-off-by: Josh Soref <jsoref@users.noreply.github.com>

* spelling: retrieved

Signed-off-by: Josh Soref <jsoref@users.noreply.github.com>

* spelling: scenarios

Signed-off-by: Josh Soref <jsoref@users.noreply.github.com>

* spelling: see all

Signed-off-by: Josh Soref <jsoref@users.noreply.github.com>

* spelling: supported

Signed-off-by: Josh Soref <jsoref@users.noreply.github.com>

* spelling: supports

Signed-off-by: Josh Soref <jsoref@users.noreply.github.com>

* spelling: temporary

Signed-off-by: Josh Soref <jsoref@users.noreply.github.com>

* spelling: tsdb

Signed-off-by: Josh Soref <jsoref@users.noreply.github.com>

* spelling: tutorial

Signed-off-by: Josh Soref <jsoref@users.noreply.github.com>

* spelling: updates

Signed-off-by: Josh Soref <jsoref@users.noreply.github.com>

* spelling: utilization

Signed-off-by: Josh Soref <jsoref@users.noreply.github.com>

* spelling: value

Signed-off-by: Josh Soref <jsoref@users.noreply.github.com>

* spelling: variables

Signed-off-by: Josh Soref <jsoref@users.noreply.github.com>

* spelling: visualize

Signed-off-by: Josh Soref <jsoref@users.noreply.github.com>

* spelling: voluntary

Signed-off-by: Josh Soref <jsoref@users.noreply.github.com>

* spelling: your

Signed-off-by: Josh Soref <jsoref@users.noreply.github.com>
---
 .travis/README.md                             |  4 ++--
 BUILD.md                                      |  6 ++---
 CHANGELOG.md                                  |  8 +++----
 HISTORICAL_CHANGELOG.md                       | 14 ++++++------
 backends/TIMESCALE.md                         |  2 +-
 backends/opentsdb/README.md                   |  2 +-
 build_external/README.md                      |  2 +-
 claim/README.md                               |  2 +-
 collectors/COLLECTORS.md                      | 10 ++++-----
 collectors/README.md                          |  4 ++--
 collectors/REFERENCE.md                       |  4 ++--
 collectors/apps.plugin/README.md              |  2 +-
 collectors/cgroups.plugin/README.md           |  2 +-
 collectors/ebpf.plugin/README.md              |  4 ++--
 collectors/freeipmi.plugin/README.md          |  2 +-
 .../node.d.plugin/stiebeleltron/README.md     |  4 ++--
 collectors/perf.plugin/README.md              |  2 +-
 collectors/plugins.d/README.md                |  2 +-
 collectors/python.d.plugin/am2320/README.md   |  6 ++---
 .../python.d.plugin/anomalies/README.md       | 14 ++++++------
 collectors/python.d.plugin/dovecot/README.md  |  4 ++--
 .../python.d.plugin/go_expvar/README.md       |  2 +-
 collectors/python.d.plugin/mongodb/README.md  |  2 +-
 collectors/python.d.plugin/mysql/README.md    |  6 ++---
 collectors/python.d.plugin/postgres/README.md |  2 +-
 collectors/python.d.plugin/proxysql/README.md |  4 ++--
 collectors/python.d.plugin/samba/README.md    |  2 +-
 .../python.d.plugin/springboot/README.md      |  2 +-
 collectors/statsd.plugin/README.md            |  4 ++--
 collectors/tc.plugin/README.md                |  2 +-
 daemon/README.md                              |  2 +-
 daemon/config/README.md                       |  4 ++--
 docs/Running-behind-apache.md                 |  2 +-
 docs/collect/container-metrics.md             |  2 +-
 docs/collect/enable-configure.md              |  6 ++---
 docs/collect/how-collectors-work.md           |  2 +-
 docs/export/external-databases.md             |  2 +-
 docs/guides/collect-apache-nginx-web-logs.md  |  2 +-
 docs/guides/longer-metrics-storage.md         |  2 +-
 docs/guides/monitor/kubernetes-k8s-netdata.md |  4 ++--
 docs/guides/monitor/pi-hole-raspberry-pi.md   |  4 ++--
 docs/guides/monitor/process.md                | 10 ++++-----
 docs/guides/step-by-step/step-05.md           |  2 +-
 docs/guides/step-by-step/step-06.md           |  2 +-
 docs/guides/step-by-step/step-08.md           |  2 +-
 docs/guides/step-by-step/step-09.md           |  2 +-
 docs/guides/step-by-step/step-10.md           |  2 +-
 .../monitor-debug-applications-ebpf.md        |  4 ++--
 docs/netdata-security.md                      |  4 ++--
 exporting/README.md                           |  6 ++---
 exporting/TIMESCALE.md                        |  2 +-
 health/REFERENCE.md                           |  6 ++---
 health/notifications/alerta/README.md         |  4 ++--
 health/notifications/awssns/README.md         |  2 +-
 health/notifications/email/README.md          |  2 +-
 health/notifications/irc/README.md            |  2 +-
 health/notifications/prowl/README.md          |  4 ++--
 health/notifications/rocketchat/README.md     |  2 +-
 health/notifications/stackpulse/README.md     |  4 ++--
 packaging/DISTRIBUTIONS.md                    |  2 +-
 packaging/docker/README.md                    |  4 ++--
 packaging/installer/REINSTALL.md              |  2 +-
 packaging/installer/UPDATE.md                 |  4 ++--
 .../installer/methods/cloud-providers.md      |  2 +-
 packaging/installer/methods/freebsd.md        |  4 ++--
 packaging/installer/methods/kickstart-64.md   |  6 ++---
 packaging/installer/methods/kickstart.md      |  2 +-
 packaging/installer/methods/kubernetes.md     |  6 ++---
 packaging/installer/methods/manual.md         |  4 ++--
 packaging/installer/methods/packages.md       |  4 ++--
 packaging/installer/methods/pfsense.md        |  2 +-
 packaging/installer/methods/source.md         | 22 +++++++++----------
 packaging/maintainers/README.md               |  2 +-
 packaging/makeself/README.md                  |  2 +-
 parser/README.md                              |  6 ++---
 streaming/README.md                           |  2 +-
 web/api/badges/README.md                      | 10 ++++-----
 web/api/health/README.md                      |  2 +-
 web/gui/custom/README.md                      |  2 +-
 web/server/static/README.md                   |  2 +-
 80 files changed, 156 insertions(+), 156 deletions(-)

diff --git a/.travis/README.md b/.travis/README.md
index 4ffdbe6ae9..8927dd4c54 100644
--- a/.travis/README.md
+++ b/.travis/README.md
@@ -66,7 +66,7 @@ Briefly our activities include:
 
 ## Artifacts validation
 At this point we know our software is building, we need to go through the a set of checks, to guarantee
-that our product meets certain epxectations. At the current stage, we are focusing on basic capabilities
+that our product meets certain expectations. At the current stage, we are focusing on basic capabilities
 like installing in different distributions, running the full lifecycle of install-run-update-install and so on.
 We are still working on enriching this with more and more use cases, to get us closer to achieving full stability of our software.
 Briefly we currently evaluate the following activities:
@@ -121,7 +121,7 @@ The following distributions are supported
   - Bionic
   - artful
 
-- Enterprise Linux versions (Covers Redhat, CentOS, and Amazon Linux with version 6)
+- Enterprise Linux versions (Covers Red Hat, CentOS, and Amazon Linux with version 6)
   - Version 8 (TBD)
   - Version 7
   - Version 6
diff --git a/BUILD.md b/BUILD.md
index deb30b37f4..049c86d3f5 100644
--- a/BUILD.md
+++ b/BUILD.md
@@ -57,7 +57,7 @@ cmake -DENABLE_DBENGINE
 
 ### Dependency detection
 
-We have a mixture of soft- and hard-depedencies on libraries. For most of these we expect
+We have a mixture of soft- and hard-dependencies on libraries. For most of these we expect
 `pkg-config` information, for some we manually probe for libraries and include files. We
 should treat all of the external dependencies consistently:
 
@@ -346,10 +346,10 @@ We should follow these steps:
 9. Deprecate / remove the autotools build-system completely (so that we can support a single
    build-system).
 
-Some smaller miscellaeneous suggestions:
+Some smaller miscellaneous suggestions:
 
 1. Remove the `_Generic` / `strerror_r` config to make the system simpler (use the technique
-   on the blog post to make the standard version re-enterant so that it is thread-safe).
+   on the blog post to make the standard version re-entrant so that it is thread-safe).
 2. Pull in jemalloc by source into the repo if it is our preferred malloc implementation.
 
 # Background
diff --git a/CHANGELOG.md b/CHANGELOG.md
index 0a183059dc..b8271a092d 100644
--- a/CHANGELOG.md
+++ b/CHANGELOG.md
@@ -33,9 +33,9 @@
 - Exclude autofs by default in diskspace plugin [\#10441](https://github.com/netdata/netdata/pull/10441) ([nabijaczleweli](https://github.com/nabijaczleweli))
 - New eBPF kernel [\#10434](https://github.com/netdata/netdata/pull/10434) ([thiagoftsm](https://github.com/thiagoftsm))
 - Update and improve the Netdata style guide [\#10433](https://github.com/netdata/netdata/pull/10433) ([joelhans](https://github.com/joelhans))
-- Change HDDtemp to report None instead of 0 [\#10429](https://github.com/netdata/netdata/pull/10429) ([slavox](https://github.com/slavox))
+- Change hddtemp to report None instead of 0 [\#10429](https://github.com/netdata/netdata/pull/10429) ([slavox](https://github.com/slavox))
 - Use bash shell as user netdata for debug [\#10425](https://github.com/netdata/netdata/pull/10425) ([Steve8291](https://github.com/Steve8291))
-- Qick and dirty fix for \#10420 [\#10424](https://github.com/netdata/netdata/pull/10424) ([skibbipl](https://github.com/skibbipl))
+- Quick and dirty fix for \#10420 [\#10424](https://github.com/netdata/netdata/pull/10424) ([skibbipl](https://github.com/skibbipl))
 - Add instructions on enabling explicitly disabled collectors [\#10418](https://github.com/netdata/netdata/pull/10418) ([joelhans](https://github.com/joelhans))
 - Change links at bottom of all install docs [\#10416](https://github.com/netdata/netdata/pull/10416) ([joelhans](https://github.com/joelhans))
 - Improve configuration docs with common changes and start/stop/restart directions [\#10415](https://github.com/netdata/netdata/pull/10415) ([joelhans](https://github.com/joelhans))
@@ -139,7 +139,7 @@
 - add `nvidia\_smi` collector data to the dashboard\_info.js [\#10230](https://github.com/netdata/netdata/pull/10230) ([ilyam8](https://github.com/ilyam8))
 - health: convert `elasticsearch\_last\_collected` alarm to template [\#10226](https://github.com/netdata/netdata/pull/10226) ([ilyam8](https://github.com/ilyam8))
 - streaming: fix a typo in the README.md [\#10225](https://github.com/netdata/netdata/pull/10225) ([ilyam8](https://github.com/ilyam8))
-- collectors/xenstat.plugin: recieved =\> received [\#10224](https://github.com/netdata/netdata/pull/10224) ([ilyam8](https://github.com/ilyam8))
+- collectors/xenstat.plugin: received =\> received [\#10224](https://github.com/netdata/netdata/pull/10224) ([ilyam8](https://github.com/ilyam8))
 - dashboard\_info.js: fix a typo \(vernemq\) [\#10223](https://github.com/netdata/netdata/pull/10223) ([ilyam8](https://github.com/ilyam8))
 - Fix chart filtering [\#10218](https://github.com/netdata/netdata/pull/10218) ([vlvkobal](https://github.com/vlvkobal))
 - Don't stop Prometheus remote write collector when data is not available for dimension formatting [\#10217](https://github.com/netdata/netdata/pull/10217) ([vlvkobal](https://github.com/vlvkobal))
@@ -226,7 +226,7 @@
 - Fix memory mode none not dropping stale dimension data [\#9917](https://github.com/netdata/netdata/pull/9917) ([mfundul](https://github.com/mfundul))
 - Fix memory mode none not marking dimensions as obsolete. [\#9912](https://github.com/netdata/netdata/pull/9912) ([mfundul](https://github.com/mfundul))
 - Fix buffer overflow in rrdr structure [\#9903](https://github.com/netdata/netdata/pull/9903) ([mfundul](https://github.com/mfundul))
-- Fix missing newline concatentation slash causing rpm build to fail [\#9900](https://github.com/netdata/netdata/pull/9900) ([prologic](https://github.com/prologic))
+- Fix missing newline concatenation slash causing rpm build to fail [\#9900](https://github.com/netdata/netdata/pull/9900) ([prologic](https://github.com/prologic))
 - installer: update go.d.plugin version to v0.22.0 [\#9898](https://github.com/netdata/netdata/pull/9898) ([ilyam8](https://github.com/ilyam8))
 - Add v2 HTTP message with compression to ACLK [\#9895](https://github.com/netdata/netdata/pull/9895) ([underhood](https://github.com/underhood))
 - Fix lock order reversal \(Coverity defect CID 361629\) [\#9888](https://github.com/netdata/netdata/pull/9888) ([mfundul](https://github.com/mfundul))
diff --git a/HISTORICAL_CHANGELOG.md b/HISTORICAL_CHANGELOG.md
index 9698a20d81..16ef78616b 100644
--- a/HISTORICAL_CHANGELOG.md
+++ b/HISTORICAL_CHANGELOG.md
@@ -164,7 +164,7 @@ netdata (1.6.0) - 2017-03-20
 
     1.  number of sensors by state
     2.  number of events in SEL
-    3.  Temperatures CELCIUS
+    3.  Temperatures CELSIUS
     4.  Temperatures FAHRENHEIT
     5.  Voltages
     6.  Currents
@@ -239,7 +239,7 @@ netdata (1.5.0) - 2017-01-22
     Vladimir Kobal (@vlvkobal) has done a magnificent work
     porting netdata to FreeBSD and MacOS.
 
-    Everyhing works: cpu, memory, disks performance, disks space,
+    Everything works: cpu, memory, disks performance, disks space,
     network interfaces, interrupts, IPv4 metrics, IPv6 metrics
     processes, context switches, softnet, IPC queues,
     IPC semaphores, IPC shared memory, uptime, etc. Wow!
@@ -382,7 +382,7 @@ netdata (1.4.0) - 2016-10-04
        cgroups,
        hddtemp,
        sensors,
-       phpfm,
+       phpfpm,
        tc (QoS)
 
  In detail:
@@ -483,7 +483,7 @@ netdata (1.3.0) - 2016-08-28
     -   hddtemp
     -   mysql
     -   nginx
-    -   phpfm
+    -   phpfpm
     -   postfix
     -   sensors
     -   squid
@@ -518,7 +518,7 @@ netdata (1.3.0) - 2016-08-28
 -   apps.plugin improvements:
 
     -   can now run with command line argument 'without-files'
-        to prevent it from enumating all the open files/sockets/pipes
+        to prevent it from enumerating all the open files/sockets/pipes
         of all running processes.
 
     -   apps.plugin now scales the collected values to match the
@@ -575,7 +575,7 @@ netdata (1.2.0) - 2016-05-16
         20% better performance for the core of netdata.
 
     -   More efficient threads locking in key components
-        contributed to the overal efficiency.
+        contributed to the overall efficiency.
 
 -   netdata now has a CENTRAL REGISTRY !
 
@@ -625,7 +625,7 @@ netdata (1.1.0) - 2016-04-20
 -   Data collection: apps.plugin: grouping of processes now support patterns
 -   Data collection: apps.plugin: now it is faster, after the new features added
 -   Data collection: better auto-detection of partitions for disk monitoring
--   Data collection: better fireqos intergation for QoS monitoring
+-   Data collection: better fireqos integration for QoS monitoring
 -   Data collection: squid monitoring now uses squidclient
 -   Data collection: SNMP monitoring now supports 64bit counters
 -   API: fixed issues in CSV output generation
diff --git a/backends/TIMESCALE.md b/backends/TIMESCALE.md
index 854c4112e8..05a3c3b470 100644
--- a/backends/TIMESCALE.md
+++ b/backends/TIMESCALE.md
@@ -27,7 +27,7 @@ TimescaleDB.
 Finally, another member of Netdata's community has built a project that quickly launches Netdata, TimescaleDB, and
 Grafana in easy-to-manage Docker containers. Rune Juhl Jacobsen's
 [project](https://github.com/runejuhl/grafana-timescaledb) uses a `Makefile` to create everything, which makes it
-perferct for testing and experimentation.
+perfect for testing and experimentation.
 
 ## Netdata&#8596;TimescaleDB in action
 
diff --git a/backends/opentsdb/README.md b/backends/opentsdb/README.md
index b9d0b9873b..5ba7b12c58 100644
--- a/backends/opentsdb/README.md
+++ b/backends/opentsdb/README.md
@@ -21,7 +21,7 @@ change the `destination = localhost:4242` line accordingly.
 
 As of [v1.16.0](https://github.com/netdata/netdata/releases/tag/v1.16.0), Netdata can send metrics to OpenTSDB using
 TLS/SSL. Unfortunately, OpenTDSB does not support encrypted connections, so you will have to configure a reverse proxy
-to enable HTTPS communication between Netdata and OpenTSBD. You can set up a reverse proxy with
+to enable HTTPS communication between Netdata and OpenTSDB. You can set up a reverse proxy with
 [Nginx](/docs/Running-behind-nginx.md).
 
 After your proxy is configured, make the following changes to `netdata.conf`:
diff --git a/build_external/README.md b/build_external/README.md
index f52f55734a..6a1e30a574 100644
--- a/build_external/README.md
+++ b/build_external/README.md
@@ -12,7 +12,7 @@ decoupled. This allows:
 -   Cross-compilation (e.g. linux development from macOS)
 -   Cross-distro (e.g. using CentOS user-land while developing on Debian)
 -   Multi-host scenarios (e.g. parent-child configurations)
--   Bleeding-edge sceneraios (e.g. using the ACLK (**currently for internal-use only**))
+-   Bleeding-edge scenarios (e.g. using the ACLK (**currently for internal-use only**))
 
 The advantage of these scenarios is that they allow **reproducible** builds and testing
 for developers. This is the first iteration of the build-system to allow the team to use
diff --git a/claim/README.md b/claim/README.md
index a2e5116c13..ade6a221f2 100644
--- a/claim/README.md
+++ b/claim/README.md
@@ -304,7 +304,7 @@ This node no longer has access to the credentials it was claimed with and cannot
 You will still be able to see this node in your War Rooms in an **unreachable** state.
 
 If you want to reclaim this node into a different Space, you need to create a new identity by adding `-id=$(uuidgen)` to
-the claiming script parameters. Make sure that you have the `uuidgen-runtime` packagen installed, as it is used to run the command `uuidgen`. For example, using the default claiming script:
+the claiming script parameters. Make sure that you have the `uuidgen-runtime` package installed, as it is used to run the command `uuidgen`. For example, using the default claiming script:
 
 ```bash
 sudo netdata-claim.sh -token=TOKEN -rooms=ROOM1,ROOM2 -url=https://app.netdata.cloud -id=$(uuidgen)
diff --git a/collectors/COLLECTORS.md b/collectors/COLLECTORS.md
index 939594e0ce..e718fd239a 100644
--- a/collectors/COLLECTORS.md
+++ b/collectors/COLLECTORS.md
@@ -222,7 +222,7 @@ configure any of these collectors according to your setup and infrastructure.
 -   [ISC DHCP (Go)](https://learn.netdata.cloud/docs/agent/collectors/go.d.plugin/modules/isc_dhcpd): Reads a
     `dhcpd.leases` file and collects metrics on total active leases, pool active leases, and pool utilization.
 -   [ISC DHCP (Python)](/collectors/python.d.plugin/isc_dhcpd/README.md): Reads `dhcpd.leases` file and reports DHCP
-    pools utiliation and leases statistics (total number, leases per pool).
+    pools utilization and leases statistics (total number, leases per pool).
 -   [OpenLDAP](/collectors/python.d.plugin/openldap/README.md): Provides statistics information from the OpenLDAP
     (`slapd`) server.
 -   [NSD](/collectors/python.d.plugin/nsd/README.md): Monitor nameserver performance metrics using the `nsd-control`
@@ -357,7 +357,7 @@ The Netdata Agent can collect these system- and hardware-level metrics using a v
 -   [BCACHE](/collectors/proc.plugin/README.md): Monitor BCACHE statistics with the the `proc.plugin` collector.
 -   [Block devices](/collectors/proc.plugin/README.md): Gather metrics about the health and performance of block
     devices using the the `proc.plugin` collector.
--   [Btrfs](/collectors/proc.plugin/README.md): Montiors Btrfs filesystems with the the `proc.plugin` collector.
+-   [Btrfs](/collectors/proc.plugin/README.md): Monitors Btrfs filesystems with the the `proc.plugin` collector.
 -   [Device mapper](/collectors/proc.plugin/README.md): Gather metrics about the Linux device mapper with the proc
     collector.
 -   [Disk space](/collectors/diskspace.plugin/README.md): Collect disk space usage metrics on Linux mount points.
@@ -445,7 +445,7 @@ The Netdata Agent can collect these system- and hardware-level metrics using a v
 -   [systemd](/collectors/cgroups.plugin/README.md): Monitor the CPU and memory usage of systemd services using the
     `cgroups.plugin` collector.
 -   [systemd unit states](https://learn.netdata.cloud/docs/agent/collectors/go.d.plugin/modules/systemdunits): See the 
-    state (active, inactive, activating, deactiviating, failed) of various systemd unit types.
+    state (active, inactive, activating, deactivating, failed) of various systemd unit types.
 -   [System processes](/collectors/proc.plugin/README.md): Collect metrics on system load and total processes running
     using `/proc/loadavg` and the `proc.plugin` collector.
 -   [Uptime](/collectors/proc.plugin/README.md): Monitor the uptime of a system using the `proc.plugin` collector.
@@ -511,10 +511,10 @@ the `go.d.plugin`.
 
 ## Third-party collectors
 
-These collectors are developed and maintined by third parties and, unlike the other collectors, are not installed by
+These collectors are developed and maintained by third parties and, unlike the other collectors, are not installed by
 default. To use a third-party collector, visit their GitHub/documentation page and follow their installation procedures.
 
--   [CyberPower UPS](https://github.com/HawtDogFlvrWtr/netdata_cyberpwrups_plugin): Polls Cyberpower UPS data using
+-   [CyberPower UPS](https://github.com/HawtDogFlvrWtr/netdata_cyberpwrups_plugin): Polls CyberPower UPS data using
     PowerPanel® Personal Linux.
 -   [Logged-in users](https://github.com/veksh/netdata-numsessions): Collect the number of currently logged-on users.
 -   [nim-netdata-plugin](https://github.com/FedericoCeratto/nim-netdata-plugin): A helper to create native Netdata
diff --git a/collectors/README.md b/collectors/README.md
index ef1f9610c1..a37a7e890a 100644
--- a/collectors/README.md
+++ b/collectors/README.md
@@ -32,7 +32,7 @@ guide](/collectors/QUICKSTART.md).
 
 [Monitor Nginx or Apache web server log files with Netdata](/docs/guides/collect-apache-nginx-web-logs.md)
 
-[Monitor CockroadchDB metrics with Netdata](/docs/guides/monitor-cockroachdb.md)
+[Monitor CockroachDB metrics with Netdata](/docs/guides/monitor-cockroachdb.md)
 
 [Monitor Unbound DNS servers with Netdata](/docs/guides/collect-unbound-metrics.md)
 
@@ -40,7 +40,7 @@ guide](/collectors/QUICKSTART.md).
 
 ## Related features
 
-**[Dashboards](/web/README.md)**: Vizualize your newly-collect metrics in real-time using Netdata's [built-in
+**[Dashboards](/web/README.md)**: Visualize your newly-collect metrics in real-time using Netdata's [built-in
 dashboard](/web/gui/README.md). 
 
 **[Backends](/backends/README.md)**: Extend our built-in [database engine](/database/engine/README.md), which supports
diff --git a/collectors/REFERENCE.md b/collectors/REFERENCE.md
index 08a405dc7b..9c6f0a61ed 100644
--- a/collectors/REFERENCE.md
+++ b/collectors/REFERENCE.md
@@ -46,7 +46,7 @@ However, there are cases that auto-detection fails. Usually, the reason is that
 allow Netdata to connect. In most of the cases, allowing the user `netdata` from `localhost` to connect and collect
 metrics, will automatically enable data collection for the application in question (it will require a Netdata restart).
 
-View our [collectors quickstart](/collectors/QUICKSTART.md) for explict details on enabling and configuring collector modules.
+View our [collectors quickstart](/collectors/QUICKSTART.md) for explicit details on enabling and configuring collector modules.
 
 ## Troubleshoot a collector
 
@@ -112,7 +112,7 @@ This section features a list of Netdata's plugins, with a boolean setting to ena
 	# charts.d = yes
 ```
 
-By default, most plugins are enabled, so you don't need to enable them explicity to use their collectors. To enable or
+By default, most plugins are enabled, so you don't need to enable them explicitly to use their collectors. To enable or
 disable any specific plugin, remove the comment (`#`) and change the boolean setting to `yes` or `no`.
 
 All **external plugins** are managed by [plugins.d](plugins.d/), which provides additional management options.
diff --git a/collectors/apps.plugin/README.md b/collectors/apps.plugin/README.md
index 5529226961..d10af1cdd3 100644
--- a/collectors/apps.plugin/README.md
+++ b/collectors/apps.plugin/README.md
@@ -59,7 +59,7 @@ Each of these sections provides the same number of charts:
     -   Pipes open (`apps.pipes`)
 -   Swap memory
     -   Swap memory used (`apps.swap`)
-    -   Major page faults (i.e. swap activiy, `apps.major_faults`)
+    -   Major page faults (i.e. swap activity, `apps.major_faults`)
 -   Network
     -   Sockets open (`apps.sockets`)
     
diff --git a/collectors/cgroups.plugin/README.md b/collectors/cgroups.plugin/README.md
index 9b26deb2ce..21dbcae83f 100644
--- a/collectors/cgroups.plugin/README.md
+++ b/collectors/cgroups.plugin/README.md
@@ -145,7 +145,7 @@ Support per distribution:
 |Fedora 25|YES|[here](http://pastebin.com/ax0373wF)||
 |Debian 8|NO||can be enabled, see below|
 |AMI|NO|[here](http://pastebin.com/FrxmptjL)|not a systemd system|
-|Centos 7.3.1611|NO|[here](http://pastebin.com/SpzgezAg)|can be enabled, see below|
+|CentOS 7.3.1611|NO|[here](http://pastebin.com/SpzgezAg)|can be enabled, see below|
 
 ### how to enable cgroup accounting on systemd systems that is by default disabled
 
diff --git a/collectors/ebpf.plugin/README.md b/collectors/ebpf.plugin/README.md
index 44e238e36d..5ea3b49514 100644
--- a/collectors/ebpf.plugin/README.md
+++ b/collectors/ebpf.plugin/README.md
@@ -221,7 +221,7 @@ The following options are available:
 -   `ports`: Define the destination ports for Netdata to monitor.
 -   `hostnames`: The list of hostnames that can be resolved to an IP address. 
 -   `ips`: The IP or range of IPs that you want to monitor. You can use IPv4 or IPv6 addresses, use dashes to define a
-    range of IPs, or use CIDR values. The default behavior is to only collect data for private IP addresess, but this
+    range of IPs, or use CIDR values. The default behavior is to only collect data for private IP addresses, but this
     can be changed with the `ips` setting.
     
 By default, Netdata displays up to 500 dimensions on network connection charts. If there are more possible dimensions, 
@@ -275,7 +275,7 @@ curl -sSL https://raw.githubusercontent.com/netdata/kernel-collector/master/tool
 
 If this script returns no output, your system is ready to compile and run the eBPF collector.
 
-If you see a warning about a missing kerkel configuration (`KPROBES KPROBES_ON_FTRACE HAVE_KPROBES BPF BPF_SYSCALL
+If you see a warning about a missing kernel configuration (`KPROBES KPROBES_ON_FTRACE HAVE_KPROBES BPF BPF_SYSCALL
 BPF_JIT`), you will need to recompile your kernel to support this configuration. The process of recompiling Linux
 kernels varies based on your distribution and version. Read the documentation for your system's distribution to learn
 more about the specific workflow for recompiling the kernel, ensuring that you set all the necessary 
diff --git a/collectors/freeipmi.plugin/README.md b/collectors/freeipmi.plugin/README.md
index 64328fc9e7..52945e3c62 100644
--- a/collectors/freeipmi.plugin/README.md
+++ b/collectors/freeipmi.plugin/README.md
@@ -25,7 +25,7 @@ The plugin creates (up to) 8 charts, based on the information collected from IPM
 
 1.  number of sensors by state
 2.  number of events in SEL
-3.  Temperatures CELCIUS
+3.  Temperatures CELSIUS
 4.  Temperatures FAHRENHEIT
 5.  Voltages
 6.  Currents
diff --git a/collectors/node.d.plugin/stiebeleltron/README.md b/collectors/node.d.plugin/stiebeleltron/README.md
index 30f51169b3..59bbf703c4 100644
--- a/collectors/node.d.plugin/stiebeleltron/README.md
+++ b/collectors/node.d.plugin/stiebeleltron/README.md
@@ -40,7 +40,7 @@ The charts are configurable, however, the provided default configuration collect
     -   Heat circuit 1 room temperature in C (set/actual)
     -   Heat circuit 2 room temperature in C (set/actual)
 
-5.  **Eletric Reheating**
+5.  **Electric Reheating**
 
     -   Dual Mode Reheating temperature in C (hot water/heating)
 
@@ -68,7 +68,7 @@ If no configuration is given, the module will be disabled. Each `update_every` i
 
 Original author: BrainDoctor (github)
 
-The module supports any metrics that are parseable with RegEx. There is no API that gives direct access to the values (AFAIK), so the "workaround" is to parse the HTML output of the ISG.
+The module supports any metrics that are parsable with RegEx. There is no API that gives direct access to the values (AFAIK), so the "workaround" is to parse the HTML output of the ISG.
 
 ### Testing
 
diff --git a/collectors/perf.plugin/README.md b/collectors/perf.plugin/README.md
index d4bb41cb60..ccd185cedb 100644
--- a/collectors/perf.plugin/README.md
+++ b/collectors/perf.plugin/README.md
@@ -64,7 +64,7 @@ enable the perf plugin, edit /etc/netdata/netdata.conf and set:
 You can use the `command options` parameter to pick what data should be collected and which charts should be
 displayed. If `all` is used, all general performance monitoring counters are probed and corresponding charts
 are enabled for the available counters. You can also define a particular set of enabled charts using the
-following keywords: `cycles`, `instructions`, `branch`, `cache`, `bus`, `stalled`, `migrations`, `alighnment`,
+following keywords: `cycles`, `instructions`, `branch`, `cache`, `bus`, `stalled`, `migrations`, `alignment`,
 `emulation`, `L1D`, `L1D-prefetch`, `L1I`, `LL`, `DTLB`, `ITLB`, `PBU`.
 
 ## Debugging
diff --git a/collectors/plugins.d/README.md b/collectors/plugins.d/README.md
index 913ad9177c..c166e11e36 100644
--- a/collectors/plugins.d/README.md
+++ b/collectors/plugins.d/README.md
@@ -79,7 +79,7 @@ Example:
 ```
 
 The setting `enable running new plugins` sets the default behavior for all external plugins. It can be 
-overriden for distinct plugins by modifying the appropriate plugin value configuration to either `yes` or `no`.
+overridden for distinct plugins by modifying the appropriate plugin value configuration to either `yes` or `no`.
 
 The setting `check for new plugins every` sets the interval between scans of the directory
 `/usr/libexec/netdata/plugins.d`. New plugins can be added any time, and Netdata will detect them in a timely manner.
diff --git a/collectors/python.d.plugin/am2320/README.md b/collectors/python.d.plugin/am2320/README.md
index c17b33dfa1..14ddaa735d 100644
--- a/collectors/python.d.plugin/am2320/README.md
+++ b/collectors/python.d.plugin/am2320/README.md
@@ -6,7 +6,7 @@ sidebar_label: "AM2320"
 
 # AM2320 sensor monitoring with netdata
 
-Displays a graph of the temperature and humity from a AM2320 sensor.
+Displays a graph of the temperature and humidity from a AM2320 sensor.
 
 ## Requirements
  - Adafruit Circuit Python AM2320 library
@@ -28,10 +28,10 @@ cd /etc/netdata   # Replace this path with your Netdata config directory, if dif
 sudo ./edit-config python.d/am2320.conf
 ```
 
-Raspbery Pi Instructions:
+Raspberry Pi Instructions:
 
 Hardware install:
-Connect the am2320 to the Raspbery Pi I2C pins
+Connect the am2320 to the Raspberry Pi I2C pins
 
 Raspberry Pi 3B/4 Pins:
 
diff --git a/collectors/python.d.plugin/anomalies/README.md b/collectors/python.d.plugin/anomalies/README.md
index 8346aa6693..1e27f3b5be 100644
--- a/collectors/python.d.plugin/anomalies/README.md
+++ b/collectors/python.d.plugin/anomalies/README.md
@@ -134,7 +134,7 @@ local:
     diffs_n: 1
 
     # What is the typical proportion of anomalies in your data on average? 
-    # This paramater can control the sensitivity of your models to anomalies. 
+    # This parameter can control the sensitivity of your models to anomalies. 
     # Some discussion here: https://github.com/yzhao062/pyod/issues/144
     contamination: 0.001
 
@@ -142,7 +142,7 @@ local:
     # just the average of all anomaly probabilities at each time step
     include_average_prob: true
 
-    # Define any custom models you would like to create anomaly probabilties for, some examples below to show how.
+    # Define any custom models you would like to create anomaly probabilities for, some examples below to show how.
     # For example below example creates two custom models, one to run anomaly detection user and system cpu for our demo servers
     # and one on the cpu and mem apps metrics for the python.d.plugin.
     # custom_models:
@@ -161,7 +161,7 @@ local:
 
 In the `anomalies.conf` file you can also define some "custom models" which you can use to group one or more metrics into a single model much like is done by default for the charts you specify. This is useful if you have a handful of metrics that exist in different charts but perhaps are related to the same underlying thing you would like to perform anomaly detection on, for example a specific app or user. 
 
-To define a custom model you would include configuation like below in `anomalies.conf`. By default there should already be some commented out examples in there. 
+To define a custom model you would include configuration like below in `anomalies.conf`. By default there should already be some commented out examples in there. 
 
 `name` is a name you give your custom model, this is what will appear alongside any other specified charts in the `anomalies.probability` and `anomalies.anomaly` charts. `dimensions` is a string of metrics you want to include in your custom model. By default the [netdata-pandas](https://github.com/netdata/netdata-pandas) library used to pull the data from Netdata uses a "chart.a|dim.1" type of naming convention in the pandas columns it returns, hence the `dimensions` string should look like "chart.name|dimension.name,chart.name|dimension.name". The examples below hopefully make this clear.
 
@@ -194,7 +194,7 @@ sudo su -s /bin/bash netdata
 /usr/libexec/netdata/plugins.d/python.d.plugin anomalies debug trace nolock
 ```
 
-## Deepdive turorial
+## Deepdive tutorial
 
 If you would like to go deeper on what exactly the anomalies collector is doing under the hood then check out this [deepdive tutorial](https://github.com/netdata/community/blob/main/netdata-agent-api/netdata-pandas/anomalies_collector_deepdive.ipynb) in our community repo where you can play around with some data from our demo servers (or your own if its accessible to you) and work through the calculations step by step.
 
@@ -206,7 +206,7 @@ If you would like to go deeper on what exactly the anomalies collector is doing
 - Python 3 is also required for the underlying ML libraries of [numba](https://pypi.org/project/numba/), [scikit-learn](https://pypi.org/project/scikit-learn/), and [PyOD](https://pypi.org/project/pyod/).
 - It may take a few hours or so (depending on your choice of `train_secs_n`) for the collector to 'settle' into it's typical behaviour in terms of the trained models and probabilities you will see in the normal running of your node.
 - As this collector does most of the work in Python itself, with [PyOD](https://pyod.readthedocs.io/en/latest/) leveraging [numba](https://numba.pydata.org/) under the hood, you may want to try it out first on a test or development system to get a sense of its performance characteristics on a node similar to where you would like to use it.
-- `lags_n`, `smooth_n`, and `diffs_n` together define the preprocessing done to the raw data before models are trained and before each prediction. This essentially creates a [feature vector](https://en.wikipedia.org/wiki/Feature_(machine_learning)#:~:text=In%20pattern%20recognition%20and%20machine,features%20that%20represent%20some%20object.&text=Feature%20vectors%20are%20often%20combined,score%20for%20making%20a%20prediction.) for each chart model (or each custom model). The default settings for these parameters aim to create a rolling matrix of recent smoothed [differenced](https://en.wikipedia.org/wiki/Autoregressive_integrated_moving_average#Differencing) values for each chart. The aim of the model then is to score how unusual this 'matrix' of features is for each chart based on what it has learned as 'normal' from the training data. So as opposed to just looking at the single most recent value of a dimension and considering how strange it is, this approach looks at a recent smoothed window of all dimensions for a chart (or dimensions in a custom model) and asks how unusual the data as a whole looks. This should be more flexibile in capturing a wider range of [anomaly types](https://andrewm4894.com/2020/10/19/different-types-of-time-series-anomalies/) and be somewhat more robust to temporary 'spikes' in the data that tend to always be happening somewhere in your metrics but often are not the most important type of anomaly (this is all covered in a lot more detail in the [deepdive tutorial](https://nbviewer.jupyter.org/github/netdata/community/blob/main/netdata-agent-api/netdata-pandas/anomalies_collector_deepdive.ipynb)).
+- `lags_n`, `smooth_n`, and `diffs_n` together define the preprocessing done to the raw data before models are trained and before each prediction. This essentially creates a [feature vector](https://en.wikipedia.org/wiki/Feature_(machine_learning)#:~:text=In%20pattern%20recognition%20and%20machine,features%20that%20represent%20some%20object.&text=Feature%20vectors%20are%20often%20combined,score%20for%20making%20a%20prediction.) for each chart model (or each custom model). The default settings for these parameters aim to create a rolling matrix of recent smoothed [differenced](https://en.wikipedia.org/wiki/Autoregressive_integrated_moving_average#Differencing) values for each chart. The aim of the model then is to score how unusual this 'matrix' of features is for each chart based on what it has learned as 'normal' from the training data. So as opposed to just looking at the single most recent value of a dimension and considering how strange it is, this approach looks at a recent smoothed window of all dimensions for a chart (or dimensions in a custom model) and asks how unusual the data as a whole looks. This should be more flexible in capturing a wider range of [anomaly types](https://andrewm4894.com/2020/10/19/different-types-of-time-series-anomalies/) and be somewhat more robust to temporary 'spikes' in the data that tend to always be happening somewhere in your metrics but often are not the most important type of anomaly (this is all covered in a lot more detail in the [deepdive tutorial](https://nbviewer.jupyter.org/github/netdata/community/blob/main/netdata-agent-api/netdata-pandas/anomalies_collector_deepdive.ipynb)).
 - You can see how long model training is taking by looking in the logs for the collector `grep 'anomalies' /var/log/netdata/error.log | grep 'training'` and you should see lines like `2020-12-01 22:02:14: python.d INFO: anomalies[local] : training complete in 2.81 seconds (runs_counter=2700, model=pca, train_n_secs=14400, models=26, n_fit_success=26, n_fit_fails=0, after=1606845731, before=1606860131).`. 
   - This also gives counts of the number of models, if any, that failed to fit and so had to default back to the DefaultModel (which is currently [HBOS](https://pyod.readthedocs.io/en/latest/_modules/pyod/models/hbos.html)).
   - `after` and `before` here refer to the start and end of the training data used to train the models.
@@ -215,8 +215,8 @@ If you would like to go deeper on what exactly the anomalies collector is doing
   - Typically ~3%-3.5% additional cpu usage from scoring, jumping to ~60% for a couple of seconds during model training.
   - About ~150mb of ram (`apps.mem`) being continually used by the `python.d.plugin`.
 - If you activate this collector on a fresh node, it might take a little while to build up enough data to calculate a realistic and useful model.
-- Some models like `iforest` can be comparatively expensive (on same n1-standard-2 system above ~2s runtime during predict, ~40s training time, ~50% cpu on both train and predict) so if you would like to use it you might be advised to set a relativley high `update_every` maybe 10, 15 or 30 in `anomalies.conf`.
-- Setting a higher `train_every_n` and `update_every` is an easy way to devote less resources on the node to anomaly detection. Specifying less charts and a lower `train_n_secs` will also help reduce resources at the expense of covering less charts and maybe a more noisey model if you set `train_n_secs` to be too small for how your node tends to behave.
+- Some models like `iforest` can be comparatively expensive (on same n1-standard-2 system above ~2s runtime during predict, ~40s training time, ~50% cpu on both train and predict) so if you would like to use it you might be advised to set a relatively high `update_every` maybe 10, 15 or 30 in `anomalies.conf`.
+- Setting a higher `train_every_n` and `update_every` is an easy way to devote less resources on the node to anomaly detection. Specifying less charts and a lower `train_n_secs` will also help reduce resources at the expense of covering less charts and maybe a more noisy model if you set `train_n_secs` to be too small for how your node tends to behave.
 
 ## Useful links and further reading
 
diff --git a/collectors/python.d.plugin/dovecot/README.md b/collectors/python.d.plugin/dovecot/README.md
index 55aeed3eb5..730b64257b 100644
--- a/collectors/python.d.plugin/dovecot/README.md
+++ b/collectors/python.d.plugin/dovecot/README.md
@@ -38,8 +38,8 @@ Module gives information with following charts:
 
 5.  **Context Switches**
 
-    -   volountary
-    -   involountary
+    -   voluntary
+    -   involuntary
 
 6.  **disk** in bytes/s
 
diff --git a/collectors/python.d.plugin/go_expvar/README.md b/collectors/python.d.plugin/go_expvar/README.md
index 66ebc0b67b..a73610e7a1 100644
--- a/collectors/python.d.plugin/go_expvar/README.md
+++ b/collectors/python.d.plugin/go_expvar/README.md
@@ -69,7 +69,7 @@ Sample output:
 ```json
 {
 "cmdline": ["./expvar-demo-binary"],
-"memstats": {"Alloc":630856,"TotalAlloc":630856,"Sys":3346432,"Lookups":27, <ommited for brevity>}
+"memstats": {"Alloc":630856,"TotalAlloc":630856,"Sys":3346432,"Lookups":27, <omitted for brevity>}
 }
 ```
 
diff --git a/collectors/python.d.plugin/mongodb/README.md b/collectors/python.d.plugin/mongodb/README.md
index 5d5295aa46..c0df123d7a 100644
--- a/collectors/python.d.plugin/mongodb/README.md
+++ b/collectors/python.d.plugin/mongodb/README.md
@@ -80,7 +80,7 @@ Number of charts depends on mongodb version, storage engine and other features (
 13. **Cache metrics** (WiredTiger):
 
     -   percentage of bytes currently in the cache (amount of space taken by cached data)
-    -   percantage of tracked dirty bytes in the cache (amount of space taken by dirty data)
+    -   percentage of tracked dirty bytes in the cache (amount of space taken by dirty data)
 
 14. **Pages evicted from cache** (WiredTiger):
 
diff --git a/collectors/python.d.plugin/mysql/README.md b/collectors/python.d.plugin/mysql/README.md
index 5b9feadd54..d8d3c1d0b1 100644
--- a/collectors/python.d.plugin/mysql/README.md
+++ b/collectors/python.d.plugin/mysql/README.md
@@ -67,7 +67,7 @@ This module will produce following charts (if data is available):
     -   immediate
     -   waited
 
-6.  **Table Select Join Issuess** in joins/s
+6.  **Table Select Join Issues** in joins/s
 
     -   full join
     -   full range join
@@ -75,7 +75,7 @@ This module will produce following charts (if data is available):
     -   range check
     -   scan
 
-7.  **Table Sort Issuess** in joins/s
+7.  **Table Sort Issues** in joins/s
 
     -   merge passes
     -   range
@@ -164,7 +164,7 @@ This module will produce following charts (if data is available):
     -   updated
     -   deleted
 
-24. **InnoDB Buffer Pool Pagess** in pages
+24. **InnoDB Buffer Pool Pages** in pages
 
     -   data
     -   dirty
diff --git a/collectors/python.d.plugin/postgres/README.md b/collectors/python.d.plugin/postgres/README.md
index 67cc8fe323..3d573d6dcc 100644
--- a/collectors/python.d.plugin/postgres/README.md
+++ b/collectors/python.d.plugin/postgres/README.md
@@ -22,7 +22,7 @@ Following charts are drawn:
 
     -   active
 
-3.  **Current Backend Processe Usage** percentage
+3.  **Current Backend Process Usage** percentage
 
     -   used
     -   available
diff --git a/collectors/python.d.plugin/proxysql/README.md b/collectors/python.d.plugin/proxysql/README.md
index 6f4ca69131..f1b369a446 100644
--- a/collectors/python.d.plugin/proxysql/README.md
+++ b/collectors/python.d.plugin/proxysql/README.md
@@ -31,7 +31,7 @@ It produces:
     -   questions: total number of queries sent from frontends
     -   slow_queries: number of queries that ran for longer than the threshold in milliseconds defined in global variable `mysql-long_query_time`
 
-3.  **Overall Bandwith (backends)**
+3.  **Overall Bandwidth (backends)**
 
     -   in
     -   out
@@ -45,7 +45,7 @@ It produces:
         -   `4=OFFLINE_HARD`: when a server is put into OFFLINE_HARD mode, the existing connections are dropped, while new incoming connections aren't accepted either. This is equivalent to deleting the server from a hostgroup, or temporarily taking it out of the hostgroup for maintenance work
         -   `-1`: Unknown status
 
-5.  **Bandwith (backends)**
+5.  **Bandwidth (backends)**
 
     -   Backends
         -   in
diff --git a/collectors/python.d.plugin/samba/README.md b/collectors/python.d.plugin/samba/README.md
index 2c86e7b609..ed26d28718 100644
--- a/collectors/python.d.plugin/samba/README.md
+++ b/collectors/python.d.plugin/samba/README.md
@@ -21,7 +21,7 @@ It produces the following charts:
 1.  **Syscall R/Ws** in kilobytes/s
 
     -   sendfile
-    -   recvfle
+    -   recvfile
 
 2.  **Smb2 R/Ws** in kilobytes/s
 
diff --git a/collectors/python.d.plugin/springboot/README.md b/collectors/python.d.plugin/springboot/README.md
index 46bc2d3568..f38e8bf05a 100644
--- a/collectors/python.d.plugin/springboot/README.md
+++ b/collectors/python.d.plugin/springboot/README.md
@@ -93,7 +93,7 @@ Please refer [Spring Boot Actuator: Production-ready Features](https://docs.spri
     -   MarkSweep
     -   ...
 
-4.  **Heap Mmeory Usage** in KB
+4.  **Heap Memory Usage** in KB
 
     -   used
     -   committed
diff --git a/collectors/statsd.plugin/README.md b/collectors/statsd.plugin/README.md
index d5bc0d1ad5..332b60e735 100644
--- a/collectors/statsd.plugin/README.md
+++ b/collectors/statsd.plugin/README.md
@@ -38,7 +38,7 @@ Netdata fully supports the statsd protocol. All statsd client libraries can be u
 
      `:value` can be omitted and statsd will assume it is `1`. `|c`, `|C` and `|m` can be omitted an statsd will assume it is `|m`. So, the application may send just `name` and statsd will parse it as `name:1|m`.
 
-     For counters use `|c` (esty/statsd compatible) or `|C` (brubeck compatible), for meters use `|m`.
+     For counters use `|c` (etsy/statsd compatible) or `|C` (brubeck compatible), for meters use `|m`.
 
      Sampling rate is supported (check below).
 
@@ -290,7 +290,7 @@ dimension = [pattern] METRIC NAME TYPE MULTIPLIER DIVIDER OPTIONS
 
 `pattern` is a keyword. When set, `METRIC` is expected to be a Netdata simple pattern that will be used to match all the statsd metrics to be added to the chart. So, `pattern` automatically matches any number of statsd metrics, all of which will be added as separate chart dimensions.
 
-`TYPE`, `MUTLIPLIER`, `DIVIDER` and `OPTIONS` are optional.
+`TYPE`, `MULTIPLIER`, `DIVIDER` and `OPTIONS` are optional.
 
 `TYPE` can be:
 
diff --git a/collectors/tc.plugin/README.md b/collectors/tc.plugin/README.md
index 70e31c236b..480076087e 100644
--- a/collectors/tc.plugin/README.md
+++ b/collectors/tc.plugin/README.md
@@ -172,7 +172,7 @@ And this is what you are going to get:
 
 ## QoS Configuration with tc
 
-First, setup the tc rules in rc.local using commands to assign different DSCP markings to different classids. You can see one such example in [github issue #4563](https://github.com/netdata/netdata/issues/4563#issuecomment-455711973). 
+First, setup the tc rules in rc.local using commands to assign different QoS markings to different classids. You can see one such example in [github issue #4563](https://github.com/netdata/netdata/issues/4563#issuecomment-455711973). 
 
 Then, map the classids to names by creating `/etc/iproute2/tc_cls`. For example:
 
diff --git a/daemon/README.md b/daemon/README.md
index 9aa483b711..ec1f1c7ccc 100644
--- a/daemon/README.md
+++ b/daemon/README.md
@@ -514,7 +514,7 @@ section(s) you need to trace.
 
 We have made the most to make Netdata crash free. If however, Netdata crashes on your system, it would be very helpful
 to provide stack traces of the crash. Without them, is will be almost impossible to find the issue (the code base is
-quite large to find such an issue by just objerving it).
+quite large to find such an issue by just observing it).
 
 To provide stack traces, **you need to have Netdata compiled with debugging**. There is no need to enable any tracing
 (`debug flags`).
diff --git a/daemon/config/README.md b/daemon/config/README.md
index 71c0c0e841..a1e2b04b5c 100644
--- a/daemon/config/README.md
+++ b/daemon/config/README.md
@@ -100,7 +100,7 @@ Additionally, there will be the following options:
 |:-----:|:-----:|:---|
 | PATH environment variable|`auto-detected`||
 | PYTHONPATH environment variable||Used to set a custom python path|
-| enable running new plugins|`yes`|When set to `yes`, Netdata will enable detected plugins, even if they are not configured explicitly. Setting this to `no` will only enable plugins explicitly configirued in this file with a `yes`|
+| enable running new plugins|`yes`|When set to `yes`, Netdata will enable detected plugins, even if they are not configured explicitly. Setting this to `no` will only enable plugins explicitly configured in this file with a `yes`|
 | check for new plugins every|60|The time in seconds to check for new plugins in the plugins directory. This allows having other applications dynamically creating plugins for Netdata.|
 | checks|`no`|This is a debugging plugin for the internal latency|
 
@@ -190,7 +190,7 @@ that is information about lines that begin with `dim`, which affect a chart's di
 You may notice some settings that begin with `dim` beneath the ones defined in the table above. These settings determine
 which dimensions appear on the given chart and how Netdata calculates them.
 
-Each dimension setting has the following structure: `dim [DIMENSION ID] [OPTION] = [VALUE]`. The available options are `name`, `algorithm`, `multipler`, and `divisor`.
+Each dimension setting has the following structure: `dim [DIMENSION ID] [OPTION] = [VALUE]`. The available options are `name`, `algorithm`, `multiplier`, and `divisor`.
 
 | Setting      | Function                                                                                                                                                                                                                                                      |
 | :----------- | :------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------ |
diff --git a/docs/Running-behind-apache.md b/docs/Running-behind-apache.md
index 8a547e7b4f..8810dc8fc5 100644
--- a/docs/Running-behind-apache.md
+++ b/docs/Running-behind-apache.md
@@ -365,7 +365,7 @@ apache logs accesses and Netdata logs them too. You can prevent Netdata from gen
 
 ## Troubleshooting mod_proxy
 
-Make sure the requests reach Netdata, by examing `/var/log/netdata/access.log`.
+Make sure the requests reach Netdata, by examining `/var/log/netdata/access.log`.
 
 1.  if the requests do not reach Netdata, your apache does not forward them.
 2.  if the requests reach Netdata but the URLs are wrong, you have not re-written them properly.
diff --git a/docs/collect/container-metrics.md b/docs/collect/container-metrics.md
index e8d7516440..b5bb9da01c 100644
--- a/docs/collect/container-metrics.md
+++ b/docs/collect/container-metrics.md
@@ -65,7 +65,7 @@ collection capabilities.
 ## Collect Kubernetes metrics
 
 We already have a few complementary tools and collectors for monitoring the many layers of a Kubernetes cluster,
-_entirely for free_. These methods work together to help you troubleshoot performance or availablility issues across
+_entirely for free_. These methods work together to help you troubleshoot performance or availability issues across
 your k8s infrastructure.
 
 -   A [Helm chart](https://github.com/netdata/helmchart), which bootstraps a Netdata Agent pod on every node in your
diff --git a/docs/collect/enable-configure.md b/docs/collect/enable-configure.md
index 16e3d8f942..33d7a7bb4f 100644
--- a/docs/collect/enable-configure.md
+++ b/docs/collect/enable-configure.md
@@ -18,10 +18,10 @@ enable or configure a collector to gather all available metrics from your system
 ## Enable a collector or its orchestrator
 
 You can enable/disable collectors individually, or enable/disable entire orchestrators, using their configuration files.
-For example, you can change the behavior of the Go orchestator, or any of its collectors, by editing `go.d.conf`.
+For example, you can change the behavior of the Go orchestrator, or any of its collectors, by editing `go.d.conf`.
 
 Use `edit-config` from your [Netdata config directory](/docs/configure/nodes.md#the-netdata-config-directory) to open
-the orchestrator's primary configuration file:
+the orchestrator primary configuration file:
 
 ```bash
 cd /etc/netdata
@@ -29,7 +29,7 @@ sudo ./edit-config go.d.conf
 ```
 
 Within this file, you can either disable the orchestrator entirely (`enabled: yes`), or find a specific collector and
-enable/disable it with `yes` and `no` settings. Uncomment any line you change to ensure the Netdata deamon reads it on
+enable/disable it with `yes` and `no` settings. Uncomment any line you change to ensure the Netdata daemon reads it on
 start.
 
 After you make your changes, restart the Agent with `service netdata restart`.
diff --git a/docs/collect/how-collectors-work.md b/docs/collect/how-collectors-work.md
index 6818e79427..5ae444a6f2 100644
--- a/docs/collect/how-collectors-work.md
+++ b/docs/collect/how-collectors-work.md
@@ -55,7 +55,7 @@ terms related to collecting metrics.
 
 -   **Modules** are a type of collector.
 -   **Orchestrators** are external plugins that run and manage one or more modules. They run as independent processes.
-    The Go orchestator is in active development.
+    The Go orchestrator is in active development.
     -   [go.d.plugin](https://learn.netdata.cloud/docs/agent/collectors/go.d.plugin/): An orchestrator for data
         collection modules written in `go`.
     -   [python.d.plugin](/collectors/python.d.plugin/README.md): An orchestrator for data collection modules written in
diff --git a/docs/export/external-databases.md b/docs/export/external-databases.md
index 3b7753903b..309b03a878 100644
--- a/docs/export/external-databases.md
+++ b/docs/export/external-databases.md
@@ -7,7 +7,7 @@ custom_edit_url: https://github.com/netdata/netdata/edit/master/docs/export/exte
 # Export metrics to external time-series databases
 
 Netdata allows you to export metrics to external time-series databases with the [exporting
-engine](/exporting/README.md). This system uses a number of **connectors** to intiate connections to [more than
+engine](/exporting/README.md). This system uses a number of **connectors** to initiate connections to [more than
 thirty](#supported-databases) supported databases, including InfluxDB, Prometheus, Graphite, ElasticSearch, and much
 more. 
 
diff --git a/docs/guides/collect-apache-nginx-web-logs.md b/docs/guides/collect-apache-nginx-web-logs.md
index 6a32c8971d..215ced3efb 100644
--- a/docs/guides/collect-apache-nginx-web-logs.md
+++ b/docs/guides/collect-apache-nginx-web-logs.md
@@ -10,7 +10,7 @@ performance of their web servers, and Netdata is taking important steps to make
 
 By parsing web server log files with Netdata, and seeing the volume of redirects, requests, or server errors over time,
 you can better understand what's happening on your infrastructure. Too many bad requests? Maybe a recent deploy missed a
-few small SVG icons. Too many requsests? Time to batten down the hatches—it's a DDoS.
+few small SVG icons. Too many requests? Time to batten down the hatches—it's a DDoS.
 
 Netdata has been capable of monitoring web log files for quite some time, thanks for the [weblog python.d
 module](/collectors/python.d.plugin/web_log/README.md), but we recently refactored this module in Go, and that effort
diff --git a/docs/guides/longer-metrics-storage.md b/docs/guides/longer-metrics-storage.md
index e58265c55f..85b397f6d4 100644
--- a/docs/guides/longer-metrics-storage.md
+++ b/docs/guides/longer-metrics-storage.md
@@ -53,7 +53,7 @@ size` and `dbengine multihost disk space`.
 
 `page cache size` sets the maximum amount of RAM (in MiB) the database engine will use for caching and indexing.
 `dbengine multihost disk space` sets the maximum disk space (again, in MiB) the database engine will use for storing
-compressed metrics. The default settings retain about two day's worth of metris on a system collecting 2,000 metrics
+compressed metrics. The default settings retain about two day's worth of metrics on a system collecting 2,000 metrics
 every second.
 
 [**See our database engine
diff --git a/docs/guides/monitor/kubernetes-k8s-netdata.md b/docs/guides/monitor/kubernetes-k8s-netdata.md
index 097cc13ac9..40af0e94ed 100644
--- a/docs/guides/monitor/kubernetes-k8s-netdata.md
+++ b/docs/guides/monitor/kubernetes-k8s-netdata.md
@@ -13,7 +13,7 @@ troubleshoot issues with your cluster.
 
 Some k8s providers, like GKE (Google Kubernetes Engine), do deploy clusters bundled with monitoring capabilities, such
 as Google Stackdriver Monitoring. However, these pre-configured solutions might not offer the depth of metrics,
-customization, or integration with your perferred alerting methods. 
+customization, or integration with your preferred alerting methods. 
 
 Without this visibility, it's like you built an entire house and _then_ smashed your way through the finished walls to
 add windows.
@@ -23,7 +23,7 @@ you actively troubleshoot anomalies or outages. Better yet, this toolkit include
 let you monitor the many layers of a Kubernetes cluster entirely for free.
 
 We already have a few complementary tools and collectors for monitoring the many layers of a Kubernetes cluster,
-_entirely for free_. These methods work together to help you troubleshoot performance or availablility issues across
+_entirely for free_. These methods work together to help you troubleshoot performance or availability issues across
 your k8s infrastructure.
 
 -   A [Helm chart](https://github.com/netdata/helmchart), which bootstraps a Netdata Agent pod on every node in your
diff --git a/docs/guides/monitor/pi-hole-raspberry-pi.md b/docs/guides/monitor/pi-hole-raspberry-pi.md
index 97db0a1bfa..a180466fb0 100644
--- a/docs/guides/monitor/pi-hole-raspberry-pi.md
+++ b/docs/guides/monitor/pi-hole-raspberry-pi.md
@@ -31,7 +31,7 @@ directly using a keyboard, mouse, and monitor.
 Netdata helps you monitor and troubleshoot all kinds of devices and the applications they run, including IoT devices
 like the Raspberry Pi and applications like Pi-hole.
 
-After a two-minute installation and with zero configuration, you'll be able to seeall of Pi-hole's metrics, including
+After a two-minute installation and with zero configuration, you'll be able to see all of Pi-hole's metrics, including
 the volume of queries, connected clients, DNS queries per type, top clients, top blocked domains, and more.
 
 With Netdata installed, you can also monitor system metrics and any other applications you might be running. By default,
@@ -107,7 +107,7 @@ walkthrough of all its features. For a more expedited tour, see the [get started
 You need to manually enable Netdata's built-in [temperature sensor
 collector](https://learn.netdata.cloud/docs/agent/collectors/charts.d.plugin/sensors) to start collecting metrics.
 
-> Netdata uses a few plugins to manage its [collectors](/collectors/REFERENCE.md), each using a different lanaguge: Go,
+> Netdata uses a few plugins to manage its [collectors](/collectors/REFERENCE.md), each using a different language: Go,
 > Python, Node.js, and Bash. While our Go collectors are undergoing the most active development, we still support the
 > other languages. In this case, you need to enable a temperature sensor collector that's written in Bash.
 
diff --git a/docs/guides/monitor/process.md b/docs/guides/monitor/process.md
index c1a2d031a8..893e6b7049 100644
--- a/docs/guides/monitor/process.md
+++ b/docs/guides/monitor/process.md
@@ -15,7 +15,7 @@ and performance of your infrastructure.
 One of these layers is the _process_. Every time a Linux system runs a program, it creates an independent process that
 executes the program's instructions in parallel with anything else happening on the system. Linux systems track the
 state and resource utilization of processes using the [`/proc` filesystem](https://en.wikipedia.org/wiki/Procfs), and
-Netdata is designed to hook into those metrics to create meaningul visualizations out of the box.
+Netdata is designed to hook into those metrics to create meaningful visualizations out of the box.
 
 While there are a lot of existing command-line tools for tracking processes on Linux systems, such as `ps` or `top`,
 only Netdata provides dozens of real-time charts, at both per-second and event frequency, without you having to write
@@ -86,7 +86,7 @@ Linux systems:
     -   Pipes open (`apps.pipes`)
 -   Swap memory
     -   Swap memory used (`apps.swap`)
-    -   Major page faults (i.e. swap activiy, `apps.major_faults`)
+    -   Major page faults (i.e. swap activity, `apps.major_faults`)
 -   Network
     -   Sockets open (`apps.sockets`)
 -   eBPF file
@@ -132,7 +132,7 @@ sudo ./edit-config apps_groups.conf
 
 Inside the file are lists of process names, oftentimes using wildcards (`*`), that the Netdata Agent looks for and
 groups together. For example, the Netdata Agent looks for processes starting with `mysqld`, `mariad`, `postgres`, and
-others, and groups them into `sql`. That makes sense, since all these procesess are for SQL databases.
+others, and groups them into `sql`. That makes sense, since all these processes are for SQL databases.
 
 ```conf
 sql: mysqld* mariad* postgres* postmaster* oracle_* ora_* sqlservr
@@ -247,7 +247,7 @@ metrics](https://user-images.githubusercontent.com/1153921/101411810-d08fb800-38
 
 ### Using Netdata's eBPF collector (`ebpf.plugin`)
 
-Netdata's eBPF collector puts its charts in two places. Of most imporance to process monitoring are the **ebpf file**,
+Netdata's eBPF collector puts its charts in two places. Of most importance to process monitoring are the **ebpf file**,
 **ebpf syscall**, **ebpf process**, and **ebpf net** sub-sections under **Applications**, shown in the above screenshot.
 
 For example, running the above workload shows the entire "story" how MySQL interacts with the Linux kernel to open
@@ -274,7 +274,7 @@ piece of data needed to discover the root cause of an incident. See our [collect
 setup](/docs/collect/enable-configure.md) doc for details.
 
 [Create new dashboards](/docs/visualize/create-dashboards.md) in Netdata Cloud using charts from `apps.plugin`,
-`ebpf.plugin`, and application-specific collectors to build targeted dashboards for monitoring key procesess across your
+`ebpf.plugin`, and application-specific collectors to build targeted dashboards for monitoring key processes across your
 infrastructure.
 
 Try running [Metric Correlations](https://learn.netdata.cloud/docs/cloud/insights/metric-correlations) on a node that's
diff --git a/docs/guides/step-by-step/step-05.md b/docs/guides/step-by-step/step-05.md
index f54545acf5..5e627632d0 100644
--- a/docs/guides/step-by-step/step-05.md
+++ b/docs/guides/step-by-step/step-05.md
@@ -11,7 +11,7 @@ To accurately monitor the health of your systems and applications, you need to k
 strange going on. Netdata's alarm and notification systems are essential to keeping you informed. 
 
 Netdata comes with hundreds of pre-configured alarms that don't require configuration. They were designed by our
-community of system adminstrators to cover the most important parts of production systems, so, in many cases, you won't
+community of system administrators to cover the most important parts of production systems, so, in many cases, you won't
 need to edit them.
 
 Luckily, Netdata's alarm and notification system are incredibly adaptable to your infrastructure's unique needs.
diff --git a/docs/guides/step-by-step/step-06.md b/docs/guides/step-by-step/step-06.md
index 4ff16136bc..bb1f23494b 100644
--- a/docs/guides/step-by-step/step-06.md
+++ b/docs/guides/step-by-step/step-06.md
@@ -34,7 +34,7 @@ underlying architecture.
 
 By default, Netdata collects a lot of metrics every second using any number of discrete collector. Collectors, in turn,
 are organized and manged by plugins. **Internal** plugins collect system metrics, **external** plugins collect
-non-system metrics, and **orchestrator** plugins group individal collectors together based on the programming language
+non-system metrics, and **orchestrator** plugins group individual collectors together based on the programming language
 they were built in.
 
 These modules are primarily written in [Go](https://learn.netdata.cloud/docs/agent/collectors/go.d.plugin/) (`go.d`) and
diff --git a/docs/guides/step-by-step/step-08.md b/docs/guides/step-by-step/step-08.md
index 3d733878d9..76a1b07750 100644
--- a/docs/guides/step-by-step/step-08.md
+++ b/docs/guides/step-by-step/step-08.md
@@ -76,7 +76,7 @@ following:
 </html>
 ```
 
-Try visiting `http://HOST:19999/custom-dashbord.html` in your browser.
+Try visiting `http://HOST:19999/custom-dashboard.html` in your browser.
 
 If you get a blank page with this text: `Access to file is not permitted: /usr/share/netdata/web/custom-dashboard.html`.
 You can fix this error by changing the dashboard file's permissions to make it owned by the `netdata` user.
diff --git a/docs/guides/step-by-step/step-09.md b/docs/guides/step-by-step/step-09.md
index 8397a96d12..636ffea1f8 100644
--- a/docs/guides/step-by-step/step-09.md
+++ b/docs/guides/step-by-step/step-09.md
@@ -143,7 +143,7 @@ Add the following section to the file:
 ```
 
 [Restart](/docs/getting-started.md#start-stop-and-restart-netdata) Netdata to enable the MongoDB exporting connector.
-Click on the **Netdata Montioring** menu and check out the **exporting my mongo instance** sub-menu. You should start
+Click on the **Netdata Monitoring** menu and check out the **exporting my mongo instance** sub-menu. You should start
 seeing these charts fill up with data about the exporting process!
 
 ![image](https://user-images.githubusercontent.com/1153921/70443852-25171200-1a56-11ea-8be3-494544b1c295.png)
diff --git a/docs/guides/step-by-step/step-10.md b/docs/guides/step-by-step/step-10.md
index 2fec3895a1..28ab47c67c 100644
--- a/docs/guides/step-by-step/step-10.md
+++ b/docs/guides/step-by-step/step-10.md
@@ -65,7 +65,7 @@ Check out [Nginx's installation
 instructions](https://docs.nginx.com/nginx/admin-guide/installing-nginx/installing-nginx-open-source/) for details on
 other Linux distributions.
 
-Certbot is a tool to help you create and renew certiciate+key pairs for your domain. Visit their
+Certbot is a tool to help you create and renew certificate+key pairs for your domain. Visit their
 [instructions](https://certbot.eff.org/instructions) to get a detailed installation process for your operating system.
 
 ### Fully qualified domain name
diff --git a/docs/guides/troubleshoot/monitor-debug-applications-ebpf.md b/docs/guides/troubleshoot/monitor-debug-applications-ebpf.md
index 1106be541d..342193c583 100644
--- a/docs/guides/troubleshoot/monitor-debug-applications-ebpf.md
+++ b/docs/guides/troubleshoot/monitor-debug-applications-ebpf.md
@@ -120,7 +120,7 @@ calls to open/close files, call functions like `do_fork`, IO activity on the VFS
 See the [eBPF collector documentation](/collectors/ebpf.plugin/README.md#integration-with-appsplugin) for the full list
 of per-application charts.
 
-Let's show some examples of how you can first identify normal eBPF patterns, then use that knowledge to idenfity
+Let's show some examples of how you can first identify normal eBPF patterns, then use that knowledge to identify
 anomalies in a few simulated scenarios.
 
 For example, the following screenshot shows the number of open files, failures to open files, and closed files on a
@@ -252,7 +252,7 @@ Debugging and troubleshooting an application takes a special combination of prac
 Netdata's eBPF metrics to back you up, you can rest assured that you see every minute detail of how your application
 interacts with the Linux kernel.
 
-If you're still trying to wrap your head aroud what we offer, be sure to read up on our accompanying documentation and
+If you're still trying to wrap your head around what we offer, be sure to read up on our accompanying documentation and
 other resources on eBPF monitoring with Netdata:
 
 -   [eBPF collector](/collectors/ebpf.plugin/README.md)
diff --git a/docs/netdata-security.md b/docs/netdata-security.md
index e020c0c8d2..50c6b05482 100644
--- a/docs/netdata-security.md
+++ b/docs/netdata-security.md
@@ -131,7 +131,7 @@ to IP addresses within the `160.1.x.x` range and that reverse DNS is setup for t
 
 #### Use an authenticating web server in proxy mode
 
-Use one web server to provide authentication in front of **all your Netdata servers**. So, you will be accessing all your Netdata with URLs like `http://{HOST}/netdata/{NETDATA_HOSTNAME}/` and authentication will be shared among all of them (you will sign-in once for all your servers). Instructions are provided on how to set the proxy configuration to have Netdata run behind [nginx](Running-behind-nginx.md), [Apache](Running-behind-apache.md), [lighthttpd](Running-behind-lighttpd.md) and [Caddy](Running-behind-caddy.md).
+Use one web server to provide authentication in front of **all your Netdata servers**. So, you will be accessing all your Netdata with URLs like `http://{HOST}/netdata/{NETDATA_HOSTNAME}/` and authentication will be shared among all of them (you will sign-in once for all your servers). Instructions are provided on how to set the proxy configuration to have Netdata run behind [nginx](Running-behind-nginx.md), [Apache](Running-behind-apache.md), [lighttpd](Running-behind-lighttpd.md) and [Caddy](Running-behind-caddy.md).
 
 To use this method, you should firewall protect all your Netdata servers, so that only the web server IP will allowed to directly access Netdata. To do this, run this on each of your servers (or use your firewall manager):
 
@@ -212,7 +212,7 @@ If sending this information to the central Netdata registry violates your securi
 Starting with v1.12, Netdata collects anonymous usage information by default and sends it to Google Analytics. Read
 about the information collected, and learn how to-opt, on our [anonymous statistics](anonymous-statistics.md) page.
 
-The usage statistics are _vital_ for us, as we use them to discover bugs and priortize new features. We thank you for
+The usage statistics are _vital_ for us, as we use them to discover bugs and prioritize new features. We thank you for
 _actively_ contributing to Netdata's future.
 
 ## Netdata directories
diff --git a/exporting/README.md b/exporting/README.md
index d10ea32aa8..933de0e075 100644
--- a/exporting/README.md
+++ b/exporting/README.md
@@ -92,7 +92,7 @@ X seconds (though, it can send them per second if you need it to).
 
 ## Configuration
 
-Here are the configruation blocks for every supported connector. Your current `exporting.conf` file may look a little
+Here are the configuration blocks for every supported connector. Your current `exporting.conf` file may look a little
 different. 
 
 You can configure each connector individually using the available [options](#options). The
@@ -234,7 +234,7 @@ Configure individual connectors and override any global settings with the follow
 
 -   `prefix = Netdata`, is the prefix to add to all metrics.
 
--   `update every = 10`, is the number of seconds between sending data to the external datanase. Netdata will add some
+-   `update every = 10`, is the number of seconds between sending data to the external database. Netdata will add some
     randomness to this number, to prevent stressing the external server when many Netdata servers send data to the same
     database. This randomness does not affect the quality of the data, only the time they are sent.
 
@@ -266,7 +266,7 @@ Configure individual connectors and override any global settings with the follow
 -   `send configured labels = yes | no` controls if labels defined in the `[host labels]` section in `netdata.conf`
     should be sent to the external database
 
--   `send automatic labels = yes | no` controls if automatially created labels, like `_os_name` or `_architecture`
+-   `send automatic labels = yes | no` controls if automatically created labels, like `_os_name` or `_architecture`
     should be sent to the external database
 
 > Starting from Netdata v1.20 the host tags (defined in the `[backend]` section of `netdata.conf`) are parsed in
diff --git a/exporting/TIMESCALE.md b/exporting/TIMESCALE.md
index 8a074b5836..c98003ed4d 100644
--- a/exporting/TIMESCALE.md
+++ b/exporting/TIMESCALE.md
@@ -39,7 +39,7 @@ TimescaleDB.
 Finally, another member of Netdata's community has built a project that quickly launches Netdata, TimescaleDB, and
 Grafana in easy-to-manage Docker containers. Rune Juhl Jacobsen's
 [project](https://github.com/runejuhl/grafana-timescaledb) uses a `Makefile` to create everything, which makes it
-perferct for testing and experimentation.
+perfect for testing and experimentation.
 
 ## Netdata&#8596;TimescaleDB in action
 
diff --git a/health/REFERENCE.md b/health/REFERENCE.md
index b7301b3a2b..bc5f40ccd5 100644
--- a/health/REFERENCE.md
+++ b/health/REFERENCE.md
@@ -21,7 +21,7 @@ You can configure the Agent's health watchdog service by editing files in two lo
     altogether, run health checks more or less often, and more. See [daemon
     configuration](/daemon/config/README.md#health-section-options) for a table of all the available settings, their
     default values, and what they control.
--   The individual `.conf` files in `health.d/`. These health entitiy files are organized by the type of metric they are
+-   The individual `.conf` files in `health.d/`. These health entity files are organized by the type of metric they are
     performing calculations on or their associated collector. You should edit these files using the `edit-config`
     script. For example: `sudo ./edit-config health.d/cpu.conf`.
 
@@ -241,7 +241,7 @@ A `calc` is designed to apply some calculation to the values or variables availa
 calculation will be made available at the `$this` variable, overwriting the value from your `lookup`, to use in warning
 and critical expressions.
 
-When paired with `lookup`, `calc` will perform the calculation just after `lookup` has retreived a value from Netdata's
+When paired with `lookup`, `calc` will perform the calculation just after `lookup` has retrieved a value from Netdata's
 database.
 
 You can use `calc` without `lookup` if you are using [other available variables](#variables).
@@ -340,7 +340,7 @@ delay: [[[up U] [down D] multiplier M] max X]
      will delay the notification by 1 minute. This is used to prevent notifications for flapping
      alarms. The default `D` is zero.
 
--   `mutliplier M` multiplies `U` and `D` when an alarm changes state, while a notification is
+-   `multiplier M` multiplies `U` and `D` when an alarm changes state, while a notification is
      delayed. The default multiplier is `1.0`.
 
 -   `max X`  defines the maximum absolute notification delay an alarm may get. The default `X`
diff --git a/health/notifications/alerta/README.md b/health/notifications/alerta/README.md
index ff8c1aabbf..8f1679a096 100644
--- a/health/notifications/alerta/README.md
+++ b/health/notifications/alerta/README.md
@@ -29,7 +29,7 @@ The easiest way to install Alerta is to use the Docker image available
 on [Docker hub][1]. Alternatively, follow the ["getting started"][2]
 tutorial to deploy Alerta to an Ubuntu server. More advanced
 configurations are out os scope of this tutorial but information
-about different deployment scenaries can be found in the  [docs][3].
+about different deployment scenarios can be found in the  [docs][3].
 
 [1]: https://hub.docker.com/r/alerta/alerta-web/
 
@@ -86,7 +86,7 @@ We can test alarms using the standard approach:
 
 Note: Netdata will send 3 alarms, and because last alarm is "CLEAR"
 you will not see them in main Alerta page, you need to select to see
-"closed" alarma in top-right lookup. A little change in `alarm-notify.sh`
+"closed" alarm in top-right lookup. A little change in `alarm-notify.sh`
 that let us test each state one by one will be useful.
 
 For more information see <https://docs.alerta.io>
diff --git a/health/notifications/awssns/README.md b/health/notifications/awssns/README.md
index 038d9bcc86..c682513792 100644
--- a/health/notifications/awssns/README.md
+++ b/health/notifications/awssns/README.md
@@ -37,7 +37,7 @@ Once that's done, you're ready to go and can specify the desired topic ARN as a
 Notes:
 
 -   Netdata's native email notification support is far better in almost all respects than it's support through Amazon SNS.  If you want email notifications, use the native support, not SNS.
-    -   If you need to change the notification format for SNS notifications, you can do so by specifying the format in `AWSSNS_MESSAGE_FORMAT` in the configuration.  This variable supports all the same vairiables you can use in custom notifications.
+    -   If you need to change the notification format for SNS notifications, you can do so by specifying the format in `AWSSNS_MESSAGE_FORMAT` in the configuration.  This variable supports all the same variables you can use in custom notifications.
     -   While Amazon SNS supports sending differently formatted messages for different delivery methods, Netdata does not currently support this functionality.
 
 [![analytics](https://www.google-analytics.com/collect?v=1&aip=1&t=pageview&_s=1&ds=github&dr=https%3A%2F%2Fgithub.com%2Fnetdata%2Fnetdata&dl=https%3A%2F%2Fmy-netdata.io%2Fgithub%2Fhealth%2Fnotifications%2Fawssns%2FREADME&_u=MAC~&cid=5792dfd7-8dc4-476b-af31-da2fdb9f93d2&tid=UA-64295674-3)](<>)
diff --git a/health/notifications/email/README.md b/health/notifications/email/README.md
index beac705439..827a9c0be0 100644
--- a/health/notifications/email/README.md
+++ b/health/notifications/email/README.md
@@ -55,7 +55,7 @@ sendmail="/usr/bin/msmtp"
 (sudo) su -s /bin/bash netdata
 ```
 - Configure `~/.msmtprc` as shown [in the documentation](https://marlam.de/msmtp/documentation/).
-- Finaly set the appropriate permissions on the `.msmtprc` file :
+- Finally set the appropriate permissions on the `.msmtprc` file :
 ```sh
 chmod 600 ~/.msmtprc
 ```
diff --git a/health/notifications/irc/README.md b/health/notifications/irc/README.md
index 8ee440397b..e7f22e1fd0 100644
--- a/health/notifications/irc/README.md
+++ b/health/notifications/irc/README.md
@@ -29,7 +29,7 @@ Set the path for `nc` in `/etc/netdata/health_alarm_notify.conf` (to edit it on
 nc="/usr/bin/nc"
 ```
 
-2.  Αn `IRC_NETWORK` to which your preffered channels belong to.   
+2.  Αn `IRC_NETWORK` to which your preferred channels belong to.   
 3.  One or more channels ( `DEFAULT_RECIPIENT_IRC` ) to post the messages to.   
 4.  An `IRC_NICKNAME` and an `IRC_REALNAME` to identify in IRC.   
 
diff --git a/health/notifications/prowl/README.md b/health/notifications/prowl/README.md
index 4c8bfe453f..7c60de2704 100644
--- a/health/notifications/prowl/README.md
+++ b/health/notifications/prowl/README.md
@@ -6,7 +6,7 @@ custom_edit_url: https://github.com/netdata/netdata/edit/master/health/notificat
 # Prowl
 
 [Prowl](https://www.prowlapp.com/) is a push notification service for iOS devices.  Netdata
-supprots delivering notifications to iOS devices through Prowl.
+supports delivering notifications to iOS devices through Prowl.
 
 Because of how Netdata integrates with Prowl, there is a hard limit of
 at most 1000 notifications per hour (starting from the first notification
@@ -20,7 +20,7 @@ the alert, directly to the chart that it triggered on.
 
 ## configuration
 
-To use this, you will need a Prowl API key, which can be rquested through
+To use this, you will need a Prowl API key, which can be requested through
 the Prowl website after registering.
 
 Once you have an API key, simply specify that as a recipient for Prowl
diff --git a/health/notifications/rocketchat/README.md b/health/notifications/rocketchat/README.md
index 34a61aaf5b..a54f5826c0 100644
--- a/health/notifications/rocketchat/README.md
+++ b/health/notifications/rocketchat/README.md
@@ -47,6 +47,6 @@ role_recipients_rocketchat[webmaster]="marketing development"
 ```
 
 The keywords `systems`, `databases`, `marketing`, `development` are RocketChat channels (they should already exist).
-Both public and private channels can be used, even if they differ from the channel configured in yout RocketChat incomming webhook.
+Both public and private channels can be used, even if they differ from the channel configured in your RocketChat incoming webhook.
 
 [![analytics](https://www.google-analytics.com/collect?v=1&aip=1&t=pageview&_s=1&ds=github&dr=https%3A%2F%2Fgithub.com%2Fnetdata%2Fnetdata&dl=https%3A%2F%2Fmy-netdata.io%2Fgithub%2Fhealth%2Fnotifications%2Frocketchat%2FREADME&_u=MAC~&cid=5792dfd7-8dc4-476b-af31-da2fdb9f93d2&tid=UA-64295674-3)](<>)
diff --git a/health/notifications/stackpulse/README.md b/health/notifications/stackpulse/README.md
index ddba585360..13d2f7235e 100644
--- a/health/notifications/stackpulse/README.md
+++ b/health/notifications/stackpulse/README.md
@@ -7,7 +7,7 @@ custom_edit_url: https://github.com/netdata/netdata/edit/master/health/notificat
 
 # Send notifications to StackPulse
 
-[StackPulse](https://stackpulse.com/) is a software-as-a-service platform for site reliablility engineering.
+[StackPulse](https://stackpulse.com/) is a software-as-a-service platform for site reliability engineering.
 It helps SREs, DevOps Engineers and Software Developers reduce toil and alert fatigue while improving reliability of 
 software services by managing, analyzing and automating incident response activities.
 
@@ -18,7 +18,7 @@ Sending Netdata alarm notifications to StackPulse allows you to create smart aut
 -   Performing triage actions and analyzing their results
 -   Orchestrating incident management and notification flows
 -   Performing automatic and semi-automatic remediation actions
--   Analzying incident data and remediation patterns to improve reliability of your services
+-   Analyzing incident data and remediation patterns to improve reliability of your services
 
 To send the notification you need:
 
diff --git a/packaging/DISTRIBUTIONS.md b/packaging/DISTRIBUTIONS.md
index 239113f9bc..ca8ab1c646 100644
--- a/packaging/DISTRIBUTIONS.md
+++ b/packaging/DISTRIBUTIONS.md
@@ -19,7 +19,7 @@ have complete visibility over the range of support.
 -   **Family**: The family that the OS belongs to
 -   **CI: Smoke Testing**: Smoke testing has been implemented on our CI, to prevent broken code reaching our users
 -   **CI: Testing**: Testing has been implemented to prevent broken or problematic code reaching our users
--   **CD**: Continious deployment support has been fully enabled for this operating system
+-   **CD**: Continuous deployment support has been fully enabled for this operating system
 -   **.DEB**: We provide a `.DEB` package for that particular operating system
 -   **.RPM**: We provide a `.RPM` package for that particular operating system
 -   **Installer**: Running netdata from source, using our installer, is working for this operating system
diff --git a/packaging/docker/README.md b/packaging/docker/README.md
index ecdde44099..e97cb0c6f3 100644
--- a/packaging/docker/README.md
+++ b/packaging/docker/README.md
@@ -15,7 +15,7 @@ Starting with v1.12, Netdata collects anonymous usage information by default and
 about the information collected, and learn how to-opt, on our [anonymous statistics](/docs/anonymous-statistics.md)
 page.
 
-The usage statistics are _vital_ for us, as we use them to discover bugs and priortize new features. We thank you for
+The usage statistics are _vital_ for us, as we use them to discover bugs and prioritize new features. We thank you for
 _actively_ contributing to Netdata's future.
 
 ## Limitations running the Agent in Docker
@@ -107,7 +107,7 @@ You can control how the health checks run by using the environment variable `NET
     correctly or not. This is sufficient to ensure that Netdata did not
     hang during startup, but does not provide a rigorous verification
     that the daemon is collecting data or is otherwise usable.
--   If set to anything else, the health check will treat the vaule as a
+-   If set to anything else, the health check will treat the value as a
     URL to check for a 200 status code on. In most cases, this should
     start with `http://localhost:19999/` to check the agent running in
     the container.
diff --git a/packaging/installer/REINSTALL.md b/packaging/installer/REINSTALL.md
index 122f1a8473..50a15d799b 100644
--- a/packaging/installer/REINSTALL.md
+++ b/packaging/installer/REINSTALL.md
@@ -11,7 +11,7 @@ Netdata Agent on your node.
 
 Before you try reinstalling Netdata, figure out which [installation method you
 used](/packaging/installer/UPDATE.md#determine-which-installation-method-you-used) if you do not already know. This will
-deterimine the reinstallation method.
+determine the reinstallation method.
 
 ## One-line installer script (`kickstart.sh`)
 
diff --git a/packaging/installer/UPDATE.md b/packaging/installer/UPDATE.md
index 468b25fa85..ba57c8cc5f 100644
--- a/packaging/installer/UPDATE.md
+++ b/packaging/installer/UPDATE.md
@@ -26,7 +26,7 @@ most installations, this is `/etc/netdata`.
 Use `cd` to navigate to the Netdata config directory, then use `ls -a` to look for a file called `.environment`.
 
 -   If the `.environment` file _does not_ exist, reinstall with your [package manager](#deb-or-rpm-packages).
--   If the `.environtment` file _does_ exist, check its contents with `less .environment`.
+-   If the `.environment` file _does_ exist, check its contents with `less .environment`.
     -   If `IS_NETDATA_STATIC_BINARY` is `"yes"`, update using the [pre-built static
         binary](#pre-built-static-binary-for-64-bit-systems-kickstart-static64sh).
     -   In all other cases, update using the [one-line installer script](#one-line-installer-script-kickstartsh).
@@ -118,7 +118,7 @@ installation instructions](/packaging/docker/README.md#create-a-new-netdata-agen
 
 ## macOS
 
-If you installed Netdata on your macOS system using Homebrew, you can explictly request an update:
+If you installed Netdata on your macOS system using Homebrew, you can explicitly request an update:
 
 ```bash
 brew upgrade netdata
diff --git a/packaging/installer/methods/cloud-providers.md b/packaging/installer/methods/cloud-providers.md
index 815eaa085c..943a649b6d 100644
--- a/packaging/installer/methods/cloud-providers.md
+++ b/packaging/installer/methods/cloud-providers.md
@@ -89,7 +89,7 @@ to create a new firewall rule.
 #### Amazon Web Services (AWS) / EC2
 
 Sign in to the [AWS console](https://console.aws.amazon.com/) and navigate to the EC2 dashboard. Click on the **Security
-Groups** link in the naviagtion, beneath the **Network & Security** heading. Find the Security Group your instance
+Groups** link in the navigation, beneath the **Network & Security** heading. Find the Security Group your instance
 belongs to, and either right-click on it or click the **Actions** button above to see a dropdown menu with **Edit
 inbound rules**.
 
diff --git a/packaging/installer/methods/freebsd.md b/packaging/installer/methods/freebsd.md
index e682e7e117..e2af417545 100644
--- a/packaging/installer/methods/freebsd.md
+++ b/packaging/installer/methods/freebsd.md
@@ -80,7 +80,7 @@ The `netdata-updater.sh` script will update your Agent.
 | `--dont-wait` | Run installation in non-interactive mode|
 | `--auto-update` or `-u` | Install netdata-updater in cron to update netdata automatically once per day|
 | `--stable-channel` | Use packages from GitHub release pages instead of GCS (nightly updates). This results in less frequent updates|
-| `--nightly-channel` | Use most recent nightly udpates instead of GitHub releases. This results in more frequent updates|
+| `--nightly-channel` | Use most recent nightly updates instead of GitHub releases. This results in more frequent updates|
 | `--disable-go` | Disable installation of go.d.plugin|
 | `--disable-ebpf` | Disable eBPF Kernel plugin (Default: enabled)|
 | `--disable-cloud` | Disable all Netdata Cloud functionality|
@@ -103,6 +103,6 @@ The `netdata-updater.sh` script will update your Agent.
 | `--disable-lto` | Disable Link-Time-Optimization. Default: enabled|
 | `--disable-x86-sse` | Disable SSE instructions. By default SSE optimizations are enabled|
 | `--zlib-is-really-here` or `--libs-are-really-here` | If you get errors about missing zlib or libuuid but you know it is available, you might have a broken pkg-config. Use this option to proceed without checking pkg-config|
-|`--disable-telemetry` | Use this flag to opt-out from our anonymous telemetry progam. (DO_NOT_TRACK=1)|
+|`--disable-telemetry` | Use this flag to opt-out from our anonymous telemetry program. (DO_NOT_TRACK=1)|
 
 [![analytics](https://www.google-analytics.com/collect?v=1&aip=1&t=pageview&_s=1&ds=github&dr=https%3A%2F%2Fgithub.com%2Fnetdata%2Fnetdata&dl=https%3A%2F%2Fmy-netdata.io%2Fgithub%2Fpackaging%2Finstaller%2Fmethods%2Ffreebsd&_u=MAC~&cid=5792dfd7-8dc4-476b-af31-da2fdb9f93d2&tid=UA-64295674-3)](<>)
diff --git a/packaging/installer/methods/kickstart-64.md b/packaging/installer/methods/kickstart-64.md
index b3f79278ef..120cc9e11b 100644
--- a/packaging/installer/methods/kickstart-64.md
+++ b/packaging/installer/methods/kickstart-64.md
@@ -37,7 +37,7 @@ The `kickstart.sh` script does the following after being downloaded and run:
     added.
 -   Installs `netdata-updater.sh` to `cron.daily` to enable automatic updates, unless you added the `--no-updates`
     option.
--   Prints a message about whether the insallation succeeded for failed for QA purposes.
+-   Prints a message about whether the installation succeeded for failed for QA purposes.
 
 If your shell fails to handle the above one-liner, you can download and run the `kickstart-static64.sh` script manually.
 
@@ -65,7 +65,7 @@ your installation. Here are a few important parameters:
     Netdata better.
 -   `--no-updates`: Prevent automatic updates of any kind.
 -   `--reinstall`: If an existing installation is detected, reinstall instead of attempting to update it. Note
-    that this cannot be used to switch betwen installation types.
+    that this cannot be used to switch between installation types.
 -   `--local-files`: Used for [offline installations](/packaging/installer/methods/offline.md). Pass four file paths:
     the Netdata tarball, the checksum file, the go.d plugin tarball, and the go.d plugin config tarball, to force
     kickstart run the process using those files. This option conflicts with the `--stable-channel` option. If you set
@@ -73,7 +73,7 @@ your installation. Here are a few important parameters:
 
 ## Verify script integrity
 
-To use `md5sum` to verify the intregity of the `kickstart-static64.sh` script you will download using the one-line
+To use `md5sum` to verify the integrity of the `kickstart-static64.sh` script you will download using the one-line
 command above, run the following:
 
 ```bash
diff --git a/packaging/installer/methods/kickstart.md b/packaging/installer/methods/kickstart.md
index f753b6f63e..f825f808e7 100644
--- a/packaging/installer/methods/kickstart.md
+++ b/packaging/installer/methods/kickstart.md
@@ -56,7 +56,7 @@ installation. Here are a few important parameters:
 
 ## Verify script integrity
 
-To use `md5sum` to verify the intregity of the `kickstart.sh` script you will download using the one-line command above,
+To use `md5sum` to verify the integrity of the `kickstart.sh` script you will download using the one-line command above,
 run the following:
 
 ```bash
diff --git a/packaging/installer/methods/kubernetes.md b/packaging/installer/methods/kubernetes.md
index 90fab3df7a..3e85928e99 100644
--- a/packaging/installer/methods/kubernetes.md
+++ b/packaging/installer/methods/kubernetes.md
@@ -89,7 +89,7 @@ applications](https://github.com/netdata/helmchart#service-discovery-and-support
 by our [generic Prometheus collector](https://learn.netdata.cloud/docs/agent/collectors/go.d.plugin/modules/prometheus).
 
 If you haven't changed listening ports, image names, or other defaults, service discovery should find your pods, create
-the proper configurations based on the service that pod runs, and begin monitoring them immediately after depolyment.
+the proper configurations based on the service that pod runs, and begin monitoring them immediately after deployment.
 
 However, if you have changed some of these defaults, you need to copy a file from the Netdata Helm chart repository,
 make your edits, and pass the changed file to `helm install`/`helm upgrade`.
@@ -188,7 +188,7 @@ Cloud](https://user-images.githubusercontent.com/1153921/94497340-c1f49880-01ab-
 ## Update/reinstall the Netdata Helm chart
 
 If you update the Helm chart's configuration, run `helm upgrade` to redeploy your Netdata service, replacing `netdata`
-with the name of the release, if you changed it upon installtion:
+with the name of the release, if you changed it upon installation:
 
 ```bash
 helm upgrade netdata netdata/netdata
@@ -203,7 +203,7 @@ Check out our [infrastructure](/docs/quickstart/infrastructure.md) for details a
 and learn more about [configuring the Netdata Agent](/docs/configure/nodes.md) to better understand the settings you
 might be interested in changing.
 
-To futher configure Netdata for your cluster, see our [Helm chart repository](https://github.com/netdata/helmchart) and
+To further configure Netdata for your cluster, see our [Helm chart repository](https://github.com/netdata/helmchart) and
 the [service discovery repository](https://github.com/netdata/agent-service-discovery/).
 
 [![analytics](https://www.google-analytics.com/collect?v=1&aip=1&t=pageview&_s=1&ds=github&dr=https%3A%2F%2Fgithub.com%2Fnetdata%2Fnetdata&dl=https%3A%2F%2Fmy-netdata.io%2Fgithub%2Fpackaging%2Finstaller%2Fmethods%2Fkubernetes&_u=MAC~&cid=5792dfd7-8dc4-476b-af31-da2fdb9f93d2&tid=UA-64295674-3)](<>)
diff --git a/packaging/installer/methods/manual.md b/packaging/installer/methods/manual.md
index 40c1a0399f..6ece952404 100644
--- a/packaging/installer/methods/manual.md
+++ b/packaging/installer/methods/manual.md
@@ -32,7 +32,7 @@ and other operating systems and is regularly tested. You can find this tool [her
 
 -   **Debian** Linux and its derivatives (including **Ubuntu**, **Mint**)
 
--   **Redhat Enterprise Linux** and its derivatives (including **Fedora**, **CentOS**, **Amazon Machine Image**)
+-   **Red Hat Enterprise Linux** and its derivatives (including **Fedora**, **CentOS**, **Amazon Machine Image**)
     -   Please note that for RHEL/CentOS you need
         [EPEL](http://www.tecmint.com/how-to-enable-epel-repository-for-rhel-centos-6-5/).
         In addition, RHEL/CentOS version 6 also need
@@ -147,7 +147,7 @@ And install the minimum required dependencies.
 ### CentOS / RHEL 8.x
 
 For CentOS / RHEL 8.x a lot of development packages have moved out into their
-own separate repositories. Some other dependeicies are either missing completely
+own separate repositories. Some other dependencies are either missing completely
 or have to be sourced by 3rd-parties.
 
 CentOS 8.x:
diff --git a/packaging/installer/methods/packages.md b/packaging/installer/methods/packages.md
index 1fa2f9cf83..cf1e33591d 100644
--- a/packaging/installer/methods/packages.md
+++ b/packaging/installer/methods/packages.md
@@ -28,8 +28,8 @@ package downloads.
 If you are using such a setup, there are a couple of ways to work around this:
 
 -   Configure your proxy to automatically pass through HTTPS connections without caching them. This is the simplest
-    solution, but means that downloads of Netdata pacakges will not be cached.
--   Mirror the respository locally on your proxy system, and use that mirror when installing on other systems. This
+    solution, but means that downloads of Netdata packages will not be cached.
+-   Mirror the repository locally on your proxy system, and use that mirror when installing on other systems. This
     requires more setup and more disk space on the caching host, but it lets you cache the packages locally.
 -   Some specific caching proxies may have alternative configuration options to deal with these issues. Find
     such options in their documentation.
diff --git a/packaging/installer/methods/pfsense.md b/packaging/installer/methods/pfsense.md
index fa490f6511..3146436c76 100644
--- a/packaging/installer/methods/pfsense.md
+++ b/packaging/installer/methods/pfsense.md
@@ -48,7 +48,7 @@ pkg add http://pkg.freebsd.org/FreeBSD:11:amd64/latest/All/py37-yaml-5.3.1.txz
 > Python from the FreeBSD repository as instructed above.
 
 > ⚠️ If you are using the `apcupsd` collector, you need to make sure that apcupsd is up before starting Netdata.
-> Otherwise a infinitely running `cat` process triggered by the default activated apcuspd charts plugin will eat up CPU
+> Otherwise a infinitely running `cat` process triggered by the default activated apcupsd charts plugin will eat up CPU
 > and RAM (`/tmp/.netdata-charts.d-*/run-*`). This also applies to `OPNsense`.
 
 ## Install Netdata
diff --git a/packaging/installer/methods/source.md b/packaging/installer/methods/source.md
index 929252426e..e0827fc17c 100644
--- a/packaging/installer/methods/source.md
+++ b/packaging/installer/methods/source.md
@@ -62,9 +62,9 @@ Netdata Cloud functionality. To prepare this library for the build system:
     of `packaging/mosquitto.version` in your Netdata sources.
 2.  Obtain the sources for that version by either:
     -   Navigating to https://github.com/netdata/mosquitto/releases and
-        donwloading and unpacking the source code archive for that release.
+        downloading and unpacking the source code archive for that release.
     -   Cloning the repository with `git` and checking out the required tag.
-3.  If building on a platfom other than Linux, prepare the mosquitto
+3.  If building on a platform other than Linux, prepare the mosquitto
     sources by running `cmake -D WITH_STATIC_LIBRARIES:boolean=YES .` in
     the mosquitto source directory.
 4.  Build mosquitto by running `make -C lib` in the mosquitto source directory.
@@ -89,9 +89,9 @@ library for the build system:
     of `packaging/libwebsockets.version` in your Netdata sources.
 2.  Obtain the sources for that version by either:
     -   Navigating to https://github.com/warmcat/libwebsockets/releases and
-        donwloading and unpacking the source code archive for that release.
+        downloading and unpacking the source code archive for that release.
     -   Cloning the repository with `git` and checking out the required tag.
-3.  Prepare the libweboskcets sources by running `cmake -D
+3.  Prepare the libwebsockets sources by running `cmake -D
     LWS_WITH_SOCKS5:bool=ON .` in the libwebsockets source directory.
 4.  Build libwebsockets by running `make` in the libwebsockets source
     directory.
@@ -112,7 +112,7 @@ you can do the following to prepare a copy for the build system:
 1.  Verify the tag that Netdata expects to be used by checking the contents
     of `packaging/jsonc.version` in your Netdata sources.
 2.  Obtain the sources for that version by either:
-    -   Navigating to https://github.com/json-c/json-c and donwloading
+    -   Navigating to https://github.com/json-c/json-c and downloading
         and unpacking the source code archive for that release.
     -   Cloning the repository with `git` and checking out the required tag.
 3.  Prepare the JSON-C sources by running `cmake -DBUILD_SHARED_LIBS=OFF .`
@@ -201,7 +201,7 @@ NPM. Once you have the required tools, do the following:
 1.  Verify the release version that Netdata expects to be used by checking
     the contents of `packaging/dashboard.version` in your Netdata sources.
 2.  Obtain the sources for that version by either:
-    -   Navigating to https://github.com/netdata/dashboard and donwloading
+    -   Navigating to https://github.com/netdata/dashboard and downloading
         and unpacking the source code archive for that release.
     -   Cloning the repository with `git` and checking out the required tag.
 3.  Run `npm install` in the dashboard source tree.
@@ -216,7 +216,7 @@ and are developed in a separate repository from the mian Netdata code.
 An installation without these collectors is still usable, but will be
 unable to collect metrics for a number of network services the system
 may be providing. You can either install a pre-built copy of these
-eollectors, or build them locally.
+collectors, or build them locally.
 
 #### Installing the pre-built Go collectors
 
@@ -229,7 +229,7 @@ we officially support. To use one of these:
     required release, and download the `go.d.plugin-*.tar.gz` file
     for your system type and CPu architecture and the `config.tar.gz`
     configuration file archive.
-3.  Extract the `go.d.plugin-*.tar.gz` archive into a temprary
+3.  Extract the `go.d.plugin-*.tar.gz` archive into a temporary
     location, and then copy the single file in the archive to
     `/usr/libexec/netdata/plugins.d` or the equivalent location for your
     build of Netdata and rename it to `go.d.plugin`.
@@ -246,12 +246,12 @@ newer. Once you have the required tools, do the following:
 1.  Verify the release version that Netdata expects to be used by checking
     the contents of `packaging/go.d.version` in your Netdata sources.
 2.  Obtain the sources for that version by either:
-    -   Navigating to https://github.com/netdata/go.d.plugin and donwloading
+    -   Navigating to https://github.com/netdata/go.d.plugin and downloading
         and unpacking the source code archive for that release.
     -   Cloning the repository with `git` and checking out the required tag.
 3.  Run `make` in the go.d.plugin source tree.
 4.  Copy `bin/godplugin` to `/usr/libexec/netdata/plugins.d` or th
-    eequivalent location for your build of Netdata and rename it to
+    equivalent location for your build of Netdata and rename it to
     `go.d.plugin`.
 5.  Copy the contents of the `config` directory to `/etc/netdata` or the
     equivalent location for your build of Netdata.
@@ -274,7 +274,7 @@ using glibc or musl. To use one of these:
     the contents of `packaging/ebpf.version` in your Netdata sources.
 2.  Go to https://github.com/netdata/kernel-collector/releases, select the
     required release, and download the `netdata-kernel-collector-*.tar.xz`
-    file for the libc variant your system uses (eithe rmusl or glibc).
+    file for the libc variant your system uses (either rmusl or glibc).
 3.  Extract the contents of the archive to a temporary location, and then
     copy all of the `.o` and `.so.*` files and the contents of the `library/`
     directory to `/usr/libexec/netdata/plugins.d` or the equivalent location
diff --git a/packaging/maintainers/README.md b/packaging/maintainers/README.md
index 72f7e61c8e..903c37d5f6 100644
--- a/packaging/maintainers/README.md
+++ b/packaging/maintainers/README.md
@@ -22,7 +22,7 @@ This page tracks the package maintainers for Netdata, for various operating syst
 | Debian | Release | @lhw @FedericoCeratto | [netdata @ debian](http://salsa.debian.org/debian/netdata) |
 | Slackware | Release | @willysr | [netdata @ slackbuilds](https://slackbuilds.org/repository/14.2/system/netdata/) |
 | Ubuntu |  |  |  |
-| Red Hat / Fedora / Centos |  |  |  |
+| Red Hat / Fedora / CentOS |  |  |  |
 | SUSE SLE / openSUSE Tumbleweed & Leap |  |  | [netdata @ SUSE OpenBuildService](https://software.opensuse.org/package/netdata) |
 
 ---
diff --git a/packaging/makeself/README.md b/packaging/makeself/README.md
index a265914cb3..c898a94049 100644
--- a/packaging/makeself/README.md
+++ b/packaging/makeself/README.md
@@ -49,6 +49,6 @@ If Netdata crashes, `valgrind` will print a stack trace of the issue. Open a git
 
 To stop Netdata while it runs under `valgrind`, press Control-C on the console.
 
-> If you omit the parameter `--undef-value-errors=no` to valgrind, you will get hundreds of errors about conditional jumps that depend on uninitialized values. This is normal. Valgrind has heuristics to prevent it from printing such errors for system libraries, but for the static Netdata binary, all the required libraries are built into Netdata. So, valgrind cannot appply its heuristics and prints them.
+> If you omit the parameter `--undef-value-errors=no` to valgrind, you will get hundreds of errors about conditional jumps that depend on uninitialized values. This is normal. Valgrind has heuristics to prevent it from printing such errors for system libraries, but for the static Netdata binary, all the required libraries are built into Netdata. So, valgrind cannot apply its heuristics and prints them.
 
 [![analytics](https://www.google-analytics.com/collect?v=1&aip=1&t=pageview&_s=1&ds=github&dr=https%3A%2F%2Fgithub.com%2Fnetdata%2Fnetdata&dl=https%3A%2F%2Fmy-netdata.io%2Fgithub%2Fmakeself%2FREADME&_u=MAC~&cid=5792dfd7-8dc4-476b-af31-da2fdb9f93d2&tid=UA-64295674-3)](<>)
diff --git a/parser/README.md b/parser/README.md
index 3bc48bc906..50d55e3d63 100644
--- a/parser/README.md
+++ b/parser/README.md
@@ -6,7 +6,7 @@ Usage
 
 1. Define a structure that will be used to share user state across calls 
 1. Initialize the parser using `parser_init`
-2. Register keywords and assosiated callback function using `parser_add_keyword`
+2. Register keywords and associated callback function using `parser_add_keyword`
 3. Register actions on the keywords 
 4. Start a loop until EOF
    1.  Fetch the next line using `parser_next`
@@ -79,11 +79,11 @@ Input
       * PARSER_RC_ERROR -- Callback failed, exit
 
 Output
-- The correspoding keyword and callback will be registered
+- The corresponding keyword and callback will be registered
   
 Returns
 - 0 maximum callbacks already registered for this keyword
-- > 0 which is the number of callbacks assosiated with this keyword.
+- > 0 which is the number of callbacks associated with this keyword.
 
    
 ----
diff --git a/streaming/README.md b/streaming/README.md
index 34d899c3da..94ab1f2e80 100644
--- a/streaming/README.md
+++ b/streaming/README.md
@@ -287,7 +287,7 @@ Now you update the list of certificates running the following, again either as `
 ```
 
 > Some Linux distributions have different methods of updating the certificate list. For more details, please read this
-> guide on [addding trusted root certificates](https://github.com/Busindre/How-to-Add-trusted-root-certificates).
+> guide on [adding trusted root certificates](https://github.com/Busindre/How-to-Add-trusted-root-certificates).
 
 Once you update your certificate list, you can set the stream parameters for Netdata to trust the parent certificate. Open `stream.conf` for editing and change the following lines:
 
diff --git a/web/api/badges/README.md b/web/api/badges/README.md
index 582c80c611..b5fc534684 100644
--- a/web/api/badges/README.md
+++ b/web/api/badges/README.md
@@ -188,10 +188,10 @@ These are options dedicated to badges:
     The following parameters specify colors of each individual part of the badge. Each parameter is documented in detail
     below.
 
-    | Area of badge      | Backgroud color parameter | Text color parameter |
-    | ---:               | :-----------------------: | :------------------: |
-    | Label (left) part  | `label_color`             | `text_color_lbl`     |
-    | Value (right) part | `value_color`             | `text_color_val`     |
+    | Area of badge      | Background color parameter | Text color parameter |
+    | ---:               | :------------------------: | :------------------: |
+    | Label (left) part  | `label_color`              | `text_color_lbl`     |
+    | Value (right) part | `value_color`              | `text_color_val`     |
 
     -   `label_color=COLOR`
 
@@ -223,7 +223,7 @@ These are options dedicated to badges:
 
         The above will set `grey` if no value exists (not collected within the `gap when lost iterations above` in
         `netdata.conf` for the chart), `green` if the value is less than 10, `yellow` if the value is less than 100, and
-        so on. Netdata will use `red` if no other conditions match. Only integers are suported as values.
+        so on. Netdata will use `red` if no other conditions match. Only integers are supported as values.
 
         The supported operators are `<`, `>`, `<=`, `>=`, `=` (or `:`), and `!=` (or `<>`).
 
diff --git a/web/api/health/README.md b/web/api/health/README.md
index b007ef1f55..b6e8b5c234 100644
--- a/web/api/health/README.md
+++ b/web/api/health/README.md
@@ -75,7 +75,7 @@ curl "http://NODE:19999/api/v1/manage/health?cmd=RESET" -H "X-Auth-Token: Mytoke
 By default access to the health management API is only allowed from `localhost`. Accessing the API from anything else will return a 403 error with the message `You are not allowed to access this resource.`. You can change permissions by editing the `allow management from` variable in `netdata.conf` within the [web] section. See [web server access lists](/web/server/README.md#access-lists) for more information.
 
 The command `RESET` just returns Netdata to the default operation, with all health checks and notifications enabled.
-If you've configured and entered your token correclty, you should see the plain text response `All health checks and notifications are enabled`.
+If you've configured and entered your token correctly, you should see the plain text response `All health checks and notifications are enabled`.
 
 ### Disable or silence all alarms
 
diff --git a/web/gui/custom/README.md b/web/gui/custom/README.md
index a85c7f5464..733ef5211d 100644
--- a/web/gui/custom/README.md
+++ b/web/gui/custom/README.md
@@ -324,7 +324,7 @@ Netdata supports a number of chart libraries. The default chart library is
 Each chart library has a number of specific settings. To learn more about them,
 you should investigate the documentation of the given chart library, or visit
 the appropriate JavaScript file that defines the library's options. These files
-are concatenated into the monolithin `dashboard.js` for deployment.
+are concatenated into the monolithic `dashboard.js` for deployment.
 
 -   [Dygraph](https://github.com/netdata/netdata/blob/5b57fc441c40959514c4e2d0863be2e6a417e352/web/gui/dashboard.js#L2034)
 -   [d3](https://github.com/netdata/netdata/blob/5b57fc441c40959514c4e2d0863be2e6a417e352/web/gui/dashboard.js#L4095)
diff --git a/web/server/static/README.md b/web/server/static/README.md
index 63edd9ba7e..e095f2ef3e 100644
--- a/web/server/static/README.md
+++ b/web/server/static/README.md
@@ -1,6 +1,6 @@
 <!--
 title: "`static-threaded` web server"
-description: "The Netdata Agent's static-threaded web server spawns a fixed nubmer of threads that listen to web requests and uses non-blocking I/O."
+description: "The Netdata Agent's static-threaded web server spawns a fixed number of threads that listen to web requests and uses non-blocking I/O."
 custom_edit_url: https://github.com/netdata/netdata/edit/master/web/server/static/README.md
 -->