The only agent that thinks for itself

Autonomous Monitoring with self-learning AI built-in, operating independently across your entire stack.

Unlimited Metrics & Logs
Machine learning & MCP
5% CPU, 150MB RAM
3GB disk, >1 year retention
800+ integrations, zero config
Dashboards, alerts out of the box
> Discover Netdata Agents

Centralized metrics streaming and storage

Aggregate metrics from multiple agents into centralized Parent nodes for unified monitoring across your infrastructure.

Stream from unlimited agents
Long-term data retention
High availability clustering
Data replication & backup
Scalable architecture
Enterprise-grade security
> Learn about Parents

Fully managed cloud platform

Access your monitoring data from anywhere with our SaaS platform. No infrastructure to manage, automatic updates, and global availability.

Zero infrastructure management
99.9% uptime SLA
Global data centers
Automatic updates & patches
Enterprise SSO & RBAC
SOC2 & ISO certified
> Explore Netdata Cloud

Deploy Netdata Cloud in your infrastructure

Run the full Netdata Cloud platform on-premises for complete data sovereignty and compliance with your security policies.

Complete data sovereignty
Air-gapped deployment
Custom compliance controls
Private network integration
Dedicated support team
Kubernetes & Docker support
> Learn about Cloud On-Premises

Powerful, intuitive monitoring interface

Modern, responsive UI built for real-time troubleshooting with customizable dashboards and advanced visualization capabilities.

Real-time chart updates
Customizable dashboards
Dark & light themes
Advanced filtering & search
Responsive on all devices
Collaboration features
> Explore Netdata UI

Monitor on the go

Native iOS and Android apps bring full monitoring capabilities to your mobile device with real-time alerts and notifications.

iOS & Android apps
Push notifications
Touch-optimized interface
Offline data access
Biometric authentication
Widget support
> Download apps

The future of infrastructure observability

See our strategic direction across AI-native observability, full-stack signals, operational intelligence, and enterprise platform maturity.

AI-native observability
Full-stack signal coverage
Operational intelligence
Enterprise platform maturity
Agent releases every 6 weeks
Cloud continuous delivery
> Explore Product Roadmap

Best energy efficiency

True real-time per-second

100% automated zero config

Centralized observability

Multi-year retention

High availability built-in

Zero maintenance

Always up-to-date

Enterprise security

Complete data control

Air-gap ready

Compliance certified

Millisecond responsiveness

Infinite zoom & pan

Works on any device

Native performance

Instant alerts

Monitor anywhere

AI-native observability

Continuous delivery

Open source foundation

80% Faster Incident Resolution

AI-powered troubleshooting from detection, to root cause and blast radius identification, to reporting.

True Real-Time and Simple, even at Scale

Linearly and infinitely scalable full-stack observability, that can be deployed even mid-crisis.

90% Cost Reduction, Full Fidelity

Instead of centralizing the data, Netdata distributes the code, eliminating pipelines and complexity.

Control Without Surrender

SOC 2 Type 2 certified with every metric kept on your infrastructure.

Integrations

800+ collectors and notification channels, auto-discovered and ready out of the box.

800+ data collectors
Auto-discovery & zero config
Cloud, infra, app protocols
Notifications out of the box
> Explore integrations
Real Results
46% Cost Reduction

Reduced monitoring costs by 46% while cutting staff overhead by 67%.

— Leonardo Antunez, Codyas

Zero Pipeline

No data shipping. No central storage costs. Query at the edge.

From Our Users
"Out-of-the-Box"

So many out-of-the-box features! I mostly don't have to develop anything.

— Simon Beginn, LANCOM Systems

No Query Language

Point-and-click troubleshooting. No PromQL, no LogQL, no learning curve.

Enterprise Ready
67% Less Staff, 46% Cost Cut

Enterprise efficiency without enterprise complexity—real ROI from day one.

— Leonardo Antunez, Codyas

SOC 2 Type 2 Certified

Zero data egress. Only metadata reaches the cloud. Your metrics stay on your infrastructure.

Full Coverage
800+ Collectors

Auto-discovered and configured. No manual setup required.

Any Notification Channel

Slack, PagerDuty, Teams, email, webhooks—all built-in.

Built for the People Who Get Paged

Because 3am alerts deserve instant answers, not hour-long hunts.

Every Industry Has Rules. We Master Them.

See how healthcare, finance, and government teams cut monitoring costs 90% while staying audit-ready.

Monitor Any Technology. Configure Nothing.

Install the agent. It already knows your stack.
From Our Users
"A Rare Unicorn"

Netdata gives more than you invest in it. A rare unicorn that obeys the Pareto rule.

— Eduard Porquet Mateu, TMB Barcelona

99% Downtime Reduction

Reduced website downtime by 99% and cloud bill by 30% using Netdata alerts.

— Falkland Islands Government

Real Savings
30% Cloud Cost Reduction

Optimized resource allocation based on Netdata alerts cut cloud spending by 30%.

— Falkland Islands Government

46% Cost Cut

Reduced monitoring staff by 67% while cutting operational costs by 46%.

— Codyas

Real Coverage
"Plugin for Everything"

Netdata has agent capacity or a plugin for everything, including Windows and Kubernetes.

— Eduard Porquet Mateu, TMB Barcelona

"Out-of-the-Box"

So many out-of-the-box features! I mostly don't have to develop anything.

— Simon Beginn, LANCOM Systems

Real Speed
Troubleshooting in 30 Seconds

From 2-3 minutes to 30 seconds—instant visibility into any node issue.

— Matthew Artist, Nodecraft

20% Downtime Reduction

20% less downtime and 40% budget optimization from out-of-the-box monitoring.

— Simon Beginn, LANCOM Systems

Pay per Node. Unlimited Everything Else.

One price per node. Unlimited metrics, logs, users, and retention. No per-GB surprises.

Free tier—forever
No metric limits or caps
Retention you control
Cancel anytime
> See pricing plans

What's Your Monitoring Really Costing You?

Most teams overpay by 40-60%. Let's find out why.

Expose hidden metric charges
Calculate tool consolidation
Customers report 30-67% savings
Results in under 60 seconds
> See what you're really paying

Your Infrastructure Is Unique. Let's Talk.

Because monitoring 10 nodes is different from monitoring 10,000.

On-prem & air-gapped deployment
Volume pricing & agreements
Architecture review for your scale
Compliance & security support
> Start a conversation

Monitoring That Sells Itself

Deploy in minutes. Impress clients in hours. Earn recurring revenue for years.

30-second live demos close deals
Zero config = zero support burden
Competitive margins & deal protection
Response in 48 hours
> Apply to partner

Per-Second Metrics at Homelab Prices

Same engine, same dashboards, same ML. Just priced for tinkerers.

Community: Free forever · 5 nodes · non-commercial
Homelab: $90/yr · unlimited nodes · fair usage
> Get the Homelab Plan

$1,000 Per Referral. Unlimited Referrals.

Your colleagues get 10% off. You get 10% commission. Everyone wins.

10% of subscriptions, up to $1,000 each
Track earnings inside Netdata Cloud
PayPal/Venmo payouts in 3-4 weeks
No caps, no complexity
> Get your referral link
Cost Proof
40% Budget Optimization

"Netdata's significant positive impact" — LANCOM Systems

Calculate Your Savings

Compare vs Datadog, Grafana, Dynatrace

Savings Proof
46% Cost Reduction

"Cut costs by 46%, staff by 67%" — Codyas

30% Cloud Bill Savings

"Reduced cloud bill by 30%" — Falkland Islands Gov

Enterprise Proof
"Better Than Combined Alternatives"

"Better observability with Netdata than combining other tools." — TMB Barcelona

Real Engineers, <24h Response

DPA, SLAs, on-prem, volume pricing

Why Partners Win
Demo Live Infrastructure

One command, 30 seconds, real data—no sandbox needed

Zero Tickets, High Margins

Auto-config + per-node pricing = predictable profit

Homelab Ready
Free Video Course

8-episode Netdata tutorial by LearnLinux.tv

76k+ GitHub Stars

3rd most starred monitoring project

Worth Recommending
Product That Delivers

Customers report 40-67% cost cuts, 99% downtime reduction

Zero Risk to Your Rep

Free tier lets them try before they buy

AI Support Assistant, Available 24/7

Nedi has access to all official documentation, source code, and resources. Ask any question about Netdata—responds in your language.

Deployment & configuration
Troubleshooting & sizing
Alerts & notifications
Evidence-based answers
> Ask Nedi now

Never Fight Fires Alone

Docs, community, and expert help—pick your path to resolution.

Learn.netdata.cloud docs
Discord, Forums, GitHub
Premium support available
> Get answers now

60 Seconds to First Dashboard

One command to install. Zero config. 850+ integrations documented.

Linux, Windows, K8s, Docker
Auto-discovers your stack
> Read our documentation

Level Up Your Monitoring

Real problems. Real solutions. 112+ guides from basic monitoring to AI observability.

76,000+ Engineers Strong

615+ contributors. 1.5M daily downloads. One mission: simplify observability.

Per-Second. 90% Cheaper. Data Stays Home.

Side-by-side comparisons: costs, real-time granularity, and data sovereignty for every major tool.

See why teams switch from Datadog, Prometheus, Grafana, and more.

> Browse all comparisons
Edge-Native Observability, Born Open Source
Per-second visibility, ML on every metric, and data that never leaves your infrastructure.
Founded in 2016
615+ contributors worldwide
Remote-first, engineering-driven
Open source first
> Read our story
Promises We Publish—and Prove
12 principles backed by open code, independent validation, and measurable outcomes.
Open source, peer-reviewed
Zero config, instant value
Data sovereignty by design
Aligned pricing, no surprises
> See all 12 principles
Edge-Native, AI-Ready, 100% Open
76k+ stars. Full ML, AI, and automation—GPLv3+, not premium add-ons.
76,000+ GitHub stars
GPLv3+ licensed forever
ML on every metric, included
Zero vendor lock-in
> Explore our open source
Build Real-Time Observability for the World
Remote-first team shipping per-second monitoring with ML on every metric.
Remote-first, fully distributed
Open source (76k+ stars)
Challenging technical problems
Your code on millions of systems
> See open roles
Meet the Team Behind Netdata
Conferences, meetups, and tradeshows where you can see Netdata in action and talk to the engineers who build it.
Live demos and deep dives
Book 1-on-1 meetings
Talks and panel sessions
Event recaps and photos
> See all events
Talk to a Netdata Human in <24 Hours
Sales, partnerships, press, or professional services—real engineers, fast answers.
Discuss your observability needs
Pricing and volume discounts
Partnership opportunities
Media and press inquiries
> Book a conversation
Your Data. Your Rules.
On-prem data, cloud control plane, transparent terms.
Trust & Scale
76,000+ GitHub Stars

One of the most popular open-source monitoring projects

SOC 2 Type 2 Certified

Enterprise-grade security and compliance

Data Sovereignty

Your metrics stay on your infrastructure

Validated
University of Amsterdam

"Most energy-efficient monitoring solution" — ICSOC 2023, peer-reviewed

ADASTEC (Autonomous Driving)

"Doesn't miss alerts—mission-critical trust for safety software"

Community Stats
615+ Contributors

Global community improving monitoring for everyone

1.5M+ Downloads/Day

Trusted by teams worldwide

GPLv3+ Licensed

Free forever, fully open source agent

Why Join?
Remote-First

Work from anywhere, async-friendly culture

Impact at Scale

Your work helps millions of systems

$ guides / docker
DOCKER · OPERATIONS PLAYBOOK

Running Docker in production, without surprises

What Docker is doing under the hood, where it tends to break, what to monitor as your operation matures, and which mistakes to avoid before they become incidents.

"

Docker is easy to start with and surprisingly easy to run blind.

A container can be "up" while the application inside it is broken. The Docker daemon can be alive but too wedged to answer docker ps. A host can look healthy until /var/lib/docker fills and every container on it starts failing at once. CPU usage can look normal while CFS throttling quietly destroys latency. Logs can grow for weeks and then take the host down in an afternoon.

These guides are written for engineers who already run Docker, not for people learning what a container is. The goal is to give you the mental model, the failure patterns, the monitoring story, and the runbooks you wish someone had handed you before your last incident.

How Docker actually runs in production

Docker is not one thing. It is a stack of cooperating components, and most production failures happen between these layers, not inside any single one of them.

01
docker CLI / API clients
Where requests come in. Your CLI, your CI/CD system, your orchestrator, anything talking to the Docker socket.
USER
02
dockerd
The management plane. Handles images, networks, volumes, lifecycle, and log routing.
DAEMON
03
containerd
The runtime manager underneath dockerd. Owns container execution and lifecycle.
RUNTIME
04
containerd-shim
Per-container supervisor. Lets your containers survive a daemon restart.
SUPERVISOR
05
runc
OCI runtime. Sets up namespaces, cgroups, mounts, and seccomp policy when a container starts.
OCI
06
Linux kernel
The real workhorse. cgroups, namespaces, overlay filesystems, bridges, iptables. Docker is mostly a friendly interface to these.
KERNEL
07
your container process
PID 1 inside the container, plus its children. The workload itself.
CONTAINER

Why this matters: a container can keep running while the daemon is hung. The daemon can answer /_ping while docker ps blocks. A memory kill comes from the kernel, not Docker. A network problem may live in iptables, in embedded DNS, in conntrack, or in the application, and each one looks different from the outside.

The failures you'll actually see

Most Docker incidents are not exotic. They cluster into a small set of recurring patterns. Recognise the shape, and triage gets dramatically faster.

IMMINENT

The disk filled silently

Images, writable layers, volumes, build cache, metadata, and (often) logs all share /var/lib/docker. When that filesystem fills, every container on the host starts failing at once.

  • new containers fail to start
  • image pulls fail
  • running containers cannot write
  • containers stuck in dead or removing state
Investigate
ACTIVE

The container death spiral

A container crashes, the restart policy brings it back, it crashes again. Logs flood, restart count climbs, and the actual root cause hides behind the restart loop.

  • high restart count
  • exit code 137 / 143 / 139 / 1
  • OOMKilled = true
  • health check failing
Investigate
CRITICAL

The daemon wedged

dockerd is up but unresponsive. systemd reports it active, the process exists, but docker ps hangs and you cannot manage anything.

  • docker ps / inspect hangs
  • running containers keep going
  • storage driver stalls
  • internal lock contention
Investigate
IMMINENT

The network black hole

Bridges, veth pairs, iptables/NAT, embedded DNS, overlay networks. They all have to line up. When one piece is wrong the container looks healthy while the path around it is broken.

  • DNS resolution fails inside containers
  • container-to-container traffic blocked
  • published ports unreachable
  • conntrack table fills under load
Investigate
WATCHFUL

The hidden CPU throttle

Average CPU usage looks normal while the kernel keeps pausing the container against its CFS quota. Latency rises; the CPU graph is calm.

  • p95 / p99 latency climbs
  • request timeouts under load
  • slow health checks
  • throttled time keeps rising
Investigate
ACTIVE

The OOM cascade

A container hits its memory limit. The kernel kills a process. Docker reports it after the fact, often after corrupted state, interrupted transactions, or restarts that hide the root cause.

  • exit code 137
  • OOMKilled = true
  • child killed, PID 1 still alive
  • approaching memory limit
Investigate

Docker monitoring maturity levels

Docker monitoring works in four practical levels. Each level is a complete operation, not a stepping stone you must climb. Pick the level that matches how much your Docker reliability matters and how much investment your team has the bandwidth for. Most production teams should aim for the second level.

Level 1: Survival

Know that something is wrong

Survival monitoring is the floor. With these signals you can answer one question: is Docker still working? You will not learn what broke or why, but you will learn that something broke before users do. Survival is enough for hobby clusters, dev environments, and small teams running a handful of containers where Docker reliability is not in the critical path. It is not enough if Docker runs your customers, your pipeline, or your revenue.

  • Docker daemon responsiveness Can the Docker API answer basic requests at all?
  • Docker data directory disk usage Is /var/lib/docker close to full?
  • Container state Are expected containers running, or exited, dead, restarting?
  • Container restart count Is anything quietly crash-looping?
  • OOMKilled status Did the kernel just kill a container for memory pressure?
  • Host CPU, memory, disk, I/O Is the host itself under pressure?

Level 2: Operational

Diagnose most incidents on your own

Operational monitoring is what most production Docker hosts should target. Once survival signals tell you something is wrong, operational signals tell you what. With this coverage your team can usually diagnose an incident on its own: restart loops, OOM kills, disk filling, network drops, daemon slowness. If you only invest in one level above survival, this is the one to invest in.

  • Container exit codes Why did a container actually stop?
  • Memory usage vs limits How close is each container to its OOM ceiling?
  • CPU usage per container Which workload is consuming the host?
  • CPU throttling Are CFS quotas the real cause of latency?
  • Container log file size Are logs about to fill the host?
  • Daemon response latency Is dockerd getting slow before it hangs?
  • Daemon error logs Storage, network, or runtime errors surfacing?
  • Image and volume disk breakdown What is actually consuming /var/lib/docker?
  • Image pull failures Are deployments blocked on registry or network?
  • Container network errors Are packets dropping or connections failing?

Level 3: Mature

Catch problems before they become incidents

Mature monitoring catches problems before they wake anyone up. Memory creeping toward a limit, daemon latency drifting upward, conntrack tables slowly filling, container start latency growing, lifecycle event rates climbing under invisible load. None of these will page you on day one. They turn into pages on day thirty if no one is watching. Mature monitoring is for teams that have already been bitten by leading-indicator failures and want to spot the next one early.

  • Memory breakdown: anon, file, slab Is growth real workload memory, cache, or kernel memory?
  • Writable layer growth Is the application writing into the container layer instead of a volume?
  • Container PID count and zombies Is PID exhaustion or missing child reaping creeping in?
  • Daemon file descriptor usage Is dockerd approaching its FD limits?
  • Daemon goroutine count Is internal concurrency growing abnormally?
  • DNS health from inside containers Does resolution work where the application lives?
  • conntrack utilisation Are new connections at risk of being dropped?
  • Storage driver health Are overlay2 errors appearing in the daemon logs?
  • Lifecycle event rate Is container churn quietly stressing the host?

Level 4: Expert

Reactive instrumentation after real incidents

Expert signals are reactive, not aspirational. Each one tends to enter your monitoring stack the day after a specific incident proved you needed it. Daemon pprof captures, conntrack auditing, Pressure Stall Information, AppArmor drift detection, sub-second cgroup analysis. Most teams never need every signal at this level. Add the ones your incident history tells you to add. Adopting the full list without a reason is a way to spend engineering time on noise.

  • Daemon pprof profiles Heap, goroutine, mutex captures during pathological events.
  • containerd shim health Per-container supervisor responsiveness and exit signals.
  • overlay2 layer count per image Layer fan-out causing storage driver pressure.
  • iptables and NAT rule count Rule scale impacting packet processing latency.
  • cgroup period-level CPU analysis Sub-second throttling patterns hidden by averages.
  • Pressure Stall Information (PSI) Kernel signal when CPU, memory, or I/O are blocking work.
  • seccomp, AppArmor, capability audit Drift in container security posture over time.
  • Docker socket access audit Who and what is talking to /var/run/docker.sock?
  • Sensitive environment variable scan Secrets that ended up in container env by accident.

Operating mistakes worth avoiding

The traps teams keep falling into. Each of these has a clear, well-known fix. Most teams only learn it after an incident.

No log rotation

The default json-file driver grows unbounded. One noisy container can fill the host. Configure rotation early, not after the first incident.

Alerting on CPU usage but not CPU throttling

High CPU is obvious. Throttling is dangerous because it looks like application latency while average CPU appears acceptable.

Treating total memory as the whole story

Container memory includes reclaimable cache. Split anonymous, file, slab, and what the application actually owns.

Restart policies as a substitute for reliability

Restart policies keep workloads alive but hide repeated failures. Restart count is a signal, not noise.

Running Docker without disk hygiene

Old images, unused volumes, build cache, stopped containers, logs. They all accumulate. Cleanup policies and disk monitoring are not optional.

Exposing the Docker socket casually

Mounting /var/run/docker.sock into a container gives that container control over the host. Treat it like root access.

Docker runbooks in this section

Each guide is a focused runbook for one symptom or topic. Pick one when you have an incident, or use the categories to learn the area.

WHERE TO GO NEXT

Setting up Docker monitoring, or putting out a fire?

If you're starting from scratch, the monitoring checklist is the path of least regret. If you're mid-incident, jump straight to the symptom that matches what you're seeing.