The only agent that thinks for itself

Autonomous Monitoring with self-learning AI built-in, operating independently across your entire stack.

Unlimited Metrics & Logs
Machine learning & MCP
5% CPU, 150MB RAM
3GB disk, >1 year retention
800+ integrations, zero config
Dashboards, alerts out of the box
> Discover Netdata Agents

Centralized metrics streaming and storage

Aggregate metrics from multiple agents into centralized Parent nodes for unified monitoring across your infrastructure.

Stream from unlimited agents
Long-term data retention
High availability clustering
Data replication & backup
Scalable architecture
Enterprise-grade security
> Learn about Parents

Fully managed cloud platform

Access your monitoring data from anywhere with our SaaS platform. No infrastructure to manage, automatic updates, and global availability.

Zero infrastructure management
99.9% uptime SLA
Global data centers
Automatic updates & patches
Enterprise SSO & RBAC
SOC2 & ISO certified
> Explore Netdata Cloud

Deploy Netdata Cloud in your infrastructure

Run the full Netdata Cloud platform on-premises for complete data sovereignty and compliance with your security policies.

Complete data sovereignty
Air-gapped deployment
Custom compliance controls
Private network integration
Dedicated support team
Kubernetes & Docker support
> Learn about Cloud On-Premises

Powerful, intuitive monitoring interface

Modern, responsive UI built for real-time troubleshooting with customizable dashboards and advanced visualization capabilities.

Real-time chart updates
Customizable dashboards
Dark & light themes
Advanced filtering & search
Responsive on all devices
Collaboration features
> Explore Netdata UI

Monitor on the go

Native iOS and Android apps bring full monitoring capabilities to your mobile device with real-time alerts and notifications.

iOS & Android apps
Push notifications
Touch-optimized interface
Offline data access
Biometric authentication
Widget support
> Download apps

The future of infrastructure observability

See our strategic direction across AI-native observability, full-stack signals, operational intelligence, and enterprise platform maturity.

AI-native observability
Full-stack signal coverage
Operational intelligence
Enterprise platform maturity
Agent releases every 6 weeks
Cloud continuous delivery
> Explore Product Roadmap

Best energy efficiency

True real-time per-second

100% automated zero config

Centralized observability

Multi-year retention

High availability built-in

Zero maintenance

Always up-to-date

Enterprise security

Complete data control

Air-gap ready

Compliance certified

Millisecond responsiveness

Infinite zoom & pan

Works on any device

Native performance

Instant alerts

Monitor anywhere

AI-native observability

Continuous delivery

Open source foundation

80% Faster Incident Resolution

AI-powered troubleshooting from detection, to root cause and blast radius identification, to reporting.

True Real-Time and Simple, even at Scale

Linearly and infinitely scalable full-stack observability, that can be deployed even mid-crisis.

90% Cost Reduction, Full Fidelity

Instead of centralizing the data, Netdata distributes the code, eliminating pipelines and complexity.

Control Without Surrender

SOC 2 Type 2 certified with every metric kept on your infrastructure.

Integrations

800+ collectors and notification channels, auto-discovered and ready out of the box.

800+ data collectors
Auto-discovery & zero config
Cloud, infra, app protocols
Notifications out of the box
> Explore integrations
Real Results
46% Cost Reduction

Reduced monitoring costs by 46% while cutting staff overhead by 67%.

— Leonardo Antunez, Codyas

Zero Pipeline

No data shipping. No central storage costs. Query at the edge.

From Our Users
"Out-of-the-Box"

So many out-of-the-box features! I mostly don't have to develop anything.

— Simon Beginn, LANCOM Systems

No Query Language

Point-and-click troubleshooting. No PromQL, no LogQL, no learning curve.

Enterprise Ready
67% Less Staff, 46% Cost Cut

Enterprise efficiency without enterprise complexity—real ROI from day one.

— Leonardo Antunez, Codyas

SOC 2 Type 2 Certified

Zero data egress. Only metadata reaches the cloud. Your metrics stay on your infrastructure.

Full Coverage
800+ Collectors

Auto-discovered and configured. No manual setup required.

Any Notification Channel

Slack, PagerDuty, Teams, email, webhooks—all built-in.

Built for the People Who Get Paged

Because 3am alerts deserve instant answers, not hour-long hunts.

Every Industry Has Rules. We Master Them.

See how healthcare, finance, and government teams cut monitoring costs 90% while staying audit-ready.

Monitor Any Technology. Configure Nothing.

Install the agent. It already knows your stack.
From Our Users
"A Rare Unicorn"

Netdata gives more than you invest in it. A rare unicorn that obeys the Pareto rule.

— Eduard Porquet Mateu, TMB Barcelona

99% Downtime Reduction

Reduced website downtime by 99% and cloud bill by 30% using Netdata alerts.

— Falkland Islands Government

Real Savings
30% Cloud Cost Reduction

Optimized resource allocation based on Netdata alerts cut cloud spending by 30%.

— Falkland Islands Government

46% Cost Cut

Reduced monitoring staff by 67% while cutting operational costs by 46%.

— Codyas

Real Coverage
"Plugin for Everything"

Netdata has agent capacity or a plugin for everything, including Windows and Kubernetes.

— Eduard Porquet Mateu, TMB Barcelona

"Out-of-the-Box"

So many out-of-the-box features! I mostly don't have to develop anything.

— Simon Beginn, LANCOM Systems

Real Speed
Troubleshooting in 30 Seconds

From 2-3 minutes to 30 seconds—instant visibility into any node issue.

— Matthew Artist, Nodecraft

20% Downtime Reduction

20% less downtime and 40% budget optimization from out-of-the-box monitoring.

— Simon Beginn, LANCOM Systems

Pay per Node. Unlimited Everything Else.

One price per node. Unlimited metrics, logs, users, and retention. No per-GB surprises.

Free tier—forever
No metric limits or caps
Retention you control
Cancel anytime
> See pricing plans

What's Your Monitoring Really Costing You?

Most teams overpay by 40-60%. Let's find out why.

Expose hidden metric charges
Calculate tool consolidation
Customers report 30-67% savings
Results in under 60 seconds
> See what you're really paying

Your Infrastructure Is Unique. Let's Talk.

Because monitoring 10 nodes is different from monitoring 10,000.

On-prem & air-gapped deployment
Volume pricing & agreements
Architecture review for your scale
Compliance & security support
> Start a conversation

Monitoring That Sells Itself

Deploy in minutes. Impress clients in hours. Earn recurring revenue for years.

30-second live demos close deals
Zero config = zero support burden
Competitive margins & deal protection
Response in 48 hours
> Apply to partner

Per-Second Metrics at Homelab Prices

Same engine, same dashboards, same ML. Just priced for tinkerers.

Community: Free forever · 5 nodes · non-commercial
Homelab: $90/yr · unlimited nodes · fair usage
> Get the Homelab Plan

$1,000 Per Referral. Unlimited Referrals.

Your colleagues get 10% off. You get 10% commission. Everyone wins.

10% of subscriptions, up to $1,000 each
Track earnings inside Netdata Cloud
PayPal/Venmo payouts in 3-4 weeks
No caps, no complexity
> Get your referral link
Cost Proof
40% Budget Optimization

"Netdata's significant positive impact" — LANCOM Systems

Calculate Your Savings

Compare vs Datadog, Grafana, Dynatrace

Savings Proof
46% Cost Reduction

"Cut costs by 46%, staff by 67%" — Codyas

30% Cloud Bill Savings

"Reduced cloud bill by 30%" — Falkland Islands Gov

Enterprise Proof
"Better Than Combined Alternatives"

"Better observability with Netdata than combining other tools." — TMB Barcelona

Real Engineers, <24h Response

DPA, SLAs, on-prem, volume pricing

Why Partners Win
Demo Live Infrastructure

One command, 30 seconds, real data—no sandbox needed

Zero Tickets, High Margins

Auto-config + per-node pricing = predictable profit

Homelab Ready
Free Video Course

8-episode Netdata tutorial by LearnLinux.tv

76k+ GitHub Stars

3rd most starred monitoring project

Worth Recommending
Product That Delivers

Customers report 40-67% cost cuts, 99% downtime reduction

Zero Risk to Your Rep

Free tier lets them try before they buy

AI Support Assistant, Available 24/7

Nedi has access to all official documentation, source code, and resources. Ask any question about Netdata—responds in your language.

Deployment & configuration
Troubleshooting & sizing
Alerts & notifications
Evidence-based answers
> Ask Nedi now

Never Fight Fires Alone

Docs, community, and expert help—pick your path to resolution.

Learn.netdata.cloud docs
Discord, Forums, GitHub
Premium support available
> Get answers now

60 Seconds to First Dashboard

One command to install. Zero config. 850+ integrations documented.

Linux, Windows, K8s, Docker
Auto-discovers your stack
> Read our documentation

Level Up Your Monitoring

Real problems. Real solutions. 112+ guides from basic monitoring to AI observability.

76,000+ Engineers Strong

615+ contributors. 1.5M daily downloads. One mission: simplify observability.

Per-Second. 90% Cheaper. Data Stays Home.

Side-by-side comparisons: costs, real-time granularity, and data sovereignty for every major tool.

See why teams switch from Datadog, Prometheus, Grafana, and more.

> Browse all comparisons
Edge-Native Observability, Born Open Source
Per-second visibility, ML on every metric, and data that never leaves your infrastructure.
Founded in 2016
615+ contributors worldwide
Remote-first, engineering-driven
Open source first
> Read our story
Promises We Publish—and Prove
12 principles backed by open code, independent validation, and measurable outcomes.
Open source, peer-reviewed
Zero config, instant value
Data sovereignty by design
Aligned pricing, no surprises
> See all 12 principles
Edge-Native, AI-Ready, 100% Open
76k+ stars. Full ML, AI, and automation—GPLv3+, not premium add-ons.
76,000+ GitHub stars
GPLv3+ licensed forever
ML on every metric, included
Zero vendor lock-in
> Explore our open source
Build Real-Time Observability for the World
Remote-first team shipping per-second monitoring with ML on every metric.
Remote-first, fully distributed
Open source (76k+ stars)
Challenging technical problems
Your code on millions of systems
> See open roles
Meet the Team Behind Netdata
Conferences, meetups, and tradeshows where you can see Netdata in action and talk to the engineers who build it.
Live demos and deep dives
Book 1-on-1 meetings
Talks and panel sessions
Event recaps and photos
> See all events
Talk to a Netdata Human in <24 Hours
Sales, partnerships, press, or professional services—real engineers, fast answers.
Discuss your observability needs
Pricing and volume discounts
Partnership opportunities
Media and press inquiries
> Book a conversation
Your Data. Your Rules.
On-prem data, cloud control plane, transparent terms.
Trust & Scale
76,000+ GitHub Stars

One of the most popular open-source monitoring projects

SOC 2 Type 2 Certified

Enterprise-grade security and compliance

Data Sovereignty

Your metrics stay on your infrastructure

Validated
University of Amsterdam

"Most energy-efficient monitoring solution" — ICSOC 2023, peer-reviewed

ADASTEC (Autonomous Driving)

"Doesn't miss alerts—mission-critical trust for safety software"

Community Stats
615+ Contributors

Global community improving monitoring for everyone

1.5M+ Downloads/Day

Trusted by teams worldwide

GPLv3+ Licensed

Free forever, fully open source agent

Why Join?
Remote-First

Work from anywhere, async-friendly culture

Impact at Scale

Your work helps millions of systems

$ guides / kubernetes
KUBERNETES · OPERATIONS PLAYBOOK

Running Kubernetes in production, without the 3 a.m. surprises

What the control plane is doing under the hood, where clusters tend to break, what to monitor as your operation matures, and which mistakes to stop making before the next incident.

"

Kubernetes is easy to demo and hard to operate.

A pod can be Running while the application inside it is broken. A node can report Ready while its kubelet is too wedged to start anything new. The API server can pass /healthz while etcd is too slow to answer real writes. A service can have endpoints and still be unreachable because conntrack is full, an iptables sync is stalled, or NetworkPolicy quietly denied the packet. A pod evicted for memory pressure tells you something — a pod evicted because another pod broke a node tells you something different.

These guides are written for engineers who already run Kubernetes, not for people learning what a pod is. The goal is to give you the mental model of the control plane, the failure patterns that keep recurring, the monitoring story that catches issues before they page anyone, and the runbooks you wish someone had handed you before your last incident.

How Kubernetes actually runs in production

Kubernetes is not one thing. It is a stack of cooperating components, and most production failures happen between these layers, not inside any single one of them.

01
kubectl / API clients
Where requests come in. Your CLI, controllers, operators, dashboards, CI systems, anything talking to the API server.
USER
02
kube-apiserver
The control-plane front door. Authenticates, authorizes, admits, validates, and writes. The only thing that talks to etcd.
API
03
etcd
Cluster state of record. Quorum-based, fsync-bound. Every Kubernetes object lives here.
STATE
04
scheduler &amp; controllers
Decision makers. Watch desired state, compute placement, drive reconciliation toward actual state.
CONTROL
05
kubelet
Per-node agent. Watches the API for assigned pods, pulls images, supervises the runtime, reports node status.
NODE
06
runtime + CNI + CSI
containerd or CRI-O runs pods; the CNI plugin wires the network; the CSI driver provides storage.
RUNTIME
07
Linux kernel
The real workhorse. cgroups, namespaces, iptables/IPVS, conntrack, overlayfs. Kubernetes is mostly a friendly interface to these.
KERNEL
08
your pod processes
PID 1 inside the container, plus children. The workload itself.
POD

Why this matters: a pod can be Running while the application is broken. NodeReady can be true while the kubelet's PLEG is stalling. The API server can answer /healthz while etcd is too slow to commit writes. A 'service not reachable' might live in kube-proxy, endpoints, DNS, NetworkPolicy, or the application — and each one looks different from the outside.

The failures you'll actually see

Most Kubernetes incidents are not exotic. They cluster into a small set of recurring patterns. Recognise the shape, and triage gets dramatically faster.

CRITICAL

The control-plane bottleneck

The API server is up but slow. kubectl responses crawl, controllers stop reconciling, the scheduler queue drains too slowly. Underneath it is usually etcd latency, a misbehaving admission webhook, or APF starvation of legitimate traffic.

  • kubectl latency climbs
  • controller queue depth grows
  • etcd backend commit duration spikes
  • APF rejected requests increase
Investigate
IMMINENT

The node death by eviction

One node hits MemoryPressure or DiskPressure. The kubelet evicts pods. Replacement pods land on other nodes, push them over their thresholds, and the cascade keeps going. The cluster looks alive while replicas vanish.

  • evictions across multiple nodes
  • NodeNotReady flapping
  • pods rescheduled onto pressure
  • PLEG response time rising
Investigate
ACTIVE

The pod that won't run

A workload never reaches Ready. The pod is stuck in Pending, ContainerCreating, ImagePullBackOff, or CrashLoopBackOff. Each state has a different root cause and a different runbook.

  • replicas stuck below desired
  • restartCount climbing
  • ImagePullBackOff or ErrImagePull
  • PodScheduled = false
Investigate
IMMINENT

The silent network black hole

Services with endpoints that aren't reachable. Pod-to-pod traffic dropping. Conntrack tables full, iptables sync stalled, CNI plugin in an inconsistent state, NetworkPolicy denying silently. Most of these look healthy from outside the data path.

  • service connections time out
  • conntrack utilisation climbs
  • iptables-restore latency rises
  • pods reach external but not cluster IPs
Investigate
WATCHFUL

The DNS chase

Resolution inside pods is slow or intermittent. Applications hit a 5-second resolver timeout. Upstream services get classified as flaky when the real problem is CoreDNS, the cluster DNS service IP, or ndots.

  • 5-second tail latency on outbound calls
  • CoreDNS request rate spikes
  • kube-dns service endpoint unhealthy
  • pod resolv.conf points wrong
Investigate
CRITICAL

The certificate clock

kubelet or API-server certs expire without warning. Nodes drop out one at a time. kubectl stops authenticating. There is no preceding load event, no obvious trigger, just a date on a certificate nobody was watching.

  • x509: certificate has expired
  • node NotReady, one at a time
  • kubelet TLS handshake errors
  • control plane refuses connections
Investigate

Kubernetes monitoring maturity levels

Kubernetes monitoring works in four practical levels. Each level is a complete operation, not a stepping stone you must climb. Pick the level that matches how much your cluster's reliability matters and how much investment your team has the bandwidth for. Most production clusters should aim for the second level.

Level 1: Survival

Know that something is wrong

Survival monitoring is the floor. With these signals you can answer one question: is the cluster still functioning? You will not learn what broke or why, but you will learn that something broke before users do. Survival is enough for hobby clusters, dev environments, and teams running stateless workloads where Kubernetes reliability is not in the critical path.

  • API server availability Does /readyz answer, and how fast?
  • etcd availability Are all members healthy and reporting?
  • Node Ready count How many nodes are flagged Ready right now?
  • Pod state distribution Running vs Pending vs Failed across the cluster.
  • Workload deployment readiness Are the critical deployments at desired replicas?
  • Node disk and memory utilisation Is any node close to eviction thresholds?

Level 2: Operational

Diagnose most incidents on your own

Operational monitoring is what most production clusters should target. Once survival signals tell you something is wrong, operational signals tell you what. With this coverage your team can usually diagnose an incident on its own: scheduling failures, evictions, control-plane latency, network drops, image pull issues. If you only invest in one level above survival, this is the one to invest in.

  • API server request latency p99 Is the control plane slowing down before it fails?
  • etcd backend commit duration Is the cluster state store healthy?
  • Pending pods by failure reason Why are pods not being scheduled?
  • Pod restart count and exit reasons Where are the crash loops, and why?
  • Eviction events with reason Memory, disk, PID, or something else?
  • Node conditions (Ready, Disk, Memory, PID) Which pressure signals are firing?
  • Service endpoint readiness Are services backed by healthy pods?
  • Image pull failure rate per node Are deployments blocked at the registry layer?
  • kubelet error log rate Runtime, network, volume issues surfacing?
  • PVC bind state Are persistent volumes attaching successfully?

Level 3: Mature

Catch problems before they become incidents

Mature monitoring catches problems before they wake anyone up. APF throttling drifting upward, kubelet certificates approaching expiry, conntrack tables filling, iptables sync time creeping, controller queue depth growing under invisible load. None of these will page you on day one. They turn into pages on day thirty if no one is watching.

  • APF throttling per priority level Is legitimate traffic being shed before failure?
  • PLEG relisting duration Is the kubelet keeping pace with pod events?
  • kubelet certificate expiry Days until silent failure?
  • conntrack utilisation per node Are new connections at risk of being dropped?
  • iptables sync time per node Is kube-proxy struggling to apply rules?
  • DNS query latency from pods Is CoreDNS slowing down inside the cluster?
  • Volume attach / mount duration How long does CSI take to bring storage up?
  • etcd disk fsync p99 Is the underlying disk fast enough for etcd?
  • Controller queue depth Are controllers keeping up with reconciliation?

Level 4: Expert

Reactive instrumentation after real incidents

Expert signals are reactive, not aspirational. Each one tends to enter your stack the day after a specific incident proved you needed it. kubelet pprof captures, scheduler attempt latency by predicate, audit log forensics, network policy hit-miss tracking, etcd compaction frequency. Most teams never need every signal at this level. Add the ones your incident history tells you to add.

  • kubelet pprof captures Heap, goroutine, mutex profiles during pathological events.
  • Scheduler attempt latency by predicate Which scheduling step is the bottleneck?
  • Webhook latency p99 Which admission webhook is slowing the API server?
  • etcd compaction and defrag history Is the keyspace growing pathologically?
  • Audit log analysis Who is changing what, when, and from where?
  • NetworkPolicy hit/miss telemetry Are policies allowing or denying as intended?
  • Pod startup phase breakdown Image pull, mount, network, sandbox, runtime — where is the time going?
  • Service account token rotation lifecycle Are bound tokens rotating and being honoured?

Operating mistakes worth avoiding

The traps teams keep falling into. Each has a clear, well-known fix. Most teams only learn it after an incident.

No resource requests on workloads

Pods without requests get evicted first under pressure and are scheduled greedily by the scheduler. Set CPU and memory requests for every production workload, not just limits.

Watching only API server availability

The API server can answer /readyz while etcd is gasping. Watch etcd commit duration and the API-server-to-etcd latency separately.

Ignoring webhook failurePolicy and timeouts

One slow admission webhook can stall the entire control plane. Set failurePolicy and timeoutSeconds explicitly, alert on webhook latency, and exclude critical namespaces from webhooks that need them.

Treating Node Ready as healthy

Ready only means the kubelet is reporting. The node may still be degraded (PLEG slow, runtime stalls, disk pressure). Watch every node condition individually, not just the Ready summary.

No NetworkPolicy default-deny

In a multi-tenant or microservice cluster, default-allow is a blast-radius hazard. Default-deny per namespace and explicitly allow the traffic each workload needs.

Skipping certificate rotation drills

Most clusters auto-rotate kubelet and apiserver certs. A few don't. The day they expire silently, you find out which kind you have. Verify rotation works on a non-production cluster before you need it.

Kubernetes runbooks in this section

Each guide is a focused runbook for one symptom or topic. Pick one when you have an incident, or use the categories to learn the area.

WHERE TO GO NEXT

Setting up Kubernetes monitoring, or putting out a fire?

If you're starting from scratch, the monitoring checklist is the path of least regret. If you're mid-incident, jump straight to the symptom that matches what you're seeing.