The first week after the AIOps rollout, paging felt better. The second week it felt haunted.
Start here: More in AIOps.
Alert volume dropped, but the remaining alerts were stranger. The model grouped symptoms differently than humans did. Escalations happened later, not earlier. By the time the on-call committed to a decision, the system had already moved on. The loop got quieter. The loop also got SLOwer.
AIOps does not usually fail because it is inaccurate. It fails because teams give it authority without boundaries and then measure the wrong thing. They celebrate fewer pages while decision latency rises, and they do not notice until incident tempo collapses.
Accuracy is not the goal
The tempting belief is that AIOps is a precision problem. Reduce false positives and the on-call will calm down.
In practice, on-call pain is usually a time-to-decision problem. The burden is not the alert count. The burden is the uncertainty you have to resolve before you can act. A model can reduce noise while increasing ambiguity, and ambiguity is what makes people hesitate.
When an AIOps model changes, what usually breaks first is decision latency. The system shifts the distribution of what arrives. Your runbooks and instincts stop matching the new shape. The first minutes of an incident turn into translation work, which is exactly what you were trying to buy back.
Two deployment modes, two cost profiles
There are two viable ways to deploy AIOps. Most teams mix them without acknowledging the trade.
In a proposal loop, the model ranks, groups, and drafts. Humans decide. The model increases operator throughput but does not touch the control plane.
In a control loop, the model pages, silences, reroutes, or remediates. This is not automatically wrong. It is expensive. You now own a second production system, which is the automation loop itself. It needs Observability, rollback, audits, and an owner who is accountable when it makes a plausible mistake.
If you cannot observe and govern the loop, what usually breaks first is trust. People route around automation they cannot predict. Once they do, you keep the overhead and lose the leverage.
How the loop gets slower, even when dashboards look better
Here is a realistic failure mode that looks like progress on the metric most teams show to leadership.
You deploy alert clustering. Page volume drops because repeated symptoms are merged. The first page the on-call receives now represents a cluster that includes multiple plausible causes. The operator starts with less certainty about what to do first, not more.
The first thing that breaks is escalation timing. People hesitate because the cluster is not clearly a single incident, and they do not want to wake a specialist for something that might dissolve. The cost is not a wrong page. The cost is delayed commitment to a hypothesis. While the operator waits for more evidence, the system keeps moving, and the set of safe interventions shrinks.
If page volume drops but time to first decision rises, incident tempo gets worse. The system still fails at the same rate. Humans arrive later.
The failure signatures your dashboards miss
If you only track precision, you will miss the signatures that actually matter. These are the ones that show up in the seat.
Noise moves rather than drops. Different alerts arrive, but stress does not change. Pages shift left rather than down. The model saves humans from duplicates but not from uncertainty. Confidence feels fake. The model is confident in ways that do not map to mechanism, so humans spend time arguing with it instead of using it.
The fastest confirmation is to audit the loop, not the model. Look at the distribution of time to first decision, not just the average. Track false action rate, which is how often an AIOps-driven recommendation or action had to be reversed. Compare the top alert causes by week, before and after deployment. If false action rate rises, the first thing that breaks is willingness to use the system at all. Operators will do the work manually to avoid an automated misstep. That is rational behavior.
Add stop rules before you add authority
A stop rule is a bounded runtime condition that disables automation when uncertainty rises. It is the safety valve for when the model is confidently wrong. Stop rules are not a governance document. They are a mechanism that forces the loop to fail safe.
Without stop rules, a model can amplify disagreement. Confidence spikes while symptoms drift, operators try to reconcile conflicting signals, and the automation keeps producing more confident output. The loop stops reducing uncertainty and starts increasing it.
A practical stop rule template
Keep this narrow. You can widen it later. The goal is not to handle every edge case. The goal is to prevent the loop from acting when it is most likely to be wrong.
Automation scope
Define the exact actions the automation can take. Then define what it cannot do without human confirmation, including paging, silencing, remediation, and routing changes that affect multiple services.
Stop triggers
Stop if model confidence increases while a canonical user-harm indicator worsens. Stop if the top driver for alerts changes abruptly inside a short window. Stop if the system recommends action without a traceable mechanism, meaning you cannot explain the causal story from evidence to action.
What happens on stop
Automation becomes advisory only. Alert routing reverts to the last known stable rules. A named human owner is paged to evaluate model drift and decide whether to continue in proposal mode, roll back the model, or gate it behind stricter constraints.
This template is intentionally strict. If you let automation keep acting during ambiguity, you are choosing speed over correctness. Sometimes that trade is worth it. You should name it as a choice and attach an owner to the consequences.
How a senior should explain this to a peer
AIOps is useful when it reduces time to a bounded decision without hijacking the control plane. We keep a canonical indicator for user harm, we constrain what automation can change, and we add stop rules so the loop fails safe when the model drifts. The win is not fewer alerts. The win is fewer minutes spent guessing.
The unresolved part is authority. Everyone wants the automation to act when it is right. Nobody wants to own the consequences when it is wrong. If you cannot name who owns the loop, you should not promote it from proposals to control.
Sanity checks
Which single metric best represents time to first decision for your rotation today? What is your stop rule, and what exact conditions disable automation when uncertainty rises? What is your false action rate, and how will you measure it without turning every review into an argument about anecdotes?
Related reading: More in AIOps. Next: Release gates that survive pressure, On-call load is a system, Agent skills in production.
Continue Reading
🤖The practitioner guide to AIOps: alert correlation, anomaly detection, LLM integration, and automated remediation.
Metrics, distributed tracing, structured logs, SLOs, and Error Budgets — and how to extend them for AI systems.
Stay Sharp
New articles on AIOps and SRE, straight to your inbox.
Practical content for practitioners. No noise, no vendor pitches.
No spam. Unsubscribe any time.


