SLOs
Burn Rate Alerts: Proactive Incident Management with Watch.dog
Master SLO burn rate alerts to identify catastrophic service failures before they exhaust your error budget. A technical guide for SRE teams.
By Watch Dog TeamPublished April 15, 202612 min read
Understanding Burn Rate Math
Symptom Log
manual_check_fail.log
[WARNING] Checking error budget for API-Gateway...
[INFO] 42% consumed in 30 days.
# Error: Missed 25% spike in the last hour because static check only runs daily.A burn rate tells you how fast you're consuming your error budget relative to your SLO window. A burn rate of 1x means you'll exhaust your budget exactly at the end of the window.
Manual monitoring often misses high-velocity spikes until it's too late to recover.
Fix: Multi-Window Alerts
Configure Watch.dog Burn Alerts with multiple time windows (e.g., 1h and 6h) to detect both spikes and slow leaks.
Fix Verification
watch_dog_burn.json
{
"status": "ALERTING",
"burn_type": "FAST_BURN",
"rate": "14.4x",
"exhaustion_est": "2.4 hours",
"impact": "High"
}Actionable Response Strategy
Symptom Log
incident_started.log
[CRITICAL] Budget exhaustion imminent. 25x burn rate detected.
[INFO] Manual intervention required...When a fast burn occurs, your automated response policy should prioritize stability over feature releases. If no action is taken, your SLO is guaranteed to fail.
Fix Verification
auto_lock.log
[ACTION] Triggering pipeline lock for project 'Checkout-API'.
[INFO] Sending emergency notification to Dev-Ops Slack channel.
[SUCCESS] Feature releases paused. Reliability restored in 12m.In Watch.dog, you can automate 'Hard Freezes' on deployments when burn rate exceeds 10x.
