DigitalOcean Redefines Availability Metrics: From Incident Counting to SLIs

NewsDigitalOcean Redefines Availability Metrics: From Incident Counting to SLIs

DigitalOcean Overhauls Availability Metrics to Better Reflect Customer Experience

DigitalOcean has announced a significant update to its platform availability metrics, aiming to provide a more accurate representation of customer experiences. This overhaul, which began in early 2025, addresses discrepancies between reported availability and actual user experiences, moving away from an incident-based measurement system that failed to capture the nuances of service performance. The new framework introduces Service Level Indicators (SLIs) that promise to deliver a clearer picture of reliability across its services.

Identifying the Problem with Traditional Metrics

At the start of 2025, DigitalOcean’s internal analysis revealed a troubling gap between reported availability figures and customer experiences. Monthly availability fluctuated between 99.5% and 99.9%, largely influenced by whether high-severity incidents were declared. This incident-based approach treated any declared incident as a complete outage while ignoring lower-severity issues that still affected users. As DigitalOcean expanded its offerings, this method became increasingly inadequate, creating a structural trap: including lower-severity incidents would artificially lower the overall availability metric.

The previous formula for calculating availability was straightforward but flawed. If no incidents were declared in a week, the platform was considered 100% available. However, if an incident lasted two hours, those minutes were deducted from the total available time. This simplistic approach led to three main issues:

  • The assumption that all customers experienced total downtime during an incident.
  • Lack of volume weighting across different products, treating low-traffic services the same as high-traffic ones.
  • A one-size-fits-all formula that did not accurately reflect individual product performance.

Introducing Control Plane and Data Plane Measurements

To rectify these shortcomings, DigitalOcean decided to separate its measurements into two distinct planes: Control Plane and Data Plane. Each plane utilizes its own SLI methodology tailored to specific operational contexts.

Control Plane Metrics

The Control Plane encompasses orchestration tasks such as API calls and Cloud Panel operations. Here, the SLI is based on the success rate of valid requests while only counting server errors (5xx) as failures. Client errors (4xx) remain in the denominator but do not count against reliability metrics since they stem from user mistakes rather than platform issues.

This change means that if users encounter errors in API or Cloud Control Panel operations due to Control Plane degradation, their ongoing workloads remain unaffected. The focus is now on measuring what truly impacts customer experience rather than simply counting incidents.

Data Plane Metrics

The Data Plane measures live product instances such as CPU and GPU Droplets, DOKS Clusters, and Serverless Inference endpoints. Unlike the Control Plane, where failures are more uniform, different products within the Data Plane fail in unique ways necessitating tailored measurement approaches.

For instance:

  • Droplet Networking: A Droplet is considered available when all connectivity probes pass simultaneously; any single failure counts against its availability.
  • DOKS Clusters: Availability is contingent upon multiple conditions being met simultaneously; failure in any component results in downtime for that minute.
  • Spaces Storage: Availability is measured by the ratio of successful responses (non-5xx) at the storage load balancer level.

This nuanced approach allows DigitalOcean to define failure states based on actual customer experiences rather than arbitrary thresholds.

Aggregation and Alerting Improvements

The next challenge was aggregating data across multiple regions with varying traffic volumes. DigitalOcean employed a weighted request average for Control Plane metrics so that smaller data centers contribute less to overall availability calculations compared to larger ones handling more traffic.

This method ensures that statistics accurately reflect service performance where it matters most—where customers are actively using resources. For Data Plane metrics, a similar magnitude-weighted approach was adopted to account for resource health across different products effectively.

To enhance alerting mechanisms, DigitalOcean transitioned from single burn rate alerts to multi-window alerts that combine short-term and long-term data analysis. This system allows engineers to distinguish between temporary spikes in error rates and ongoing issues requiring immediate attention.

Error Budgets as Decision-Making Tools

The introduction of error budgets—essentially an allowance for acceptable failures—has transformed how teams prioritize their work at DigitalOcean. By defining thresholds for error budgets based on SLIs, teams can make informed decisions about deploying new features or focusing on stability improvements based on their current standing within defined zones ranging from green (healthy) to red (critical).

  • Green (0-60%): Normal operations allowed.
  • Yellow (61%-80%): Caution advised; verify no impact on dependencies.
  • Orange (81-100%): Increased risk; large rollouts paused for reliability work.
  • Red (>100%): All rollouts halted except for critical fixes.

What This Means for Customers

This comprehensive overhaul of DigitalOcean’s availability metrics signifies a commitment to transparency and improved customer experience. By implementing SLIs tailored to specific operational contexts and employing sophisticated alerting mechanisms alongside error budgets as decision-making tools, DigitalOcean aims not only to improve internal performance but also enhance trust among its users. As newer products like GPU Droplets and AI agents roll out under this framework, customers can expect more reliable service backed by metrics that genuinely reflect their experiences rather than just theoretical calculations of uptime.

For more information, read the original report here.

Neil S
Neil S
Neil is a highly qualified Technical Writer with an M.Sc(IT) degree and an impressive range of IT and Support certifications including MCSE, CCNA, ACA(Adobe Certified Associates), and PG Dip (IT). With over 10 years of hands-on experience as an IT support engineer across Windows, Mac, iOS, and Linux Server platforms, Neil possesses the expertise to create comprehensive and user-friendly documentation that simplifies complex technical concepts for a wide audience.
Watch & Subscribe Our YouTube Channel
YouTube Subscribe Button

Latest From Hawkdive

You May like these Related Articles

LEAVE A REPLY

Please enter your comment!
Please enter your name here

This site uses Akismet to reduce spam. Learn how your comment data is processed.