Scalr incident

API Performance Degradation

Critical Resolved View vendor source →
Started
Apr 28, 2026, 09:52 AM UTC
Resolved
Apr 28, 2026, 09:52 AM UTC
Duration
Detected by Pingoru
Apr 28, 2026, 09:52 AM UTC

Affected components

Scalr PlatformScalr Worker

Update timeline

  1. investigating Apr 28, 2026, 08:53 AM UTC

    We are currently investigating this issue.

  2. investigating Apr 28, 2026, 09:00 AM UTC

    We are continuing to investigate this issue.

  3. investigating Apr 28, 2026, 09:07 AM UTC

    We are continuing to investigate this issue.

  4. investigating Apr 28, 2026, 09:08 AM UTC

    We are continuing to investigate this issue.

  5. identified Apr 28, 2026, 09:18 AM UTC

    The issue has been identified and a fix is being implemented.

  6. monitoring Apr 28, 2026, 09:21 AM UTC

    A fix has been implemented and we are monitoring the results.

  7. resolved Apr 28, 2026, 09:52 AM UTC

    This incident has been resolved.

  8. postmortem May 01, 2026, 02:38 PM UTC

    **Summary** A backend optimization deployed on April 24 inadvertently triggered a database performance issue that caused a platform-wide outage. The change has been fully reverted and the platform is stable. We are addressing the underlying query design before re-attempting the optimization. **What Happened** The optimization changed how policy code is delivered during runs, moving from an inline payload to a download from blob storage. This introduced an additional authorization check on each policy download that was not present before. That authorization check relied on an existing database query with significant hidden complexity: under normal conditions it goes unnoticed, but at production policy-check volume it joined approximately 2 million rows per call, taking around 25 seconds to complete. The increased call frequency exposed this latency, drove the database to 100% CPU utilization, and exhausted the connection pool, making the platform unavailable. **Resolution** The optimization was fully reverted, restoring normal platform behavior. No data was lost or corrupted. **What We're Doing Next** The authorization query is being redesigned to use a direct indexed lookup, which eliminates the row-scan behavior that caused the spike. The optimization will not be re-released until this redesign is complete and validated. We apologize for the disruption. If you have questions or are still experiencing issues, please contact our support team.

Looking to track Scalr downtime and outages?

Pingoru polls Scalr's status page every 5 minutes and alerts you the moment it reports an issue — before your customers do.

  • Real-time alerts when Scalr reports an incident
  • Email, Slack, Discord, Microsoft Teams, and webhook notifications
  • Track Scalr alongside 5,000+ providers in one dashboard
  • Component-level filtering
  • Notification groups + maintenance calendar
Start monitoring Scalr for free

5 free monitors · No credit card required