SecurID incident

RSA ID Plus Service Incident (ANZ Region)

Critical Resolved View vendor source →
Started
Apr 01, 2026, 08:15 AM UTC
Resolved
Apr 01, 2026, 11:02 AM UTC
Duration
2h 47m
Detected by Pingoru
Apr 01, 2026, 08:15 AM UTC

Affected components

access-anz Authentication Service

Update timeline

  1. investigating Apr 01, 2026, 08:15 AM UTC

    We have detected an issue affecting RSA ID Plus. SaaS Operations is investigating the issue and will post updates as they become available.

  2. investigating Apr 01, 2026, 08:52 AM UTC

    The SecurID SaaS Operations Team continues to investigate this issue. We will provide an update as soon as more information is available. Thank you, The SecurID Team

  3. monitoring Apr 01, 2026, 09:25 AM UTC

    The issue affecting RSA ID Plus has been corrected. The SaaS Operations team is monitoring the fix. We will post a root cause analysis as soon as it is available.

  4. resolved Apr 01, 2026, 11:02 AM UTC

    After monitoring the fix, SaaS Operations has determined that the incident affecting RSA ID Plus has been resolved. We will post a root cause analysis as soon as it is available.

  5. postmortem Apr 02, 2026, 10:11 PM UTC

    **Summary** On 04/01/2026, RSA ID Plus experienced a service outage in the ANZ region, impacting authentication workflows and related services. The issue was caused by a performance issue at the datastore layer caused by a non-optimal query execution pattern within one of the service data tiers, which led to elevated resource utilization and increased latency. This resulted in an authentication service outage. Service was restored through a combination of capacity adjustments and regional failover, and stability has been maintained since recovery. **Preliminary** **Root Cause** The incident was attributed to a query optimizer behavior that resulted in a suboptimal execution plan within the service data tier under specific runtime conditions. Under these conditions, the optimizer selected an inefficient query execution strategy, which led to: * Increased resource consumption within the data tier * Elevated query execution times * Resource contention impacting dependent services These conditions drove increased latency and service instability within authentication workflows. The behavior is consistent with a query plan regression scenario, where the optimizer generates a plan that is not optimal for the current data distribution or workload characteristics. **Recovery** During the incident, the team followed a controlled recovery approach aligned with the platform’s resilience design. RSA ID Plus is architected with strong in-region resilience, including redundancy and scaling capabilities across service layers, as well as a warm secondary region available for failover. In line with this design, the initial response focused on stabilizing the primary region by addressing resource contention within the data tier. Failover to the secondary region is available and was successfully executed as part of the recovery. Based on real-time impact assessment and recovery progress, it was determined that in-region remediation would not restore service within acceptable thresholds. A controlled failover to the secondary region was then initiated. Following stabilization, corrective changes were implemented within the primary region to address the underlying query optimization behavior. Once these changes were validated, traffic was safely transitioned back to the primary region in a controlled manner. All systems have remained stable under continued monitoring since recovery. **Mitigation and Resolution** The following actions were taken to restore service: * **Capacity Adjustment:** Increased available resources within the data tier to reduce contention * **Regional Failover:** Traffic was redirected to a warm region to restore service responsiveness * **Service Stabilization:** Performance returned to normal following these actions **Preventive Actions** **Immediate Actions** * Identification and correction of impacted query behavior * Stabilization of query execution patterns to prevent recurrence **Vendor Engagement and Safeguard Review** As part of the ongoing investigation and prevention efforts, RSA has engaged with our data service provider to further evaluate query optimization behavior observed during the incident. This includes a focused review of existing safeguards and protective mechanisms designed to detect and mitigate suboptimal query execution patterns. While these controls are in place, this event identified conditions under which they did not intervene as expected. This review is actively in progress, and findings will be incorporated into follow-up corrective actions.

Looking to track SecurID downtime and outages?

Pingoru polls SecurID's status page every 5 minutes and alerts you the moment it reports an issue — before your customers do.

  • Real-time alerts when SecurID reports an incident
  • Email, Slack, Discord, Microsoft Teams, and webhook notifications
  • Track SecurID alongside 5,000+ providers in one dashboard
  • Component-level filtering
  • Notification groups + maintenance calendar
Start monitoring SecurID for free

5 free monitors · No credit card required