- Detected by Pingoru
- May 02, 2026, 01:38 AM UTC
- Resolved
- May 02, 2026, 02:47 AM UTC
- Duration
- 1h 9m
Affected: Support PortalTicketingPublic Issues Tracker
Timeline · 4 updates
-
investigating May 02, 2026, 01:38 AM UTC
We are investigating an incident affecting Jira Software (viewBoard, viewIssue, createIssue). Our team is working to identify the cause and restore service to normal levels. We will provide the next update within 60 minutes.
-
investigating May 02, 2026, 01:38 AM UTC
We are actively investigating reports of a service disruption affecting the issue viewing experience in Jira. This is also impacting the accessibility of support tickets. We will share updates here as more information becomes available.
-
investigating May 02, 2026, 02:00 AM UTC
We are actively investigating reports of a service disruption affecting the Work Item viewing experience in Jira. This is also impacting the accessibility of support tickets. We will share updates here as more information becomes available.
-
resolved May 02, 2026, 02:47 AM UTC
The Jira Work Item View experience has been restored to normal service. Our teams are continuing to investigate the root cause affecting this issue. We will provide more details once we identify the root cause
Read the full incident report →
- Detected by Pingoru
- May 02, 2026, 12:55 AM UTC
- Resolved
- May 02, 2026, 02:48 AM UTC
- Duration
- 1h 52m
Affected: Viewing contentCreate and edit
Timeline · 4 updates
-
investigating May 02, 2026, 12:55 AM UTC
We are investigating an incident affecting Jira Software (viewBoard, viewIssue, createIssue). Our team is working to identify the cause and restore service to normal levels. We will provide the next update within 60 minutes.
-
investigating May 02, 2026, 01:38 AM UTC
We are actively investigating reports of a service disruption affecting the issue viewing experience in Jira. This is also impacting the accessibility of support tickets. We will share updates here as more information becomes available.
-
investigating May 02, 2026, 02:00 AM UTC
We are actively investigating reports of a service disruption affecting the Work Item viewing experience in Jira. This is also impacting the accessibility of support tickets. We will share updates here as more information becomes available.
-
resolved May 02, 2026, 02:48 AM UTC
The Jira Work Item View experience has been restored to normal service. Our teams are continuing to investigate the root cause affecting this issue. We will provide more details once we identify the root cause
Read the full incident report →
- Detected by Pingoru
- May 02, 2026, 12:55 AM UTC
- Resolved
- May 02, 2026, 02:48 AM UTC
- Duration
- 1h 52m
Affected: Service PortalOpsgenie Incident Flow
Timeline · 4 updates
-
investigating May 02, 2026, 12:55 AM UTC
We are investigating an incident affecting Jira Software (viewBoard, viewIssue, createIssue). Our team is working to identify the cause and restore service to normal levels. We will provide the next update within 60 minutes.
-
investigating May 02, 2026, 01:38 AM UTC
We are actively investigating reports of a service disruption affecting the issue viewing experience in Jira. This is also impacting the accessibility of support tickets. We will share updates here as more information becomes available.
-
investigating May 02, 2026, 02:00 AM UTC
We are actively investigating reports of a service disruption affecting the Work Item viewing experience in Jira. This is also impacting the accessibility of support tickets. We will share updates here as more information becomes available.
-
resolved May 02, 2026, 02:48 AM UTC
The Jira Work Item View experience has been restored to normal service. Our teams are continuing to investigate the root cause affecting this issue. We will provide more details once we identify the root cause
Read the full incident report →
- Detected by Pingoru
- May 02, 2026, 12:55 AM UTC
- Resolved
- May 02, 2026, 02:48 AM UTC
- Duration
- 1h 52m
Affected: Service PortalOpsgenie Incident Flow
Timeline · 4 updates
-
investigating May 02, 2026, 12:55 AM UTC
We are investigating an incident affecting Jira Software (viewBoard, viewIssue, createIssue). Our team is working to identify the cause and restore service to normal levels. We will provide the next update within 60 minutes.
-
investigating May 02, 2026, 01:38 AM UTC
We are actively investigating reports of a service disruption affecting the issue viewing experience in Jira. This is also impacting the accessibility of support tickets. We will share updates here as more information becomes available.
-
investigating May 02, 2026, 02:00 AM UTC
We are actively investigating reports of a service disruption affecting the Work Item viewing experience in Jira. This is also impacting the accessibility of support tickets. We will share updates here as more information becomes available.
-
resolved May 02, 2026, 02:48 AM UTC
The Jira Work Item View experience has been restored to normal service. Our teams are continuing to investigate the root cause affecting this issue. We will provide more details once we identify the root cause
Read the full incident report →
- Detected by Pingoru
- May 02, 2026, 12:55 AM UTC
- Resolved
- May 02, 2026, 02:48 AM UTC
- Duration
- 1h 52m
Affected: Viewing contentCreate and edit
Timeline · 4 updates
-
investigating May 02, 2026, 12:55 AM UTC
We are investigating an incident affecting Jira Software (viewBoard, viewIssue, createIssue). Our team is working to identify the cause and restore service to normal levels. We will provide the next update within 60 minutes.
-
investigating May 02, 2026, 01:38 AM UTC
We are actively investigating reports of a service disruption affecting the issue viewing experience in Jira. This is also impacting the accessibility of support tickets. We will share updates here as more information becomes available.
-
investigating May 02, 2026, 02:00 AM UTC
We are actively investigating reports of a service disruption affecting the Work Item viewing experience in Jira. This is also impacting the accessibility of support tickets. We will share updates here as more information becomes available.
-
resolved May 02, 2026, 02:48 AM UTC
The Jira Work Item View experience has been restored to normal service. Our teams are continuing to investigate the root cause affecting this issue. We will provide more details once we identify the root cause
Read the full incident report →
- Detected by Pingoru
- May 02, 2026, 12:55 AM UTC
- Resolved
- May 02, 2026, 02:48 AM UTC
- Duration
- 1h 52m
Affected: Viewing contentCreate and edit
Timeline · 4 updates
-
investigating May 02, 2026, 12:55 AM UTC
We are investigating an incident affecting Jira Software (viewBoard, viewIssue, createIssue). Our team is working to identify the cause and restore service to normal levels. We will provide the next update within 60 minutes.
-
investigating May 02, 2026, 01:38 AM UTC
We are actively investigating reports of a service disruption affecting the issue viewing experience in Jira. This is also impacting the accessibility of support tickets. We will share updates here as more information becomes available.
-
investigating May 02, 2026, 02:00 AM UTC
We are actively investigating reports of a service disruption affecting the Work Item viewing experience in Jira. This is also impacting the accessibility of support tickets. We will share updates here as more information becomes available.
-
resolved May 02, 2026, 02:48 AM UTC
The Jira Work Item View experience has been restored to normal service. Our teams are continuing to investigate the root cause affecting this issue. We will provide more details once we identify the root cause
Read the full incident report →
- Detected by Pingoru
- Apr 30, 2026, 08:59 AM UTC
- Resolved
- Apr 30, 2026, 02:56 PM UTC
- Duration
- 5h 57m
Affected: Bitbucket Cloud APIsApp listing managementApp DeploymentApp listingsConfluence Cloud APIsArtifactory (Maven repository)Atlassian Support contact formApp pricingJira Cloud APIsCreate and manage appsDeveloper communityDeveloper documentationDeveloper service deskApp submissionsAuthentication and user managementProduct EventsMarketplace service deskUser APIsCategory landing pagesForge App InstallationWebhooksEvaluations and purchasesAtlassian SupportForge CDN (Custom UI)In-product Marketplace and app installation (Cloud)Forge Function InvocationWeb TriggersVulnerability management [AMS]In-product Marketplace and app installation (Server)aui-cdn.atlassian.comNotificationsForge App LogsPrivate listingsForge App MonitoringReporting APIs and dashboardsDeveloper consoleSearchForge direct app distributionVendor managementHosted storageVendor Home PageForge CLIEnd-user consentForge App AlertsApp Data ResidencyForge SQLForge MonetisationForge Feature Flags Evaluations
Timeline · 4 updates
-
investigating Apr 30, 2026, 08:59 AM UTC
We are investigating reports of delays in webhook delivery. Customers may experience intermittent delays when receiving webhooks. Our engineering teams are actively investigating and working to resolve the issue. We will provide updates as more information becomes available.
-
identified Apr 30, 2026, 10:05 AM UTC
We have identified the cause of the delayed Jira webhook delivery and are deploying a fix. We will continue to provide updates as more information becomes available.
-
monitoring Apr 30, 2026, 01:36 PM UTC
A fix has been deployed to address the intermittent delays in Jira webhook delivery. We are actively monitoring the situation to confirm full recovery. We will continue to provide updates as available.
-
resolved Apr 30, 2026, 02:56 PM UTC
The issue causing intermittent delays in our webhook delivery has been resolved. A fix was deployed to address a resource constraint in our systems. Our monitoring shows webhook delivery is now operating normally.
Read the full incident report →
- Detected by Pingoru
- Apr 30, 2026, 08:59 AM UTC
- Resolved
- Apr 30, 2026, 02:56 PM UTC
- Duration
- 5h 57m
Affected: Viewing contentCreate and editAuthentication and User ManagementSearchNotificationsAdministrationMarketplaceMobilePurchasing & LicensingSignupAutomation for Jira
Timeline · 4 updates
-
investigating Apr 30, 2026, 08:59 AM UTC
We are investigating reports of delays in webhook delivery. Customers may experience intermittent delays when receiving webhooks. Our engineering teams are actively investigating and working to resolve the issue. We will provide updates as more information becomes available.
-
identified Apr 30, 2026, 10:05 AM UTC
We have identified the cause of the delayed Jira webhook delivery and are deploying a fix. We will continue to provide updates as more information becomes available.
-
monitoring Apr 30, 2026, 01:36 PM UTC
A fix has been deployed to address the intermittent delays in Jira webhook delivery. We are actively monitoring the situation to confirm full recovery. We will continue to provide updates as available.
-
resolved Apr 30, 2026, 02:56 PM UTC
The issue causing intermittent delays in our webhook delivery has been resolved. A fix was deployed to address a resource constraint in our systems. Our monitoring shows webhook delivery is now operating normally.
Read the full incident report →
- Detected by Pingoru
- Apr 29, 2026, 07:00 AM UTC
- Resolved
- Apr 29, 2026, 07:50 AM UTC
- Duration
- 50m
Affected: RecordingViewingAuthenticationLoom AI
Timeline · 2 updates
-
investigating Apr 29, 2026, 07:00 AM UTC
We are investigating degraded performance issues affecting Loom's core functionalities - authentication, recording, viewing and editing looms. We will provide more details within the next hour.
-
resolved Apr 29, 2026, 07:50 AM UTC
The Team has identified the root cause and the issue has now been resolved, and the impacted services are operating normally. Team will continue to monitor the impacted services.
Read the full incident report →
- Detected by Pingoru
- Apr 28, 2026, 06:28 PM UTC
- Resolved
- Apr 28, 2026, 09:50 PM UTC
- Duration
- 3h 22m
Affected: Bitbucket Cloud APIsApp listing managementApp DeploymentApp listingsConfluence Cloud APIsArtifactory (Maven repository)Atlassian Support contact formApp pricingJira Cloud APIsCreate and manage appsDeveloper communityDeveloper documentationDeveloper service deskApp submissionsAuthentication and user managementProduct EventsMarketplace service deskUser APIsCategory landing pagesForge App InstallationWebhooksEvaluations and purchasesAtlassian SupportForge CDN (Custom UI)In-product Marketplace and app installation (Cloud)Forge Function InvocationWeb TriggersVulnerability management [AMS]In-product Marketplace and app installation (Server)aui-cdn.atlassian.comNotificationsForge App LogsPrivate listingsForge App MonitoringReporting APIs and dashboardsDeveloper consoleSearchForge direct app distributionVendor managementHosted storageVendor Home PageForge CLIEnd-user consentForge App AlertsApp Data ResidencyForge SQLForge MonetisationForge Feature Flags Evaluations
Timeline · 5 updates
-
investigating Apr 28, 2026, 06:28 PM UTC
We are actively investigating reports of performance degradation affecting Forge webhook delivery in prod-euwest region. We will share updates here as more information is available.
-
investigating Apr 28, 2026, 06:34 PM UTC
We are actively investigating reports of performance degradation affecting Forge triggers in prod-euwest region. We will share updates here as more information is available.
-
identified Apr 28, 2026, 08:16 PM UTC
We have identified the likely cause of the issue, and our teams are diligently working on a mitigation. Affected users in prod-euwest region may experience performance degradation affecting Forge triggers. We will continue to share additional updates here as more information is available.
-
monitoring Apr 28, 2026, 09:34 PM UTC
The performance degradation of Forge triggers has been resolved, and services are now operating normally for all affected customers. We'll continue to monitor performance closely to confirm stability.
-
resolved Apr 28, 2026, 09:50 PM UTC
The issue has now been resolved, and the service is operating normally for all affected customers.
Read the full incident report →
- Detected by Pingoru
- Apr 22, 2026, 06:57 AM UTC
- Resolved
- Apr 23, 2026, 04:14 PM UTC
- Duration
- 1d 9h
Affected: Bitbucket Cloud APIsApp listing managementApp DeploymentApp listingsConfluence Cloud APIsArtifactory (Maven repository)Atlassian Support contact formApp pricingJira Cloud APIsCreate and manage appsDeveloper communityDeveloper documentationDeveloper service deskApp submissionsAuthentication and user managementProduct EventsMarketplace service deskUser APIsCategory landing pagesForge App InstallationWebhooksEvaluations and purchasesAtlassian SupportForge CDN (Custom UI)In-product Marketplace and app installation (Cloud)Forge Function InvocationWeb TriggersVulnerability management [AMS]In-product Marketplace and app installation (Server)aui-cdn.atlassian.comNotificationsForge App LogsPrivate listingsForge App MonitoringReporting APIs and dashboardsDeveloper consoleSearchForge direct app distributionVendor managementHosted storageVendor Home PageForge CLIEnd-user consentForge App AlertsApp Data ResidencyForge SQLForge MonetisationForge Feature Flags Evaluations
Timeline · 6 updates
-
investigating Apr 22, 2026, 06:57 AM UTC
We are aware that some Forge apps are experiencing an increased number of 429 Too Many Requests errors when making calls to the Jira REST API. This is causing disruption for end users of affected apps. Confluence APIs are not impacted. If your app is impacted by this issue, the 429 response will contain the following headers: X-Failure-Category: FAILURE CLIENT RATE LIMITED X-Ratelimit-Limit: 360000 X-Ratelimit-Remaining: 0 Note that this gateway-level rate limiting does not return a Retry-After header. You can use the X-Ratelimit-Reset header to determine when the rate limit window resets and implement a backoff strategy accordingly. Our teams are investigating this issue with urgency, and we will provide an update within two hours.
-
identified Apr 22, 2026, 08:53 AM UTC
Our teams have implemented initial remediation by increasing certain rate limits, and we will continue to monitor the situation closely to confirm stability. We will share further updates within the next 8 hours, or sooner if we have any meaningful developments to report.
-
identified Apr 22, 2026, 04:21 PM UTC
After the initial remediation, we are seeing improvements and our teams continue to monitor the situation closely to confirm stability. We will share further updates within the next 8 hours, or sooner if there are any major developments.
-
identified Apr 22, 2026, 11:29 PM UTC
We continue to see improvements in error rates, and our teams are evaluating further mitigation to expedite resolution.
-
monitoring Apr 23, 2026, 03:27 PM UTC
The performance degradation has been resolved, and services are now operating normally for all affected customers. We will continue to monitor performance closely to confirm stability.
-
resolved Apr 23, 2026, 04:14 PM UTC
The issue has now been resolved, and the service is operating normally for all affected customers.
Read the full incident report →
- Detected by Pingoru
- Apr 16, 2026, 08:23 PM UTC
- Resolved
- Apr 16, 2026, 08:44 PM UTC
- Duration
- 21m
Affected: WebsiteAPIGit via SSHAuthentication and user managementGit via HTTPSWebhooksSource downloadsPipelinesGit LFSEmail deliveryPurchasing & LicensingSignup
Timeline · 3 updates
-
monitoring Apr 16, 2026, 08:23 PM UTC
We are investigating cases of degraded performance for Atlassian Bitbucket Pipelines customers. We mitigated an issue preventing new pipelines from starting, and have scaled up services are processing a backlog. Customers should see recovery shortly.
-
monitoring Apr 16, 2026, 08:44 PM UTC
We are investigating cases of degraded performance for Atlassian Bitbucket Pipelines customers. We mitigated an issue preventing new pipelines from starting, and have scaled up services are processing a backlog. Customers should see recovery shortly.
-
resolved Apr 16, 2026, 08:44 PM UTC
On April 16 at 8:00PM UTC Bitbucket Pipelines users may have experienced performance degradation running new pipelines. The issue has now been resolved, and the service is operating normally for all affected customers.
Read the full incident report →
- Detected by Pingoru
- Apr 14, 2026, 12:25 PM UTC
- Resolved
- Apr 14, 2026, 04:11 PM UTC
- Duration
- 3h 46m
Affected: View ContentiOS AppCreate and EditAndroid AppCommentsAuthentication and User ManagementSearchAdministrationNotificationsMarketplace AppsPurchasing & LicensingSignupConfluence AutomationsCloud to Cloud Migrations - Copy Product DataServer to Cloud Migrations - Copy Product Data
Timeline · 5 updates
-
investigating Apr 14, 2026, 12:25 PM UTC
We are actively investigating reports of a partial service disruption affecting Rovo, specifically for customers using automation rules that invoke Rovo agents. Some customers may find that these automations are not completing as expected. We'll share updates here as more information becomes available.
-
investigating Apr 14, 2026, 12:25 PM UTC
We have identified that this incident also affects automation rules in Confluence that use Rovo agents. We are investigating and will provide updates as we learn more.
-
identified Apr 14, 2026, 02:11 PM UTC
We have identified the issue, and our teams are working to resolve it and restore normal operations as quickly as possible. We will provide further updates as they become available.
-
monitoring Apr 14, 2026, 03:46 PM UTC
The issue has been resolved, and services are now operating normally for all affected customers. We will continue to monitor closely to confirm stability.
-
resolved Apr 14, 2026, 04:11 PM UTC
On April 14, 2026, affected users may have experienced some service disruption with automation rules that use Rovo agents. The issue has now been resolved, and the service is operating normally for all affected customers.
Read the full incident report →
- Detected by Pingoru
- Apr 14, 2026, 12:12 PM UTC
- Resolved
- Apr 14, 2026, 04:11 PM UTC
- Duration
- 3h 58m
Affected: Viewing contentCreate and editAuthentication and User ManagementSearchNotificationsAdministrationMarketplaceMobilePurchasing & LicensingSignupAutomation for Jira
Timeline · 5 updates
-
investigating Apr 14, 2026, 12:12 PM UTC
We are actively investigating reports of a partial service disruption affecting Rovo, specifically for customers using automation rules that invoke Rovo agents. Some customers may find that these automations are not completing as expected. We'll share updates here as more information becomes available.
-
investigating Apr 14, 2026, 12:25 PM UTC
We have identified that this incident also affects automation rules in Confluence that use Rovo agents. We are investigating and will provide updates as we learn more.
-
identified Apr 14, 2026, 02:11 PM UTC
We have identified the issue, and our teams are working to resolve it and restore normal operations as quickly as possible. We will provide further updates as they become available.
-
monitoring Apr 14, 2026, 03:46 PM UTC
The issue has been resolved, and services are now operating normally for all affected customers. We will continue to monitor closely to confirm stability.
-
resolved Apr 14, 2026, 04:11 PM UTC
On April 14, 2026, affected users may have experienced some service disruption with automation rules that use Rovo agents. The issue has now been resolved, and the service is operating normally for all affected customers.
Read the full incident report →
- Detected by Pingoru
- Apr 14, 2026, 12:12 PM UTC
- Resolved
- Apr 14, 2026, 04:11 PM UTC
- Duration
- 3h 58m
Affected: Viewing contentCreate and editAuthentication and User ManagementSearchNotificationsAdministrationMarketplaceMobilePurchasing & LicensingSignupAutomation for Jira
Timeline · 5 updates
-
investigating Apr 14, 2026, 12:12 PM UTC
We are actively investigating reports of a partial service disruption affecting Rovo, specifically for customers using automation rules that invoke Rovo agents. Some customers may find that these automations are not completing as expected. We'll share updates here as more information becomes available.
-
investigating Apr 14, 2026, 12:25 PM UTC
We have identified that this incident also affects automation rules in Confluence that use Rovo agents. We are investigating and will provide updates as we learn more.
-
identified Apr 14, 2026, 02:11 PM UTC
We have identified the issue, and our teams are working to resolve it and restore normal operations as quickly as possible. We will provide further updates as they become available.
-
monitoring Apr 14, 2026, 03:46 PM UTC
The issue has been resolved, and services are now operating normally for all affected customers. We will continue to monitor closely to confirm stability.
-
resolved Apr 14, 2026, 04:11 PM UTC
On April 14, 2026, affected users may have experienced some service disruption with automation rules that use Rovo agents. The issue has now been resolved, and the service is operating normally for all affected customers.
Read the full incident report →
- Detected by Pingoru
- Apr 14, 2026, 12:12 PM UTC
- Resolved
- Apr 14, 2026, 04:11 PM UTC
- Duration
- 3h 58m
Affected: Jira Service Management WebService PortalOpsgenie Incident FlowOpsgenie Alert FlowOpsgenie Incident FlowOpsgenie Alert FlowJira Service Management Email RequestsAuthentication and User ManagementPurchasing & LicensingSignupAutomation for JiraAssist
Timeline · 5 updates
-
investigating Apr 14, 2026, 12:12 PM UTC
We are actively investigating reports of a partial service disruption affecting Rovo, specifically for customers using automation rules that invoke Rovo agents. Some customers may find that these automations are not completing as expected. We'll share updates here as more information becomes available.
-
investigating Apr 14, 2026, 12:25 PM UTC
We have identified that this incident also affects automation rules in Confluence that use Rovo agents. We are investigating and will provide updates as we learn more.
-
identified Apr 14, 2026, 02:11 PM UTC
We have identified the issue, and our teams are working to resolve it and restore normal operations as quickly as possible. We will provide further updates as they become available.
-
monitoring Apr 14, 2026, 03:46 PM UTC
The issue has been resolved, and services are now operating normally for all affected customers. We will continue to monitor closely to confirm stability.
-
resolved Apr 14, 2026, 04:11 PM UTC
On April 14, 2026, affected users may have experienced some service disruption with automation rules that use Rovo agents. The issue has now been resolved, and the service is operating normally for all affected customers.
Read the full incident report →
- Detected by Pingoru
- Apr 14, 2026, 12:12 PM UTC
- Resolved
- Apr 14, 2026, 04:11 PM UTC
- Duration
- 3h 58m
Affected: Jira Service Management WebService PortalOpsgenie Incident FlowOpsgenie Alert FlowOpsgenie Incident FlowOpsgenie Alert FlowJira Service Management Email RequestsAuthentication and User ManagementPurchasing & LicensingSignupAutomation for JiraAssist
Timeline · 5 updates
-
investigating Apr 14, 2026, 12:12 PM UTC
We are actively investigating reports of a partial service disruption affecting Rovo, specifically for customers using automation rules that invoke Rovo agents. Some customers may find that these automations are not completing as expected. We'll share updates here as more information becomes available.
-
investigating Apr 14, 2026, 12:25 PM UTC
We have identified that this incident also affects automation rules in Confluence that use Rovo agents. We are investigating and will provide updates as we learn more.
-
identified Apr 14, 2026, 02:11 PM UTC
We have identified the issue, and our teams are working to resolve it and restore normal operations as quickly as possible. We will provide further updates as they become available.
-
monitoring Apr 14, 2026, 03:46 PM UTC
The issue has been resolved, and services are now operating normally for all affected customers. We will continue to monitor closely to confirm stability.
-
resolved Apr 14, 2026, 04:11 PM UTC
On April 14, 2026, affected users may have experienced some service disruption with automation rules that use Rovo agents. The issue has now been resolved, and the service is operating normally for all affected customers.
Read the full incident report →
- Detected by Pingoru
- Apr 14, 2026, 12:12 PM UTC
- Resolved
- Apr 14, 2026, 04:11 PM UTC
- Duration
- 3h 58m
Affected: Viewing contentCreate and editAuthentication and User ManagementSearchNotificationsAdministrationMarketplaceMobilePurchasing & LicensingSignupAutomation for Jira
Timeline · 5 updates
-
investigating Apr 14, 2026, 12:12 PM UTC
We are actively investigating reports of a partial service disruption affecting Rovo, specifically for customers using automation rules that invoke Rovo agents. Some customers may find that these automations are not completing as expected. We'll share updates here as more information becomes available.
-
investigating Apr 14, 2026, 12:25 PM UTC
We have identified that this incident also affects automation rules in Confluence that use Rovo agents. We are investigating and will provide updates as we learn more.
-
identified Apr 14, 2026, 02:11 PM UTC
We have identified the issue, and our teams are working to resolve it and restore normal operations as quickly as possible. We will provide further updates as they become available.
-
monitoring Apr 14, 2026, 03:46 PM UTC
The issue has been resolved, and services are now operating normally for all affected customers. We will continue to monitor closely to confirm stability.
-
resolved Apr 14, 2026, 04:11 PM UTC
On April 14, 2026, affected users may have experienced some service disruption with automation rules that use Rovo agents. The issue has now been resolved, and the service is operating normally for all affected customers.
Read the full incident report →
- Detected by Pingoru
- Apr 13, 2026, 10:58 AM UTC
- Resolved
- Apr 13, 2026, 03:11 PM UTC
- Duration
- 4h 13m
Affected: Bitbucket Cloud APIsApp listing managementApp DeploymentApp listingsConfluence Cloud APIsArtifactory (Maven repository)Atlassian Support contact formApp pricingJira Cloud APIsCreate and manage appsDeveloper communityDeveloper documentationDeveloper service deskApp submissionsAuthentication and user managementProduct EventsMarketplace service deskUser APIsCategory landing pagesForge App InstallationWebhooksEvaluations and purchasesAtlassian SupportForge CDN (Custom UI)In-product Marketplace and app installation (Cloud)Forge Function InvocationWeb TriggersVulnerability management [AMS]In-product Marketplace and app installation (Server)aui-cdn.atlassian.comNotificationsForge App LogsPrivate listingsForge App MonitoringReporting APIs and dashboardsDeveloper consoleSearchForge direct app distributionVendor managementHosted storageVendor Home PageForge CLIEnd-user consentForge App AlertsApp Data ResidencyForge SQLForge MonetisationForge Feature Flags Evaluations
Timeline · 3 updates
-
investigating Apr 13, 2026, 10:58 AM UTC
We are aware that transaction-related information for Marketplace apps has not been available in partner reports since April 10, 2026. Our team is actively investigating the root cause and working to restore the data. Partners may notice missing transaction records for this period. We will provide updates as we have more information. We apologize for the inconvenience.
-
identified Apr 13, 2026, 01:32 PM UTC
Transaction information for Marketplace apps has been restored and is now appearing again in partner reports. You may still see up to a 3-hour delay before new transactions show in your reports while we work to bring this back to normal. We are closely monitoring the situation and will continue to share updates.
-
resolved Apr 13, 2026, 03:11 PM UTC
The issue has now been resolved, and the service is operating normally for all affected customers.
Read the full incident report →
- Detected by Pingoru
- Apr 13, 2026, 07:29 AM UTC
- Resolved
- Apr 13, 2026, 10:17 AM UTC
- Duration
- 2h 47m
Affected: Domain ClaimsData ClassificationSAML-based SSOData Security PoliciesUser ProvisioningGuard DetectAccount ManagementAudit LogsSignupAPI tokens
Timeline · 4 updates
-
monitoring Apr 13, 2026, 07:29 AM UTC
Our team is aware that some users were unable to log in to Atlassian products with their Atlassian accounts. While we believe this issue is now resolved, we are continuing to monitor all products and services for any ongoing impact. Our team is investigating with urgency, and we will provide an update within 1 hour.
-
monitoring Apr 13, 2026, 08:33 AM UTC
Atlassian account login services are now operating as expected, and we are not observing new errors. We continue to closely monitor our systems to ensure they remain stable. We will provide another update within 60 minutes or sooner if we detect a change in status.
-
resolved Apr 13, 2026, 10:17 AM UTC
On April 13, 2026, between 05:49 a.m. and 06:25 a.m. UTC, some users were unable to log in to Atlassian products with their Atlassian accounts. The underlying issue has been addressed, and authentication services have remained stable with no new impact observed.
-
postmortem Apr 21, 2026, 02:35 AM UTC
### Summary On April 13, 2026, between 05:49 and 06:29 UTC, customers experienced failures when attempting to log in, sign up, reset passwords, and complete multi-factor authentication flows across Atlassian cloud products. Approximately 90% of authentication requests failed during the peak impact window, affecting users in the US East and EU regions. The incident was mitigated within 40 minutes through manual intervention, and full service was restored by 06:29 UTC. ### **IMPACT** * **Duration**: ~40 minutes \(05:49–06:29 UTC, April 13, 2026\) * **Affected regions**: US East and EU \(authentication infrastructure serves EU traffic from US East, with traffic primarily from EU at this time of day\). * **Affected products**: All Atlassian cloud products requiring authentication, including Jira, Confluence, Jira Service Management, and Trello. * **Customer experience**: Users attempting to log in, sign up, reset passwords, or complete MFA flows received errors. Users already logged in with active sessions were unaffected. ### **ROOT CAUSE** This incident had several contributing factors that combined to produce a failure that the system could not recover from without manual intervention. **The primary cause** was a recently enabled change that caused our authentication infrastructure to retry requests to a downstream identity service when those requests were slow to respond. This retry behaviour was rolled out to 100% of traffic earlier the same day. Under normal conditions this would be benign, but it meant that any slowness in the downstream service was amplified. Since multiple upstream services were also independently retrying their own failed requests, the amplification compounded further into a retry storm. **The trigger** was a burst of legitimate user traffic. A pattern of many parallel link preview requests for a single user caused a concentrated load spike on a downstream identity service, pushing its response times above the retry threshold. On its own, this kind of spike had occurred many times before and always recovered. With the retry amplification now in effect, the spike instead created a runaway feedback loop: slow responses caused retries, retries increased load, increased load caused slower responses, preventing recovery. The incident was mitigated by manually scaling up the downstream identity service to provide sufficient capacity to absorb the amplified load. Once scaled, the service recovered immediately, bringing authentication error rates to zero within one minute. **REMEDIAL ACTIONS PLAN & NEXT STEPS** We are taking the following actions designed to prevent recurrence and improve our resilience: 1. **Immediate**: The retry-on-timeout change has been disabled. 2. **Load shedding and self-healing**: We are adding load shedding capabilities to our authentication services so that they can automatically shed excess load and self-recover during traffic spikes, without requiring action before automatic scaling starts. 3. **Reducing request fan-out**: We are reviewing patterns where a single user action can generate many parallel downstream requests, and will introduce methods where possible to reduce the amplification potential. We apologize to customers whose services were interrupted by this incident and we are taking immediate steps to improve the platform’s reliability. Thanks, Atlassian Customer Support
Read the full incident report →
- Detected by Pingoru
- Apr 13, 2026, 07:29 AM UTC
- Resolved
- Apr 13, 2026, 10:17 AM UTC
- Duration
- 2h 47m
Affected: Domain ClaimsData ClassificationSAML-based SSOData Security PoliciesUser ProvisioningGuard DetectAccount ManagementAudit LogsSignupAPI tokens
Timeline · 4 updates
-
monitoring Apr 13, 2026, 07:29 AM UTC
Our team is aware that some users were unable to log in to Atlassian products with their Atlassian accounts. While we believe this issue is now resolved, we are continuing to monitor all products and services for any ongoing impact. Our team is investigating with urgency, and we will provide an update within 1 hour.
-
monitoring Apr 13, 2026, 08:33 AM UTC
Atlassian account login services are now operating as expected, and we are not observing new errors. We continue to closely monitor our systems to ensure they remain stable. We will provide another update within 60 minutes or sooner if we detect a change in status.
-
resolved Apr 13, 2026, 10:17 AM UTC
On April 13, 2026, between 05:49 a.m. and 06:25 a.m. UTC, some users were unable to log in to Atlassian products with their Atlassian accounts. The underlying issue has been addressed, and authentication services have remained stable with no new impact observed.
-
postmortem Apr 21, 2026, 02:35 AM UTC
### Summary On April 13, 2026, between 05:49 and 06:29 UTC, customers experienced failures when attempting to log in, sign up, reset passwords, and complete multi-factor authentication flows across Atlassian cloud products. Approximately 90% of authentication requests failed during the peak impact window, affecting users in the US East and EU regions. The incident was mitigated within 40 minutes through manual intervention, and full service was restored by 06:29 UTC. ### **IMPACT** * **Duration**: ~40 minutes \(05:49–06:29 UTC, April 13, 2026\) * **Affected regions**: US East and EU \(authentication infrastructure serves EU traffic from US East, with traffic primarily from EU at this time of day\). * **Affected products**: All Atlassian cloud products requiring authentication, including Jira, Confluence, Jira Service Management, and Trello. * **Customer experience**: Users attempting to log in, sign up, reset passwords, or complete MFA flows received errors. Users already logged in with active sessions were unaffected. ### **ROOT CAUSE** This incident had several contributing factors that combined to produce a failure that the system could not recover from without manual intervention. **The primary cause** was a recently enabled change that caused our authentication infrastructure to retry requests to a downstream identity service when those requests were slow to respond. This retry behaviour was rolled out to 100% of traffic earlier the same day. Under normal conditions this would be benign, but it meant that any slowness in the downstream service was amplified. Since multiple upstream services were also independently retrying their own failed requests, the amplification compounded further into a retry storm. **The trigger** was a burst of legitimate user traffic. A pattern of many parallel link preview requests for a single user caused a concentrated load spike on a downstream identity service, pushing its response times above the retry threshold. On its own, this kind of spike had occurred many times before and always recovered. With the retry amplification now in effect, the spike instead created a runaway feedback loop: slow responses caused retries, retries increased load, increased load caused slower responses, preventing recovery. The incident was mitigated by manually scaling up the downstream identity service to provide sufficient capacity to absorb the amplified load. Once scaled, the service recovered immediately, bringing authentication error rates to zero within one minute. **REMEDIAL ACTIONS PLAN & NEXT STEPS** We are taking the following actions designed to prevent recurrence and improve our resilience: 1. **Immediate**: The retry-on-timeout change has been disabled. 2. **Load shedding and self-healing**: We are adding load shedding capabilities to our authentication services so that they can automatically shed excess load and self-recover during traffic spikes, without requiring action before automatic scaling starts. 3. **Reducing request fan-out**: We are reviewing patterns where a single user action can generate many parallel downstream requests, and will introduce methods where possible to reduce the amplification potential. We apologize to customers whose services were interrupted by this incident and we are taking immediate steps to improve the platform’s reliability. Thanks, Atlassian Customer Support
Read the full incident report →
- Detected by Pingoru
- Apr 13, 2026, 07:29 AM UTC
- Resolved
- Apr 13, 2026, 10:17 AM UTC
- Duration
- 2h 47m
Affected: Support PortalTicketingKnowledge BaseCommunityAPI DocsPublic Issues TrackerTrainingBlogsCareersContent DeliverySearchDownloads AccessPreferences Center
Timeline · 4 updates
-
monitoring Apr 13, 2026, 07:29 AM UTC
Our team is aware that some users were unable to log in to Atlassian products with their Atlassian accounts. While we believe this issue is now resolved, we are continuing to monitor all products and services for any ongoing impact. Our team is investigating with urgency, and we will provide an update within 1 hour.
-
monitoring Apr 13, 2026, 08:33 AM UTC
Atlassian account login services are now operating as expected, and we are not observing new errors. We continue to closely monitor our systems to ensure they remain stable. We will provide another update within 60 minutes or sooner if we detect a change in status.
-
resolved Apr 13, 2026, 10:17 AM UTC
On April 13, 2026, between 05:49 a.m. and 06:25 a.m. UTC, some users were unable to log in to Atlassian products with their Atlassian accounts. The underlying issue has been addressed, and authentication services have remained stable with no new impact observed.
-
postmortem Apr 21, 2026, 02:35 AM UTC
### Summary On April 13, 2026, between 05:49 and 06:29 UTC, customers experienced failures when attempting to log in, sign up, reset passwords, and complete multi-factor authentication flows across Atlassian cloud products. Approximately 90% of authentication requests failed during the peak impact window, affecting users in the US East and EU regions. The incident was mitigated within 40 minutes through manual intervention, and full service was restored by 06:29 UTC. ### **IMPACT** * **Duration**: ~40 minutes \(05:49–06:29 UTC, April 13, 2026\) * **Affected regions**: US East and EU \(authentication infrastructure serves EU traffic from US East, with traffic primarily from EU at this time of day\). * **Affected products**: All Atlassian cloud products requiring authentication, including Jira, Confluence, Jira Service Management, and Trello. * **Customer experience**: Users attempting to log in, sign up, reset passwords, or complete MFA flows received errors. Users already logged in with active sessions were unaffected. ### **ROOT CAUSE** This incident had several contributing factors that combined to produce a failure that the system could not recover from without manual intervention. **The primary cause** was a recently enabled change that caused our authentication infrastructure to retry requests to a downstream identity service when those requests were slow to respond. This retry behaviour was rolled out to 100% of traffic earlier the same day. Under normal conditions this would be benign, but it meant that any slowness in the downstream service was amplified. Since multiple upstream services were also independently retrying their own failed requests, the amplification compounded further into a retry storm. **The trigger** was a burst of legitimate user traffic. A pattern of many parallel link preview requests for a single user caused a concentrated load spike on a downstream identity service, pushing its response times above the retry threshold. On its own, this kind of spike had occurred many times before and always recovered. With the retry amplification now in effect, the spike instead created a runaway feedback loop: slow responses caused retries, retries increased load, increased load caused slower responses, preventing recovery. The incident was mitigated by manually scaling up the downstream identity service to provide sufficient capacity to absorb the amplified load. Once scaled, the service recovered immediately, bringing authentication error rates to zero within one minute. **REMEDIAL ACTIONS PLAN & NEXT STEPS** We are taking the following actions designed to prevent recurrence and improve our resilience: 1. **Immediate**: The retry-on-timeout change has been disabled. 2. **Load shedding and self-healing**: We are adding load shedding capabilities to our authentication services so that they can automatically shed excess load and self-recover during traffic spikes, without requiring action before automatic scaling starts. 3. **Reducing request fan-out**: We are reviewing patterns where a single user action can generate many parallel downstream requests, and will introduce methods where possible to reduce the amplification potential. We apologize to customers whose services were interrupted by this incident and we are taking immediate steps to improve the platform’s reliability. Thanks, Atlassian Customer Support
Read the full incident report →
- Detected by Pingoru
- Apr 13, 2026, 07:29 AM UTC
- Resolved
- Apr 13, 2026, 10:17 AM UTC
- Duration
- 2h 47m
Affected: WebsiteAPIGit via SSHAuthentication and user managementGit via HTTPSWebhooksSource downloadsPipelinesGit LFSEmail deliveryPurchasing & LicensingSignup
Timeline · 4 updates
-
monitoring Apr 13, 2026, 07:29 AM UTC
Our team is aware that some users were unable to log in to Atlassian products with their Atlassian accounts. While we believe this issue is now resolved, we are continuing to monitor all products and services for any ongoing impact. Our team is investigating with urgency, and we will provide an update within 1 hour.
-
monitoring Apr 13, 2026, 08:33 AM UTC
Atlassian account login services are now operating as expected, and we are not observing new errors. We continue to closely monitor our systems to ensure they remain stable. We will provide another update within 60 minutes or sooner if we detect a change in status.
-
resolved Apr 13, 2026, 10:17 AM UTC
On April 13, 2026, between 05:49 a.m. and 06:25 a.m. UTC, some users were unable to log in to Atlassian products with their Atlassian accounts. The underlying issue has been addressed, and authentication services have remained stable with no new impact observed.
-
postmortem Apr 21, 2026, 02:32 AM UTC
### Summary On April 13, 2026, between 05:49 and 06:29 UTC, customers experienced failures when attempting to log in, sign up, reset passwords, and complete multi-factor authentication flows across Atlassian cloud products. Approximately 90% of authentication requests failed during the peak impact window, affecting users in the US East and EU regions. The incident was mitigated within 40 minutes through manual intervention, and full service was restored by 06:29 UTC. ### **IMPACT** * **Duration**: ~40 minutes \(05:49–06:29 UTC, April 13, 2026\) * **Affected regions**: US East and EU \(authentication infrastructure serves EU traffic from US East, with traffic primarily from EU at this time of day\). * **Affected products**: All Atlassian cloud products requiring authentication, including Jira, Confluence, Jira Service Management, and Trello. * **Customer experience**: Users attempting to log in, sign up, reset passwords, or complete MFA flows received errors. Users already logged in with active sessions were unaffected. ### **ROOT CAUSE** This incident had several contributing factors that combined to produce a failure that the system could not recover from without manual intervention. **The primary cause** was a recently enabled change that caused our authentication infrastructure to retry requests to a downstream identity service when those requests were slow to respond. This retry behaviour was rolled out to 100% of traffic earlier the same day. Under normal conditions this would be benign, but it meant that any slowness in the downstream service was amplified. Since multiple upstream services were also independently retrying their own failed requests, the amplification compounded further into a retry storm. **The trigger** was a burst of legitimate user traffic. A pattern of many parallel link preview requests for a single user caused a concentrated load spike on a downstream identity service, pushing its response times above the retry threshold. On its own, this kind of spike had occurred many times before and always recovered. With the retry amplification now in effect, the spike instead created a runaway feedback loop: slow responses caused retries, retries increased load, increased load caused slower responses, preventing recovery. The incident was mitigated by manually scaling up the downstream identity service to provide sufficient capacity to absorb the amplified load. Once scaled, the service recovered immediately, bringing authentication error rates to zero within one minute. **REMEDIAL ACTIONS PLAN & NEXT STEPS** We are taking the following actions designed to prevent recurrence and improve our resilience: 1. **Immediate**: The retry-on-timeout change has been disabled. 2. **Load shedding and self-healing**: We are adding load shedding capabilities to our authentication services so that they can automatically shed excess load and self-recover during traffic spikes, without requiring action before automatic scaling starts. 3. **Reducing request fan-out**: We are reviewing patterns where a single user action can generate many parallel downstream requests, and will introduce methods where possible to reduce the amplification potential. We apologize to customers whose services were interrupted by this incident and we are taking immediate steps to improve the platform’s reliability. Thanks, Atlassian Customer Support
Read the full incident report →
- Detected by Pingoru
- Apr 13, 2026, 07:29 AM UTC
- Resolved
- Apr 13, 2026, 10:17 AM UTC
- Duration
- 2h 47m
Affected: Jira Service Management WebService PortalOpsgenie Incident FlowOpsgenie Alert FlowOpsgenie Incident FlowOpsgenie Alert FlowJira Service Management Email RequestsAuthentication and User ManagementPurchasing & LicensingSignupAutomation for JiraAssist
Timeline · 4 updates
-
monitoring Apr 13, 2026, 07:29 AM UTC
Our team is aware that some users were unable to log in to Atlassian products with their Atlassian accounts. While we believe this issue is now resolved, we are continuing to monitor all products and services for any ongoing impact. Our team is investigating with urgency, and we will provide an update within 1 hour.
-
monitoring Apr 13, 2026, 08:33 AM UTC
Atlassian account login services are now operating as expected, and we are not observing new errors. We continue to closely monitor our systems to ensure they remain stable. We will provide another update within 60 minutes or sooner if we detect a change in status.
-
resolved Apr 13, 2026, 10:17 AM UTC
On April 13, 2026, between 05:49 a.m. and 06:25 a.m. UTC, some users were unable to log in to Atlassian products with their Atlassian accounts. The underlying issue has been addressed, and authentication services have remained stable with no new impact observed.
-
postmortem Apr 21, 2026, 02:33 AM UTC
### Summary On April 13, 2026, between 05:49 and 06:29 UTC, customers experienced failures when attempting to log in, sign up, reset passwords, and complete multi-factor authentication flows across Atlassian cloud products. Approximately 90% of authentication requests failed during the peak impact window, affecting users in the US East and EU regions. The incident was mitigated within 40 minutes through manual intervention, and full service was restored by 06:29 UTC. ### **IMPACT** * **Duration**: ~40 minutes \(05:49–06:29 UTC, April 13, 2026\) * **Affected regions**: US East and EU \(authentication infrastructure serves EU traffic from US East, with traffic primarily from EU at this time of day\). * **Affected products**: All Atlassian cloud products requiring authentication, including Jira, Confluence, Jira Service Management, and Trello. * **Customer experience**: Users attempting to log in, sign up, reset passwords, or complete MFA flows received errors. Users already logged in with active sessions were unaffected. ### **ROOT CAUSE** This incident had several contributing factors that combined to produce a failure that the system could not recover from without manual intervention. **The primary cause** was a recently enabled change that caused our authentication infrastructure to retry requests to a downstream identity service when those requests were slow to respond. This retry behaviour was rolled out to 100% of traffic earlier the same day. Under normal conditions this would be benign, but it meant that any slowness in the downstream service was amplified. Since multiple upstream services were also independently retrying their own failed requests, the amplification compounded further into a retry storm. **The trigger** was a burst of legitimate user traffic. A pattern of many parallel link preview requests for a single user caused a concentrated load spike on a downstream identity service, pushing its response times above the retry threshold. On its own, this kind of spike had occurred many times before and always recovered. With the retry amplification now in effect, the spike instead created a runaway feedback loop: slow responses caused retries, retries increased load, increased load caused slower responses, preventing recovery. The incident was mitigated by manually scaling up the downstream identity service to provide sufficient capacity to absorb the amplified load. Once scaled, the service recovered immediately, bringing authentication error rates to zero within one minute. **REMEDIAL ACTIONS PLAN & NEXT STEPS** We are taking the following actions designed to prevent recurrence and improve our resilience: 1. **Immediate**: The retry-on-timeout change has been disabled. 2. **Load shedding and self-healing**: We are adding load shedding capabilities to our authentication services so that they can automatically shed excess load and self-recover during traffic spikes, without requiring action before automatic scaling starts. 3. **Reducing request fan-out**: We are reviewing patterns where a single user action can generate many parallel downstream requests, and will introduce methods where possible to reduce the amplification potential. We apologize to customers whose services were interrupted by this incident and we are taking immediate steps to improve the platform’s reliability. Thanks, Atlassian Customer Support
Read the full incident report →
- Detected by Pingoru
- Apr 13, 2026, 07:29 AM UTC
- Resolved
- Apr 13, 2026, 10:17 AM UTC
- Duration
- 2h 47m
Affected: Jira Service Management WebService PortalOpsgenie Incident FlowOpsgenie Alert FlowOpsgenie Incident FlowOpsgenie Alert FlowJira Service Management Email RequestsAuthentication and User ManagementPurchasing & LicensingSignupAutomation for JiraAssist
Timeline · 4 updates
-
monitoring Apr 13, 2026, 07:29 AM UTC
Our team is aware that some users were unable to log in to Atlassian products with their Atlassian accounts. While we believe this issue is now resolved, we are continuing to monitor all products and services for any ongoing impact. Our team is investigating with urgency, and we will provide an update within 1 hour.
-
monitoring Apr 13, 2026, 08:33 AM UTC
Atlassian account login services are now operating as expected, and we are not observing new errors. We continue to closely monitor our systems to ensure they remain stable. We will provide another update within 60 minutes or sooner if we detect a change in status.
-
resolved Apr 13, 2026, 10:17 AM UTC
On April 13, 2026, between 05:49 a.m. and 06:25 a.m. UTC, some users were unable to log in to Atlassian products with their Atlassian accounts. The underlying issue has been addressed, and authentication services have remained stable with no new impact observed.
-
postmortem Apr 21, 2026, 02:33 AM UTC
### Summary On April 13, 2026, between 05:49 and 06:29 UTC, customers experienced failures when attempting to log in, sign up, reset passwords, and complete multi-factor authentication flows across Atlassian cloud products. Approximately 90% of authentication requests failed during the peak impact window, affecting users in the US East and EU regions. The incident was mitigated within 40 minutes through manual intervention, and full service was restored by 06:29 UTC. ### **IMPACT** * **Duration**: ~40 minutes \(05:49–06:29 UTC, April 13, 2026\) * **Affected regions**: US East and EU \(authentication infrastructure serves EU traffic from US East, with traffic primarily from EU at this time of day\). * **Affected products**: All Atlassian cloud products requiring authentication, including Jira, Confluence, Jira Service Management, and Trello. * **Customer experience**: Users attempting to log in, sign up, reset passwords, or complete MFA flows received errors. Users already logged in with active sessions were unaffected. ### **ROOT CAUSE** This incident had several contributing factors that combined to produce a failure that the system could not recover from without manual intervention. **The primary cause** was a recently enabled change that caused our authentication infrastructure to retry requests to a downstream identity service when those requests were slow to respond. This retry behaviour was rolled out to 100% of traffic earlier the same day. Under normal conditions this would be benign, but it meant that any slowness in the downstream service was amplified. Since multiple upstream services were also independently retrying their own failed requests, the amplification compounded further into a retry storm. **The trigger** was a burst of legitimate user traffic. A pattern of many parallel link preview requests for a single user caused a concentrated load spike on a downstream identity service, pushing its response times above the retry threshold. On its own, this kind of spike had occurred many times before and always recovered. With the retry amplification now in effect, the spike instead created a runaway feedback loop: slow responses caused retries, retries increased load, increased load caused slower responses, preventing recovery. The incident was mitigated by manually scaling up the downstream identity service to provide sufficient capacity to absorb the amplified load. Once scaled, the service recovered immediately, bringing authentication error rates to zero within one minute. **REMEDIAL ACTIONS PLAN & NEXT STEPS** We are taking the following actions designed to prevent recurrence and improve our resilience: 1. **Immediate**: The retry-on-timeout change has been disabled. 2. **Load shedding and self-healing**: We are adding load shedding capabilities to our authentication services so that they can automatically shed excess load and self-recover during traffic spikes, without requiring action before automatic scaling starts. 3. **Reducing request fan-out**: We are reviewing patterns where a single user action can generate many parallel downstream requests, and will introduce methods where possible to reduce the amplification potential. We apologize to customers whose services were interrupted by this incident and we are taking immediate steps to improve the platform’s reliability. Thanks, Atlassian Customer Support
Read the full incident report →