- Detected by Pingoru
- Apr 28, 2026, 01:45 PM UTC
- Resolved
- Apr 28, 2026, 07:33 PM UTC
- Duration
- 5h 48m
Affected: Gradient AI
Timeline · 4 updates
-
investigating Apr 28, 2026, 01:45 PM UTC
Serverless inference customers are experiencing elevated 5xx errors, including “context canceled” responses. This may result in intermittent request failures. Our team is actively investigating and will provide updates as more information becomes available.
-
identified Apr 28, 2026, 03:59 PM UTC
We have identified an issue affecting our service and are currently working to implement a fix. Our team is actively investigating and taking the necessary steps to restore normal operations as quickly as possible. We appreciate your patience and will provide updates as soon as more information becomes available.
-
monitoring Apr 28, 2026, 07:00 PM UTC
Service for Serverless Inference has been restored. We’ve implemented tighter rate limits to help prevent recurrence and are closely monitoring system performance. Some users may still experience intermittent latency as we complete final stabilization efforts. Our team remains actively engaged to ensure full recovery. We appreciate your patience and will provide further updates as needed.
-
resolved Apr 28, 2026, 07:33 PM UTC
All services are operating normally. We will continue to monitor the system to ensure ongoing reliability. Thank you for your patience while we worked to resolve this issue.
Read the full incident report →
- Detected by Pingoru
- Apr 27, 2026, 10:38 AM UTC
- Resolved
- Apr 27, 2026, 11:37 AM UTC
- Duration
- 59m
Affected: Gradient AI
Timeline · 3 updates
-
investigating Apr 27, 2026, 10:38 AM UTC
We are currently investigating an issue impacting DigitalOcean Serverless Inference for all Anthropic models. Customers may experience an elevated rate of failed requests, including HTTP 429 responses. Our engineering team is actively working to identify the cause and mitigate the impact. We apologize for the inconvenience and will share another update as soon as more information is available.
-
monitoring Apr 27, 2026, 11:07 AM UTC
We identified the cause of intermittent HTTP 429 responses affecting some customers using Anthropic models on DigitalOcean Serverless Inference and applied a mitigation. Service has recovered, and we are monitoring stability.
-
resolved Apr 27, 2026, 11:37 AM UTC
The issue is resolved, and service is operating normally.
Read the full incident report →
- Detected by Pingoru
- Apr 23, 2026, 10:26 PM UTC
- Resolved
- Apr 23, 2026, 11:51 PM UTC
- Duration
- 1h 25m
Affected: Gradient AI
Timeline · 2 updates
-
investigating Apr 23, 2026, 10:26 PM UTC
As of 21:53 UTC, our Engineering team is investigating reports of increased internal errors for hosted models in the ATL1 region, impacting Serverless Inference. At this point, users with models hosted in ATL1 may experience intermittent errors when using Serverless Inference. We apologize for the inconvenience and will share an update once we have more information.
-
resolved Apr 23, 2026, 11:51 PM UTC
This incident has been resolved.
Read the full incident report →
- Detected by Pingoru
- Apr 23, 2026, 08:08 AM UTC
- Resolved
- Apr 23, 2026, 10:09 AM UTC
- Duration
- 2h 1m
Affected: Global
Timeline · 4 updates
-
investigating Apr 23, 2026, 08:08 AM UTC
Our Engineering team is currently investigating reports of build failures on App Platform. Users may experience errors when attempting to build their applications, resulting in failed deployments. Our Engineering team is working to fix the issue and will share an update once we have more information. We apologize for the inconvenience this issue may be causing and appreciate your patience as we work to resolve it.
-
investigating Apr 23, 2026, 08:31 AM UTC
Our Engineering team is currently investigating reports of build failures on App Platform. During this time, some users may encounter errors while building their applications, which may result in failed deployments. In addition, we are observing an issue where Kubernetes (DOKS) nodes are being marked as unhealthy by load balancers, which may impact traffic routing for affected services. Our Engineering team is actively working to resolve these issues and will share an update as soon as more information becomes available. We apologize for the inconvenience this may be causing and appreciate your patience.
-
monitoring Apr 23, 2026, 09:52 AM UTC
Our Engineering team has implemented a fix to address the issue causing in App Platform deployments and Kubernetes (DOKS) nodes. We are actively monitoring the situation to ensure overall stability. Users may already notice improvements while deploying apps and DOKS nodes. We appreciate your patience throughout the process and will provide a further update once the issue is fully confirmed to be resolved.
-
resolved Apr 23, 2026, 10:09 AM UTC
Our Engineering team has confirmed that the issue impacting App Platform deployments and Kubernetes (DOKS) nodes has been fully resolved at 09:22 UTC. Users may already notice improvements while deploying apps and DOKS nodes. All App Platform deployments are now succeeding as expected. Customers who previously encountered build failures should now be able to deploy their applications without further issues. If you continue to experience any problems, please open a ticket with our support team. Thank you for your patience, and we apologize for any inconvenience.
Read the full incident report →
- Detected by Pingoru
- Apr 23, 2026, 07:01 AM UTC
- Resolved
- Apr 23, 2026, 08:46 AM UTC
- Duration
- 1h 44m
Affected: Global
Timeline · 4 updates
-
investigating Apr 23, 2026, 07:01 AM UTC
Our Engineering team is currently investigating an issue impacting Managed Kubernetes UI across all regions. During this time, some users may experience the Kubernetes UI page not loading in DigitalOcean Onboarding. We apologize for the inconvenience and will share more information as soon as it's available.
-
investigating Apr 23, 2026, 07:14 AM UTC
Our Engineering team is currently investigating an issue impacting the Create Managed Kubernetes UI across all regions. During this time, users with a Member role may experience the Kubernetes UI page not loading in the cloud console. As a workaround, the DigitalOcean API and doctl (CLI) continue to function normally, and you can use them to manage your Kubernetes resources in the meantime. We apologize for the inconvenience and will share more information as soon as it becomes available.
-
investigating Apr 23, 2026, 07:52 AM UTC
Our Engineering team is currently investigating an issue impacting the Managed Kubernetes UI across all regions. During this time, users with a Member role may experience the Kubernetes UI page not loading in the cloud console. As a workaround, the DigitalOcean API and doctl (CLI) continue to function normally, and you can use them to manage your Kubernetes resources in the meantime. We apologize for the inconvenience and will share more information as soon as it becomes available
-
resolved Apr 23, 2026, 08:46 AM UTC
Our Engineering team has confirmed full resolution of the issue with Cloud UI for Managed Kubernetes at 09:19 UTC. All the services should be functioning as expected. If you continue to experience problems, please open a ticket with our support team. Thank you for your patience throughout this incident!
Read the full incident report →
- Detected by Pingoru
- Apr 22, 2026, 11:06 AM UTC
- Resolved
- Apr 22, 2026, 11:47 AM UTC
- Duration
- 40m
Affected: APICloud Control Panel
Timeline · 3 updates
-
investigating Apr 22, 2026, 11:06 AM UTC
Our Engineering team is currently investigating an issue affecting the Cloud Control Panel and API. During this time, API requests to create, destroy, or trigger events on droplets may not succeed. Additionally, the Droplets page in the Cloud Control Panel may not load properly, and users could experience issues while reviewing the droplet listing. We are actively looking into the root cause and will provide updates as soon as more information becomes available. We apologize for the inconvenience and appreciate your patience.
-
monitoring Apr 22, 2026, 11:25 AM UTC
Our Engineering team has implemented a fix for the issue affecting the DigitalOcean API and the Cloud Control Panel. We are now actively monitoring the system to ensure full stability and will provide a final update once the issue is completely resolved. We sincerely appreciate your patience and understanding as we work to restore normal service.
-
resolved Apr 22, 2026, 11:47 AM UTC
Our Engineering team has resolved the issue that was impacting the DigitalOcean API and the Cloud Control Panel. The system is now operating normally. During the incident, users may have experienced difficulties accessing droplets, viewing the Droplets page, or performing CRUD (Create, Read, Update, Delete) operations via the API. We have verified that the issue has been fully resolved. We apologize for the inconvenience this may have caused. If you continue to experience any issues, please submit a support ticket through the Cloud Control Panel so our team can assist you further.
Read the full incident report →
- Detected by Pingoru
- Apr 17, 2026, 09:22 PM UTC
- Resolved
- Apr 18, 2026, 02:10 PM UTC
- Duration
- 16h 47m
Affected: DNS
Timeline · 3 updates
-
identified Apr 17, 2026, 09:22 PM UTC
Our Engineering team is aware of a widespread, external issue affecting .co top-level domain (TLD). While this incident originates outside of DigitalOcean's infrastructure, you may experience errors when querying a .co domain, regardless of the DNS resolver being used. Our Engineers are actively deploying temporary backend mitigations to help minimize the impact on our customers. We will continue to monitor the situation closely and post updates as more information becomes available.
-
monitoring Apr 17, 2026, 10:53 PM UTC
Our Engineering team has implemented a fix to resolve the issue that was affecting the DNS resolution of .co top-level domain (TLD). During this time, users who are using DigitalOcean DNS resolvers in their resources should no longer experience issues related to the DNS resolution. Our Engineers are currently monitoring the situation. We will post an update as soon as the issue is fully resolved.
-
resolved Apr 18, 2026, 02:10 PM UTC
Our Engineering team has confirmed that the issue affecting the DNS resolution of .co top-level domain (TLD) has been resolved. The DNS resolution for .co domains is now working as expected. We apologize for the inconvenience. If you continue to face any issues, please open a support ticket from within your account.
Read the full incident report →
- Detected by Pingoru
- Apr 17, 2026, 06:27 PM UTC
- Resolved
- Apr 17, 2026, 06:27 PM UTC
- Duration
- —
Affected: NYC3
Timeline · 1 update
-
resolved Apr 17, 2026, 06:27 PM UTC
From 17:22 to 17:46 UTC, our Engineering team observed an issue impacting Spaces availability in the NYC3 region. During this time, customers may have encountered 500 errors and degraded performance while accessing Spaces buckets. The issue has now been fully resolved. If you continue to experience problems, please open a ticket with our support team from within your Cloud Control Panel. We apologize for any inconvenience caused.
Read the full incident report →
- Detected by Pingoru
- Apr 15, 2026, 01:07 AM UTC
- Resolved
- Apr 15, 2026, 02:56 AM UTC
- Duration
- 1h 48m
Affected: Gradient AI
Timeline · 3 updates
-
investigating Apr 15, 2026, 01:07 AM UTC
Our Engineering team is investigating an issue causing elevated error rates due to service instability and terminating open connections which causes some 500s. Some requests may fail while we work to resolve it. We apologize for the inconvenience and will share an update once we have more information.
-
monitoring Apr 15, 2026, 02:24 AM UTC
Our Engineering team has implemented a fix for the issue causing elevated error rates due to service instability. We are currently monitoring the situation to ensure stability and confirm that error rates, including HTTP 500 responses, have returned to normal levels. We will provide a further update once we confirm the issue is fully resolved.
-
resolved Apr 15, 2026, 02:56 AM UTC
From 23:20 UTC to 02:00 UTC, users may have experienced elevated error rates due to service instability, which resulted in intermittent HTTP 500 errors and terminated connections. Our Engineering team has confirmed full resolution of the issue, and all systems are now operating normally. If you continue to experience any issues, please open a ticket with our support team. We apologize for any inconvenience caused.
Read the full incident report →
- Detected by Pingoru
- Apr 14, 2026, 12:53 PM UTC
- Resolved
- Apr 14, 2026, 03:19 PM UTC
- Duration
- 2h 26m
Affected: GlobalGlobal
Timeline · 3 updates
-
investigating Apr 14, 2026, 12:53 PM UTC
Our engineering team is investigating an issue impacting resize operations for Managed Databases. During this time, users may experience error when attempting to resize Managed Database via Cloud Control Panel and API in all regions. We apologize for the inconvenience and will share an update once we have more information.
-
monitoring Apr 14, 2026, 02:30 PM UTC
Our Engineering team has taken action to mitigate the issue with resize operations for Managed Databases and implemented a fix . We are monitoring the situation and will post an update as soon as we confirm that the issue is fully resolved.
-
resolved Apr 14, 2026, 03:19 PM UTC
Our Engineering team has resolved the issue with resize operations for Managed Databases and should now be operating normally. If you continue to experience problems, please open a ticket with our support team from within your Cloud Control Panel. We apologize for any inconvenience.
Read the full incident report →
- Detected by Pingoru
- Apr 10, 2026, 08:32 PM UTC
- Resolved
- Apr 10, 2026, 09:06 PM UTC
- Duration
- 33m
Affected: APIGlobalGlobalCloud Control Panel
Timeline · 2 updates
-
monitoring Apr 10, 2026, 08:32 PM UTC
Subject: Droplet Availability in All Regions Our Engineering team has identifed an issue with Droplet creates in all regions. A root cause has been found, a fix has been put in place and we are currently monitoring the situation to ensure full resolution. Users should be able to create new Droplets at this time. We will continue to monitor and we will post an update as soon as it is fully resolved. We apologize for the inconvenience.
-
resolved Apr 10, 2026, 09:06 PM UTC
Our Engineering team has confirmed full resolution of the issue with creating Droplets in all regions. Users should be able to create Droplets without issue. We apologize for the inconvenience. If you continue to face any issues, please open a support ticket from within your account.
Read the full incident report →
- Detected by Pingoru
- Apr 07, 2026, 05:49 PM UTC
- Resolved
- Apr 07, 2026, 05:49 PM UTC
- Duration
- —
Affected: APICloud Control Panel
Timeline · 1 update
-
resolved Apr 07, 2026, 05:49 PM UTC
Our Engineering team has resolved the control plane disruption that occurred from 17:06 to 17:18 UTC. During this time, users may have experienced intermittent issues with managing their resources through the Cloud Control Panel or DigitalOcean API. The root cause of the disruption was identified and addressed, and all services are now operating normally. If you continue to experience any problems, please open a ticket with our Support team. We apologize for any inconvenience this may have caused.
Read the full incident report →
- Detected by Pingoru
- Apr 07, 2026, 12:49 PM UTC
- Resolved
- Apr 07, 2026, 03:50 PM UTC
- Duration
- 3h 1m
Affected: Gradient AI
Timeline · 3 updates
-
investigating Apr 07, 2026, 12:49 PM UTC
Serverless inference for alibaba-qwen3-32b (Qwen 3 32B) in tor1 is experiencing high error rates starting at 10:46 UTC.
-
identified Apr 07, 2026, 12:55 PM UTC
We are currently investigating reports of elevated latency affecting requests to this model when using Serverless Inference and Agents. Earlier observations indicated increased error rates for the open-source Qwen 3 32B model. The Ray dashboard also showed multiple workers in a pending state, suggesting capacity constraints. Our analysis determined that the model was experiencing higher-than-expected request volume without sufficient resources to scale accordingly. To address this, the node pool size has been increased to improve available capacity. However, there are still insufficient nodes to fully support the desired number of model replicas. Following the node pool expansion, a new pod-related error has been identified. Our Engineering team is actively working to resolve this issue and restore full service performance.
-
resolved Apr 07, 2026, 03:50 PM UTC
Service has been fully restored, and the model is now operating normally. We have implemented improvements to enhance stability and reduce the likelihood of similar issues in the future.
Read the full incident report →
- Detected by Pingoru
- Apr 06, 2026, 12:28 PM UTC
- Resolved
- Apr 06, 2026, 06:02 PM UTC
- Duration
- 5h 34m
Affected: Gradient AI
Timeline · 3 updates
-
investigating Apr 06, 2026, 12:28 PM UTC
Our Engineering team is investigating an issue with Serverless inference. At this time, users may experience high error rates for open source models (llama 3.3 70b). We apologize for the inconvenience and will share an update once we have more information.
-
monitoring Apr 06, 2026, 03:15 PM UTC
A fix has been implemented and we are monitoring the results.
-
resolved Apr 06, 2026, 06:02 PM UTC
This incident has been resolved.
Read the full incident report →
- Detected by Pingoru
- Mar 20, 2026, 10:32 AM UTC
- Resolved
- Mar 20, 2026, 12:01 PM UTC
- Duration
- 1h 28m
Affected: Frankfurt
Timeline · 3 updates
-
investigating Mar 20, 2026, 10:32 AM UTC
Our engineers are currently investigating an issue impacting new App Platform deployments using Dedicated Egress IP in FRA1 region. During this time, some users may experience delay when creating new App Platform apps or deploying existing apps. Existing apps are not affected and should continue to function normally. We apologize for any inconvenience, and we'll share more information as it becomes available.
-
monitoring Mar 20, 2026, 11:14 AM UTC
Our Engineering team has deployed a fix to resolve the issue impacting new App Platform deployments using Dedicated Egress IP in FRA1 region. We are actively monitoring the situation to ensure stability and will provide an update once the incident has been fully resolved. Thank you for your patience and we apologize for the inconvenience.
-
resolved Mar 20, 2026, 12:01 PM UTC
The issue impacting delays in App Platform deployments has been confirmed to be resolved. Between approximately 00:08am UTC & 11:46am UTC, users may have noticed delays while creating or updating apps, or may have encountered failed deployments. For failed deployments, please trigger a redeploy, which should successfully resolve the issue. We confirmed that the service is functioning as expected. Once again, we sincerely apologize for the inconvenience caused and appreciate your understanding. However, if you continue to experience any issues, please don't hesitate to raise a support ticket for further investigation. We'll be happy to assist you.
Read the full incident report →
- Detected by Pingoru
- Mar 20, 2026, 08:50 AM UTC
- Resolved
- Mar 20, 2026, 02:14 PM UTC
- Duration
- 5h 24m
Affected: Gradient AI
Timeline · 5 updates
-
investigating Mar 20, 2026, 08:50 AM UTC
We are currently investigating issue affecting the accessibility of agents and services on the Gradient AI Platform. Users may experience failures or unresponsiveness when attempting to use these features. Our engineering team is actively working to identify the root cause and restore full functionality. We apologize for the inconvenience and will share an update once we have more information.
-
identified Mar 20, 2026, 09:51 AM UTC
We’ve identified the issue and are currently working on restoring the services. We’ll continue to provide updates as progress is made.
-
identified Mar 20, 2026, 11:05 AM UTC
We've identified the issue and are actively working to restore the affected services. We're making steady progress and closely monitoring the situation. Further updates will be shared as they become available.
-
monitoring Mar 20, 2026, 02:04 PM UTC
A fix has been implemented and services have been restored. We are continuing to monitor the system to ensure stability. We will provide further updates if needed.
-
resolved Mar 20, 2026, 02:14 PM UTC
Our Engineering team has implemented a fix, the issues impacting Gradient AI Platform have been resolved. All agents are back up and healthy. Service has been fully restored.
Read the full incident report →
- Detected by Pingoru
- Mar 17, 2026, 03:00 PM UTC
- Resolved
- Mar 17, 2026, 07:49 PM UTC
- Duration
- 4h 48m
Affected: Gradient AI
Timeline · 2 updates
-
investigating Mar 17, 2026, 03:00 PM UTC
Our Engineering team is investigating reports of Gradient AI model availability issues impacting multiple models. Users may experience issues with models availability, including Llama3.1-8b and Qwen3-32b, as well as embedding models such as GTE Large (v1.5), All-MiniLM-L6-v2, Multi-QA-mpnet-base-dot-v1, and Qwen3 Embedding 0.6B. Additionally, Guardrails are not available, affecting associated agents, and users attempting to run inference on the Llama3.3-70b model will see degraded performance. We apologize for the inconvenience and will share an update once we have more information.
-
resolved Mar 17, 2026, 07:49 PM UTC
Our Engineering team has implemented a fix, the issues impacting model availability and performance have been resolved. All models, including those previously degraded, are back up and healthy. Service has been fully restored.
Read the full incident report →
- Detected by Pingoru
- Mar 15, 2026, 02:55 AM UTC
- Resolved
- Mar 15, 2026, 03:31 AM UTC
- Duration
- 36m
Affected: Gradient AI
Timeline · 2 updates
-
investigating Mar 15, 2026, 02:55 AM UTC
Our Engineering team is investigating an issue related to all Gradient AI agents and serverless inference that require BYOK Anthropic modles. Impacted users may experience degraded performance. We will provide an update as soon as possible
-
resolved Mar 15, 2026, 03:31 AM UTC
The issue is now resolved, all Anthropic BYOK models in Gradient AI should work normally. Contact support if issues persist.
Read the full incident report →
- Detected by Pingoru
- Mar 13, 2026, 10:01 PM UTC
- Resolved
- Mar 14, 2026, 01:47 AM UTC
- Duration
- 3h 46m
Affected: Global
Timeline · 4 updates
-
monitoring Mar 13, 2026, 09:30 PM UTC
Starting at 20:40 UTC, users may have seen delays with deploying new apps on App Platform. At this time, our Engineering team is seeing signs of recovery, and users should be able to deploy new apps without issue. We're currently monitoring the situation to ensure full recovery. We apologize for the inconvenience. We'll post an update once the issue has been confirmed to be resolved.
-
identified Mar 13, 2026, 10:01 PM UTC
Our Engineering team is starting to see delays once again with new App Platform deployments. During this time, users may still experience delays with deploying new apps. We're working with our upstream provider to resolve the issue. We again apologize for the inconvenience. We will post further updates once we have more information.
-
monitoring Mar 13, 2026, 11:39 PM UTC
After working with our upstream provider, our Engineering team has implemented a fix to resolve the issue that was causing delays in the deployment of new apps, and they are currently monitoring the situation. During this time, users should no longer experience issues with creating new apps and all the stalled creation events should provision completely. We will post an update as soon as the issue is fully resolved.
-
resolved Mar 14, 2026, 01:47 AM UTC
As of 23:00 UTC, our Engineering team has confirmed that the issue causing delays in App Platform deployments has been fully resolved. The fix implemented earlier has been successful, and we are no longer seeing any delays or errors with deployments. Users should now be able to deploy their apps successfully and without any issues. We apologize again for the inconvenience caused. However, if you continue to experience any issues, please don't hesitate to raise a support ticket for further investigation.
Read the full incident report →
- Detected by Pingoru
- Mar 13, 2026, 11:26 AM UTC
- Resolved
- Mar 13, 2026, 04:35 PM UTC
- Duration
- 5h 8m
Affected: Global
Timeline · 4 updates
-
investigating Mar 13, 2026, 11:26 AM UTC
Our Engineering team is investigating an issue impacting newly provisioned Managed Kubernetes nodes. During this time, new nodes may experience DNS timeouts, which could temporarily affect cluster services. We apologize for the inconvenience and will share an update once we have more information.
-
identified Mar 13, 2026, 12:32 PM UTC
Our Engineering team is investigating an issue impacting newly provisioned Managed Kubernetes nodes. During this time, Only customers who run a NAT Gateway in their VPC and a VPC-native clusters are affected and may experience DNS timeouts. We apologize for the inconvenience and will share an update once we have more information.
-
monitoring Mar 13, 2026, 01:55 PM UTC
Our Engineering team has implemented a fix to address the issue causing DNS timeouts for newly provisioned Managed Kubernetes nodes. Further investigation has confirmed that this issue primarily affected customers utilizing a NAT Gateway within their VPC and running a VPC-native cluster. We are actively monitoring the situation to ensure overall stability. We appreciate your patience and will provide a further update once the issue is fully confirmed to be resolved.
-
resolved Mar 13, 2026, 04:35 PM UTC
Our Engineering team has confirmed the resolution of the issue impacting DNS timeouts for newly provisioned Managed Kubernetes nodes. At this time all cluster services should now be functioning normally. If you continue to experience problems, please open a ticket with our support team. We apologize for any inconvenience.
Read the full incident report →
- Detected by Pingoru
- Mar 09, 2026, 07:23 PM UTC
- Resolved
- Mar 09, 2026, 07:23 PM UTC
- Duration
- —
Affected: Global
Timeline · 1 update
-
resolved Mar 09, 2026, 07:23 PM UTC
From 17:50 to 19:06 UTC, Our Engineering team observed an issue with mirrors.digitalocean.com. During this time, users may have experienced errors when trying to update packages on Debian and Ubuntu Images. Our team has fully resolved the issues as of 19:06 UTC. If you continue to experience problems, please open a ticket with our support team from within your Cloud Control Panel. We apologize for any inconvenience caused.
Read the full incident report →
- Detected by Pingoru
- Mar 06, 2026, 09:22 PM UTC
- Resolved
- Mar 06, 2026, 09:22 PM UTC
- Duration
- —
Affected: Global
Timeline · 1 update
-
resolved Mar 06, 2026, 09:22 PM UTC
Our Engineering team identified an issue affecting the App Platform. During the incident, users may have experienced HTTP 522 (Connection Timed Out) errors when accessing their apps. The issue seems to be resolved now. We apologize for the inconvenience caused. If you continue to experience any related errors, please contact our Support team by opening a ticket at https://www.digitalocean.com/support/contact/.
Read the full incident report →
- Detected by Pingoru
- Mar 05, 2026, 12:23 AM UTC
- Resolved
- Mar 05, 2026, 01:52 AM UTC
- Duration
- 1h 28m
Affected: Global
Timeline · 3 updates
-
investigating Mar 05, 2026, 12:23 AM UTC
Our Engineering team is investigating an issue affecting Internal Load Balancers. Customers may experience connectivity loss between Internal Load Balancers and their associated target droplets. We apologize for the inconvenience and will share an update as soon as more information becomes available.
-
monitoring Mar 05, 2026, 01:19 AM UTC
Our Engineering team has implemented mitigation measures to address the connectivity issues affecting Internal Load Balancers and their associated target droplets. We are actively monitoring the situation to ensure stability and to prevent any recurrence. We will provide a further update once we confirm the issue is fully resolved.
-
resolved Mar 05, 2026, 01:52 AM UTC
From 19:57 UTC to 01:03 UTC, customers may have experienced connectivity issues between Internal Load Balancers and their associated target droplets, which could have resulted in service disruption or traffic routing failures. Our Engineering team has confirmed full resolution of the issue, and Internal Load Balancers should now be functioning normally. If you continue to experience any problems, please open a ticket with our Support team. We apologize for any inconvenience caused.
Read the full incident report →
- Detected by Pingoru
- Feb 26, 2026, 04:00 PM UTC
- Resolved
- Feb 26, 2026, 10:23 PM UTC
- Duration
- 6h 23m
Affected: Gradient AI
Timeline · 3 updates
-
investigating Feb 26, 2026, 07:38 PM UTC
We are currently investigating an issue affecting the Llama 3.3-70B model. Symptoms: Users may encounter intermittent errors when making serverless inference requests via APIs and Agents. Current Status: Our engineering team is actively investigating the issue to determine the root cause.
-
monitoring Feb 26, 2026, 09:52 PM UTC
Fix deployed. Monitoring resources related to the Llama 3.3-70B. Users should no longer experience intermittent errors when making serverless inference requests via APIs and Agents . Awaiting confirmation before closure.
-
resolved Feb 26, 2026, 10:23 PM UTC
Issue resolved. Cause: A few requests made to the Llama 3.3-70B model caused issues. Impact: Intermittent errors when interacting with the model through serverless inference and/or with agents created using this model. Contact support if issues persist.
Read the full incident report →
- Detected by Pingoru
- Feb 19, 2026, 05:34 PM UTC
- Resolved
- Feb 19, 2026, 06:42 PM UTC
- Duration
- 1h 7m
Affected: Cloud Control Panel
Timeline · 3 updates
-
investigating Feb 19, 2026, 05:34 PM UTC
Our team is currently investigating an issue impacting the visibility of the Control Panel. Users may notice unexpected behavior, such as being prompted for login credentials when accessing the console, being unable to select radio buttons for any plans during resize, columns appearing squished, etc. We apologize for the inconvenience and will share further updates as soon as more information becomes available.
-
monitoring Feb 19, 2026, 06:05 PM UTC
Our team has implemented a fix to address the issue affecting the visibility of the Cloud Panel. We are actively monitoring the situation to ensure overall stability. Users should no longer encounter abnormalities when accessing the console, resizing the Droplet, misalignments within the Cloud Panel, etc. We will provide a further update once the issue is fully confirmed to be resolved.
-
resolved Feb 19, 2026, 06:42 PM UTC
The issue impacting the visibility of Cloud Panel has been confirmed to be resolved. Between approximately 16:08 PM & 17:56 PM UTC, users may have noticed unusual behavior when accessing the console, performing resizing operations, or experiencing issues with the Cloud Panel visibility. Our team has taken necessary corrective measures to restore the service, and we can confirm that it is now functioning as expected. We sincerely apologize for the inconvenience caused and truly appreciate your understanding throughout this process. However, if you continue to experience any issues, please do not hesitate to raise a support ticket for further investigation. We’ll be happy to assist you.
Read the full incident report →