Balena Outage History

Balena is up right now

There were 7 Balena outages since February 4, 2026 totaling 604h 38m of downtime. Each is summarised below — incident details, duration, and resolution information.

Source: https://status.balena.io

Minor March 31, 2026

Elevated GIT/Application Builder Errors

Detected by Pingoru
Mar 31, 2026, 12:55 PM UTC
Resolved
Apr 21, 2026, 04:30 PM UTC
Duration
21d 3h
Affected: Application Builder
Timeline · 5 updates
  1. identified Mar 31, 2026, 12:55 PM UTC

    We're experiencing an elevated level of errors in our application builder infrastructure and are currently looking into the issue.

  2. monitoring Apr 08, 2026, 07:51 PM UTC

    A fix has been implemented and we are monitoring the results.

  3. monitoring Apr 08, 2026, 07:51 PM UTC

    We are continuing to monitor for any further issues.

  4. resolved Apr 21, 2026, 04:30 PM UTC

    This incident has been resolved.

  5. postmortem Apr 21, 2026, 05:06 PM UTC

    Starting around March 11, some cloud builds began failing intermittently with no such image errors. The failures were non-deterministic and affected all architectures. At peak, some users saw around 50% failure rates. We identified and fixed several bugs in the builder's image garbage collector that caused it to over-count freed disk space and run too aggressively, eventually deleting images that in-progress builds still needed. Fixes were deployed between March 19 and April 14, with build failure rates dropping to near-zero after the final deploy. We're continuing to monitor and working on additional safeguards to prevent the garbage collector from targeting images that active builds depend on.

Read the full incident report →

Major March 24, 2026

Elevated API Errors

Detected by Pingoru
Mar 24, 2026, 12:43 PM UTC
Resolved
Mar 25, 2026, 02:07 PM UTC
Duration
1d 1h
Affected: APIApplication BuilderApplication RegistryDashboardDelta Image DownloadsDevice URLsGitCloudlink (VPN)
Timeline · 4 updates
  1. identified Mar 24, 2026, 12:43 PM UTC

    We're experiencing an elevated level of API errors and are currently looking into the issue.

  2. monitoring Mar 24, 2026, 12:58 PM UTC

    A fix has been implemented and we are monitoring the results.

  3. monitoring Mar 24, 2026, 01:21 PM UTC

    We are continuing to monitor for any further issues.

  4. resolved Mar 25, 2026, 02:07 PM UTC

    We experienced degraded API performance due to an internal configuration change that unintentionally increased system load, resulting in slower response times and reduced request capacity. Our team identified the issue and rolled back the change to restore stability. A root cause fix has been implemented and deployed aswell. The system has now fully recovered, and services are operating normally.

Read the full incident report →

Minor March 23, 2026

Builder Degraded performance

Detected by Pingoru
Mar 23, 2026, 04:57 PM UTC
Resolved
Mar 25, 2026, 01:18 PM UTC
Duration
1d 20h
Affected: Application Builder
Timeline · 4 updates
  1. investigating Mar 23, 2026, 04:57 PM UTC

    We are seeing several builds intermitently failing with 404 errors - No such image during builds and are investigating.

  2. monitoring Mar 23, 2026, 06:55 PM UTC

    A fix has been implemented and we are monitoring the results.

  3. resolved Mar 25, 2026, 01:18 PM UTC

    This incident has been resolved.

  4. postmortem Mar 25, 2026, 03:38 PM UTC

    Between March 11 and March 25, some cloud builds experienced intermittent failures with "no such image" errors. The issue was non-deterministic and did not affect all builds. We've identified a likely contributing factor and deployed mitigations that have stabilized build reliability. We're continuing to investigate the underlying cause to prevent recurrence. If you experienced build failures during this window, re-running your build should succeed. We appreciate your patience while we worked through this, and we apologize for the disruption.

Read the full incident report →

Minor March 3, 2026

Elevated Dashboard Errors

Detected by Pingoru
Mar 03, 2026, 05:44 PM UTC
Resolved
Mar 03, 2026, 08:17 PM UTC
Duration
2h 33m
Affected: Dashboard
Timeline · 4 updates
  1. investigating Mar 03, 2026, 05:44 PM UTC

    We're experiencing an issue where the Dashboard may redirect to an unexpected page on initial load, which can prevent access to certain account and billing pages.

  2. monitoring Mar 03, 2026, 07:45 PM UTC

    A fix has been implemented and we are monitoring the results.

  3. resolved Mar 03, 2026, 08:17 PM UTC

    This incident has been resolved.

  4. postmortem Mar 09, 2026, 01:01 PM UTC

    We identified an issue in Dashboard v32.2.0, released on March 2, 2026, where opening the dashboard via a direct link to certain pages \(such as billing or other account management pages\) could result in being unexpectedly redirected to the fleets overview. This was caused by a race condition in our access control logic that made a routing decision before all authorization data had finished loading. The issue was resolved on March 3, 2026 with a fix that ensures the dashboard waits for all access information to be available before determining whether a user can view a page. We understand this was frustrating, particularly for users trying to manage billing or account settings via bookmarked or shared links. We apologize for the disruption and are adding test coverage for direct-link navigation to prevent similar regressions in the future.

Read the full incident report →

Minor February 26, 2026

Elevated Device SSH Errors

Detected by Pingoru
Feb 26, 2026, 09:13 PM UTC
Resolved
Feb 27, 2026, 03:39 AM UTC
Duration
6h 25m
Affected: SSH proxy
Timeline · 6 updates
  1. investigating Feb 26, 2026, 07:20 PM UTC

    We're experiencing an elevated level of device SSH errors and are currently looking into the issue.

  2. investigating Feb 26, 2026, 09:13 PM UTC

    We are continuing to investigate this issue. It only affects the direct SSH to device connection for example established via the balena-cli with balena device ssh . Web Terminal via the dashboard and standard tunnelling is not affected.

  3. identified Feb 27, 2026, 02:41 AM UTC

    The issue has been identified and a fix is being implemented.

  4. monitoring Feb 27, 2026, 03:02 AM UTC

    A fix has been implemented and we are monitoring the results.

  5. resolved Feb 27, 2026, 03:39 AM UTC

    This incident has been resolved.

  6. postmortem Mar 17, 2026, 11:47 AM UTC

    On February 26, 2026 at approximately 19:00 UTC, users began experiencing failures when connecting to their devices via balena device ssh. The command returned a generic "host error" with no further detail. The issue affected all CLI-based SSH access through the balena proxy. Web terminal access and balena device tunnel remained functional throughout the incident and were available as workarounds. Devices themselves were online and healthy; the failure occurred at the authentication stage between the proxy and the device. **Root cause** The balena proxy runs several internal processes. As a resource-saving measure, the service that handles web terminal connections had been disabled in the proxy pods dedicated to CLI SSH, since those pods do not serve web terminal sessions. What was not recognized at the time was that this service also carried the responsibility for loading SSH identity keys into the shared SSH agent used by the proxy. With it disabled, the proxy pods had no keys loaded and could not authenticate with devices, causing every SSH connection attempt to fail at the authentication stage. **Resolution** The service that handles web terminal connections was temporarily re-enabled in the SSH proxy pods while a permanent fix was being developed, restoring SSH key availability and resolving the authentication failures. A permanent fix has since been deployed. **Follow-up** We have decoupled the SSH key loading from the service that handles web terminal connections so that it is an independent initialization process in the proxy. We also added additional monitoring to ensure that SSH access through the proxy is adequately checked. We sincerely apologize for the impact this caused. We are committed to improving our processes to prevent issues like these.

Read the full incident report →

Major February 10, 2026

Elevated Delta Errors

Detected by Pingoru
Feb 10, 2026, 09:33 AM UTC
Resolved
Feb 11, 2026, 12:02 AM UTC
Duration
14h 28m
Affected: Delta Image Downloads
Timeline · 4 updates
  1. investigating Feb 10, 2026, 09:33 AM UTC

    Some delta generation requests are encountering errors and failing. We are currently investigating this issue.

  2. monitoring Feb 10, 2026, 10:25 AM UTC

    We have identified the potential cause and have rolled back the changes.

  3. resolved Feb 11, 2026, 12:02 AM UTC

    This incident has been resolved.

  4. postmortem Feb 11, 2026, 12:10 AM UTC

    v2 delta generation service experienced failures from ~21:15 UTC Feb 9 to ~10:00 UTC Feb 10, 2026, due to a missing configuration dependency during a logic change. **Impact:** * v2 delta generation requests failed to complete * No data loss or security impact **Root Cause:** Recent logic changes were deployed without the required accompanying configuration update, preventing the service from completing v2 delta requests. **Resolution:** The logic changes were rolled back, restoring the service to its previous stable state. **Follow-up Actions:** * Prepare and deploy the permanent fix We apologize for the disruption and any inconvenience this caused. We are committed to improving our processes to prevent similar issues in the future.

Read the full incident report →

Critical February 10, 2026

Elevated Cloudlink Errors

Detected by Pingoru
Feb 10, 2026, 04:05 AM UTC
Resolved
Feb 10, 2026, 07:56 AM UTC
Duration
3h 51m
Affected: Cloudlink (VPN)
Timeline · 5 updates
  1. investigating Feb 10, 2026, 04:05 AM UTC

    We're experiencing an elevated level of errors in our Cloudlink infrastructure and are currently looking into the issue.

  2. identified Feb 10, 2026, 06:39 AM UTC

    The issue has been identified and a fix is being implemented.

  3. monitoring Feb 10, 2026, 07:17 AM UTC

    A fix has been implemented and we are monitoring the results.

  4. resolved Feb 10, 2026, 07:56 AM UTC

    This incident has been resolved.

  5. postmortem Feb 10, 2026, 11:52 AM UTC

    Balena devices were unable to connect to Cloudlink on February 10, 2026, from approximately 02:26 GMT to 07:11 GMT due to an expired server certificate. Devices that were already connected to Cloudlink were unaffected unless the connection was terminated. **Root Cause:** The Cloudlink servers were using an expired certificate that was due for replacement. Consequently, incoming Cloudlink connections failed with a certificate verification error. **Resolution:** The certificate has been replaced, and Cloudlink servers were restarted to use the new certificate. Balena devices are expected to reconnect to Cloudlink within a few minutes after being disconnected due to the restart. **Follow-up Actions:** * Expand certificate expiry monitoring coverage to include all active certificates * Automate the certificate renewal process for Cloudlink We apologize for any disruption this caused and appreciate your patience as we continue improving our processes and operations.

Read the full incident report →

Looking to track Balena downtime and outages?

Pingoru polls Balena's status page every 5 minutes and alerts you the moment it reports an issue — before your customers do.

  • Real-time alerts when Balena reports an incident
  • Email, Slack, Discord, Microsoft Teams, and webhook notifications
  • Track Balena alongside 5,000+ providers in one dashboard
  • Component-level filtering
  • Notification groups + maintenance calendar
Start monitoring Balena for free

5 free monitors · No credit card required