GitHub Outage History

GitHub is up right now

There were 46 GitHub outages since February 27, 2026 totaling 33h 27m of downtime. Each is summarised below — incident details, duration, and resolution information.

Source: https://www.githubstatus.com

Major March 19, 2026

Issues with Copilot Coding Agent

Detected by Pingoru
Mar 19, 2026, 01:44 PM UTC
Resolved
Mar 19, 2026, 02:32 PM UTC
Duration
48m
Affected: Copilot
Timeline · 5 updates
  1. investigating Mar 19, 2026, 01:44 PM UTC

    We are investigating reports of impacted performance for some GitHub services.

  2. investigating Mar 19, 2026, 01:45 PM UTC

    Copilot is experiencing degraded performance. We are continuing to investigate.

  3. investigating Mar 19, 2026, 02:02 PM UTC

    We are investigating reports that Copilot Coding Agent session logs are not available in the UI.

  4. investigating Mar 19, 2026, 02:06 PM UTC

    Copilot is operating normally.

  5. resolved Mar 19, 2026, 02:32 PM UTC

    On March 19, 2026, between 01:05 UTC and 02:52 UTC, and again on March 20, 2026, between 00:42 UTC and 01:58 UTC, the Copilot Coding Agent service was degraded and users were unable to start new Copilot Agent sessions or view existing ones. During the first incident, the average error rate was ~53% and peaked at ~93% of requests to the service. During the second incident, the average error rate was ~99%% and peaked at ~100%% of requests with significant retry amplification. Both incidents were caused by the same underlying system authentication issue that prevented the service from connecting to its backing datastore. We mitigated each incident by rotating the affected credentials, which restored connectivity and returned error rates to normal. The mitigation time was 01:24. The second occurrence was due to an incomplete remediation of the first. We are implementing automated monitoring for credential lifecycle events and improving operational processes to reduce our time to detection and mitigation of issues like this one in the future.

Read the full incident report →

Minor March 19, 2026

Disruption with Copilot Coding Agent sessions

Detected by Pingoru
Mar 19, 2026, 02:05 AM UTC
Resolved
Mar 19, 2026, 02:52 AM UTC
Duration
46m
Timeline · 4 updates
  1. investigating Mar 19, 2026, 02:05 AM UTC

    We are investigating reports of impacted performance for some GitHub services.

  2. investigating Mar 19, 2026, 02:25 AM UTC

    We are seeing widespread issues starting and viewing Copilot Agent sessions. We have a hypothesis for the cause and are working on remediation.

  3. investigating Mar 19, 2026, 02:46 AM UTC

    We have rolled out our mitigation and are seeing recovery for Copilot Coding Agent sessions

  4. resolved Mar 19, 2026, 02:52 AM UTC

    On March 19, 2026, between 01:05 UTC and 02:52 UTC, and again on March 20, 2026, between 00:42 UTC and 01:58 UTC, the Copilot Coding Agent service was degraded and users were unable to start new Copilot Agent sessions or view existing ones. During the first incident, the average error rate was ~53% and peaked at ~93% of requests to the service. During the second incident, the average error rate was ~99%% and peaked at ~100%% of requests with significant retry amplification. Both incidents were caused by the same underlying system authentication issue that prevented the service from connecting to its backing datastore. We mitigated each incident by rotating the affected credentials, which restored connectivity and returned error rates to normal. The mitigation time was 01:24. The second occurrence was due to an incomplete remediation of the first. We are implementing automated monitoring for credential lifecycle events and improving operational processes to reduce our time to detection and mitigation of issues like this one in the future.

Read the full incident report →

Minor March 18, 2026

Webhook delivery is delayed

Detected by Pingoru
Mar 18, 2026, 06:51 PM UTC
Resolved
Mar 18, 2026, 07:46 PM UTC
Duration
55m
Affected: Webhooks
Timeline · 3 updates
  1. investigating Mar 18, 2026, 06:51 PM UTC

    We are investigating reports of degraded performance for Webhooks

  2. investigating Mar 18, 2026, 07:25 PM UTC

    We are seeing recovery and are continuing to monitor the latency for webhook deliveries

  3. resolved Mar 18, 2026, 07:46 PM UTC

    On March 18, 2026, between 18:18 UTC and 19:46 UTC all webhook deliveries experienced elevated latency. During this time, average delivery latency increased from a baseline of approximately 5 seconds to a peak of approximately 160 seconds. This was due to resource constraints in the webhook delivery pipeline, which caused queue backlog growth and increased delivery latency. We mitigated the incident by shifting traffic and adding capacity, after which webhook delivery latency returned to normal. We are working to improve capacity management and detection in the webhook delivery pipeline to help prevent similar issues in the future.

Read the full incident report →

Minor March 16, 2026

Errors starting and connecting to Codespaces

Detected by Pingoru
Mar 16, 2026, 03:01 PM UTC
Resolved
Mar 16, 2026, 03:28 PM UTC
Duration
27m
Timeline · 4 updates
  1. investigating Mar 16, 2026, 03:01 PM UTC

    We are investigating reports of impacted performance for some GitHub services.

  2. investigating Mar 16, 2026, 03:06 PM UTC

    We are investigating reports of users experiencing errors when starting or connecting to Codespaces. Some users may be unable to access their development environments during this time. We are working to identify the root cause and will implement a fix as soon as possible.

  3. investigating Mar 16, 2026, 03:27 PM UTC

    Errors starting or resuming Codespaces have resolved.

  4. resolved Mar 16, 2026, 03:28 PM UTC

    On 16 March 2026, between 14:16 UTC and 15:18 UTC, Codespaces users encountered a download failure error message when starting newly created or resumed codespaces. At peak, 96% of the created or resumed codespaces were impacted. Active codespaces with a running VSCode environment were not affected. The error was a result of an API deployment issue with our VS Code remote experience dependency and was resolved by rolling back that deployment. We are working with our partners to reduce our incident engagement time, improve early detection before they impact our customers, and ensure safe rollout of similar changes in the future.

Read the full incident report →

Minor March 13, 2026

Degraded performance for various services

Detected by Pingoru
Mar 13, 2026, 03:12 PM UTC
Resolved
Mar 13, 2026, 04:15 PM UTC
Duration
1h 2m
Affected: IssuesActionsPackages
Timeline · 6 updates
  1. investigating Mar 13, 2026, 03:12 PM UTC

    We are investigating reports of degraded performance for Actions and Issues

  2. investigating Mar 13, 2026, 03:14 PM UTC

    We are investigating reports of issues with service(s): Actions, Feeds, Issues, Profiles, Registry Metadata, Star, User Dashboard. We will continue to keep users updated on progress towards mitigation.

  3. investigating Mar 13, 2026, 03:20 PM UTC

    Packages is experiencing degraded performance. We are continuing to investigate.

  4. investigating Mar 13, 2026, 03:47 PM UTC

    We are investigating intermittent performance degradation affecting Actions, Feeds, Issues, Package Registry, Profiles, Registry Metadata, Star, and User Dashboard. Users may experience elevated error rates and slower response times when accessing these services. We have identified a potential cause and are implementing mitigations to restore normal service. We'll post another update by 16:15 UTC.

  5. investigating Mar 13, 2026, 04:02 PM UTC

    We have deployed mitigations and are actively monitoring for recovery. We'll post another update by 17:00 UTC.

  6. resolved Mar 13, 2026, 04:15 PM UTC

    On March 13, 2026, between 13:35 UTC and 16:02 UTC, a configuration change to an internal authorization service reduced its processing capacity below what was needed during peak traffic. This caused intermittent timeouts when other GitHub services checked user permissions, resulting in four to five waves of errors over roughly two hours and forty minutes. In total, 0.4% of users were denied access to actions they were authorized to perform. The root cause was a resource right-sizing change deployed to the authorization service the previous day. It reduced CPU allocation below what was required at peak, causing the service's network gateway to throttle under load. Because the change was deployed after peak traffic on March 12, the reduced capacity wasn't surfaced until the next day's peak. The incident was mitigated by manually scaling up the authorization service and reverting the configuration change. To prevent recurrence, we are adding further resource utilization monitors across our entire stack to detect throttling and improving error handling so transient infrastructure timeouts are distinguished from authorization failures, enabling quicker detection of the root issue.

Read the full incident report →

Minor March 12, 2026

Degraded Codespaces experience

Detected by Pingoru
Mar 12, 2026, 01:06 PM UTC
Resolved
Mar 12, 2026, 06:53 PM UTC
Duration
5h 46m
Affected: Codespaces
Timeline · 9 updates
  1. investigating Mar 12, 2026, 01:06 PM UTC

    We are investigating reports of degraded performance for Codespaces

  2. investigating Mar 12, 2026, 01:07 PM UTC

    We're investigating an issue where extensions fail to install in newly created Codespaces. Users can still create and access Codespaces, but extensions will not be operational, resulting in a degraded development experience. Our team is actively working to identify and resolve the root cause. We'll post another update by 14:00 UTC.

  3. investigating Mar 12, 2026, 01:50 PM UTC

    We are continuing to investigate an issue where extensions fail to install in newly created Codespaces. Users can create and access Codespaces, but extensions will not be operational, resulting in a degraded experience. The team is working on a fix. All newly created Codespaces are affected. We'll post another update by 15:00 UTC.

  4. investigating Mar 12, 2026, 02:29 PM UTC

    We have deployed a fix for the issue affecting extension installation in newly created Codespaces. New Codespaces are now being created with working extensions. We'll post another update by 15:30 UTC.

  5. investigating Mar 12, 2026, 03:08 PM UTC

    We're seeing partial recovery for the issue affecting extension installation in newly created Codespaces. Some users may still experience degraded functionality where extensions hit errors. The team continues to investigate the root cause while monitoring the recovery.

  6. investigating Mar 12, 2026, 04:09 PM UTC

    We're seeing intermittent failures downloading from the extension marketplace from codespaces and are investigating.

  7. investigating Mar 12, 2026, 05:20 PM UTC

    We're seeing intermittent failures downloading from the extension marketplace from codespaces, caused by IP blocks for some codespaces. We're working to remove those blocks.

  8. investigating Mar 12, 2026, 05:59 PM UTC

    Codespaces IPs are no longer being blocked from Visual Studio Marketplace operations and we are monitoring for full recovery

  9. resolved Mar 12, 2026, 06:53 PM UTC

    On March 12, 2026, between 01:00 UTC and 18:53 UTC, users saw failures downloading extensions within created or resumed codespaces. Users would see an error when attempting to use an extension within VS Code. Active codespaces with extensions already downloaded were not impacted. The extensions download failures were the result of a change introduced in our extension dependency and was resolved by updating the configuration of how those changes affect requests from Codespaces. We are enhancing observability and alerting of critical issues within regular codespace operations to better detect and mitigate similar issues in the future.

Read the full incident report →

Minor March 12, 2026

Actions failures to download (401 Unauthorized)

Detected by Pingoru
Mar 12, 2026, 04:46 AM UTC
Resolved
Mar 12, 2026, 06:02 AM UTC
Duration
1h 15m
Affected: Actions
Timeline · 4 updates
  1. investigating Mar 12, 2026, 04:46 AM UTC

    We are investigating reports of degraded performance for Actions

  2. investigating Mar 12, 2026, 05:40 AM UTC

    We are continuing investigation of reports of degraded performance for Actions and GitHub Apps

  3. monitoring Mar 12, 2026, 06:02 AM UTC

    Actions is operating normally.

  4. resolved Mar 12, 2026, 06:02 AM UTC

    On March 12, 2026 between 02:30 and 06:02 UTC some GitHub Apps were unable to mint server to server tokens, resulting in 401 Unauthorized errors. During the outage window, ~1.3% of requests resulted in 401 errors incorrectly. This manifested in GitHub Actions jobs failing to download tarballs, as well as failing to mint fine-grained tokens. During this period, approximately 5% of Actions jobs were impacted The root cause was a failure with the authentication service’s token cache layer, a newly created secondary cache layer backed by Redis – caused by Kubernetes control plane instability, leading to an inability to read certain tokens which resulted in 401 errors. The mitigation was to fallback reads to the primary cache layer backed by mysql. As permanent mitigations, we have made changes to how we deploy redis to not rely on the Kubernetes control plane and maintain service availability during similar failure modes. We also improved alerting to reduce overall impact time from similar failures.

Read the full incident report →

Minor March 11, 2026

Incident with API Requests

Detected by Pingoru
Mar 11, 2026, 02:37 PM UTC
Resolved
Mar 11, 2026, 03:02 PM UTC
Duration
24m
Affected: API Requests
Timeline · 3 updates
  1. investigating Mar 11, 2026, 02:37 PM UTC

    We are investigating reports of degraded performance for API Requests

  2. investigating Mar 11, 2026, 03:02 PM UTC

    We are investigating elevated timeouts that affected GitHub API requests. The incident began at 14:37 UTC. Some users experienced slower response times and request failures. System metrics have returned to normal levels, and we are now investigating the root cause to prevent recurrence.

  3. resolved Mar 11, 2026, 03:02 PM UTC

    On March 11, 2026, between 14:25 UTC and 14:34 UTC, the REST API platform was degraded, resulting in increased error rates and request timeouts. REST API 5xx error rates peaked at ~5% during the incident window with two distinct spikes: the first impacting REST services broadly, and the second driven by sustained timeouts on a subset of endpoints. The incident was caused by a performance degradation in our data layer, which resulted in increased query latency across dependent services. Most services recovered quickly after the initial spike, but resource contention caused sustained 5xx errors due to how certain endpoints responded to the degraded state. A fix addressing the behavior that prolonged impact has already been shipped. We are continuing to work to resolve the primary contributing factor of the degradation and to implement safeguards against issues causing cascading impact in the future.

Read the full incident report →

Minor March 11, 2026

Degraded experience with Copilot Code Review

Detected by Pingoru
Mar 11, 2026, 02:25 PM UTC
Resolved
Mar 11, 2026, 03:53 PM UTC
Duration
1h 27m
Timeline · 5 updates
  1. investigating Mar 11, 2026, 02:25 PM UTC

    We are investigating reports of impacted performance for some GitHub services.

  2. monitoring Mar 11, 2026, 02:28 PM UTC

    We are investigating degraded performance with Copilot Code Review. Customers may experience extended review times or occasional failures. We are seeing signs of improvement as our team works to restore normal service. We'll post another update by 15:30 UTC.

  3. monitoring Mar 11, 2026, 03:31 PM UTC

    We experienced degraded performance with Copilot Code Review starting at 14:01 UTC. Customers experienced extended review times and occasional failures. Some extended processing times may continue briefly. We are monitoring for full recovery. We'll post another update by 16:30 UTC.

  4. monitoring Mar 11, 2026, 03:53 PM UTC

    Copilot Code Review queue processing has returned to normal levels.

  5. resolved Mar 11, 2026, 03:53 PM UTC

    On March 11, 2026, between 13:00 UTC and 15:23 UTC the Copilot Code Review service was degraded and experienced longer than average review times. On average, Copilot Code Review requests took 4 minutes and peaked at just under 8 minutes. This was due to hitting worker capacity limits and CPU throttling. We mitigated the incident by increasing partitions, and we are improving our resource monitoring to identify potential issues sooner.

Read the full incident report →

Notice March 10, 2026

Incident With Webhooks

Detected by Pingoru
Mar 10, 2026, 11:00 PM UTC
Resolved
Mar 10, 2026, 11:00 PM UTC
Duration
Timeline · 1 update
  1. resolved Mar 18, 2026, 04:04 PM UTC

    On March 10, 2026, between 23:00 UTC and 23:40 UTC, the Webhooks service was degraded and ~6% of users experienced intermittent errors when accessing webhook delivery history, retrying webhook deliveries, and listing webhooks via the UI and API. Approximately 0.37% of requests resulted in errors, while at peak 0.5% of requests resulted in errors. This was due to unhealthy infrastructure. We mitigated the incident by redeploying affected services, after which service health returned to normal. We are working to improve detection of unhealthy infrastructure and strengthen service safeguards to reduce time to detect and mitigate similar issues in the future.

Read the full incident report →

Minor March 9, 2026

Incident with Webhooks

Detected by Pingoru
Mar 09, 2026, 03:50 PM UTC
Resolved
Mar 09, 2026, 05:03 PM UTC
Duration
1h 13m
Affected: Webhooks
Timeline · 4 updates
  1. investigating Mar 09, 2026, 03:50 PM UTC

    We are investigating reports of degraded performance for Webhooks

  2. investigating Mar 09, 2026, 03:56 PM UTC

    We are experiencing latency on the API and UI endpoints. We are working to resolve the issue.

  3. investigating Mar 09, 2026, 05:03 PM UTC

    Webhooks is operating normally.

  4. resolved Mar 09, 2026, 05:03 PM UTC

    On March 9, 2026, between 15:03 and 20:52 UTC, the Webhooks API experienced was degraded, resulted in higher average latency on requests and in certain cases error responses. Approximately 0.6% of total requests exceeded the normal latency threshold of 3s, while 0.4% of requests resulted in 500 errors. At peak, 2.0% experienced latency greater than 3 seconds and 2.8% of requests returned 500 errors. The issue was caused by a noisy actor that led to resource contention on the Webhooks API service. We mitigated the issue initially by increasing CPU resources for the Webhooks API service, and ultimately applied lower rate limiting thresholds to the noisy actor to prevent further impact to other users. We are working to improve monitoring to more quickly ascertain noisy traffic and will continue to improve our rate-limiting mechanisms to help prevent similar issues in the future.

Read the full incident report →

Major March 5, 2026

Actions is experiencing degraded availability

Detected by Pingoru
Mar 05, 2026, 10:53 PM UTC
Resolved
Mar 05, 2026, 11:55 PM UTC
Duration
1h 1m
Affected: Actions
Timeline · 7 updates
  1. investigating Mar 05, 2026, 10:53 PM UTC

    We are investigating reports of degraded performance for Actions

  2. investigating Mar 05, 2026, 10:54 PM UTC

    Actions is experiencing degraded availability. We are continuing to investigate.

  3. investigating Mar 05, 2026, 11:00 PM UTC

    We applied a mitigation and we should see a recovery soon.

  4. investigating Mar 05, 2026, 11:15 PM UTC

    Actions and dependent services, including Pages, are recovering.

  5. investigating Mar 05, 2026, 11:37 PM UTC

    Actions is experiencing degraded performance. We are continuing to investigate.

  6. investigating Mar 05, 2026, 11:40 PM UTC

    We are close to full recovery. Actions and dependent services should be functioning normally now.

  7. resolved Mar 05, 2026, 11:55 PM UTC

    On March 5, between 22:39 and 23:55 UTC, Actions was degraded due to a repeat of an incident a few hours prior. In this case, a Redis cluster topology change made as a follow-up to the earlier incident caused a repeat of the earlier degradation of Actions jobs. Details of both incidents and the follow-ups are shared at https://www.githubstatus.com/incidents/g5gnt5l5hf56.

Read the full incident report →

Major March 5, 2026

Multiple services are affected, service degradation

Detected by Pingoru
Mar 05, 2026, 04:35 PM UTC
Resolved
Mar 05, 2026, 07:30 PM UTC
Duration
2h 55m
Affected: WebhooksActions
Timeline · 11 updates
  1. investigating Mar 05, 2026, 04:35 PM UTC

    We are investigating reports of degraded performance for Actions

  2. investigating Mar 05, 2026, 04:41 PM UTC

    Actions is experiencing degraded availability. We are continuing to investigate.

  3. investigating Mar 05, 2026, 04:47 PM UTC

    Webhooks is experiencing degraded availability. We are continuing to investigate.

  4. investigating Mar 05, 2026, 04:52 PM UTC

    We are observing delays in queuing Actions workflow runs. We’re still investigating the causes of these delays.

  5. investigating Mar 05, 2026, 05:25 PM UTC

    We have applied mitigations for connection failures across backend resources and we are observing a recovery in queueing Actions workflow runs.

  6. investigating Mar 05, 2026, 05:48 PM UTC

    We are back to queueing Actions workflow runs at nominal rates and we are monitoring the clearing of queued runs during the incident.

  7. investigating Mar 05, 2026, 06:15 PM UTC

    The queue of requested Actions jobs continues to make progress. Job delays are now approximately 6 minutes and continuing to decrease.

  8. investigating Mar 05, 2026, 06:59 PM UTC

    Actions is now fully recovered.

  9. investigating Mar 05, 2026, 07:05 PM UTC

    Actions is operating normally.

  10. investigating Mar 05, 2026, 07:17 PM UTC

    Webhooks is operating normally.

  11. resolved Mar 05, 2026, 07:30 PM UTC

    On Mar 5, 2026, between 16:24 UTC and 19:30 UTC, Actions was degraded. During this time, 95% of workflow runs failed to start within 5 minutes with an average delay of 30 minutes and 10% workflow runs failed with an infrastructure error. This was due to Redis infrastructure updates that were being rolled out to production to improve our resiliency. These changes introduced a set of incorrect configuration change into our Redis load balancer causing internal traffic to be routed to an incorrect host leading to two incidents. We mitigated this incident by correcting the misconfigured load balancer. Actions jobs were running successfully starting at 17:24 UTC. The remaining time until we closed the incident was burning through the queue of jobs. We immediately rolled back the updates that were a contributing factor and have frozen all changes in this area until we have completed follow-up work from this. We are working to improve our automation to ensure incorrect configuration changes are not able to propagate through our infrastructure. We are also working on improved alerting to catch misconfigured load balancers before it becomes an incident. Additionally, we are updating the Redis client configuration in Actions to improve resiliency to brief cache interruptions.

Read the full incident report →

Minor March 5, 2026

Some OpenAI models degraded in Copilot

Detected by Pingoru
Mar 05, 2026, 12:47 AM UTC
Resolved
Mar 05, 2026, 01:13 AM UTC
Duration
25m
Affected: Copilot
Timeline · 4 updates
  1. investigating Mar 05, 2026, 12:47 AM UTC

    We are investigating reports of degraded performance for Copilot

  2. investigating Mar 05, 2026, 12:53 AM UTC

    We are experiencing degraded availability for the gpt-5.3-codex model in Copilot Chat, VS Code and other Copilot products. This is due to an issue with an upstream model provider. We are working with them to resolve the issue.

  3. investigating Mar 05, 2026, 01:13 AM UTC

    The issues with our upstream model provider have been resolved, and gpt-5.3-codex is once again available in Copilot Chat and across IDE integrations. We will continue monitoring to ensure stability, but mitigation is complete.

  4. resolved Mar 05, 2026, 01:13 AM UTC

    On March 5th, 2026, between approximately 00:26 and 00:44 UTC, the Copilot service experienced a degradation of the GPT 3.5 Codex model due to an issue with our upstream provider. Users encountered elevated error rates when using GPT 3.5 Codex, impacting approximately 30% of requests. No other models were impacted. The issue was resolved by a mitigation put in place by our provider.

Read the full incident report →

Minor March 3, 2026

Claude Opus 4.6 Fast not appearing for some Copilot users

Detected by Pingoru
Mar 03, 2026, 08:31 PM UTC
Resolved
Mar 03, 2026, 09:11 PM UTC
Duration
39m
Affected: Copilot
Timeline · 3 updates
  1. investigating Mar 03, 2026, 08:31 PM UTC

    We are investigating reports of degraded performance for Copilot

  2. investigating Mar 03, 2026, 09:05 PM UTC

    We believe that all expected users still have access to Claude Opus 4.6. We confirm that no users have lost access.

  3. resolved Mar 03, 2026, 09:11 PM UTC

    On March 3, 2026, between 19:44 UTC and 21:05 UTC, some GitHub Copilot users reported that the Claude Opus 4.6 Fast model was no longer available in their IDE model selection. After investigation, we confirmed that this was caused by enterprise administrators adjusting their organization's model policies, which correctly removed the model for users in those organizations. No users outside the affected organizations lost access. We confirmed that the Copilot settings were functioning as designed, and all expected users retained access to the model. The incident was resolved once we verified that the change was intentional and no platform regression had occurred.

Read the full incident report →

Major March 3, 2026

Incident with all GitHub services

Detected by Pingoru
Mar 03, 2026, 06:59 PM UTC
Resolved
Mar 03, 2026, 08:09 PM UTC
Duration
1h 9m
Affected: Git OperationsWebhooksAPI RequestsIssuesPull RequestsActionsCodespacesCopilot
Timeline · 25 updates
  1. investigating Mar 03, 2026, 06:59 PM UTC

    We are investigating reports of degraded availability for Actions, Copilot and Issues

  2. investigating Mar 03, 2026, 07:00 PM UTC

    API Requests is experiencing degraded availability. We are continuing to investigate.

  3. investigating Mar 03, 2026, 07:00 PM UTC

    Pull Requests is experiencing degraded performance. We are continuing to investigate.

  4. investigating Mar 03, 2026, 07:02 PM UTC

    Webhooks is experiencing degraded performance. We are continuing to investigate.

  5. investigating Mar 03, 2026, 07:03 PM UTC

    We're seeing some service degradation across GitHub services. We're currently investigating impact.

  6. investigating Mar 03, 2026, 07:04 PM UTC

    Webhooks is experiencing degraded availability. We are continuing to investigate.

  7. investigating Mar 03, 2026, 07:05 PM UTC

    Pull Requests is experiencing degraded availability. We are continuing to investigate.

  8. investigating Mar 03, 2026, 07:11 PM UTC

    Codespaces is experiencing degraded performance. We are continuing to investigate.

  9. investigating Mar 03, 2026, 07:14 PM UTC

    API Requests is experiencing degraded performance. We are continuing to investigate.

  10. investigating Mar 03, 2026, 07:15 PM UTC

    API Requests is operating normally.

  11. investigating Mar 03, 2026, 07:17 PM UTC

    We've identified the issue and have applied a mitigation. We're seeing recovery of services. We continue to montitor for full recovery.

  12. investigating Mar 03, 2026, 07:23 PM UTC

    Issues is experiencing degraded performance. We are continuing to investigate.

  13. investigating Mar 03, 2026, 07:24 PM UTC

    Webhooks is experiencing degraded performance. We are continuing to investigate.

  14. investigating Mar 03, 2026, 07:25 PM UTC

    Codespaces is operating normally.

  15. investigating Mar 03, 2026, 07:25 PM UTC

    Webhooks is operating normally.

  16. investigating Mar 03, 2026, 07:27 PM UTC

    Issues is operating normally.

  17. investigating Mar 03, 2026, 07:28 PM UTC

    Pull Requests is experiencing degraded performance. We are continuing to investigate.

  18. investigating Mar 03, 2026, 07:31 PM UTC

    Pull Requests is operating normally.

  19. investigating Mar 03, 2026, 07:31 PM UTC

    Copilot is operating normally.

  20. investigating Mar 03, 2026, 07:33 PM UTC

    We are seeing recovery across multiple services. Impact is mostly isolated to git operations at this point, we continue to investigate

  21. investigating Mar 03, 2026, 07:36 PM UTC

    Git Operations is experiencing degraded availability. We are continuing to investigate.

  22. investigating Mar 03, 2026, 07:54 PM UTC

    Git Operations is operating normally.

  23. investigating Mar 03, 2026, 07:55 PM UTC

    Actions is operating normally.

  24. investigating Mar 03, 2026, 08:06 PM UTC

    We're seeing recovery across all services. We're continuing to monitor for full recovery.

  25. resolved Mar 03, 2026, 08:09 PM UTC

    On March 3, 2026, between 18:46 UTC and 20:09 UTC, GitHub experienced a period of degraded availability impacting GitHub.com, the GitHub API, GitHub Actions, Git operations, GitHub Copilot, and other dependent services. At the peak of the incident, GitHub.com request failures reached approximately 40%. During the same period, approximately 43% of GitHub API requests failed. Git operations over HTTP had an error rate of approximately 6%, while SSH was not impacted. GitHub Copilot requests had an error rate of approximately 21%. GitHub Actions experienced less than 1% impact. This incident shared the same underlying cause as an incident in early February where we saw a large volume of writes to the user settings caching mechanism. While deploying a change to reduce the burden of these writes, a bug caused every user’s cache to expire, get recalculated, and get rewritten. The increased load caused replication delays that cascaded down to all affected services. We mitigated this issue by immediately rolling back the faulty deployment. We understand these incidents disrupted the workflows of developers. While we have made substantial, long-term investments in how GitHub is built and operated to improve resilience, we acknowledge we have more work to do. Getting there requires deep architectural work that is already underway, as well as urgent, targeted improvements. We are taking the following immediate steps: - We have added a killswitch and improved monitoring to the caching mechanism to ensure we are notified before there is user impact and can respond swiftly. - We are moving the cache mechanism to a dedicated host, ensuring that any future issues will solely affect services that rely on it.

Read the full incident report →

Minor March 2, 2026

Delayed visibility of newly added issues on project boards

Detected by Pingoru
Mar 02, 2026, 11:10 PM UTC
Resolved
Mar 03, 2026, 05:54 AM UTC
Duration
6h 43m
Affected: Issues
Timeline · 13 updates
  1. investigating Mar 02, 2026, 11:10 PM UTC

    We are investigating reports of impacted performance for some GitHub services.

  2. investigating Mar 02, 2026, 11:11 PM UTC

    Issues is experiencing degraded performance. We are continuing to investigate.

  3. investigating Mar 02, 2026, 11:12 PM UTC

    Newly added issues can take up to 30 minutes to appear on project boards. We're investigating the cause of this delay.

  4. investigating Mar 02, 2026, 11:46 PM UTC

    Newly added issues are taking 30–60 minutes to appear on project boards, compared to the normal near-real-time behavior. We're investigating the root cause and possible mitigations.

  5. investigating Mar 03, 2026, 12:05 AM UTC

    The impact extends beyond adding issues to project boards. Adding pull requests and updating fields such as "Status" may also be affected. We're continuing to investigate the root cause.

  6. investigating Mar 03, 2026, 12:52 AM UTC

    Project board updates — including adding issues, pull requests, and changing fields such as "Status" — are currently delayed by 1–2 hours. Normal behavior is near-real-time. We're actively investigating the root cause.

  7. investigating Mar 03, 2026, 01:40 AM UTC

    The delay for project board updates has increased to up to 3 hours. We've identified a potential cause and are working on remediation.

  8. investigating Mar 03, 2026, 02:27 AM UTC

    We're deploying a fix targeting the increased delay in GitHub Projects updates. The rollout should complete within 60 minutes. If successful, the current delay of up to 4 hours should begin to decrease.

  9. investigating Mar 03, 2026, 03:22 AM UTC

    The fix is still building and is expected to deploy within 60 minutes. The current delay for GitHub Projects updates has increased to up to 5 hours.

  10. investigating Mar 03, 2026, 04:17 AM UTC

    The fix has been deployed and processing speeds have returned to normal. There is a backlog of delayed updates that will continue to be worked through — we're estimating how long that will take and will provide an update in the next 60 minutes.

  11. investigating Mar 03, 2026, 04:36 AM UTC

    The backlog of delayed updates is expected to fully clear within approximately 1 hour, after which project board updates will return to near-real-time.

  12. investigating Mar 03, 2026, 05:53 AM UTC

    This incident has been resolved. Project board updates are now processing in near-real-time.

  13. resolved Mar 03, 2026, 05:54 AM UTC

    Between March 2, 21:42 UTC and March 3, 05:54 UTC project board updates, including adding new issues, PRs, and draft items to boards, were delayed from 30 minutes to over 2 hours, as a large backlog of messages accumulated in the Projects data denormalization pipeline. The incident was caused by an anomalously large event that required longer processing time than expected. Processing this message exceeded the Kafka consumer heartbeat timeout, triggering repeated consumer group rebalances. As a result, the consumer group was unable to make forward progress, creating head-of-line blocking that delayed processing of subsequent project board updates. We mitigated the issue by deploying a targeted fix that safely bypassed the offending message and allowed normal message consumption to resume. Consumer group stability recovered at 04:10 UTC, after which the backlog began draining. All queued messages were fully processed by 05:53 UTC, returning project board updates to normal processing latency. We have identified several follow-up improvements to reduce the likelihood and impact of similar incidents in the future, including improved monitoring and alerting, as well as introducing limits for unusually large project events.

Read the full incident report →

Minor March 2, 2026

Incident with Pull Requests /pulls

Detected by Pingoru
Mar 02, 2026, 07:11 PM UTC
Resolved
Mar 02, 2026, 10:04 PM UTC
Duration
2h 53m
Affected: Pull Requests
Timeline · 6 updates
  1. investigating Mar 02, 2026, 07:11 PM UTC

    We are investigating reports of degraded performance for Pull Requests

  2. investigating Mar 02, 2026, 07:23 PM UTC

    We are seeing a degraded experience when attempting to filter the /pulls dashboard. We are working on a mitigation.

  3. investigating Mar 02, 2026, 08:02 PM UTC

    We are experiencing issues with the Pull Requests dashboard that prevent users from filtering their pull requests. We have identified a mitigation and are deploying a fix. We'll post another update by 21:00 UTC.

  4. investigating Mar 02, 2026, 09:04 PM UTC

    We're deploying a fix for pull request filtering. Full rollout across all regions is expected within 60 minutes.

  5. investigating Mar 02, 2026, 10:04 PM UTC

    The issue on https://github.com/pulls is now fully resolved. All tabs are working again.

  6. resolved Mar 02, 2026, 10:04 PM UTC

    On March 2nd, 2026, between 7:10 UTC and 22:04 UTC the pull requests service was degraded. Users navigating between tabs on the pull requests dashboard were met with 404 errors or blank pages. This was due to a configuration change deployed on February 27th at 11:03 PM UTC. We mitigated the incident by reverting the change. We’re working to improve monitoring for the page to automatically detect and alert us to routing failures.

Read the full incident report →

Looking to track GitHub downtime and outages?

Pingoru polls GitHub's status page every 5 minutes and alerts you the moment it reports an issue — before your customers do.

  • Real-time alerts when GitHub reports an incident
  • Email, Slack, Discord, Microsoft Teams, and webhook notifications
  • Track GitHub alongside 5,000+ providers in one dashboard
  • Component-level filtering
  • Notification groups + maintenance calendar
Start monitoring GitHub for free

5 free monitors · No credit card required