GitHub March 2026 Outages: Four Incidents, Copilot Hit Hard
GitHub experienced four incidents in March 2026, including Copilot Agent outages with 99% error rates and Actions delays averaging 30 minutes. Here's what broke and what GitHub is doing about it.
TL;DR
- Four incidents in March degraded GitHub services, with Copilot taking the hardest hits (21–99% error rates)
- Root causes: cache expiration bug, Redis misconfiguration, credential rotation failure, and upstream dependency outage
- GitHub is adding killswitches, improving monitoring, and isolating infrastructure to prevent cascades
What Dropped
GitHub published its March 2026 availability report, detailing four separate incidents that impacted github.com, GitHub Actions, Copilot, and Teams integrations. The worst hit Copilot's Agent service, which saw error rates spike to 99–100% during a credential authentication failure.
The Dev Angle
If you were building on GitHub in March, you likely felt at least one of these. The March 3 incident tanked github.com requests (40% failure rate) and API calls (43% failure), though SSH Git operations stayed clean. GitHub Actions had minimal impact that day but got hammered on March 5 with 95% of workflow runs delayed by an average of 30 minutes due to Redis load balancer misconfiguration.
The Copilot Agent outages on March 19–20 are the ones worth noting if you're using GitHub Copilot's cloud agent features. Users couldn't start new sessions or access existing ones, with error rates hitting 99% at peak. The Teams integration outage on March 24 broke GitHub-to-Teams notifications for 19% of integration installs.
GitHub's postmortems are refreshingly specific about what went wrong. The March 3 cache bug cascaded because a deployment expired every user's cache simultaneously, forcing recalculation and rewrite under load. The March 5 Redis update introduced incorrect configuration that routed traffic to the wrong host. March 19–20 was a credential rotation failure. March 24 was an upstream dependency outage—not GitHub's fault, but still your problem.
Should You Care?
Yes, if you rely on GitHub Actions for CI/CD or use Copilot Agent for coding tasks. The March 5 Actions incident is the one that stings—30-minute delays on workflow runs kill productivity. If you're on GitHub's free or pro tier, you have no SLA recourse, so these incidents just happen.
GitHub's mitigation steps are solid but reactive. They're adding killswitches to the cache layer, moving it to dedicated infrastructure, improving Redis client resilience, and automating credential monitoring. The real question: how long until the "deep architectural work" they mention actually ships? For now, expect occasional degradation and plan your CI/CD accordingly. If you're running mission-critical deployments, consider multi-region failover or hybrid CI/CD setups.
The Copilot outages are less predictable—credential rotation failures and upstream dependency issues are harder to prevent. If you're betting on Copilot Agent for core workflows, keep a fallback ready.
Source: GitHub Blog