Back to Blog

When the Lights Go Out: Major Software Outages of 2024–2026

By AutoSmoke Team

The modern software stack is a marvel of interconnected systems. It is also, as recent history keeps proving, profoundly fragile. Over the past two years, some of the most trusted names in tech—Microsoft, AWS, Google, CrowdStrike, Snowflake—have experienced outages that disrupted millions of businesses and users worldwide.

These aren't just embarrassing postmortems. They're a window into how complex systems fail, and what every engineering team can learn from them.

CrowdStrike (July 2024): The Update That Broke the World

No outage in recent memory had the sheer visual impact of the CrowdStrike incident. On July 19, 2024, approximately 8.5 million Windows machines around the world simultaneously displayed the Blue Screen of Death.

The cause was deceptively simple: a faulty configuration file shipped as part of a routine update to CrowdStrike's Falcon sensor. The file contained a logic error that caused the sensor—which runs at the kernel level—to trigger a crash on boot. Because the update was deployed automatically to CrowdStrike's entire user base at once, the impact was instantaneous and global.

Airlines grounded flights. Hospitals reverted to paper records. Banks and broadcasters went dark. The estimated economic damage ran into the billions of dollars.

What made recovery so painful was the manual nature of the fix. Affected machines couldn't boot, so IT teams had to physically access each one—often compounded by BitLocker encryption requiring recovery keys just to reach the command prompt. For organizations with thousands of endpoints across multiple locations, this took days.

The core failure: A content update bypassed the staged rollout process that software code changes would normally go through. No canary deployment. No gradual rollout. One bad file, deployed everywhere, simultaneously.

AWS US-East-1 (October 2025): A DNS Error, $650M in Losses

On October 20, 2025, AWS experienced one of its most damaging outages in years. A DNS error in the US-East-1 data center (Northern Virginia) prevented applications from resolving DynamoDB's endpoint. Since DynamoDB underpins a vast number of AWS-hosted services, the failure cascaded rapidly.

The outage lasted roughly 15 hours and affected over 4 million users and more than 1,000 companies—including Snapchat, Reddit, and numerous payment and financial trading platforms. Economic losses were estimated between $500 million and $650 million for US companies alone.

The concentration of workloads in us-east-1 has long been a known risk. The region hosts a disproportionate share of internet infrastructure, which means any instability there doesn't stay in us-east-1—it ripples outward through dependent services worldwide.

The core failure: Single-region dependency, combined with a DNS layer that became a single point of failure for service discovery.

Google Cloud (June 2025): 54 Services, 7+ Hours

On June 12, 2025, Google Cloud suffered a global outage lasting over seven hours that disrupted 54 services simultaneously—including API Gateway, App Engine, Cloud Run, and the Vertex Gemini API.

The root cause was an invalid automated quota update pushed to the API management system, which caused external API requests to be rejected globally. The automation that was supposed to manage resource limits became the mechanism of failure.

The blast radius extended far beyond Google's own products. Cloudflare, Spotify, Snapchat, and Discord all experienced cascading failures as their Google Cloud dependencies went dark. Within the same month, Cloudflare itself experienced a two-and-a-half-hour outage linked to a third-party cloud provider—a reminder that the dependencies of dependencies matter too.

The core failure: An automated system making configuration changes at global scope, with no circuit breaker to limit blast radius.

Microsoft Azure / M365 (Multiple, 2025–2026)

Microsoft had an unusually turbulent stretch across 2025 and into 2026.

January 2025 brought a 50-hour Azure East US2 outage caused by networking configuration issues—notable both for its duration and for being the first major outage of the year.

October 2025 saw another Azure disruption affecting Office 365, Teams, Outlook, and Xbox Live, preceded by a September incident in the same quarter.

Then, in January 2026, Microsoft 365 suffered a nine-hour outage affecting Outlook, Exchange Online, SharePoint, OneDrive, and Teams across North America. The cause: elevated service load during a maintenance window for a subset of North American infrastructure, compounded by a load balancing configuration change that worsened traffic distribution rather than improving it. Also in early 2026, an inadvertent configuration change to Azure Front Door (AFD)—a global networking layer—caused failures across all Azure regions simultaneously, affecting Entra, Defender, Purview, and downstream customers including Alaska Airlines. Recovery required rolling back to a "last known good" configuration.

The recurring pattern: Configuration changes—applied too broadly, without sufficient validation or rollback safeguards—as the initiating event for cascading failures.

Snowflake (December 2025): A Schema Change Knocks Out 10 Regions

On December 16, 2025, Snowflake pushed a backwards-incompatible database schema change that caused a 13-hour outage spanning 10 of its 23 global regions, across AWS, Azure, and GCP simultaneously.

Customers saw SQL execution internal error messages and were unable to query data or ingest files. For data-dependent businesses—analytics pipelines, dashboards, reporting workflows—the disruption was severe.

The multi-cloud nature of the outage underscored something important: the assumption that distributing workloads across cloud providers protects against outages only holds if the failure doesn't originate in a shared layer. A schema change in Snowflake's own infrastructure can simultaneously affect all three of AWS, Azure, and GCP environments.

The core failure: A breaking schema change deployed without compatibility validation or phased rollout across regions.

Intercom (January 2026): 71 Minutes of Total Darkness

Intercom's US region experienced a complete service blackout on January 9, 2026—71 minutes during which Inbox, Messenger, and all APIs were entirely unavailable.

The cause traced back to a logic bug in the Vitess database routing layer. During a routine table-move operation, rollback logic incorrectly applied an empty routing configuration (VSchema), effectively disconnecting the application from its data shards. The application had no route to its own data.

Seventy-one minutes is short compared to the other incidents in this list, but for a customer communications platform where businesses rely on Intercom to handle live support conversations, the impact on trust was real.

The core failure: Rollback logic that could apply a destructive (empty) state, with no validation that the resulting configuration was valid before applying it.


What These Outages Have in Common

Across all of these incidents—different companies, different stacks, different failure modes—certain patterns keep appearing:

1. Configuration changes are the new deployment risk. In most of these cases, no new code was shipped. A configuration update, a schema change, a routing map, a quota setting—these are the initiating events. Yet configuration changes often receive far less testing scrutiny than code changes.

2. Automation amplifies blast radius. Automated updates, automated quota management, automated rollbacks—these systems are valuable precisely because they operate at scale and speed. But when they fail, they fail at scale and speed too. The CrowdStrike and Google Cloud incidents are the clearest examples: automation that was meant to help became the mechanism of global failure.

3. Staged rollouts are non-negotiable. CrowdStrike's content update skipped the canary/staged deployment process. Snowflake's schema change hit 10 regions simultaneously. Had either been deployed incrementally—to one region, one cohort of machines, one data center first—the failure would have been detected before it became catastrophic.

4. Dependencies create hidden coupling. When Google Cloud goes down, so does Cloudflare's infrastructure that depends on it. When AWS DNS fails, every service using DynamoDB fails with it. The real blast radius of any outage is measured not just by the failing system, but by everything downstream of it.

5. Recovery is harder than prevention. The CrowdStrike incident required manual, physical intervention on millions of machines. The Azure Front Door incident required a full configuration rollback across all global regions. These recovery processes take orders of magnitude longer than the root cause change that triggered them.


What This Means for Your Team

If you're running a product on top of any of these platforms—and most teams are—these outages aren't just cautionary tales about other companies. They're a reminder that your users' experience depends on layers of infrastructure you don't control.

The practical response isn't to eliminate cloud dependencies (you can't). It's to build for the assumption that any dependency can fail:

  • Know your critical paths. Which user journeys break when your cloud provider, database, or third-party service goes down? If you don't have a list, start there.
  • Test your fallbacks, not just your happy paths. Circuit breakers, graceful degradation, and fallback UIs only work if they've been tested. A fallback that's never been exercised is probably broken.
  • Monitor what users experience, not just what servers report. A server can report healthy while the user-facing flow is completely broken. Synthetic monitoring of actual user journeys catches what infrastructure metrics miss.
  • Practice recovery. The teams that recovered fastest from CrowdStrike were the ones that had tested their business continuity plans before they needed them. Disaster recovery is a skill that degrades without practice.

The outages of the past two years haven't been caused by exotic zero-days or unprecedented failure modes. They've been caused by configuration changes, schema updates, and deployment automation—the same categories of change that every engineering team ships every week.

The difference between an incident and a catastrophe is usually not the change itself. It's the blast radius, the detection time, and the recovery plan.


AutoSmoke helps teams catch critical failures before users do—with AI-powered end-to-end tests that run continuously against your production environment. Get started free.