⚠️ SupplyStatus

Global Supply Chain Incident Tracker

AWS US-EAST-1 DynamoDB DNS Outage - October 2025

critical resolved technical failure
📅
Start DateOctober 20, 2025
Resolution DateOctober 20, 2025
🌍
LocationAshburn, United States
🏭
SupplierAmazon Web Services (AWS)
📦
SectorCloud Computing / Information Technology
🎯
Impacted ClientOver 1,000 companies globally including Snapchat, Fortnite, Roblox, Signal, Robinhood, Coinbase, Ring, Alexa, Slack, Zoom, Delta Airlines, United Airlines, Lloyds Bank, Halifax, HMRC
⚙️
Critical ComponentDynamoDB DNS management system (DNS Planner and DNS Enactor components), Network Load Balancer health check system, EC2 DropletWorkflow Manager
💰
Financial Impact$750,000,000
⏱️
Duration1 day

On October 20, 2025, Amazon Web Services experienced one of its most significant service disruptions in recent years, affecting the US-EAST-1 region in Northern Virginia. This critical incident began at approximately 00:11 Pacific Daylight Time when AWS engineers detected elevated error rates and increased latency across multiple cloud services operating within the region. The root cause of this widespread disruption was traced to a critical design flaw within DynamoDB's automated Domain Name System management infrastructure. Specifically, a race condition emerged between two independent components of the DNS management system: the DNS Planner, responsible for monitoring load balancer health and creating DNS plans, and the DNS Enactor, which applies configuration changes through Amazon Route 53.

The technical sequence of events unfolded when one DNS Enactor component experienced unusual delays in applying legacy DNS records. During this delay period, a second DNS Enactor proceeded to implement the most recent DNS records and initiated a cleanup operation. Shortly thereafter, the delayed DNS Enactor overwrote these newer records, effectively bypassing critical verification checks. Almost immediately following this action, the automated cleanup process deleted the outdated DNS records, resulting in a catastrophic situation where all Internet Protocol addresses were removed from DynamoDB's primary DNS records. This created an empty DNS record for the service's regional endpoint, rendering automatic recovery mechanisms inoperative.

The DNS resolution failure prevented applications and services from locating DynamoDB's API endpoints, effectively severing the connection between countless systems and this foundational AWS database service. Since DynamoDB serves as a critical dependency for numerous AWS offerings, the DNS issue rapidly cascaded throughout the infrastructure. More than seventy AWS services experienced disruption, including essential components such as Elastic Compute Cloud instance launches, Lambda serverless functions, Simple Queue Service messaging, CloudWatch monitoring, CloudTrail logging, Virtual Private Cloud PrivateLink connections, Global Accelerator, Security Token Service authentication, and CloudFront content delivery.

The cascading failures extended beyond AWS's internal services to impact thousands of external applications and platforms worldwide. Major social media platforms experienced connectivity issues, with Snapchat users reporting widespread service disruptions, Signal messaging experiencing outages, and WhatsApp users encountering connection problems. The gaming industry saw significant impact as players attempting to access Fortnite and Roblox found themselves completely locked out of these popular platforms.

Financial services suffered extensive disruption during the incident. Trading platforms Robinhood and Coinbase reported service interruptions that prevented users from accessing their investment portfolios and cryptocurrency holdings during critical trading hours. Payment application Venmo faced transaction processing failures, while digital banking platform Chime experienced difficulties with account access and payment processing. In the United Kingdom, traditional banking institutions including Lloyds, Halifax, and Bank of Scotland reported login issues that potentially affected millions of customers attempting to access their accounts or conduct financial transactions.

Amazon's own consumer-facing services were not immune to the outage. The Alexa voice assistant fell silent in homes worldwide, Ring video surveillance systems stopped recording and streaming, and Prime Video subscribers encountered playback errors. The disruption extended to productivity tools with collaboration platforms like Slack experiencing messaging delays, video conferencing service Zoom reporting connection issues, and graphic design tool Canva facing temporary service suspension.

Government services also experienced disruption, with the United Kingdom's Her Majesty's Revenue and Customs website becoming inaccessible, hindering taxpayers from filing returns or accessing essential services. Educational platform Duolingo reported service interruptions affecting language learners globally. Even entertainment services like popular New York Times games including Wordle, Connections, and Strands became temporarily unavailable, frustrating users attempting to maintain their daily streaks.

The aviation industry encountered operational challenges as major carriers Delta Airlines and United Airlines experienced system failures affecting their check-in processes and operational workflows. Ride-sharing service Lyft reported application downtime affecting thousands of users across the United States who were unable to request or complete trips.

AWS engineers identified the DNS resolution problem at approximately 00:26 PDT and implemented mitigation strategies. The underlying DNS issue was fully resolved by 02:24 PDT, approximately two hours after initial detection. However, recovery proved more complex than simply fixing the DNS records. Once DynamoDB connectivity was restored at 02:25 PDT, the DropletWorkflow Manager, which maintains leases for physical servers hosting EC2 instances, attempted to re-establish leases across the entire EC2 fleet simultaneously, creating a secondary wave of complications.

Following the DynamoDB DNS resolution, AWS services began the recovery process, but encountered subsequent impairments within EC2's internal subsystem responsible for launching new instances. This dependency on DynamoDB meant that EC2 instance launch capabilities remained degraded. AWS engineers temporarily implemented throttling measures on certain impaired operations, including EC2 instance launches, to facilitate controlled recovery and prevent overwhelming the recovering infrastructure.

Additionally, Network Load Balancer health checks became impaired during the recovery phase, resulting in network connectivity issues affecting multiple services including Lambda, DynamoDB itself, and CloudWatch. These health checks were incorrectly marking healthy nodes as unhealthy, causing unnecessary DNS failovers and further complicating the recovery process. Engineers suspended automatic failover operations at 01:36 PDT on October 21 to stabilize the situation.

By 12:28 PDT on October 20, many AWS customers and services were experiencing significant recovery, though some systems continued processing backlogs of delayed requests and events. AWS gradually reduced throttling on EC2 instance launch operations while working to mitigate remaining impacts. Complete normalization of all AWS services was finally achieved by 15:01 PDT, approximately fifteen hours after the initial incident began.

The financial impact of this extensive outage proved substantial and far-reaching. Industry analysis estimated that AWS downtime costs enterprises between five thousand six hundred and nine thousand dollars per minute, depending on operational scale. Specific to this incident, estimates suggested that United States companies alone faced losses approaching seventy-five million dollars per hour during the disruption. Cumulative financial losses across all affected platforms and businesses globally were estimated to reach hundreds of millions of dollars, with some analyses suggesting potential total economic impact extending into hundreds of billions when considering indirect costs, lost productivity, missed trading opportunities, and reputational damage.

More than eleven million outage reports were registered on Downdetector, the widely-used service monitoring platform, with over four million of these reports originating from affected users worldwide. At the incident's peak, approximately twenty-five hundred companies simultaneously experienced service disruptions. Even hours after AWS declared the primary issue resolved, nearly four hundred companies continued reporting residual problems.

The disruption affected organizations across virtually every economic sector. Communication platforms, financial institutions, gaming services, e-commerce websites, streaming entertainment providers, educational applications, government services, transportation systems, and Internet of Things devices all experienced varying degrees of impact. This widespread disruption highlighted the extensive dependencies that modern digital infrastructure has developed on centralized cloud service providers.

Investigation revealed that this incident was not the result of malicious activity or cyberattack. Cybersecurity experts confirmed that the outage stemmed from internal infrastructure failure, specifically a latent design defect within AWS's automated systems rather than external threat actor involvement. The race condition that triggered the failure had existed undetected within the system architecture, waiting for the precise sequence of events that would expose the vulnerability.

The US-EAST-1 region holds particular significance within AWS's global infrastructure network. As Amazon's oldest and largest data center cluster, located in Ashburn and surrounding areas of Northern Virginia, US-EAST-1 functions as a critical hub supporting an enormous portion of internet traffic and cloud computing workloads. This concentration of infrastructure and services makes the region particularly sensitive to incidents with potentially catastrophic impact. Previous significant outages in twenty twenty-one, twenty twenty-three, and now twenty twenty-five have all originated from this same regional infrastructure.

In response to this incident, AWS implemented several corrective measures and preventive safeguards. The company immediately disabled the DynamoDB DNS Planner and DNS Enactor automation systems globally, suspending these operations until comprehensive fixes could be developed and protective mechanisms implemented to prevent recurrence of the race condition. AWS committed to not resuming automated DNS management operations until thorough testing confirmed the elimination of this critical vulnerability.

Additional remediation efforts included implementing mechanisms to limit the number of servers that Network Load Balancers would disconnect when health check failures occur, preventing the cascade of unnecessary isolations that occurred during this incident. AWS also committed to strengthening recovery testing procedures for the DropletWorkflow Manager system that manages EC2 infrastructure, and improving mechanisms for limiting processing operations during periods of high load to prevent overwhelming recovering systems.

The incident sparked renewed discussion within the technology industry regarding the inherent risks associated with concentrated reliance on a small number of hyperscale cloud infrastructure providers. AWS commands approximately thirty percent of the worldwide cloud computing market, with Microsoft Azure and Google Cloud comprising the majority of the remaining market share. This high degree of market concentration means that failures at any of these major providers can have disproportionate global impact.

While cloud services offer undeniable advantages including massive scalability, operational flexibility, and cost efficiency, the incident demonstrated the systemic risks introduced when vast portions of digital infrastructure depend on single providers or specific geographic regions. The cascading nature of the failure, where a DNS issue in one foundational service rapidly propagated across dependent services and then to thousands of external applications, illustrated the complex interdependencies characteristic of modern cloud architecture.

For businesses and organizations affected by the outage, the incident served as a stark reminder of the importance of architectural resilience and disaster recovery planning. Best practices highlighted by this event include implementing multi-region architectures that distribute workloads across geographically separate AWS regions, adopting multi-cloud strategies that leverage multiple providers to avoid single-provider dependency, maintaining robust disaster recovery plans with regularly tested failover procedures, implementing comprehensive monitoring and alerting systems to detect issues rapidly, and ensuring business continuity plans address extended cloud service disruptions.

Insurance considerations also emerged as a significant topic following the outage. Many businesses discovered that their cyber insurance policies included limitations or exclusions related to cloud service disruptions, with some policies requiring outages to persist for eight hours or more before coverage activation. The gap between actual operational exposure and insurance response left many organizations facing uncompensated financial losses.

The October 2025 AWS outage provided valuable lessons for the cloud computing industry and dependent organizations. The incident underscored that even the most sophisticated and well-resourced infrastructure providers remain vulnerable to software defects, design flaws, and operational challenges. The criticality of DNS as a foundational internet technology was once again demonstrated, reinforcing the industry adage that many outages ultimately trace back to DNS issues.

Organizations evaluating their cloud strategies in the aftermath must balance cost considerations against resilience requirements. While implementing multi-region or multi-cloud architectures increases complexity and operational expenses, the potential costs of extended downtime during major provider outages can far exceed these investments. The business decision requires careful analysis of acceptable downtime tolerances, financial exposure during outages, customer expectations and contractual obligations, regulatory compliance requirements, and competitive positioning within respective industries.

The incident also highlighted potential challenges related to workforce expertise and institutional knowledge within large technology organizations. Industry observers noted concerns about experienced engineers departing major cloud providers, potentially taking decades of accumulated operational wisdom regarding complex infrastructure systems. Maintaining deep expertise in managing systems at massive scale represents an ongoing challenge for organizations operating critical internet infrastructure.

Looking forward, the incident will likely influence cloud architecture practices, provider operational procedures, and potentially regulatory approaches to critical digital infrastructure. Discussions regarding the need for greater transparency in cloud provider operations, standardization of incident response and communication protocols, potential regulatory oversight of systemically important cloud infrastructure, and industry-wide collaboration on resilience best practices are expected to intensify.

For AWS specifically, rebuilding customer confidence will require demonstrating that implemented safeguards effectively prevent recurrence of similar incidents, maintaining transparency regarding operational challenges and improvement initiatives, and continuing investment in redundancy and resilience capabilities across its global infrastructure network. The company's reputation for generally high availability and reliability will be tested by how effectively it prevents future incidents of comparable magnitude.

This incident serves as a crucial case study in the evolution of cloud computing and internet infrastructure. As organizations continue migrating critical workloads to cloud platforms and digital services become increasingly essential to daily life, understanding the risks, implementing appropriate safeguards, and maintaining realistic expectations about availability become ever more important. The October 2025 AWS outage demonstrated that no infrastructure is invulnerable, emphasizing the ongoing need for vigilance, planning, and continuous improvement in pursuit of digital resilience.

💡 Alternative Solution

Multi-region AWS architecture deployment, multi-cloud strategy implementation, enhanced disaster recovery procedures, geographic workload distribution across multiple availability zones

Published on October 24, 2025