One week after the massive Amazon Web Services (AWS) outage, Microsoft Azure experienced another significant failure on Wednesday, October 29th that took down both Microsoft’s own services and third-party applications. The issue, which began at 15:45 UTC (16:45 in Canary Islands / 17:45 in mainland Spain) and was mitigated by 00:05 UTC on October 30th, originated in Azure Front Door (AFD), Azure’s global content and application delivery network. The immediate result was latencies, timeouts, and cascading errors affecting Microsoft 365, Xbox, Minecraft, administrative portals, and a range of enterprise applications depending on Azure.
Although the impact did not reach the level of the AWS outage last week, a series of incidents involving two of the major cloud providers reopens the debate about Internet resilience and the global dependence on a handful of hyperscalers to operate critical business and services.
What happened and why
Microsoft published a Preliminary Incident Review (PIR) pointing to an unintentional configuration change in Azure Front Door as the trigger for the failure. This deployment caused an invalid or inconsistent state in numerous AFD nodes, which stopped loading correctly. As these “unhealthy” nodes exited the global pool, traffic was unequally distributed among the remaining nodes, amplifying the unavailability with spikes in errors and timeouts.
Engineering response involved immediately blocking any new configuration changes, reverting AFD to the “last known good state”, and gradually recovering nodes, rebalancing traffic to avoid overloading the system. Microsoft attributes the root cause to a software defect in the deployment validation mechanisms, which allowed the faulty configuration to slip through despite existing safeguards.
Summary Timeline:
- 15:45 UTC on 10/29: impact begins for clients.
- 16:18–16:20 UTC: initial public communications and alerts sent to clients via Azure Service Health.
- 17:30 UTC: configuration change lockout.
- 17:40–18:45 UTC: global deployment of corrected configuration and manual recovery of nodes, with gradual routing to healthy nodes.
- 00:05 UTC on 10/30: Microsoft declares AFD impact mitigated, with residual “long tail” effects on a small number of clients.
Which services were affected?
The impact was broad. In addition to Microsoft 365 — with Word, Excel, PowerPoint, and Outlook inaccessible to many during the critical hours — OneDrive and Teams experienced intermittent outages. There were also update failures in Windows Defender and issues affecting Copilot in certain environments. In the entertainment ecosystem, Xbox Live and Minecraft suffered outages. On the Azure layer, problems extended to App Service, Azure SQL Database, Container Registry, Azure Databricks, Azure Virtual Desktop, Media Services, Azure Maps, Azure Portal, Microsoft Entra ID (identity and management components), Microsoft Sentinel (Threat Intelligence), Microsoft Purview, Microsoft Copilot for Security, Healthcare APIs, among others.
Beyond Microsoft, the domino effect affected companies relying on applications hosted with AFD or services running on Azure. During the incident hours, airlines, retailers, and service providers reported disruptions: Alaska Airlines faced check-in issues; Starbucks and other brands experienced transaction failures; Heathrow and Vodafone UK are among the names cited in international coverage. The spikes in reports on Downdetector for Azure and Microsoft 365 underscored the event’s scale.
Is it resolved?
By 00:05 UTC on 10/30, Microsoft reported that errors and latency had returned to pre-incident levels, with a small backlog of clients still experiencing issues as final mitigations continued. Client configuration changes on AFD remain temporarily blocked while system stabilization progresses; Microsoft will notify of the lifting of the lock via Azure Service Health.
Why is AFD failure so impactful?
Azure Front Door is, in essence, the frontline for many services: a CDN and global application front that serves content and routes requests to backends across distributed regions, providing functions like load balancing, acceleration, and security. If AFD fails or quickly loses capacity globally, not only is content delivery degraded, but access to portals and APIs critical for other Azure services becomes congested or unavailable. This is why a problem in AFD spills over multiple layers: from the end user unable to open a document online, to the admin unable to access the Azure portal to troubleshoot.
Coincidence with AWS: two shocks in one week
On October 20th, AWS experienced a prolonged outage centered in US-EAST-1, caused by a DNS issue related to an automation problem in their ecosystem (an empty record mistakenly generated in the DNS management system linked to DynamoDB). The outage took down thousands of applications and services globally and required a temporary disablement of affected automation while additional safeguards were deployed.
Having two hyperscalers encounter significant incidents within just seven days is uncommon. Yet, the technical explanation converges: complexity and large-scale automation are advantages, but also risks when change validation (or the lack thereof) fails in critical systems.
What organizations can do (beyond the headlines)
No provider—be it public cloud, private, or on-premises—is immune to incidents. The key is to design with failure in mind. Below are some best practices recommended by the industry, which this incident highlights again:
- Control and data separation. If the control plane (portals/APIs) fails, having alternative programmatic methods (CLI/PowerShell, automation outside portals) avoids operational blindness.
- True multi-region deployment. Services deployed across multiple regions with automatic or manual guided failover — plus regular “game day” testing — shorten downtime.
- Explicit dependencies. Map which services rely on AFD (or similar CDN/WAF solutions in other providers) and reduce monocultures: for example, employ multi-CDN for high-criticality public sites.
- Caches and degraded modes. For transactional websites, enable reduced modes that allow core functions to operate if the backend responds slowly or fails (e.g., cached catalog or content reads).
- Backups and continuity. Maintain immutable copies (WORM), snapshots, and tested DR plans. In productivity suites like Microsoft 365, understanding offline modes (e.g., cached Outlook, local files in OneDrive) mitigates impact.
- Health alerts. Configure Azure Service Health (and equivalent tools in other providers) to receive alerts via email, SMS, push, or webhooks and trigger automated playbooks when degradations are detected.
Engineering (and product) lessons
- Configuration guardrails. Validation processes in deployment pipelines should be blocking and include double controls for multi-tenant or global changes. An innocent tenant change should never proliferate unchecked to a network serving millions of users.
- Rapid reversions and always accessible “last good state”. Maintain configuration snapshots and circuit breakers to cut propagation before issues escalate.
- Gradual recovery. Reintegrate nodes into a global pool step-by-step to prevent re-falls caused by overloads. Although slower, this approach is more stable.
- Effective communication. Incident reports with timelines, technical causes, actions, and preventive measures build client trust and inform internal improvements.
Context: cloud remains, but with eyes wide open
The advantages of cloud — elasticity, service catalog, scalability — don’t vanish because of two incidents. However, these episodes remind us that outsourcing infrastructure doesn’t mean outsourcing responsibility: architecture and continuity planning remain each company’s responsibility. For many organizations, a hybrid model (combining on-prem, private cloud, and public cloud) offers balanced control, cost, and agility.
Frequently Asked Questions
What exactly is Azure Front Door, and why does its failure disable so many services?
Azure Front Door is a global frontend (CDN + load balancer + security layer) that intermediates millions of requests towards Microsoft and client services. When AFD enters an invalid state at scale, access routes break and latencies/errors spike across portals, APIs, and dependent applications.
How long did the incident last, and what did Microsoft do to fix it?
Impact started at 15:45 UTC on October 29th and was mitigated by 00:05 UTC on October 30th. Microsoft blocked changes, reverted to the last known good configuration, manually recovered nodes, and gradually rebalanced traffic to re-stabilize the system.
Is this related to last week’s AWS outage?
Not directly. On October 20th, AWS’s issue was caused by a DNS problem related to automation in US-EAST-1 (an erroneous empty record in DNS linked to DynamoDB). Azure’s October 29th problem stemmed from a configuration change in Azure Front Door that bypassed safeguards due to a software defect. Both share a pattern: automation and insufficient change validation in critical systems.
What can companies do to mitigate these events?
Adopt true multi-region, alternative programmatic methods (CLI/PowerShell), multi-CDN for critical sites, degraded modes, caches, immutable backups, tested disaster recovery (DR) plans, and health alerts (like Azure Service Health) that trigger automated playbooks.
Sources (selected):

