Cloudflare activates “Code Orange” after multiple global outages: the plan to make the Internet “fail small”

For years, Cloudflare’s promise has been almost invisible to the end user: speeding up websites, blocking attacks, and making the Internet “just work.” That’s precisely why, when the company fails, the impact feels like a rare and disconcerting blackout: blank pages, 5xx errors, services that “won’t load,” and a cascade of complaints on social media.

That effect repeated itself on Friday, December 19, 2025, when multiple sites began returning 502 (Bad Gateway) errors, and the problem spread globally. In Spain, tech outlets like El Chapuzas Informático reported how the incident started in the middle of the afternoon and lasted for hours, affecting both websites and online services reliant on the company’s network.

The striking thing is that, on that same December 19, Cloudflare published an unusual article in tone: direct, self-critical, with a clear message of urgency. The company announced an internal initiative labeled “emergency”: “Code Orange: Fail Small”. In other words, acknowledge that errors happen, but prevent them from escalating into a global outage.

Two prior incidents and a “hot” week on the network

The context matters. Cloudflare had experienced two serious incidents in a few weeks:

  • November 18, 2025: The network suffered significant failures lasting roughly 2 hours and 10 minutes. According to the post-mortem, the root cause was an error in generating a functionality file related to Bot Management, which ended up affecting numerous services. The typical user-facing symptoms were pages returning errors and traffic not reaching its destination.
  • December 5, 2025: The second hit was a shorter outage (around 25 minutes), but with a chilling data point: about 28% of HTTP traffic served by Cloudflare was affected. This time, the trigger was related to an urgent security change addressing a critical vulnerability in the React ecosystem.

Given this recent history, the incident on December 19 was not perceived as “a rough afternoon” but as a sign that something structural needed to change: less blind deployment speed, more control… even if that means sacrificing some “instantaneous” performance.

What “Code Orange” means at Cloudflare

Cloudflare describes “Code Orange” as a state of maximum internal priority: resilience strengthening is prioritized above almost anything else, enabling teams across different areas to collaborate seamlessly and suspending less urgent initiatives. The company emphasizes that they had only declared something similar once before.

The core goal boils down to a simple idea—almost civil engineering applied to networking: if something breaks, let it break small. Don’t let it take down the control plane, the management panel, or hundreds of cities at once.

The critical point: configuration changes spreading in seconds

The root pattern Cloudflare identifies as particularly troubling is intrinsic to its DNA: the ability to deploy global configuration changes in a matter of seconds.

According to the company, their internal system (called Quicksilver) allows a change—such as a rule, security adjustment, or configuration—to reach most servers rapidly. That speed is useful for threat response but also transforms mistakes into worldwide problems “at the speed of light.”

This introduces the first major promise of “Fail Small”: treat configuration like code.

Cloudflare already deploys software versions through a controlled, monitored method with “gates” that allow halting or reverting if something goes wrong. Changes are first tested internally, then rolled out in phases, with the system capable of performing rollbacks automatically when anomalies appear—no manual improvisation needed.

What it had not previously done—and now commits to doing—is applying that same rigor to configuration changes. The goal is for any modification that could affect how traffic is served to undergo an update process equivalent to software updates, relying on their Health Mediated Deployments (HMD) framework.

Failing well: sensible defaults and controlled degradation

The second pillar of the plan is less glamorous but often makes the difference between inconvenience and catastrophe: defining reasonable failure modes.

Cloudflare admits that recent incidents involved errors in a component that affected a large part of their stack, including the control plane used by clients for operations. In a concrete example, the company explains that if a module (like Bot Management) fails, the default behavior should not be to “block traffic” or “break” but to degrade: use validated default values, allow traffic passage with less precise classification, and avoid internal panics that take services down (“panics”).

This is a classic philosophy in critical systems: when everything is failing, a less-than-perfect service is better than no service.

“Break glass” and circular dependencies: when even clients cannot access

The third focus addresses a pain point many administrators have experienced: during an emergency, the worst is when security mechanisms prevent you from fixing the problem.

Cloudflare discusses its “break glass” procedures—a mechanism allowing controlled privilege elevation to resolve critical issues. They acknowledge that during incidents, they took too long to resolve some problems because internal security layers and dependencies hampered access to necessary tools.

Furthermore, they cite a particularly delicate case: during the November 18 incident, Turnstile (their CAPTCHA-free anti-bot solution) went down, and because Cloudflare uses it in the login flow, some clients could not access their dashboard precisely when they needed it most.

Part of the plan involves eliminating circular dependencies, enabling controlled emergency access, and increasing internal drills so that, when the next incident occurs, issues aren’t discovered “in production.”

Why the debate is bigger than Cloudflare itself

The significance of these outages cannot be understood without the broader reality: Cloudflare is a massive part of modern Internet infrastructure. The company has publicly stated that it helps serve and protect approximately 20% of all websites. When something like this fails, it’s not just an app that crashes—it’s an entire layer of the web infrastructure that wobbles.

As users track outages on X, Telegram, or Reddit, there’s also a parallel phenomenon in Spain: the rise of niche media focused on technology, social media, and AI. Websites like Noticias Inteligencia Artificial (noticias.ai) or RevistaCloud.com show how audiences seek quick, specialized explanations when “Internet breaks,” even if only timidly in the face of media giants.

Cloudflare aims to close the loop: they apologize, admit to being “embarrassed” about the impact, and promise measurable changes in the short term. Their declared objective is that, before the end of Q1 2026, critical systems will be covered by monitored configuration deployments, improved failure modes, and more effective emergency procedures.

The lingering question is not if incidents will recur—on networks of this size, it’s naïve to think otherwise—but whether the next error will be contained… or become a new global trending topic.


Frequently Asked Questions

What is a 502 (Bad Gateway) error, and why does it appear when Cloudflare fails?
It generally indicates that an intermediary server (gateway) couldn’t get a valid response from the origin server or an intermediate component. During a network outage or degradation, it may manifest as 502/5xx errors or blank pages.

What does “Code Orange: Fail Small” mean, and how can it affect website users?
It’s a resilience plan to prevent failures from propagating globally: phased configuration rollouts, automatic rollbacks if anomalies are detected, and default behaviors prioritizing traffic continuity.

How can a digital media or e-commerce site prepare for CDN or security provider outages?
By implementing continuity plans: independent monitoring, contingency pages, multi-provider strategies (when feasible), internal procedures for urgent changes, and prepared access tokens/APIs to act even if the main dashboard is unavailable.

Why is there increasing concern about “centralization” during Cloudflare outages?
Because many websites and services concentrate performance and security in few providers. While this improves efficiency, it amplifies impact when a key component experiences an issue.

via: blog.cloudflare

Scroll to Top