Optus BGP Outage 2023: How a Routing Change Took Down an Entire National Carrier

On November 8, 2023, Optus -- Australia's second-largest telecommunications provider and a subsidiary of Singapore Telecommunications (Singtel) -- suffered a nationwide network outage that lasted over 14 hours. Approximately 10 million customers lost access to mobile, fixed-line, and internet services. The outage knocked out emergency triple-zero (000) calls for hundreds of thousands of people, disrupted hospitals, public transit payment systems, and businesses across the country. The root cause was a BGP routing change that propagated through Optus's network in a way that the company's routers could not handle, triggering a cascading failure of its core routing infrastructure.

The Optus outage of 2023 stands as one of the most severe telecommunications failures in Australian history. It prompted a formal investigation by the Australian Communications and Media Authority (ACMA), a Senate inquiry, and significant regulatory action. The incident demonstrates how a single BGP routing event can take down an entire national carrier, and it exposed critical gaps in Australia's telecommunications resilience -- particularly regarding access to emergency services during network failures.

Background: Optus and the Australian Telecommunications Landscape

Optus operates AS4804 (Optus Internet) and AS7474 (Optus Backbone), among other AS numbers. It is the second-largest carrier in Australia after Telstra and serves roughly 10 million mobile subscribers and 1.5 million fixed broadband customers. Optus's network spans the continent, providing mobile coverage in urban and regional areas, fixed-line services via its own fiber and HFC infrastructure, and enterprise connectivity.

Optus is a wholly owned subsidiary of Singtel, Singapore's largest telecommunications company, which operates AS7473 and other AS numbers. The Singtel/Optus network is interconnected with the global internet through peering relationships at major Internet Exchange Points and through transit from upstream providers. This international connectivity becomes relevant to understanding how the outage-triggering event entered Optus's network.

What Happened: The Technical Root Cause

According to Optus's public statements and the subsequent ACMA investigation, the outage was triggered by changes to routing information received from an international peering network. Specifically, a Singtel internet exchange in its international network propagated BGP routing table changes that included an unusually large number of route updates. These route updates were passed to Optus's core routers.

The key failure was in how Optus's routers handled this routing information:

  1. Route update propagation. On the morning of November 8, Optus's border routers received a set of BGP route updates from the Singtel peering exchange. The nature of these updates -- likely involving a large number of prefixes or AS path changes -- was unusual but not inherently malicious.
  2. Routing table overflow / processing failure. The volume or nature of the route updates exceeded the processing capacity or routing table limits of Optus's core Cisco-based routers. Optus had not implemented adequate filters or safety mechanisms to prevent such updates from propagating into the core network. When the routers could not process the routing information correctly, they entered a failure state.
  3. Cascading router failures. As core routers failed or became unstable, they withdrew their own BGP sessions, causing adjacent routers to recalculate routes and propagate further instability. This positive feedback loop -- routers failing, causing route withdrawals, causing other routers to fail -- spread across Optus's entire national backbone within minutes.
  4. Complete network isolation. With the core routing fabric down, Optus's mobile base stations, fixed-line aggregation points, and internet gateways all lost connectivity to each other and to the internet. The network was effectively partitioned into isolated segments with no ability to route traffic between them.
Optus Outage: Cascade of Failures Singtel Peering Exchange BGP route updates propagated to Optus border routers large update Optus Border Routers No adequate prefix filters Updates pass into core unfiltered MISSING SAFEGUARDS - No max-prefix limits - No route damping / filtering Optus Core Routing Fabric Routers overwhelmed, enter failure state BGP sessions drop, route withdrawals cascade Mobile Network ~10M subscribers offline no calls, SMS, data Fixed / Internet broadband, enterprise links all down nationwide Emergency Services 000 calls unable to connect life-safety risk 14+ hours of nationwide outage 04:05 AEDT outage begins ~12:00 AEDT partial recovery starts ~18:30 AEDT services restored

Timeline of the Outage

All times are in Australian Eastern Daylight Time (AEDT), which is UTC+11.

Why Recovery Took 14 Hours

A critical question raised by the Optus outage is why recovery took so long. A BGP misconfiguration can typically be rolled back in minutes -- simply filter the offending routes, reset the BGP sessions, and let the routing table reconverge. The extended recovery time in Optus's case was due to several compounding factors:

Impact on Emergency Services

The most serious consequence of the Optus outage was the disruption to Australia's emergency call service (Triple Zero / 000). Under Australian law, all mobile phones -- even those without an active SIM card -- must be able to call 000. The system works by routing the call through whatever mobile network is available, not just the subscriber's own carrier.

However, when Optus's entire mobile network was down, Optus subscribers' phones could not connect to the Optus network at all. In theory, these phones should have been able to roam to Telstra or Vodafone's networks to place emergency calls. In practice, many handsets were not configured for emergency roaming, or the handshake to connect to an alternate network failed. The result was that hundreds of thousands of Australians could not call emergency services for hours.

Optus reported that approximately 228 calls to 000 failed to connect during the outage. While this number may seem small, each of those calls could have been a life-threatening emergency. The Australian government subsequently mandated that all carriers implement emergency call roaming capabilities to prevent a similar failure from endangering lives.

Regulatory Response

The Optus outage triggered significant regulatory and political action:

Technical Lessons

BGP Route Filtering Is Not Optional

The Optus outage was preventable. Standard BGP best practices -- max-prefix limits, route filtering, and prefix validation via RPKI -- exist precisely to prevent scenarios where an unexpected volume or type of route updates can destabilize a network. The MANRS (Mutually Agreed Norms for Routing Security) initiative, which Optus had not fully implemented, provides a framework for these protections.

At a minimum, every BGP session should have:

Management Plane Separation

When the core routing fabric failed, Optus lost the ability to remotely manage its routers. This is a critical design flaw. The management plane (SSH, SNMP, telemetry) should be accessible via an out-of-band network that does not depend on the production routing fabric. Many large carriers use dedicated management networks, console servers with cellular backup, or satellite-based out-of-band access to ensure they can always reach their equipment -- even during a total routing failure.

Emergency Service Resilience

The failure of emergency call access exposed a systemic vulnerability in how mobile networks handle emergency roaming. When a subscriber's home network is completely unavailable, the handset should automatically attempt to register on any available network for emergency calls. This capability exists in the GSM/LTE specifications (emergency call without registration), but it was not reliably implemented across all handset and network combinations in Australia at the time of the outage.

Comparison with Other BGP Incidents

The Optus outage shares characteristics with several other major BGP-related incidents:

Explore the Routing Infrastructure

The BGP routing tables that Optus's routers process are the same routing tables visible through looking glass tools. You can examine Optus's AS4804 and AS7474 routes, explore their peering relationships, and look at how Australian internet traffic is routed today. Use the god.ad BGP Looking Glass to look up any autonomous system, IP address, or network prefix to see how the global routing table connects networks around the world.

See BGP routing data in real time

Open Looking Glass
More Articles
Azure DNS Outage 2023: How a Misconfigured Deployment Took Down Microsoft Services
Comcast BGP Route Leak 2017: How a Filtering Failure Hijacked Traffic
What is BGP? The Internet's Routing Protocol Explained
What is an Autonomous System (AS)?
What is a BGP Looking Glass?
How to Look Up an IP Address's BGP Route