How 5G Works: NR, mmWave, and Network Slicing

5G NR (New Radio) is the fifth generation of cellular network technology, standardized by 3GPP beginning with Release 15 in 2018. Unlike the incremental jump from 3G to 4G LTE, 5G represents a fundamental redesign of the radio interface, the core network, and the service architecture. It introduces a flexible OFDM-based air interface, massive MIMO antenna arrays, network slicing, and a cloud-native core -- all engineered to serve three distinct service categories: enhanced mobile broadband (eMBB), ultra-reliable low-latency communications (URLLC), and massive machine-type communications (mMTC). Understanding how 5G actually works -- beneath the marketing claims of "100x speed" -- requires examining the radio layer, spectrum strategy, antenna physics, core network architecture, and the engineering tradeoffs that determine real-world performance.

Spectrum: mmWave vs Sub-6 GHz vs Low-Band

5G operates across three distinct spectrum tiers, each with fundamentally different propagation characteristics and capacity profiles:

5G NR Spectrum Tiers Frequency 600 MHz 2.5 GHz 3.7 GHz 28 GHz 39 GHz Low-Band < 1 GHz BW: 5-10 MHz Speed: 50-250 Mbps Range: 5-10 km Great coverage Indoor penetration Limited capacity Mid-Band (Sub-6) 1-6 GHz BW: 40-100 MHz Speed: 300 Mbps - 1 Gbps Range: 1-3 km Best speed/coverage balance Capacity for dense areas Moderate indoor reach n41 n77/78 CBRS mmWave 24-100 GHz BW: 400-800 MHz Speed: 1-4+ Gbps Range: 100-300 m (LoS) Extreme capacity Blocked by walls, foliage, rain Requires dense small cells n260 (39G) n261 (28G) n262 (47G)

The spectrum strategy a carrier chooses fundamentally determines the 5G experience their customers get. T-Mobile (AS21928) led with 2.5 GHz mid-band from Sprint's legacy holdings, giving it a nationwide mid-band advantage. AT&T (AS20057) initially focused on mmWave and low-band, then aggressively deployed C-band after the 2021 auction. Verizon bet heavily on mmWave for urban areas and has been backfilling with C-band. In practice, most carriers use dynamic spectrum sharing (DSS) to run 4G LTE and 5G NR simultaneously on the same frequencies, allowing a gradual transition without forklift upgrades.

The 5G NR Air Interface

5G NR uses CP-OFDM (Cyclic Prefix Orthogonal Frequency Division Multiplexing) for the downlink and both CP-OFDM and DFT-s-OFDM (Discrete Fourier Transform-spread OFDM) for the uplink. DFT-s-OFDM has a lower peak-to-average power ratio (PAPR), which extends handset battery life and improves uplink coverage at cell edges. This dual approach is a pragmatic engineering decision: OFDM provides the highest spectral efficiency, while DFT-s-OFDM conserves power in the device where battery life matters most.

The most important innovation in the NR physical layer is flexible numerology. Unlike LTE, which uses a fixed 15 kHz subcarrier spacing and 1 ms subframe, NR defines multiple numerologies indexed by μ:

Wider subcarrier spacing reduces slot duration, which directly reduces latency. A μ=3 mmWave deployment has an air-interface latency floor of ~0.125 ms per slot, compared to 1 ms for μ=0. But wider spacing also reduces the number of subcarriers per channel, which is acceptable at mmWave frequencies where the channel bandwidth is already enormous (400+ MHz). The numerology choice trades spectral granularity against latency -- a fundamental physical constraint, not a software knob you can tune without consequences.

5G NR also supports mini-slots -- transmissions shorter than a full slot (as few as 2 OFDM symbols vs. the standard 14). Mini-slots enable preemption: a URLLC packet can interrupt an ongoing eMBB transmission mid-slot, reducing worst-case latency to under 1 ms at the radio layer. This mechanism is critical for industrial automation and autonomous vehicle use cases, though it comes at the cost of reduced eMBB throughput when URLLC traffic is present.

Massive MIMO and Beamforming

Traditional LTE base stations use 2-8 antenna elements. 5G NR base stations deploy massive MIMO (Multiple Input, Multiple Output) arrays with 32, 64, or even 128 antenna elements in a single panel. The mathematical basis is straightforward: with N antenna elements, you can form up to N independent spatial beams, and the theoretical capacity scales linearly with min(Ntx, Nrx). In practice, the gains are more modest due to spatial correlation and channel estimation overhead, but massive MIMO routinely delivers 3-5x the sector capacity of 4-antenna LTE. Massive MIMO achieves its greatest gains with TDD (Time Division Duplex) rather than FDD, because TDD allows the base station to exploit channel reciprocity — the uplink and downlink use the same frequency, so the base station can estimate the downlink channel from uplink measurements without requiring explicit CSI feedback from every user. This eliminates most of the feedback overhead that would otherwise scale linearly with the number of antenna elements and connected users.

Beamforming in 5G NR operates at two levels:

Beam management in NR is a multi-step process. The base station periodically transmits SSB (Synchronization Signal Block) beams sweeping across the cell coverage area -- typically 4-8 beams at sub-6 GHz or 64 beams at mmWave. The UE (user equipment) measures these beams and reports the best ones. The base station then refines the beam direction using CSI-RS (Channel State Information Reference Signals) for finer-grained tracking. When a user moves and the current beam degrades, beam failure recovery kicks in -- the UE initiates a random-access procedure on an alternative beam. This entire beam management framework adds latency and signaling overhead that simply does not exist in LTE, which is one reason why the "zero latency" marketing around 5G is misleading.

SA vs NSA Architecture

3GPP defined two fundamental deployment modes for 5G:

Non-Standalone (NSA, Option 3x) was designed for rapid early deployment. In NSA, the 5G NR radio connects to the existing 4G EPC (Evolved Packet Core). The LTE eNB acts as the master node, managing the control plane (signaling, mobility, session management), while the 5G gNB acts as a secondary node providing additional data capacity. The UE maintains simultaneous connections to both LTE and NR via EN-DC (E-UTRA/NR Dual Connectivity).

NSA has significant limitations. All control signaling flows through LTE, so connection setup latency is bounded by LTE's control plane -- typically 50-100 ms. The UE cannot camp on 5G NR alone; it must first connect to LTE, then be handed to the 5G carrier. This means 5G coverage is artificially limited by LTE coverage. Network slicing, URLLC, and other advanced 5G features require the 5G Core and cannot work over the legacy EPC. NSA is a crutch, not a destination.

Standalone (SA) connects 5G NR directly to the 5G Core (5GC), a cloud-native service-based architecture. SA eliminates the LTE dependency entirely. The gNB handles both the control plane and data plane natively. Connection setup is faster (target: 10-20 ms). Network slicing becomes possible. Registration and session management use 5G-native signaling. SA is the target architecture, but deployment has been slower because it requires building out the 5GC from scratch -- a massive infrastructure investment that carriers have been reluctant to make while the LTE core still works.

As of 2025, most major US carriers operate a mix of NSA and SA. T-Mobile was the first to deploy nationwide SA, while AT&T and Verizon have been transitioning progressively. The practical impact for users: SA enables faster connection setup, lower latency, and features like VoNR (Voice over New Radio) -- 5G-native voice calls without falling back to LTE's VoLTE.

5G Core (5GC): Service-Based Architecture

The 5G Core is a clean-sheet redesign that breaks from the monolithic architecture of the 4G EPC. Instead of purpose-built network elements (MME, SGW, PGW), the 5GC decomposes network functions into microservices that communicate over HTTP/2 and TCP-based service APIs. The key network functions are:

The service-based architecture (SBA) means 5GC network functions communicate via RESTful APIs using JSON over HTTP/2, not the diameter/GTP-C protocols of LTE. This makes the 5GC deployable on standard cloud infrastructure -- Kubernetes, containers, commodity x86 servers -- rather than purpose-built telecom hardware. In theory, this enables operators to scale network functions independently, deploy updates without downtime, and leverage the cloud-native ecosystem. In practice, telecom vendors have been slow to deliver truly cloud-native implementations, and many "5GC" deployments are thinly containerized monoliths.

5G End-to-End Architecture (SA Mode) UE Device NR Uu gNB gNB-CU + gNB-DU Massive MIMO Beamforming RAN (NG-RAN) NG Interface Backhaul 5G Core (5GC) -- Service-Based Architecture Service-Based Interface (SBI) -- HTTP/2 + JSON NRF NSSF AUSF UDM PCF NEF AMF Access & Mobility N2 (ctrl) SMF Session Mgmt N11 UPF User Plane Forwarding N3 (data) N4 Internet Data Network (DN / N6) N6 UPF can be at edge (MEC) Control plane User plane Internal signaling

Network Slicing

Network slicing is arguably 5G's most architecturally significant innovation, though it remains one of the least deployed. A network slice is a logically isolated end-to-end network built on shared physical infrastructure. Each slice has its own set of network functions (or shared instances with isolation guarantees), its own QoS policies, and potentially its own UPF placement.

3GPP defines slices using an S-NSSAI (Single Network Slice Selection Assistance Information), which consists of:

The NSSF selects the appropriate slice during UE registration. A single UE can connect to up to 8 slices simultaneously, each with different QoS characteristics routed through different SMF/UPF instances. The isolation is enforced at the UPF level: traffic from different slices flows through different forwarding pipelines, potentially on different physical paths.

The enterprise use case for slicing is compelling on paper: a factory could run its robotic control system on a URLLC slice with guaranteed sub-5ms latency and 99.9999% reliability, while employee smartphones use a standard eMBB slice on the same cell. In practice, slicing adoption has been slow because: (1) it requires SA mode, (2) carrier billing and provisioning systems were not designed for slice-level SLAs, (3) the additional operational complexity of managing per-slice network functions is substantial, and (4) most enterprises have not yet identified use cases that justify the premium pricing.

Edge Computing (MEC)

Multi-access Edge Computing (MEC) places compute and storage resources at the edge of the mobile network -- physically close to the radio access network, rather than in a centralized data center hundreds of miles away. In the 5G architecture, this is implemented by deploying a local UPF at the edge site (a cell tower aggregation point, a carrier's central office, or a small data center in a metro area). The SMF steers traffic matching certain criteria (destination IP ranges, application identifiers) to the local UPF, which routes it to co-located application servers rather than sending it through the backhaul to a distant core.

The latency reduction from MEC is real but often overstated. The air interface contributes 1-10 ms of latency depending on numerology and scheduling. The backhaul from the gNB to the core adds 5-20 ms. If the core is in a distant city and the application server is in yet another location, total RTT can exceed 50 ms. MEC can reduce the network component to 2-5 ms by keeping traffic local, but the air interface latency remains unchanged -- and for many applications, the processing time on the application server dominates anyway.

MEC is most valuable for applications that are both latency-sensitive and bandwidth-intensive: augmented reality, real-time video analytics, cloud gaming, and industrial IoT. The challenge is economic: every edge site needs power, cooling, physical security, and operational management. Carriers are partnering with cloud providers (AWS Wavelength, Azure Edge Zones, Google Distributed Cloud Edge) to share the infrastructure burden, essentially renting rack space and connectivity at the carrier's edge sites to hyperscalers who bring the compute platform and application ecosystem.

Backhaul and Fronthaul Requirements

5G's increased radio capacity creates a corresponding increase in transport network requirements. The RAN architecture has been split into three components: the RU (Radio Unit) at the antenna, the DU (Distributed Unit) handling real-time Layer 1/2 processing, and the CU (Centralized Unit) handling non-real-time Layer 2/3 functions. This split creates two distinct transport segments:

The transport network is often the most expensive part of a 5G deployment. A single macro cell site with massive MIMO needs multiple 25G or even 100G fiber links. Densifying the network with small cells at mmWave further multiplies the fiber requirement -- every small cell needs its own fiber or high-capacity wireless backhaul. This is why 5G buildout is concentrated in areas with existing dense fiber infrastructure and why rural 5G deployments lag significantly behind urban ones. The economics are similar to those facing submarine cable operators: the physical plant dominates the cost structure.

Open RAN (O-RAN)

Traditionally, the RAN is a vertically integrated system: the radio hardware, baseband processing, and software all come from a single vendor (Ericsson, Nokia, Samsung, or Huawei). Open RAN, driven by the O-RAN Alliance, aims to disaggregate these components through open interfaces and standardized APIs. The key elements are:

Open RAN promises vendor diversity, cost reduction, and innovation through a multi-vendor ecosystem. The reality is more nuanced: interoperability between vendors remains challenging, performance of virtualized DU/CU on COTS hardware lags behind purpose-built ASICs (particularly for massive MIMO beamforming computations), and the integration burden shifts from the vendor to the operator. Nonetheless, carriers like Dish Network built their entire 5G network on Open RAN principles, and NTIA has funded $1.5 billion for Open RAN R&D. The long-term trajectory is clear, even if the short-term execution is messy.

Real-World Performance vs Marketing

5G marketing claims need aggressive calibration against reality. Here is what the numbers actually look like in deployed networks:

The honest assessment: 5G mid-band delivers a genuine 2-3x improvement in speed and latency over LTE for typical users. mmWave delivers transformative speeds in very limited areas. Low-band 5G is LTE with a new icon on your phone. The real value of 5G lies not in consumer mobile broadband but in the architectural capabilities -- slicing, MEC, URLLC -- that enable new enterprise and industrial use cases. These are materializing slowly.

5G and the IP Layer

From the perspective of the wider internet, 5G is just another access technology. The UPF assigns the UE an IP address (IPv4, IPv6, or both via dual-stack), and from that point, traffic follows the same BGP-routed paths as any other packet. Carrier 5G traffic exits their network through peering points at internet exchange points and private interconnects, traveling across autonomous systems to reach its destination.

One notable difference: 5G carriers are increasingly deploying CGNAT (Carrier-Grade NAT) for IPv4 traffic due to address exhaustion, while offering native IPv6 addresses to 5G devices. T-Mobile has been particularly aggressive with IPv6 -- most T-Mobile 5G devices receive a /64 IPv6 prefix natively. This has implications for peer-to-peer connectivity, gaming, and any application that requires inbound connections.

The transport protocols users choose matter more than ever on 5G. QUIC/HTTP3's connection migration is particularly valuable on mobile: when a device moves between cells or switches from 5G to WiFi, QUIC can migrate the connection seamlessly using connection IDs rather than IP addresses, avoiding the TCP RST that would otherwise terminate the session. Similarly, TCP's congestion control algorithms need tuning for 5G's variable bandwidth characteristics -- the bandwidth can fluctuate dramatically as beams track, cells hand over, or the user moves from mmWave to sub-6 GHz coverage.

Spectrum Allocation and Regulatory Reality

5G spectrum policy varies dramatically by country, which directly affects network performance. The US has followed an auction model, with the C-band (3.7-3.98 GHz) auction alone generating $81 billion in revenue. The FCC allocated only 280 MHz of C-band for 5G; other countries allocated far more mid-band spectrum. China assigned 100-160 MHz of 3.5 GHz and 2.6 GHz spectrum to each of its three carriers. Germany, Japan, and South Korea allocated similar ranges in the 3.4-3.8 GHz band.

The US also faces unique coordination challenges. The C-band abuts aviation radar altimeters (4.2-4.4 GHz), leading to the FAA's extended fight with the FCC over interference risks -- resulting in reduced power levels and exclusion zones around airports that persist to this day. The 3.5 GHz CBRS band requires dynamic spectrum sharing with incumbent military radar through an SAS (Spectrum Access System), adding latency and reducing the spectrum's value for latency-sensitive applications. The 12 GHz band (12.2-12.7 GHz), potentially valuable for 5G, is contested between SpaceX Starlink and wireless carriers.

These regulatory realities mean that the 5G experience varies enormously by country, carrier, and even neighborhood. A user on mid-band 5G in Seoul (where carriers have 80-100 MHz of contiguous mid-band) will have a fundamentally different experience than a user in rural Montana on low-band 5G with 10 MHz of 600 MHz spectrum. The technology is the same; the spectrum is not.

5G Fixed Wireless Access

One of 5G's most commercially successful applications has been FWA (Fixed Wireless Access) -- using 5G as a home broadband replacement for wireline connections like cable or DSL. T-Mobile and Verizon together serve millions of FWA customers in the US, typically using mid-band spectrum with an outdoor or window-mounted CPE (Customer Premises Equipment).

FWA works because mid-band 5G can deliver 100-300 Mbps to a fixed location with a directional antenna, which is competitive with cable internet in many markets. The economics work for carriers because the cell sites are already built for mobile service -- FWA fills unused capacity during off-peak hours (home broadband usage peaks in the evening when mobile usage is lower). The risk is capacity: as FWA subscriber counts grow, the shared cell capacity starts to degrade mobile user experience. Carriers manage this through traffic shaping, deprioritization during congestion, and cell splitting.

What Comes Next

3GPP Release 18 (5G-Advanced) introduces further refinements: AI/ML integration into the RAN for predictive beam management and traffic steering, sidelink enhancements for direct device-to-device communication (critical for V2X), NR on unlicensed spectrum (NR-U, similar to WiFi's use of unlicensed bands), reduced capability (RedCap) devices for IoT that need more than NB-IoT but less than full NR, and non-terrestrial networks (NTN) for satellite-based 5G. 6G research is underway targeting the 2030 timeframe, exploring sub-THz frequencies (100-300 GHz), reconfigurable intelligent surfaces, and joint communication-sensing -- but 5G's deployment cycle has years left to run.

The honest take: 5G is a significant engineering achievement that delivers meaningful improvements in speed, latency, and capacity over LTE. It also introduces genuinely new architectural capabilities -- slicing, MEC, URLLC -- that have no LTE equivalent. But the gap between the marketed vision (instant downloads, remote surgery, autonomous vehicles) and deployed reality (somewhat faster phones, some home internet replacement) remains wide. The technology works; the business models and use cases are still catching up.

Explore Carrier Networks

5G devices connect to carrier autonomous systems that peer with the rest of the internet via BGP. You can explore how major 5G carriers connect to the global internet using the BGP looking glass:

See BGP routing data in real time

Open Looking Glass
More Articles
How DOCSIS Works: Cable Internet Technology Explained
How DSL Works: Internet Over Telephone Lines
How Submarine Cables Work: The Physical Internet
How Rate Limiting Works
How Fiber to the Home (FTTH) Works
How WiFi Works: 802.11 from Radio to Router