$16.5B Worldwide optical hardware revenue, full-year 2025 — a record high
30% Cloud & Colo share of worldwide optical hardware — up from single digits five years ago
71% CAGR of standalone WDM pluggable optics since volume shipments began in 2021
$8.8B Long-haul WDM market forecast for 2026, growing 14% — far ahead of the metro segment
Section 1

Introduction

Optical transport networks are infrastructure that most people never think about — until an AI model somewhere needs to move a trillion tokens between data centers at the speed of light. That moment of reckoning arrived decisively in 2024 and 2025, and the optical hardware industry has not been the same since.

For three decades, the primary buyers of long-haul dense wavelength division multiplexing (DWDM) equipment were telecommunications carriers connecting cities and countries. Hyperscalers — the operators of large-scale cloud computing infrastructure — participated in this market, but mainly at the metro layer. That relationship has been inverted. As of 2026, cloud operators and co-location providers account for 30% of worldwide optical hardware spending, and in North America that share reaches 62%. They are now the dominant force driving product roadmaps, procurement volumes, and the architectural evolution of the entire optical transport industry.

Artificial intelligence is the engine behind this shift. Training large language models and other foundation models requires keeping thousands — soon tens of thousands — of Graphics Processing Units (GPUs) in near-constant communication with one another. Moving the resulting volumes of data between GPU clusters, across data centers, and over long-haul backbone routes demands optical capacity at a scale and with a performance profile that the industry had not previously been asked to deliver. The hardware is responding: new form factors, new optics generations, new line system architectures, and a stressed but expanding supply chain are all part of the story.

This article examines what AI workloads actually demand from optical networks, traces the structural market shifts that have followed, and explains the technology changes — coherent pluggable optics, multi-rail line systems, optical circuit switching, and disaggregated compact modular hardware — that are enabling this new era of transport.

Section 2

What AI Workloads Actually Demand from Optical Networks

Understanding the optical transport implications of AI requires separating the distinct traffic types that AI workloads generate. Each has very different latency, bandwidth, and resilience requirements, and each places a different demand on the network.

2.1 Training Traffic: Intra-Cluster and Cross-Site

Large-scale model training distributes computation across many GPUs simultaneously. During the training process, GPUs exchange gradient updates and synchronization data in all-reduce operations — a pattern of communication in which every GPU must send data to and receive data from every other GPU in the cluster. This type of traffic is extremely latency-sensitive: even microseconds of additional delay translate directly into GPU idle time, wasted compute cycles, and longer training runs.

At the intra-cluster level (within a single data center), bandwidth requirements reach 400 to 800 Gbps per GPU for very large models. Latency targets are measured in sub-microseconds. When training is distributed across multiple data centers — a practice increasingly adopted as model sizes exceed what a single facility can host — checkpoint transfers and model-state synchronization create large, periodic bursts of traffic between sites. These flows are delay-tolerant relative to intra-cluster synchronization but still require very high bandwidth because the payloads (model weights and optimizer states for frontier models) can be hundreds of gigabytes per checkpoint event.