Akamai Launches AI Grid for Distributed Inference Across 4,400 Edge Locations

New platform leverages NVIDIA AI Grid to intelligently route AI workloads across Akamai's global network.

Mar. 16, 2026 at 9:19pm

Akamai Technologies has unveiled the first global-scale implementation of NVIDIA® AI Grid, integrating NVIDIA AI infrastructure into its own network and leveraging intelligent workload orchestration to move beyond isolated AI factories toward a unified, distributed grid for AI inference. The Akamai Inference Cloud platform will provide enterprises with the ability to run real-time AI applications at the edge, with the responsiveness of local compute and the scale of Akamai's global web footprint.

Why it matters

This move represents a significant step in the evolution of AI infrastructure, shifting from centralized 'AI factories' to a distributed, global grid for AI inference. By bringing AI processing closer to the point of user contact, Akamai aims to enable new real-time applications in industries like gaming, finance, media, and retail that require low-latency, high-performance AI capabilities.

The details

At the heart of the AI Grid is an intelligent orchestrator that acts as a real-time broker for AI requests, optimizing 'tokenomics' by improving cost per token, time-to-first-token, and throughput. Akamai is rolling out thousands of NVIDIA RTX PRO 6000 Blackwell Server Edition GPUs across its 4,400+ edge locations, as well as dedicated GPU clusters in its core cloud infrastructure, to provide a continuum of compute from the edge to the core.

  • Akamai unveiled the Akamai Inference Cloud platform on March 16, 2026.
  • Akamai is rolling out the NVIDIA AI Grid implementation across its global network in 2026.

The players

Akamai Technologies

An American cybersecurity and cloud computing company that powers and protects businesses online.

NVIDIA

An American multinational technology company that designs graphics processing units (GPUs) for the gaming and professional markets, as well as system on a chip units (SoCs) for the mobile computing and automotive market.

Adam Karon

Chief Operating Officer and General Manager of the Cloud Technology Group at Akamai.

Chris Penrose

Global VP of Business Development - Telco at NVIDIA.

Got photos? Submit your photos here. ›

What they’re saying

“AI factories have been purpose-built for training and frontier model workloads — and centralized infrastructure will continue to deliver the best tokenomics for those use cases. But real-time video, physical AI, and highly concurrent personalized experiences demand inference at the point of contact, not a round trip to a centralized cluster. Our AI Grid intelligent orchestration gives AI factories a way to scale inference outward — leveraging the same distributed architecture that revolutionized content delivery to route AI workloads across 4,400 locations, at the right cost, at the right time.”

— Adam Karon, Chief Operating Officer and General Manager, Cloud Technology Group, Akamai (Akamai)

“New AI-native applications demand predictable latency and better cost efficiency at planetary scale. By operationalizing the NVIDIA AI Grid, Akamai is building the connective tissue for generative, agentic, and physical AI, moving intelligence directly to the data to unlock the next wave of real-time applications.”

— Chris Penrose, Global VP - Business Development - Telco at NVIDIA (Akamai)

What’s next

Akamai Inference Cloud is available today for qualified enterprise customers, and the company will be showcasing the platform at NVIDIA GTC 2026 in San Jose, California from March 16-19, 2026.

The takeaway

Akamai's AI Grid represents a shift from centralized 'AI factories' to a distributed, global platform for AI inference, enabling new real-time applications that require low-latency, high-performance AI capabilities at the edge. This move could help drive the next wave of AI-powered innovation across industries.