Akamai Cloud Is Built for What Cloud Has Become (Updated May 2026)

Headshot of Shawn Michels

May 06, 2026

Shawn Michels

Headshot of Shawn Michels

Written by

Shawn Michels

Shawn Michels, Vice President of Product Management at Akamai, is responsible for driving the strategy and execution of the company’s cloud computing product portfolio. Shawn leads a global team charged with delivering a single platform that allows developers to build, secure, and deliver applications across the entire continuum of compute. Over the course of his career, Shawn's products and services have been used by the world’s largest companies and brands, including Gogo Vision, the industry’s first solution to deliver video directly to consumers over Wi-Fi while in flight.

Share

Executive summary

  • Akamai brings AI inference to production: Investment in NVIDIA RTX PRO™Pro 6000 Blackwell GPUs enable customers to run real-time AI closer to users for faster, more consistent performance. 

  • Systems come together to create a complete, balanced platform: GPUs, dedicated CPUs, and Akamai Functions work together to optimize performance and efficiency across edge and core. 

  • Core platform improvements strengthen operations at scale: Enhanced access controls, improved visibility, latest Kubernetes version updates, and an expanded object storage footprint make Akamai Cloud enterprise-ready.

 

A few years ago, Akamai set out to build a cloud for developers that could grow with enterprise-grade production workloads.

That decision has shaped how we’ve built Akamai Cloud, a distributed cloud for agentic AI, combining GPU-powered inference with edge native compute at global scale. From the beginning, the goal has been to combine cloud infrastructure with Akamai’s global network to support applications that are more distributed, more latency-sensitive, and increasingly real time.

Over time, we’ve watched the customers’ use of the platform evolve. What started with simpler workloads is now turning into more complex systems running in production, including AI inference that needs to perform consistently and respond in real time.

In this blog series, we’ll be sharing regular updates on how we see the platform evolving, and how the pieces are coming together to better serve our customers. We’ll focus on what we shipped, how it performs in real systems, and where it fits in the direction we’re taking the platform. In this blog post, we share our updates from Q1 2026.

Strengthening core compute and data infrastructure 

Up to this point, most of that work has been focused on building the core infrastructure layer across compute, storage, networking, and cloud native services.

With the introduction of Akamai Functions, powered by our acquisition of Fermyon, we’re adding a developer layer on top of that infrastructure. At the same time, we’re expanding into AI infrastructure as inference becomes a real part of production systems.

We launched NVIDIA RTX PRO™ 6000 Blackwell GPUs, completed the acquisition of Fermyon to bring Akamai Functions to the platform, and continued to strengthen the core across compute, observability, data, and delivery.

These changes expand what teams can run in production and improve how consistently those systems perform.

Why use distributed AI inference for real-time performance? 

AI is moving into production systems. Inference now sits directly in the path of the application, where latency, consistency, and cost all matter simultaneously.

In Q1 2026, we expanded our GPU offering with the launch of NVIDIA RTX PRO™ 6000 Blackwell GPUs by building on our existing RTX 4000 Ada Generation instances.

Customers are running more demanding workloads in production across computer vision, speech, multimodal applications, and real-time inference. In a market where high-end GPU capacity is constrained, access to that infrastructure matters.

The GPU is only part of the story; where that compute runs matters just as much.

When inference can run closer to where requests originate and where data already exists, system behavior changes in measurable ways:

  • Latency drops.

  • Cross-region traffic is reduced.

  • Response times become more consistent.

  • Use cases such as real-time assistants, personalization, and decision systems move from prototypes into production systems.

Adding new options for CPU

GPUs handle inference, but everything around those GPUs still depends on CPU performance.

APIs, orchestration, data processing, and integration layers all rely on consistency. When that layer is unpredictable, the entire system is affected.

With our G8 dedicated plans, built on 5th Gen AMD EPYC™ processors and backed by dedicated hardware, we deliver consistent, non-oversubscribed performance. No noisy neighbors. And no guessing how workloads will behave under pressure.

As these systems come together, a more balanced model takes shape: Run inference on distributed GPUs where it makes sense; run the rest of the application on CPU infrastructure that delivers consistent results.

How Akamai Functions accelerates edge native serverless

Up to this point, the platform has focused on infrastructure; that is, the compute, storage, networking, and the services needed to run distributed systems.

With Akamai Functions, our edge native serverless platform built on WebAssembly, we’re adding a developer layer on top.

Modern applications don’t run in one place. They require decisions about where logic executes and how requests are handled. Akamai Functions gives developers a way to run code across both edge and core without managing the underlying infrastructure.

Akamai EdgeWorkers already provides low-latency execution at the edge. Akamai Functions extends this with broader language support (e.g., SDKs for languages like Rust, Go, JavaScript, and Python) and more control over where workloads run.

In practice, this means handling requests closer to users, keeping heavier logic near data and compute, and using GPUs more efficiently by offloading everything around inference. 

Akamai Functions is part of how we keep the experience simple without limiting what can be built.

Enhancing enterprise governance with RBAC and observability 

In addition to expanding into GPUs and Akamai Functions, we’ve recently focused on improving access control, visibility, and operational consistency.

We introduced role-based access control (RBAC) to move from coarse permissions to a more granular, standardized model. This makes it easier to manage access across teams, handle onboarding and offboarding, and meet compliance requirements without adding operational overhead.

We also expanded observability. Real-time metrics are now available for Databases, Object Storage, and NodeBalancers. Audit logs are available across core regions, with alerts for Object Storage and Databases, and broader coverage expanding over time.

For teams running across edge services, delivery, and security, TrafficPeak extends that visibility across the full stack.

Strengthening Kubernetes and platform operations

Kubernetes continues to be a primary way that teams run and operate applications on Akamai Cloud.

LKE, Akamai’s managed Kubernetes engine, continues to stay current with ongoing version updates and infrastructure improvements. Recent releases include upgrades across multiple Kubernetes versions, infrastructure improvements, and foundational networking updates such as Konnectivity.

At the application layer, updates like App Platform v4.13.0 focus on making the platform more reliable and easier to operate. Moving to a modern operator-based deployment model, removing deprecated components, and improving upgrade stability and Argo CD operations reduces the operational overhead for teams running applications in production.

Data, storage, and global access

We expanded our object storage footprint with new E3 endpoints in Chicago, Frankfurt, Los Angeles, and Tokyo. E3 endpoints are designed for high-throughput access to object storage, so placing them in more regions gives teams more options to keep data closer to where it’s being read and written, reducing latency and avoiding unnecessary cross-region movement.

We also added support for metrics and audit logs, so teams can see how data is being accessed and track changes across environments without stitching together external tooling.

That same focus on global data access and predictable cost models is starting to show up externally as well, including in the Forrester Object Storage Solutions Landscape.

At the same time, we continued to evolve the underlying data and networking layer. Valkey Managed Database is now in beta for teams looking for Redis-compatible performance and APIs without the licensing constraints, making it easier to run high-throughput, in-memory workloads without adding cost or complexity.

Managed Databases also now support dual-stack virtual private clouds (VPCs) in limited availability, allowing workloads to run across both IPv4 and IPv6 environments without reworking how services connect as networks evolve.

Looking ahead: The future of distributed AI applications 

AI inference is changing how applications are built and run. More of the work is happening in real time, closer to users, and across more distributed systems.

That shift introduces real constraints. Latency budgets tighten, workloads become more fragmented, and systems have to coordinate across compute, data, and execution in ways that centralized models were not designed for.

At the same time, reliability, observability, and core infrastructure performance remain baseline requirements for running these systems in production.

We are building the platform to support that model. 

Stay tuned

That’s the update for this quarter. We’ll keep building and share our progress in the next quarter.

In the meantime, you can try Akamai Cloud for yourself.

FAQ

Akamai Cloud offers NVIDIA RTX PRO™ 6000 Blackwell GPUs, enabling high-performance, distributed AI inference for real-time applications.

Akamai Functions is an edge native serverless platform built on WebAssembly that allows developers to run code across edge and core environments without managing infrastructure.

Running inference closer to users reduces latency, improves consistency, and lowers cross-region traffic, making real-time AI applications more reliable in production.

Headshot of Shawn Michels

May 06, 2026

Shawn Michels

Headshot of Shawn Michels

Written by

Shawn Michels

Shawn Michels, Vice President of Product Management at Akamai, is responsible for driving the strategy and execution of the company’s cloud computing product portfolio. Shawn leads a global team charged with delivering a single platform that allows developers to build, secure, and deliver applications across the entire continuum of compute. Over the course of his career, Shawn's products and services have been used by the world’s largest companies and brands, including Gogo Vision, the industry’s first solution to deliver video directly to consumers over Wi-Fi while in flight.

Tags

Share

Related Blog Posts

Cloud
AI Survey: 50% of Organizations Struggle to Maintain Latency at Scale
May 06, 2026
The Akamai State of AI Inference report captures real data from the field that describes how AI inference is being built and scaled in production today.
Cloud
AI Inference at the Edge: How Akamai’s Evolution Made It Possible
November 26, 2025
Read how Akamai, a pioneer in CDNs, is undergoing a profound strategic transformation into a diversified cloud and cybersecurity provider in the AI era.
Cloud
Akamai Cloud: New G8 Dedicated Hardware and Performance VM Shapes
December 17, 2025
New compute plans powered by 5th Gen AMD EPYC™ processors deliver predictable performance, transparent pricing, and flexible options for modern workloads.