Scale Transcoding and AI Workloads with GPU Kubernetes Clusters

Hanna Jeddy

Mar 07, 2025

Hana Jeddy

Hanna Jeddy

Written by

Hana Jeddy

Hana Jeddy is a Senior Product Marketing Manager at Akamai.

Share

The intersection of container orchestration and GPU computing represents a powerful frontier for organizations seeking to optimize their performance. Running managed Kubernetes clusters on GPUs isn't just a technical choice—it's a strategic decision that can transform how enterprises handle their most demanding workloads.

The demand for GPU-accelerated workloads is driven by the explosion in AI and ML initiatives, increased demand for real-time data processing and the rising need for high-performance media processing and streaming.  

Media and streaming applications are constantly adapting to fulfill demand. Sometimes a surge in traffic or demand is predictable, like livestreaming for a major sporting event, but not always. Edge-native applications leverage Kubernetes to ensure that an application’s underlying infrastructure can scale to meet peak demand while maintaining expected performance, and without paying for infrastructure resources that would otherwise go unused.

Performant transcoding is an essential component of a scalable media application, especially for live streaming. Now, we’re making that easier than ever for our customers with GPU node pools in managed Kubernetes clusters.

Announcing GPU Support for Linode Kubernetes Engine: Adding NVIDIA RTX 4000 Ada Generation GPUs to K8s Clusters

We’re excited to announce that Linode Kubernetes Engine now supports NVIDIA RTX 4000 Ada Generation GPUs. Our RTX 4000 Ada Generation GPU plans are optimized for media use cases with each card containing dedicated 2x encode, 2x decode, and 1x AV1 encode engines, but are right-sized for a range of workloads and applications. The RTX 4000 Ada Generation plans start at $0.52 per hour for 1 GPU, 4 CPUs, and 16GB of RAM.

Getting started is simple: while setting up your Kubernetes cluster, select your preferred GPU plan and the quantity of the node pool to add to your cluster.

Note: This requires selecting a region where GPUs are offered. RTX 4000 Ada Generation GPUs are available in the following regions:

  • Chicago, USA (us-ord)
  • Seattle, USA (us-sea)
  • Frankfurt Expansion (de-fra-2)
  • Paris, FR (fr-par)
  • Osaka, JP (jp-osa)
  • Singapore Expansion (sg-sin-2)

Fastest Path to Kubernetes Value

For developers who want to reduce the complexity of building and managing workloads on Kubernetes, our recently launched Akamai App Platform can also run on GPUs. Pairing the accelerated deployment of K8s that App Platform delivers with the powerful compute of GPUs, enables the perfect storm for high performance applications like media and AI at better cost, performance and scale. 

To try it yourself, create an account and browse our Kubernetes documentation to get started, or reach out to our cloud computing consultants for assistance. 

Note: App Platform is currently only available in Beta so will need to be activated through our Beta program page before it will be visible for deployment in your Kubernetes cluster.

Hanna Jeddy

Mar 07, 2025

Hana Jeddy

Hanna Jeddy

Written by

Hana Jeddy

Hana Jeddy is a Senior Product Marketing Manager at Akamai.

Tags

Share

Related Blog Posts

Developers
Introducing Akamai Cloud Pulse: Observability for Your Cloud Infrastructure – Now in Open Beta
July 15, 2025
Akamai Cloud Pulse is now entering Open Beta for all Akamai Managed Database customers. Following successful closed beta testing, we are ready to provide you with real-time insights into your database performance and resource utilization.
Cloud
No Lag, All Frag: Level Up Your Gaming with Xonotic, K3s, and Edge Computing
March 20, 2025
Let’s set the scene for a gamer: you’re having the game of your life (and you wish you were streaming today of all days). You’re lining up the perfect Level up your gaming with Xonotic, K3s, and edge computing! Discover how to host a high-performance, low-latency game server using Akamai Cloud’s distributed compute regions. Say goodbye to lag and hello to seamless gameplay.
Cloud
An Inside Look at our Next Gen Object Storage Launch
August 28, 2025