As AI builders focused on real-world B2B applications, we closely follow infrastructure shifts that reshape how teams train, fine-tune, and deploy models at scale. Akamai’s recent announcement to deploy thousands of NVIDIA Blackwell GPUs, including the RTX PRO 6000 Blackwell Server Edition, marks a significant move toward distributed AI compute. By combining these advanced GPUs with BlueField-3 DPUs and Akamai’s massive edge network of over 4,400 global locations, the company is creating one of the world’s most widely distributed AI platforms.
This architecture intelligently routes inference workloads to localized compute resources, reducing the latency and data egress issues inherent in centralized hyperscaler data centers. The result is a unified system optimized for the inference era, where real-time processing and low-latency decision-making are critical for agentic and physical AI applications.
Why This Matters: The Shift from Centralized Training to Distributed Inference
Traditional hyperscalers excel at massive centralized training, but inference demands a different approach. As models move into production serving millions of users, latency becomes the bottleneck. Akamai addresses this by treating the global network as a low-latency backplane, routing workloads to the nearest optimized GPU clusters.
Key benefits include up to 2.5x reduced latency for inference and up to 86% cost savings compared to traditional hyperscaler infrastructure. This efficiency stems from minimizing data movement across long distances and leveraging edge proximity for faster responses.
For B2B teams, this is particularly valuable in scenarios requiring localized processing. Fine-tuning on proprietary data with data privacy and compliance in mind becomes feasible without shipping sensitive information to distant central facilities. Post-training optimization and rapid iteration gain speed advantages when compute is closer to where data originates and decisions are executed.
Akamai / Target Use Cases: Agentic AI and Physical AI at Real-World Scale
The platform targets workloads that demand production-grade performance in dynamic environments. Agentic AI systems, which act autonomously across complex tasks, benefit from low-latency inference to make timely decisions. Physical AI applications, such as robotics in manufacturing, autonomous delivery fleets, smart grid management, and surgical robotics, require compute that matches the speed of real-world interactions.
In these domains, centralized clouds often introduce unacceptable delays or high egress costs. Akamai’s distributed model enables inference-optimized compute to live closer to endpoints, supporting use cases like fraud prevention in financial services or real-time optimization in logistics, where milliseconds matter for revenue and safety.
COO Adam Karon captured the vision: Akamai focuses on the unique demands of the inference era by distributing compute across the global fabric, moving AI from laboratory to street corner and hospital bed where ROI is realized.
Akamai / Implications for B2B AI Builders in Emerging Markets
This announcement highlights the growing importance of distributed infrastructure in democratizing access to high-performance AI. For teams in Vietnam and the region, where proximity to users and data sovereignty concerns are key, edge-distributed platforms reduce dependency on distant hyperscalers and lower operational costs.
The trend toward hybrid edge-cloud AI aligns with the inference boom, where over 80% of future compute demand will come from serving models rather than training them. Builders who can leverage such architectures gain advantages in speed, cost predictability, and compliance, allowing focus on product innovation rather than infrastructure constraints.
While Akamai’s scale is enterprise-level, the underlying principle applies broadly: inference wins when compute is distributed intelligently. This validates the value of flexible, low-latency access to advanced GPUs without massive centralized commitments.
Explore GPU solutions for AI teams at: https://gpu4ai.cloud/
NVIDIA-New AI Server: 10x Performance Boost Leads the Way for Future AI Infrastructure
The GPU Challenge When AI Teams Transition from R&D to Business-Driven Operations

