Akamai Launches Cloud Inference: Revolutionizing AI Performance with Edge Computing

5 Sources

Akamai Technologies introduces Cloud Inference, a distributed AI inference platform promising improved performance, lower latency, and cost savings compared to traditional cloud infrastructures.

News article

Akamai Unveils Cloud Inference: A Game-Changer for AI Performance

Akamai Technologies, a leader in cybersecurity and cloud computing, has launched Akamai Cloud Inference, a groundbreaking service designed to revolutionize AI inference capabilities 12. This new offering leverages Akamai's globally distributed network to address the limitations of centralized cloud models, promising significant improvements in AI application performance and efficiency.

The Power of Distributed AI Inference

Akamai Cloud Inference runs on Akamai Cloud, touted as the world's most distributed cloud infrastructure platform. With over 4,100 points of presence across more than 1,200 networks in over 130 countries, Akamai's platform is uniquely positioned to bring AI inference closer to end-users and data sources 14.

Adam Karon, Chief Operating Officer and General Manager of Akamai's Cloud Technology Group, explains the significance: "Getting AI data closer to users and devices is hard and it's where legacy clouds struggle. While the heavy lifting of training LLMs will continue to happen in big hyperscale datacenters, the actionable work of inference will take place at the edge" 3.

Key Features and Benefits

Akamai Cloud Inference offers a comprehensive suite of tools for developers and platform engineers:

  1. Compute Resources: A range of options from CPUs for fine-tuned inference to powerful GPUs and ASICs, integrated with Nvidia's AI Enterprise ecosystem 4.

  2. Data Management: Partnership with VAST Data for a cutting-edge data fabric optimized for AI workloads, complemented by scalable object storage and integration with vector database vendors 4.

  3. Containerization: Leveraging Kubernetes for improved scalability, resilience, and portability of AI workloads 14.

  4. Edge Compute: WebAssembly capabilities for executing LLM inference directly from serverless apps at the network edge 4.

Performance and Cost Advantages

The distributed nature of Akamai Cloud Inference translates into tangible benefits:

  • 3x improvement in throughput
  • Up to 2.5x reduction in latency
  • Potential cost savings of up to 86% on AI inference workloads compared to traditional hyperscaler infrastructure 24

Shifting Focus from Training to Inference

As AI adoption matures, there's a growing recognition that the emphasis on large language models (LLMs) may have overshadowed more practical AI solutions. Akamai's platform caters to this shift, enabling enterprises to leverage lightweight AI models optimized for specific business problems 4.

The Future of Distributed AI

Gartner predicts that by 2025, 75% of data will be generated outside of centralized data centers or cloud regions. This trend underscores the importance of Akamai's approach, which processes data closer to its point of origin 4.

Akamai Cloud Inference represents a significant step towards more efficient, responsive, and cost-effective AI applications. By bringing inference capabilities to the edge of the network, Akamai is positioning itself at the forefront of the next wave of AI innovation, promising to deliver faster, smarter, and more personalized experiences for users across the globe.

Explore today's top stories

Apple Considers Partnering with OpenAI or Anthropic to Boost Siri's AI Capabilities

Apple is reportedly in talks with OpenAI and Anthropic to potentially use their AI models to power an updated version of Siri, marking a significant shift in the company's AI strategy.

TechCrunch logoThe Verge logoTom's Hardware logo

29 Sources

Technology

22 hrs ago

Apple Considers Partnering with OpenAI or Anthropic to

Cloudflare Launches Pay-Per-Crawl Feature to Monetize AI Bot Access

Cloudflare introduces a new tool allowing website owners to charge AI companies for content scraping, aiming to balance content creation and AI innovation.

Ars Technica logoTechCrunch logoMIT Technology Review logo

10 Sources

Technology

6 hrs ago

Cloudflare Launches Pay-Per-Crawl Feature to Monetize AI

Elon Musk's xAI Secures $10 Billion in Funding, Intensifying AI Competition

Elon Musk's AI company, xAI, has raised $10 billion in a combination of debt and equity financing, signaling a major expansion in AI infrastructure and development amid fierce industry competition.

TechCrunch logoReuters logoCNBC logo

5 Sources

Business and Economy

14 hrs ago

Elon Musk's xAI Secures $10 Billion in Funding,

Google Unveils Comprehensive AI Tools for Education with Gemini and NotebookLM

Google announces a major expansion of AI tools for education, including Gemini for Education and NotebookLM, aimed at enhancing learning experiences for students and supporting educators in classroom management.

TechCrunch logoThe Verge logoAndroid Police logo

8 Sources

Technology

22 hrs ago

Google Unveils Comprehensive AI Tools for Education with

NVIDIA's GB300 Blackwell Ultra AI Servers Set to Revolutionize AI Computing in Late 2025

NVIDIA's upcoming GB300 Blackwell Ultra AI servers, slated for release in the second half of 2025, are poised to become the most powerful AI servers globally. Major Taiwanese manufacturers are vying for production orders, with Foxconn securing the largest share.

TweakTown logoWccftech logo

2 Sources

Technology

14 hrs ago

NVIDIA's GB300 Blackwell Ultra AI Servers Set to
TheOutpost.ai

Your Daily Dose of Curated AI News

Don’t drown in AI news. We cut through the noise - filtering, ranking and summarizing the most important AI news, breakthroughs and research daily. Spend less time searching for the latest in AI and get straight to action.

© 2025 Triveous Technologies Private Limited
Twitter logo
Instagram logo
LinkedIn logo