TechSignal.news
SaaS Infrastructure

AWS Projects $600B Revenue by 2036 as AI Inference Locks Buyers Into Hyperscalers

AWS CEO forecasts revenue doubling to $600B by 2036 on enterprise AI workloads. Cloud buyers face multi-year capacity deals as hyperscalers prioritize GPU access over cost flexibility.

TechSignal.news AI4 min read

AWS Bets $200B on AI Infrastructure to Lock in Enterprise Spend

AWS CEO Andy Jassy projected the cloud provider's annual revenue could hit $600 billion by 2036—roughly double prior estimates—driven by enterprises moving AI from pilots to production inference. The forecast, disclosed in a recent investor discussion reported by Reuters, ties directly to AWS committing over $200 billion in cumulative capital expenditures through the decade for AI-specific data centers, networking, and custom silicon designed to reduce dependency on Nvidia GPUs.

The projection matters because it signals a fundamental shift in cloud buying dynamics. Enterprises running AI agents, chatbots, or inference-heavy workloads now prioritize sustained GPU access and low-latency compute over traditional cost or geographic flexibility. AWS holds 28% market share versus Microsoft Azure's 21% and Google Cloud's 14%, but all three hyperscalers are repositioning around AI capacity rather than commodity storage or VMs. Buyers who locked in multi-year commitments in 2024-2025 to secure scarce GPU clusters are finding themselves anchored to a single provider as inference costs become the dominant line item.

The $200 billion AWS capex target funds custom chips like Trainium and Inferentia to offer alternatives to Nvidia's H100 and upcoming Blackwell accelerators. For buyers, this means evaluating not just cost-per-inference but also silicon roadmaps—betting on AWS custom chips versus Azure's partnership with AMD or Google's TPUs. The choice carries multi-year lock-in risk: migrating trained models and tuned inference pipelines between clouds remains expensive and time-intensive, even with containerized workloads.

Cloud Spending Jumps 30% as AI Moves to Production

Global cloud infrastructure spending reached $119 billion in Q4 2025, up 30% or $29 billion year-over-year, pushing full-year 2025 spend past $400 billion. Omdia forecasts 27% growth to over $500 billion in 2026, with AWS, Azure, and Google Cloud controlling 63% combined share. The acceleration stems from enterprises shifting AI workloads from testing to production—running continuous inference for customer-facing applications rather than batch training jobs.

This shift pressures smaller cloud providers without the capital to build AI-optimized infrastructure at scale. Oracle Cloud Infrastructure and IBM Cloud lack the GPU inventory or power/cooling capacity to compete for sustained inference deals, forcing buyers toward the top three. Enterprises now face a tradeoff: accept vendor lock-in with a hyperscaler to guarantee capacity, or risk performance and availability gaps on second-tier clouds.

Buyers are responding with FinOps discipline. According to recent data, 82% of enterprises prioritize cloud spend management, with FinOps adopters cutting waste by 40% and boosting ROI 2.5x over organizations without cost controls. For AI workloads, this means tracking inference cost-per-request, optimizing batch sizes, and negotiating reserved capacity discounts. The alternative—on-premises AI infrastructure—remains capital-intensive and slow to scale, making hyperscaler deals the default for most buyers despite higher multi-year commitments.

Security Spending Follows Workloads Into AI-Heavy Clouds

Upwind Security raised $250 million to expand cloud-native cybersecurity for AI infrastructure, reflecting buyer anxiety over securing compute-intensive environments. The funding positions Upwind against Palo Alto Networks Prisma Cloud and Wiz, valued at $12 billion post-funding, in a market where 79% of enterprises cite security as a top cloud challenge. As AI workloads deepen cloud usage—89% of organizations now run hybrid or multi-cloud—specialized threat detection for agentic AI and model inference becomes critical.

For buyers, this means budgeting 5-10% increases in IT security spend to cover cloud-native tools that monitor GPU workloads, detect model poisoning, and secure API endpoints feeding AI agents. Traditional perimeter security misses threats specific to AI pipelines, like training data exfiltration or inference manipulation. Vendors offering integrated security for AI infrastructure—embedding protections directly into compute layers—gain advantage over bolt-on tools requiring separate agents or proxies.

What to Watch

Buyers evaluating cloud deals in 2026 should model total cost of ownership for AI workloads across three years, not single-year pricing. AWS, Azure, and Google Cloud are using capacity scarcity to push longer commitments with steeper penalties for early exit. Negotiate SKU-level flexibility within reserved capacity agreements—lock GPU hours, but retain the ability to shift between instance types as custom silicon matures. Track hyperscaler silicon roadmaps closely: AWS Trainium 2, Google TPU v6, and Azure Maia 2 all promise cost-per-inference improvements, but deliver on different timelines. Security budgets must expand in parallel—factor 8-12% of cloud AI spend for threat detection, compliance monitoring, and model integrity controls, especially if running customer-facing inference at scale.

cloud-infrastructureAWSAI-workloadshyperscalerscloud-security

Technology decisions, clearly explained.

Weekly analysis of the tools, platforms, and strategies that matter to B2B technology buyers. No fluff, no vendor spin.

More in SaaS Infrastructure