Meta will add tens of millions of AWS Graviton cores, a sign that the AI infrastructure race is no longer just about GPUs. The company argues that agentic AI is inflating CPU-heavy work such as planning, orchestration, and data movement, making Graviton5 a strategic fit.
#infrastructure
RSS FeedGoogle DeepMind’s new training stack matters because datacenter boundaries are turning into frontier bottlenecks. Decoupled DiLoCo trained a 12B Gemma model across four U.S. regions on 2-5 Gbps links, more than 20x faster than conventional synchronization while holding 64.1% average accuracy versus a 64.4% baseline.
Google is signaling that enterprise AI is moving from demos to operational scale. In its April 22 Cloud Next update, the company said customer API traffic has risen to more than 16 billion tokens per minute and that just over half of its 2026 machine-learning compute investment will go to the Cloud business.
Cerebras is taking another run at public markets after its 2024 IPO effort was delayed and withdrawn. TechCrunch reports the AI chip startup logged $510M in 2025 revenue and has demand signals tied to AWS data centers and an OpenAI deal reportedly worth more than $10B.
Anthropic said on April 6, 2026 that it secured multiple gigawatts of next-generation TPU capacity from Google and Broadcom starting in 2027. The deal pairs infrastructure scale with surging demand, as run-rate revenue has passed $30 billion and million-dollar customers have doubled since February.
On April 8, 2026, OpenAI said enterprise now accounts for more than 40% of its revenue and could reach parity with consumer by the end of 2026. The company framed its next phase around OpenAI Frontier and a unified AI superapp for company-wide agent deployment.
Microsoft announced a $10 billion Japan investment on April 3, 2026 spanning AI infrastructure, cybersecurity, and workforce training. The plan combines in-country GPU access, public-private security partnerships, and AI skilling for more than one million engineers and developers by 2030.
OpenAI said it closed a $122 billion funding round on March 31, 2026 at an $852 billion post-money valuation. The company tied the raise to compute expansion, product development, and deeper enterprise and developer adoption.
Microsoft said it will invest more than US$1 billion in Thailand’s cloud and AI infrastructure from 2026 to 2028. The company paired the infrastructure commitment with regulatory engagement, an e-commerce generative AI feasibility study, and workforce and startup collaboration.
Microsoft said it will invest $10 billion in Japan from 2026 through 2029 across AI infrastructure, cybersecurity partnerships, and workforce development. The plan includes in-country GPU options through domestic partners and training programs aimed at more than one million engineers, developers, and workers by 2030.
Anthropic said it has signed a new agreement with Google and Broadcom for multiple gigawatts of next-generation TPU capacity that will begin coming online in 2027. The company framed it as its largest compute commitment so far, tied to surging Claude demand and a rapid jump in large enterprise customers.
On April 8, 2026, Anthropic highlighted a new engineering post describing Managed Agents, its hosted service for long-running agent work on the Claude Platform. Anthropic says the system separates session, harness, and sandbox layers so agents can recover more cleanly from failure and connect to customer infrastructure with fewer assumptions.