South Korea is no longer treating AI infrastructure as a private-sector side quest. A ₩400 billion loan from the Financial Services Commission will expand Naver’s Gak Sejong facility, bankroll GPU deployment and give HyperCLOVA X a bigger domestic base just as AI sovereignty becomes industrial policy.
#ai-infrastructure
RSS FeedAlphabet just rewired the AI capital race: $10 billion goes to Anthropic now at a $350 billion valuation, with another $30 billion tied to performance targets. Coming days after Amazon’s own pledge, the deal shows that frontier labs are no longer raising money in rounds so much as pre-buying compute at planetary scale.
Google has redesigned its TPU roadmap around agent workloads instead of one-size-fits-all acceleration. TPU 8t targets giant training runs with nearly 3x per-pod compute and 121 exaflops, while TPU 8i focuses on low-latency inference with 19.2 Tb/s interconnect and up to 5x lower on-chip latency for collectives.
This is less about one more cloud partnership and more about the infrastructure shape of the next agent wave. NVIDIA and Google Cloud say A5X Rubin systems can scale to 80,000 GPUs per site and 960,000 across multisite clusters, while cutting inference cost per token and boosting token throughput per megawatt by up to 10x versus the prior generation.
HN treated TPU 8t and 8i as more than giant datacenter numbers. The thread focused on the bigger shift: agent-era infrastructure is splitting training and inference into separate hardware bets.
HN latched onto the RAM shortage because the uncomfortable link is physical: HBM demand for AI data centers is now shaping prices for phones, laptops, and handhelds.
Why it matters: AI infrastructure is moving from single accelerator rentals to managed clusters that resemble supercomputers. Google Cloud said A4X Max bare-metal instances support up to 50,000 GPUs and twice the network bandwidth of earlier generations.
HN treated rising GPU costs as more than infrastructure trivia. If frontier access tightens and inference gets pricier, startups may have to compete on procurement, routing, caching, evaluation, and smaller-model strategy rather than assuming abundant calls to the strongest model.
Anthropic said on April 7, 2026 that it has signed a deal with Google and Broadcom for multiple gigawatts of next-generation TPU capacity coming online from 2027. The company also said run-rate revenue has surpassed 30 billion dollars and more than 1,000 business customers are now spending over 1 million dollars annually.
A `r/singularity` post highlighted reporting that roughly half of planned U.S. data center projects have been delayed or canceled because transformers, switchgear, batteries, and related power equipment remain supply constrained. The story resonated because it reframes AI expansion as a grid and industrial logistics problem, not only a chip problem.
OpenAI said on March 31, 2026 that it closed a $122 billion funding round at an $852 billion post-money valuation. The company used the announcement to present consumer reach, enterprise growth, API usage, Codex adoption, and compute access as one reinforcing AI platform flywheel.
On March 17, 2026, NVIDIADC described Groq 3 LPX on X as a new rack-scale low-latency inference accelerator for the Vera Rubin platform. NVIDIA’s March 16 press release and technical blog say LPX brings 256 LPUs, 128 GB of on-chip SRAM, and 640 TB/s of scale-up bandwidth into a heterogeneous inference path with Vera Rubin NVL72 for agentic AI workloads.