Meta will add tens of millions of AWS Graviton cores, a sign that the AI infrastructure race is no longer just about GPUs. The company argues that agentic AI is inflating CPU-heavy work such as planning, orchestration, and data movement, making Graviton5 a strategic fit.
AI
RSS FeedGoogle DeepMind’s new training stack matters because datacenter boundaries are turning into frontier bottlenecks. Decoupled DiLoCo trained a 12B Gemma model across four U.S. regions on 2-5 Gbps links, more than 20x faster than conventional synchronization while holding 64.1% average accuracy versus a 64.4% baseline.
Anthropic’s new agent-market experiment matters because it turns model quality into money. In a 69-person office marketplace, Claude agents closed 186 deals worth just over $4,000, and Opus-backed users got better prices without noticing.
Hacker News treated the Bitwarden CLI compromise as the sort of GitHub Actions failure that becomes far more serious when the package sits near secrets, tokens, and password-manager workflows. By crawl time on April 25, 2026, the thread had 855 points and 416 comments.
Google has redesigned its TPU roadmap around agent workloads instead of one-size-fits-all acceleration. TPU 8t targets giant training runs with nearly 3x per-pod compute and 121 exaflops, while TPU 8i focuses on low-latency inference with 19.2 Tb/s interconnect and up to 5x lower on-chip latency for collectives.
HN did not read this as a simple cleanup patch. The thread blew up because maintainers are removing old networking code to escape AI-generated security-report overload, and commenters split over whether the real scandal is spam or years of pretending dead code was maintained.
Alphabet’s planned investment is enormous even by 2026 AI standards: $10 billion committed now, with another $30 billion tied to performance targets. Reuters says the deal comes as Anthropic’s run-rate revenue tops $30 billion and the company races to lock in more computing capacity after parallel deals with Amazon, Broadcom, and CoreWeave.
This is less about one more cloud partnership and more about the infrastructure shape of the next agent wave. NVIDIA and Google Cloud say A5X Rubin systems can scale to 80,000 GPUs per site and 960,000 across multisite clusters, while cutting inference cost per token and boosting token throughput per megawatt by up to 10x versus the prior generation.
OpenAI is moving from generic chat to a healthcare-specific workspace, and the timing is clear: 72% of physicians now report AI use in clinical practice. The new product is free to verified U.S. physicians, NPs, PAs, and pharmacists, and OpenAI says doctors rated 99.6% of tested responses safe and accurate across 6,924 conversations.
Privacy tooling usually breaks at scale or forces raw text onto a server. OpenAI’s 1.5B open-weight Privacy Filter runs locally, handles 128,000-token inputs, and posts 97.43% F1 on a corrected PII-Masking-300k benchmark.
This is not just another AI funding round. TechCrunch reports Google will put in $10 billion now at a $350 billion valuation, with as much as $30 billion more tied to Anthropic targets and 5 gigawatts of fresh compute over five years.
HN did not read Google’s TorchTPU post as another cloud pitch. The real question in the thread was whether a PyTorch user can really switch to `tpu` without falling back into the old PyTorch/XLA pain cave.