Google introduced Veo 3.1 Lite as its most cost-effective video generation model, priced at less than 50% of Veo 3.1 Fast while keeping the same speed. The model is rolling out through the paid tier of the Gemini API and Google AI Studio, broadening access to higher-volume video app use cases.
AI
RSS FeedMicrosoft said it will invest $10 billion in Japan from 2026 through 2029 across AI infrastructure, cybersecurity partnerships, and workforce development. The plan includes in-country GPU options through domestic partners and training programs aimed at more than one million engineers, developers, and workers by 2030.
Anthropic said it has signed a new agreement with Google and Broadcom for multiple gigawatts of next-generation TPU capacity that will begin coming online in 2027. The company framed it as its largest compute commitment so far, tied to surging Claude demand and a rapid jump in large enterprise customers.
OpenAI said a malicious Axios 1.14.1 package was executed in a GitHub Actions workflow used for macOS app signing. The company says it found no evidence of user-data exposure or tampered apps, but it is rotating certificates and requiring macOS users to update ChatGPT Desktop, Codex App, Codex CLI, and Atlas before May 8, 2026.
Cloudflare said on April 10, 2026 that its global network passed 500 Tbps of external capacity across 330+ cities and now protects more than 20% of the web. The company frames the milestone as both DDoS headroom and a response to the changing traffic mix of AI crawlers and autonomous agents.
A Hacker News thread drew attention to Instant 1.0, an open-source backend for AI-coded apps built around a Postgres multi-tenant database, a Clojure sync engine, and reactive query invalidation.
Vercel said AI Gateway can now enforce team-wide Zero Data Retention across model providers, extending compliance controls without code changes. The linked post says the feature routes only to ZDR-capable providers and pairs team-wide policy with request-level controls and prompt-training opt-outs.
GitHub used X to point developers to a roadmap that hardens Actions across dependency locking, policy-based execution, and runner network controls. The plan includes workflow-level dependency locks, ruleset-based execution protections, and a native egress firewall for GitHub-hosted runners.
PyTorch said on April 8 that MXFP8 and NVFP4 quantization with Diffusers and TorchAO can cut diffusion latency on NVIDIA B200 GPUs, with NVFP4 reaching up to 1.68x speedups. The accompanying blog frames selective quantization and regional compilation as the practical recipe for better latency-memory tradeoffs.
A high-scoring LocalLLaMA thread amplified AISLE's claim that smaller open or low-cost models reproduced much of the vulnerability analysis Anthropic highlighted for Mythos. The central Reddit pushback was that reasoning over an isolated vulnerable function is very different from autonomously finding that bug inside a large codebase.
A Hacker News thread pushed a GitHub repo claiming it can detect and weaken Gemini image SynthID watermarks using signal processing alone. The more important debate was not the headline claim itself, but whether the project had been validated against Google's own detector and what that says about watermark-based provenance overall.
On April 9, 2026, PyTorch said on X that Safetensors and Helion have joined the PyTorch Foundation as foundation-hosted projects. The move gives the foundation a stronger role in model distribution safety and low-level kernel tooling across the open-source AI stack.