A March 26, 2026 r/LocalLLaMA post linking NVIDIA's `gpt-oss-puzzle-88B` model card reached 284 points and 105 comments at crawl time. NVIDIA says the 88B MoE model uses its Puzzle post-training NAS pipeline to cut parameters and KV-cache costs while keeping reasoning accuracy near or above the parent model.
#nvidia
RSS FeedThinking Machines Lab said it signed a multi-year strategic partnership with NVIDIA to deploy at least one gigawatt of next-generation Vera Rubin systems. The companies also plan to co-design training and serving systems and widen access to frontier AI and open models for enterprises, research institutions, and the scientific community.
A r/LocalLLaMA thread spread reports that NVIDIA could spend $26 billion over five years on open-weight AI models, but the real discussion centered on strategy rather than headline alone. NVIDIA’s March 2026 Nemotron 3 Super release gives the clearest evidence that the company wants open models, tooling, and Blackwell-optimized deployment to move together.
Meta says a new multi-year deal with NVIDIA will support AI-optimized data centers for training, inference, and core workloads. The announcement also connects privacy, networking, and future Vera Rubin clusters to the same infrastructure roadmap.
GamesRadar says Jensen Huang told Lex Fridman he understands the criticism around Nvidia's DLSS 5 demos and that he does not love “AI slop” either, framing the tech as an artist-controlled generative AI tool.
NVIDIA said on March 25, 2026 that Nemotron Nano 12B v2 VL delivers on-prem video understanding and, in NVIDIA's telling, performs near 30B-class alternatives on the MediaPerf benchmark at less than half the footprint. NVIDIA's model card describes it as a commercially usable multimodal model for multi-image reasoning, video understanding, visual Q&A, and summarization.
OpenAI said on February 27, 2026 that it had secured $110B in new funding at a $730B pre-money valuation. The announcement pairs capital with concrete infrastructure deals, including an Amazon partnership and 5 GW of NVIDIA-backed compute split between inference and training.
At KubeCon Europe, NVIDIA moved its GPU Dynamic Resource Allocation driver into the CNCF and upstream Kubernetes ecosystem. The company also tied the donation to confidential containers support, KAI Scheduler progress, and new tools for large-scale AI cluster orchestration.
NVIDIA introduced OpenShell on March 23, 2026. The company says the open source runtime isolates each autonomous agent in its own sandbox and keeps policy enforcement at the infrastructure layer instead of relying only on model or application safeguards.
NVIDIA and Emerald AI said on March 23, 2026 that they are working with AES, Constellation, Invenergy, NextEra Energy, Nscale Energy & Power, and Vistra on power-flexible AI factories. The concept combines Vera Rubin DSX infrastructure with DSX Flex so AI campuses can connect faster and behave more like grid assets than passive loads.
NVIDIA unveiled Vera CPU on March 23, 2026. The company says it is the first CPU purpose-built for the age of agentic AI and reinforcement learning, delivering 50% faster results and twice the efficiency of traditional rack-scale CPUs.
On March 16, 2026, NVIDIA launched the Nemotron Coalition, an open-model collaboration with Black Forest Labs, Cursor, LangChain, Mistral AI, Perplexity, Reflection AI, Sarvam, and Thinking Machines Lab. The first coalition model will be trained on NVIDIA DGX Cloud and serve as the basis for the upcoming Nemotron 4 family.