Why it matters: Moonshot is turning “agent swarm” from a demo phrase into an execution claim with real scale numbers. The Kimi post says one run can coordinate 300 sub-agents across 4,000 steps and return 100-plus files instead of chat transcripts.
#open-models
RSS FeedPrismML is testing whether smaller open models can stay useful by changing the weight format, not only the architecture. Ternary Bonsai ships 8B, 4B and 1.7B models at 1.58 bits, with the 8B variant listed at 1.75GB.
Why it matters: NVIDIA is turning quantum calibration and error correction into an open model-and-tooling stack instead of a lab-only workflow. The April 14 tweet framed Ising as an open suite, and NVIDIA’s technical post says Ising Calibration 1 scored 14.5% above GPT-5.4 and 3.27% above Gemini 3.1 Pro on QCalEval.
NVIDIA is turning quantum chip calibration and error correction into an open AI stack, with one model family that beats GPT 5.4 on QCalEval and another that speeds decoding by 2.25x. If those gains travel outside NVIDIA's own workflow, one of quantum computing's nastiest software bottlenecks just moved closer to something teams can actually deploy.
Google's AI Edge team said on April 2, 2026 that Gemma 4 is bringing multi-step agentic workflows to phones, desktops, and edge hardware under an Apache 2.0 license. The launch combines open models, Agent Skills, and LiteRT-LM deployment tooling.
In a 1247-point Hacker News thread, AISLE argued that small open-weight models can recover much of Mythos-style exploit analysis when the context is tightly scoped, and the comments pushed back hard on the methodology.
An AISLE post that surged on Hacker News argues that Anthropic’s Mythos launch proves the category, but not an exclusive moat. In AISLE’s tests, small and open models recovered major parts of the showcased vulnerability work once the right code path was isolated.
A high-scoring LocalLLaMA thread amplified AISLE's claim that smaller open or low-cost models reproduced much of the vulnerability analysis Anthropic highlighted for Mythos. The central Reddit pushback was that reasoning over an isolated vulnerable function is very different from autonomously finding that bug inside a large codebase.
On April 9, 2026, Google DeepMind said on X that Gemma 4 crossed 10M downloads in its first week and that the Gemma family overall has topped 500M downloads. Google positions Gemma 4 as an open model family built for reasoning, agentic workflows, and efficient deployment on local hardware.
Google DeepMind introduced Gemma 4 on X as a family of open models designed to run on developers’ own hardware. Its April 2, 2026 developer post ties that launch to on-device agentic workflows, support for more than 140 languages, and deployment paths through AICore, AI Edge Gallery, and LiteRT-LM.
A detailed r/MachineLearning post is drawing interest to Dante-2B, a 2.1B dense Italian/English model trained from scratch on 2×H200 GPUs. The project emphasizes tokenizer efficiency for Italian, a 300B token corpus, and a fully open release of weights, tokenizer, and training pipeline after phase 2.
Google DeepMind’s April 2, 2026 X thread introduced Gemma 4 as a new open model family built for reasoning and agentic workflows. Google says the lineup spans E2B, E4B, 26B MoE, and 31B Dense, and adds native function calling, structured JSON output, and longer context windows.