A high-scoring LocalLLaMA thread surfaced Sarvam AI's release of two Apache 2.0 reasoning models, Sarvam 30B and Sarvam 105B. The company says both were trained from scratch in India, use Mixture-of-Experts designs, and target reasoning, coding, agentic workflows, and Indian-language performance.
#reasoning-models
LLM Reddit 4d ago 2 min read
LLM Hacker News Feb 25, 2026 2 min read
Inception Labs introduced Mercury 2 and claims a diffusion-based architecture can deliver reasoning quality at much lower latency. The launch emphasizes parallel token refinement, OpenAI-compatible APIs, and enterprise-ready throughput targets.
LLM Reddit Feb 17, 2026 1 min read
A high-ranking r/singularity post shared Google’s Gemini 3 Deep Think update. The announcement includes benchmark claims such as 48.4% on Humanity’s Last Exam (without tools), 84.6% on ARC-AGI-2, and Codeforces Elo 3455, plus Gemini API early access.