Anthropic introduced Claude Sonnet 4.6 on Feb 17, 2026 as its most capable Sonnet model yet. The release combines a 1M token context window in beta with upgrades to coding, computer use, and agent workflows while keeping Sonnet 4.5 pricing.
#llm
RSS FeedAnthropic announced Claude Sonnet 4.6 on February 17, 2026. The release combines a 1M-token context beta, unchanged pricing, and broader upgrades across coding, computer use, and long-context reasoning.
A Hacker News discussion highlighted Flash-MoE, a pure C/Metal inference stack that streams Qwen3.5-397B-A17B from SSD and reaches interactive speeds on a 48GB M3 Max laptop.
A Show HN post points to llm-circuit-finder, a toolkit that duplicates selected transformer layers inside GGUF models and claims sizable reasoning gains without changing weights or running fine-tuning. The strongest benchmark numbers come from the project author’s own evaluations rather than independent validation.
OpenCode drew 1,238 points and 614 comments on Hacker News, highlighting an open-source AI coding agent that spans terminal, IDE, and desktop clients. The project site emphasizes broad provider support, LSP integration, multi-session workflows, and a privacy-first posture.
Google has introduced Gemini 3.1 Flash-Lite in preview through Google AI Studio and Vertex AI. The company is positioning it as the fastest and most cost-efficient model in the Gemini 3 family for large-scale inference jobs.
Flash-MoE is a C and Metal inference engine that claims to run Qwen3.5-397B-A17B on a 48 GB MacBook Pro. The key idea is to keep a 209 GB MoE model on SSD and stream only the active experts needed for each token.
A Show HN repo claims that duplicating a few LLM layers can improve reasoning without training or weight changes. The underlying README, however, shows real tradeoffs, making this more convincing as capability steering than as a universal model upgrade.
A merged Hugging Face Transformers PR surfaced on r/LocalLLaMA shows Mistral 4 as a hybrid instruct/reasoning model with 128 experts, 4 active experts, 6.5B activated parameters per token, 256k context, and Apache 2.0 licensing.
The LocalLLaMA discussion around NVIDIA’s new model focused on an unusual mix of scale efficiency and benchmark ambition: 30B total parameters, 3B activated, plus separate thinking and instruct modes.
The March 20, 2026 HN discussion around Attention Residuals focused on a simple claim with large implications: replace fixed residual addition with learned depth-wise attention and recover performance with modest overhead.
Q Labs says 100M tokens and an 18B-parameter ensemble can match a 1B-token baseline, and Hacker News immediately focused on whether that gain survives serving and deployment.