On February 26, 2026 (UTC), Google DeepMind said on X that Nano Banana 2 can turn instructions into data-rich infographics and educational diagrams. The post also emphasized Gemini world knowledge and real-time web-grounded generation.
On February 26, 2026 (UTC), Google DeepMind said on X that Nano Banana 2 can turn instructions into data-rich infographics and educational diagrams. The post also emphasized Gemini world knowledge and real-time web-grounded generation.
On February 25, 2026 (UTC), Anthropic said on X that Opus 3 will write on Substack for at least the next three months. The post drew strong traction with roughly 1.22M views and more than 4,000 likes.
Azure posted on February 25, 2026 that three new Azure OpenAI models are rolling out in Microsoft Foundry. Microsoft positions the release for low-latency voice systems and long-running engineering workflows with published pricing and performance claims.
Perplexity announced on February 26, 2026 that `pplx-embed-v1` and `pplx-embed-context-v1` are now available in 0.6B and 4B variants. The company positions the release as retrieval-first infrastructure with quantized embeddings and benchmark-focused performance claims.
A r/LocalLLaMA post reports a from-scratch 144M-parameter Spiking Neural Network language model experiment named Nord. The author claims 97-98% inference sparsity, STDP-based online updates, and better prompt-level topic retention than GPT-2 Small on limited examples, while clearly noting current loss and benchmark limitations.
OpenAI and Paradigm launched EVMbench, a benchmark for AI agent performance on smart contract detection, patching, and exploitation tasks. OpenAI reports GPT-5.3-Codex scored 72.2% in exploit mode versus 31.9% for GPT-5.
OpenAI and Figma launched a new integration that links Codex directly with Figma through an MCP-based workflow. The goal is to reduce context loss between implementation and design by enabling continuous code-to-canvas roundtrips.
A trending Reddit post in r/singularity points to OpenAI's statement that it no longer evaluates on SWE-bench Verified, citing at least 16.4% flawed test cases. The announcement reframes how coding-model benchmark scores should be interpreted in production decision-making.
Anthropic announced it is acquiring Vercept to strengthen Claude's computer use stack. The move pairs model-level capability gains with deeper perception-and-interaction expertise for multi-step execution inside live software environments.
A trending r/LocalLLaMA thread highlighted the DualPath paper on KV-Cache bottlenecks in disaggregated inference systems. The arXiv abstract reports up to 1.87x offline throughput and 1.96x average online throughput gains while meeting SLO.
A high-engagement Hacker News thread (388 points, 535 comments) on Benedict Evans' OpenAI analysis focused on defensibility beyond raw model quality. Users debated stickiness, distribution leverage, and enterprise integration as the real battleground.
Google announced on 2026-02-25 that Gemini in Android will begin handling multi-step tasks in beta. The rollout starts on Pixel 10 devices and Samsung Galaxy S26 series, initially in the U.S. and Korea.