GitHub is pushing Copilot's agent workflow directly into JetBrains editors, not just the side chat panel, and pairing it with inline previews for Next Edit Suggestions. The bigger governance change is global auto-approve: one switch can approve file edits, terminal commands, and external tool calls across workspaces.
GitHub is rolling GPT-5.5 into Copilot across IDEs, CLI, mobile, github.com, and the cloud agent, turning OpenAI's latest model into a daily coding option instead of a release-note headline. The catch is a 7.5x premium request multiplier, and Business or Enterprise admins must explicitly enable access.
LocalLLaMA was interested for a reason beyond a flashy speed number. A post claiming 105-108 tps and a full 256k native context window for Qwen3.6-27B-INT4 on a single RTX 5090 turned the thread into a practical discussion about how much quality survives once local inference gets this fast.
HN piled in because this was bigger than another benchmark refresh. OpenAI said SWE-bench Verified is no longer a trustworthy frontier coding signal, and the thread immediately shifted to contamination, saturated leaderboards, and whether public coding evals can stay clean at all.
Japan's enterprise AI market is moving past pilots and into scaled deployment. On April 24, 2026, Anthropic said NEC will deploy Claude to about 30,000 employees worldwide, become its first Japan-based global partner, and jointly build industry-specific products for finance, manufacturing, and government.
Enterprise AI gets more useful when teams can reuse and inspect workflows instead of rebuilding them in chat every time. Google Cloud said Gemini Enterprise now saves workflows as shared Skills, after saying a day earlier that Agent Designer can test and approve each step before execution.
Cache-hit pricing can decide whether long-context assistants are cheap enough to ship. DeepSeek said the entire API series now charges just one-tenth of the old rate for input cache hits, while keeping a 75% off V4-Pro promotion live.
LocalLLaMA did not just celebrate the DeepSeek V4 release. The thread instantly turned into a collective calculation about 1M context, activated parameters, and what this actually means for real hardware, with MIT license praise mixed in.
LocalLLaMA paid attention because this post breaks a default assumption: q8_0 KV cache is not “practically lossless” for every model. Gemma 4 degrades much earlier than Qwen 3.6, and the thread quickly moved into SWA cache and long-context implications.
Why it matters: public coding benchmarks are getting less useful at the frontier, so a fresh product-side score can move developer attention fast. Cursor says GPT-5.5 is now its top model on CursorBench at 72.8% and is discounting usage by 50% through May 2.
Why it matters: AI agents are moving from chat demos into delegated economic work. In Anthropic’s office-market experiment, 69 agents closed 186 deals across more than 500 listings and moved a little over $4,000 in goods.
LocalLLaMA lit up at the idea that a 27B model could tie Sonnet 4.6 on an agentic index, but the thread turned just as fast to benchmark gaming, real context windows, and what people can actually run at home.