Anthropic put hard numbers on Claude's biology capability claims instead of vague lab hype. In 99 real-data bioinformatics problems, the company says experts were stumped on 23 and recent Claude models solved roughly 30% of that hardest slice.
Anthropic put hard numbers on Claude's biology capability claims instead of vague lab hype. In 99 real-data bioinformatics problems, the company says experts were stumped on 23 and recent Claude models solved roughly 30% of that hardest slice.
The subreddit jumped straight past the headline and into the hard question: was this finally something other than pattern replay? A Scientific American report on a 23-year-old using GPT-5.4 Pro on a 60-year-old Erdos problem sparked debate over novelty, expert cleanup, and whether messy model output can still contain a real mathematical idea.
Google DeepMind is tying frontier models directly to a national research agenda. On April 27, 2026, it said Korea will get a new AI Campus in Seoul plus joint work with SNU, KAIST, and three AI Bio Innovation Hubs.
HN read this math story less as another "AI did it" headline and more as a case where a model pointed at a route humans had not tried. The part that stuck was the expert cleanup work after the GPT-5.4 Pro draft, not the one-shot prompt itself.
This turns Google DeepMind’s science stack into a named national partnership instead of another generic AI pledge. The plan starts with a Seoul AI Campus, work with SNU and KAIST, and an AlphaFold base already used by more than 85,000 researchers in Korea.
Hacker News latched onto this paper because it was not selling a new benchmark or model, but a bigger claim: deep learning may finally be mature enough for a real scientific theory. That mix of excitement and skepticism kept the thread moving.
HN did not treat the Erdős headline as proof of autonomous math genius; the thread kept circling back to expert cleanup, problem selection, and whether the new method generalizes.
r/MachineLearning pushed this paper up because it did not promise a miracle. It argued that deep learning theory is finally accumulating enough converging evidence to resemble a genuine scientific program, and commenters liked the paper's concrete framing more than another grand AI manifesto.
r/MachineLearning found the 1,200-paper list useful, but the thread immediately separated “has a link” from “can reproduce the result.” Comments pointed to missing papers, 404s, and the gap between public code and runnable research.
The paper drew attention because it challenges today’s data appetite, but the comments quickly tested the comparison to children.
OpenAI put GPT-Rosalind into research preview for qualified life-science teams, pairing a domain model with a Codex plugin that connects to more than 50 tools and data sources. The strongest signal is not the branding: OpenAI says best-of-ten submissions ranked above the 95th percentile of human experts on one Dyno Therapeutics RNA prediction task.
r/MachineLearning reacted because the sample was small but painfully familiar: one user said 4 of 7 paper claims they checked this year did not reproduce, with 2 still sitting as unresolved GitHub issues. The comments moved from resignation about reviewers not running code to concrete demands for submission-time reproducibility reports.