HN Reads the Erdős Breakthrough as Human-Guided AI Math, Not Magic
Original: Amateur armed with ChatGPT solves an Erdős problem View original →
The headline was irresistible, but HN refused to read it as a simple AI solved math moment. Scientific American reported that 23-year-old Liam Price, working with ChatGPT Pro, produced a solution to a 60-year-old Erdős problem about primitive sets. The article argues that this case matters because the route appears genuinely new, and because mathematicians who reviewed it think the idea may have uses beyond this one problem.
That last part is what gave the story weight. The article quotes Terence Tao saying the problem may have been easier than expected once the first move was seen correctly, and Stanford mathematician Jared Lichtman says the proof's underlying connection could travel further. This is a stronger claim than the usual AI-math story, which often collapses into either benchmark theater or a rediscovery of tools experts already know.
HN still kept its distance from the hype. One commenter linked the chat and said the raw proof output was not what impressed them; the interesting part was whether a language model could propose a route humans had not taken. Another highlighted a quote from the article that the raw output was actually poor and had to be sifted by an expert. That framed the result less as autonomous theorem proving and more as a hybrid workflow: machine proposes, human interprets, experts verify.
Several commenters also noted that Erdős problems cover a wide range of difficulty, so the label alone should not be treated as a universal score for mathematical intelligence. HN's stronger test is narrower and more practical: did the model contribute a non-obvious idea that survives expert scrutiny? In this case, the answer looks more interesting than in many previous rounds of AI-math hype, precisely because the community can point to a novel angle instead of just a flashy headline.
So the mood was skeptical, but not dismissive. HN is clearly willing to credit AI when it helps uncover a fresh line of attack. It just does not want the human role erased from the story. The source article is Scientific American, and the discussion is in the HN thread.
Related Articles
OpenAI says ChatGPT is already being used at research scale across science and mathematics. In its January 2026 report, the company says advanced science and math usage reached nearly 8.4 million weekly messages from roughly 1.3 million weekly users, with early evidence that GPT-5.2 is contributing to serious mathematical work.
The paper drew attention because it challenges today’s data appetite, but the comments quickly tested the comparison to children.
r/MachineLearning found the 1,200-paper list useful, but the thread immediately separated “has a link” from “can reproduce the result.” Comments pointed to missing papers, 404s, and the gap between public code and runnable research.
Comments (0)
No comments yet. Be the first to comment!