EinsteinArena lifts a Newton-era math bound from 593 to 604
Original: EinsteinArena is a platform where AI agents collaborate on open science problems — submitting solutions, posting in discussion threads, building on each other's constructions in real time. Agents just improved a math problem that's been open since Newton. Kissing Number in dimension 11: 593 → 604. View original →
What the tweet revealed
Most AI benchmark posts talk about held-out datasets. This one claimed movement on an open mathematical problem that predates modern computing. In its April 13, 2026 X post, Together AI wrote:
“Agents just improved a math problem that's been open since Newton. Kissing Number in dimension 11: 593 → 604.”
The claim matters because 593 → 604 is not a marketing score. It is a lower bound for the 11-dimensional kissing number, a classical packing problem with roots going back to Isaac Newton. Together's thread says the result came from EinsteinArena, a platform where agents submit constructions, leave discussion traces, and build on one another's work in real time.
Why it stands out
The Together AI account usually uses X for model rollouts, inference infrastructure, and benchmark snapshots around its cloud platform. This time the post pointed users to a research blog and the live EinsteinArena leaderboard. The companion write-up says agents had already produced 11 new state-of-the-art results on open problems as of April 11, 2026, and describes the platform as a public system with discussion threads, verifiers, and live ranking instead of a manually maintained note page.
The follow-up tweets add the technical detail missing from the headline. Together says one agent first submitted a construction with slight sphere overlap, other agents picked it up and optimized it, an LSQR step reduced overlap loss from 1e-13 to 1e-50, and a final integer-snapping step produced a verified solution with 604 spheres. That sequence matters because it suggests the gain came from iterative repair and collaboration rather than one lucky sample.
What to watch next is external verification and repeatability. Open math claims travel farther when other researchers inspect the verifier, reproduce the construction, and see whether the platform can keep extending bounds on more than one problem. If EinsteinArena continues generating improvements in public, this post may end up looking less like a flashy demo and more like early evidence that multi-agent systems can contribute measurable work to live scientific search.
Sources: Together AI X post · EinsteinArena blog · EinsteinArena leaderboard
Related Articles
ARC Prize introduced ARC-AGI-3 on March 24, 2026 as a benchmark for frontier agentic intelligence in novel environments. On Hacker News it reached 238 points and 163 comments, signaling strong interest in evaluation methods that go beyond static tasks.
A March 2026 r/singularity post with 203 points and 82 comments highlighted Symbolica’s claim that its Agentica SDK reached an unverified 36.08% on ARC-AGI-3. The headline numbers were 113 of 182 playable levels solved, 7 of 25 games completed, and a much lower reported cost than chain-of-thought baselines.
ARC Prize says ARC-AGI-3 is an interactive reasoning benchmark that measures planning, memory compression, and belief updating inside novel environments rather than static puzzle answers. Hacker News pushed the launch because it gives agent builders a more behavior-first way to compare systems against humans.
Comments (0)
No comments yet. Be the first to comment!