The paper drew attention because it challenges today’s data appetite, but the comments quickly tested the comparison to children.
#machine-learning
RSS Feedr/MachineLearning reacted because the sample was small but painfully familiar: one user said 4 of 7 paper claims they checked this year did not reproduce, with 2 still sitting as unresolved GitHub issues. The comments moved from resignation about reviewers not running code to concrete demands for submission-time reproducibility reports.
r/MachineLearning did not latch onto this as empty safety branding; the hook was a concrete trade-off claim around calibration, abstention, and base accuracy. The author says HALO-Loss keeps CIFAR-10/100 accuracy roughly flat, cuts ECE to 1.5%, and drops SVHN FPR@95 from 22.08% to 10.27%.
A Reddit discussion in r/MachineLearning highlighted TorchLean, a framework that aligns neural network execution and verification semantics in Lean 4. The approach combines a PyTorch-style verified API, explicit Float32 modeling, and IBP/CROWN-style certificate-backed verification for safety-critical ML workflows.
Researchers have demonstrated that transformer models with fewer than 100 parameters can add two 10-digit numbers with 100% accuracy using digit tokenization, challenging assumptions about the minimum complexity needed for arithmetic reasoning.
Researchers have demonstrated that transformer models with fewer than 100 parameters can add two 10-digit numbers with 100% accuracy. The key ingredient is digit tokenization rather than treating numbers as opaque strings — a finding with implications for mathematical reasoning in larger LLMs.
Professor Zico Kolter's 10-202: Introduction to Modern AI at Carnegie Mellon University is now available online for free, including lecture videos, assignments, and autograded submissions — with a 2-week delay from the in-person course.
Professor Zico Kolter's 10-202: Introduction to Modern AI at Carnegie Mellon University is now available online for free, including lecture videos, assignments, and autograded submissions — with a 2-week delay from the in-person course.
MLU-Explain's interactive visualization demonstrates why decision trees remain one of the most powerful and interpretable tools in ML, showing how simple nested if-else rules form the foundation of modern ensemble methods.
A highly upvoted r/MachineLearning thread debates whether skyrocketing acceptance rates at top venues like CVPR and ICLR are diluting the academic value of conference publication, raising concerns about review quality.
A high-engagement Reddit post summarized 2025 ML competition patterns across major platforms. The author reports tracking roughly 400 contests and first-place solution details for 73, highlighting shifts in tooling, model choices, and compute budgets.