Tiny Transformers with Under 100 Parameters Achieve 100% Accuracy on 10-Digit Addition

Original: [R] Tiny transformers (<100 params) can add two 10-digit numbers to 100% accuracy View original →

Read in other languages: 한국어日本語
LLM Mar 3, 2026 By Insights AI (Reddit) 1 min read 5 views Source

Surprising Math Ability from Tiny Models

A fascinating research project shared on r/MachineLearning (144 upvotes) demonstrates that transformer models with fewer than 100 parameters can achieve 100% accuracy when adding two 10-digit numbers. The work, published as the AdderBoard project on GitHub, challenges assumptions about model scale and arithmetic capability.

The Key: Digit Tokenization

The secret behind this performance is digit tokenization — treating each individual digit as a separate token rather than processing numbers as whole units. This allows the model to learn the carry-over rules of arithmetic addition much more effectively, as each positional step becomes a learnable unit. Community members noted that this representation choice is essential: floating-point arithmetic would be dramatically harder.

Why This Matters

Large language models with billions of parameters frequently make errors on simple arithmetic. The fact that a model with under 100 parameters can perfectly solve 10-digit addition highlights that scale is not the only variable that matters — how data is represented and what the model is asked to learn are equally critical design choices.

Limitations and Future Work

The researchers note that while this approach works extremely well for integer addition, floating-point arithmetic presents a much harder challenge due to the increased complexity of number representation. This work opens new directions for understanding how to make AI models more reliably numerically accurate at minimal parameter cost.

Share:

Related Articles

LLM Reddit Mar 2, 2026 1 min read

Researchers have demonstrated that transformer models with fewer than 100 parameters can add two 10-digit numbers with 100% accuracy. The key ingredient is digit tokenization rather than treating numbers as opaque strings — a finding with implications for mathematical reasoning in larger LLMs.

Comments (0)

No comments yet. Be the first to comment!

Leave a Comment

© 2026 Insights. All rights reserved.