Google says Gemini in Sheets reaches state-of-the-art 70.48% on SpreadsheetBench
Original: R to @GoogleWorkspace: While we don't have favorites, the evolution of Gemini in Google Sheets might be our most impressive yet. Gemini in Google Sheets has achieved a state-of-the-art benchmark, achieving a 70.48% success rate on the full SpreadsheetBench dataset. This performance not only exceeds competitors but nears human expert ability. We accomplished this by equipping Gemini with better verbalization and enhanced coding capabilities. With these, Gemini can now natively build complex models and dashboards, solve your most complex optimization problems, and verify its own work for expert-level precision. Read more about Sheets here: https://blog.google/products-and-platforms/products/workspace/gemini-google-sheets-state-of-the-art View original →
What Google highlighted on X
On March 10, 2026, Google AI said Gemini in Google Sheets reached 70.48% on the full SpreadsheetBench dataset. Google characterized that result as state of the art and said it comes close to human expert performance, which is a more ambitious claim than the usual spreadsheet-assistant language around formulas or autofill.
The company also used the X thread to place Sheets inside a broader Workspace strategy. In the lead post, Google said Gemini had to be tailored to each product's data model and user needs across Docs, Sheets, Slides, and Drive. The Sheets-specific post suggests Google is pushing beyond generic chat inside productivity apps toward domain-tuned behavior for structured business work.
What the Google blog adds
Google's supporting blog post says the 70.48% score was achieved by equipping Gemini in Sheets with stronger verbalization and coding capabilities. According to Google, that allows the model to natively build complex models and dashboards, solve optimization problems, and verify its own work rather than only generate spreadsheet formulas on request.
- Google says the result exceeds competing systems on the full SpreadsheetBench benchmark.
- The company also says the score approaches human expert ability on spreadsheet tasks.
- Google frames the improvement as part of product-specific model adaptation inside Workspace, not just a general Gemini model upgrade.
Why this matters for enterprise workflows
Spreadsheet tooling is one of the clearest enterprise tests for whether AI can move from assistant behavior to delegated analytical work. Formula suggestions are useful, but real business value appears when a system can understand a table, reason through a goal, construct a model, and check whether the output is actually coherent. That is the transition Google is signaling here.
If the benchmark gains hold up in production, the implication is broader than Sheets alone. Product-specific tuning inside Workspace suggests Google sees AI productivity gains as a data-model problem as much as a language-model problem. In practice, that means future competition may revolve less around generic model IQ and more around how deeply vendors integrate with the native structures of documents, spreadsheets, presentations, and knowledge stores.
Sources: Google AI X post, Google Workspace blog
Related Articles
Google announced a major Gemini 3 Deep Think upgrade with stronger reasoning benchmarks and early API access for researchers and enterprises.
Google DeepMind said on March 3, 2026 that Gemini 3.1 Flash-Lite delivers faster performance at a lower price than Gemini 2.5 Flash. Google is rolling the model out in preview via Google AI Studio and Vertex AI for high-volume, latency-sensitive workloads.
Google AI shared practical Gemini 3.1 Flash-Lite examples, including high-volume image sorting and business automation scenarios. The thread also points developers to preview access via Gemini API, Google AI Studio, and Vertex AI.
Comments (0)
No comments yet. Be the first to comment!