Reddit Reconsiders Function Calling After a Manus Engineer's Unix-Style Agent Post
Original: I was backend lead at Manus. After building agents for 2 years, I stopped using function calling entirely. Here's what I use instead. View original →
The core argument
A high-signal r/LocalLLaMA post from a former Manus backend lead makes a sharp claim: after two years building agents, the author stopped using typed function catalogs and prefers a single run(command="...") tool backed by Unix-style commands. The post connects two older design traditions. Unix assumes everything important can move through text streams, while LLMs only reason over tokens. From that angle, a shell-like interface is not a workaround. It is the interface model most native to current models.
The author argues that a large function inventory forces the model to spend effort on tool selection rather than task completion. A single command runner instead turns tool use into string composition inside one namespace. The post gives concrete examples: instead of separate read, search, and count tools, one command chain can read a log, filter errors, and count matches. To make that practical, the runtime adds parsing for |, &&, ||, and ;, so one tool call can represent a complete workflow rather than a tiny atomic action.
Why the post resonated
The discussion became more interesting because the author did not stop at philosophy. The post describes progressive help discovery, error messages that tell the agent what to do next, consistent output footers such as exit codes and duration, and a two-layer architecture that separates raw Unix execution from what is actually shown back to the model. That presentation layer adds binary detection, overflow files for very large outputs, and stderr visibility when commands fail. In other words, the thesis is not simply give the model a terminal and hope for the best. It is that CLI becomes effective when you shape the surrounding ergonomics for model cognition.
Readers in the thread responded to both the power and the risk of that idea. Several commenters said the shell may indeed end up being the most natural agent framework. Others immediately pointed at the obvious tradeoff: once everything funnels through a terminal-like runner, sandboxing and permission boundaries become even more critical. That concern does not contradict the main thesis. It reinforces it. If the command line is the most expressive interface, then the safety model around that interface becomes the real product.
Why this matters beyond one Reddit post
The author links the argument to open-source code in agent-clip, including command routing, chain parsing, memory, browser attachment, and stderr handling. That makes the post more useful than a generic opinion thread. It is a field report from someone who has seen production failures and reworked the interface around them. The broader takeaway is that the current function-calling debate is no longer just about schema quality. It is about what kind of operating environment best matches how LLMs actually perceive tools, state, and failure.
Source post: r/LocalLLaMA thread. Primary source referenced in the post: agent-clip.
Related Articles
Anthropic introduced Claude Sonnet 4.6 on February 17, 2026, adding a beta 1M token context window while keeping API pricing at $3/$15 per million tokens. The company says the new default model improves coding, computer use, and long-context reasoning enough to cover more work that previously pushed users toward Opus-class models.
A high-engagement r/LocalLLaMA thread tracked the MiniMax-M2.5 release on Hugging Face. The model card emphasizes agentic coding/search benchmarks, runtime speedups, and aggressive cost positioning.
OpenAI announced an Operator upgrade adding Google Drive slides creation/editing and Jupyter-mode code execution in Browser. It also said Operator availability expanded to 20 additional regions in recent weeks, with new country additions including Korea and several European markets.
Comments (0)
No comments yet. Be the first to comment!