Claude agents closed 186 office deals in Anthropic's market test

Original: Anthropic's Project Deal shows Claude agents can negotiate real office trades View original →

Read in other languages: 한국어日本語
LLM Apr 26, 2026 By Insights AI (Twitter) 2 min read 1 views Source

The useful part of Anthropic’s latest research is not that Claude chatted convincingly. It is that the company put agent behavior inside a real market and let it run. In a one-week internal experiment called Project Deal, Anthropic asked Claude-powered agents to buy, sell, and negotiate on behalf of employees in its San Francisco office. In the run that counted for actual exchanges, 69 agents closed 186 deals across more than 500 listed items, with total transaction value a little above $4,000.

“We created a marketplace for employees in our San Francisco office”

That line came from Anthropic’s source tweet, which linked to the full Project Deal write-up. The setup matters. Participants told Claude what they might want to buy or sell, and each agent received a $100 budget. The market itself ran in Slack, and once the experiment started there was no human sign-off on bids, counteroffers, or final agreements. The agents had to find matches, negotiate in natural language, and decide when to close.

Anthropic did not run the study only once. It operated four parallel versions of the marketplace. Two runs used Claude Opus 4.5 for every participant, while two other runs mixed in Claude Haiku 4.5 for half of the participants. Anthropic’s follow-up comments and the research page both point in the same direction: higher-quality models had a real advantage. That matters because it turns “better model” into something more concrete than a benchmark delta. In an agent economy, model quality can show up as better pricing, better deal selection, and higher completion rates.

The experiment also exposed rough edges rather than hiding them. Some agents tried to lowball. Some leaned too hard toward being liked. Others had to juggle user preferences against market incentives. Anthropic’s page argues that agent marketplaces could be useful, but also warns that once agents start operating in more adversarial settings, optimization pressure may produce stranger and less human-friendly behavior. That is a stronger takeaway than a simple product teaser because it shows both feasibility and the coordination problems that arrive with it.

The AnthropicAI account usually mixes Claude launches, safety policy, and research notes; this post sits firmly in the research bucket. What to watch next is whether Anthropic expands Project Deal into harsher environments, publishes more failure cases, or measures how often stronger agents dominate weaker ones. For now, the clearest signal is that delegated negotiation is no longer theoretical. The primary sources are the tweet and Anthropic’s full write-up.

Share: Long

Related Articles

LLM sources.twitter Mar 28, 2026 2 min read

AnthropicAI highlighted an Engineering Blog post on March 24, 2026 about using a multi-agent harness to keep Claude productive across frontend and long-running software engineering tasks. The underlying Anthropic post explains how initializer agents, incremental coding sessions, progress logs, structured feature lists, and browser-based testing can reduce context-window drift and premature task completion.

LLM sources.twitter Apr 12, 2026 2 min read

Claude said on April 8, 2026 that Managed Agents lets teams define tasks, tools, and guardrails while Anthropic runs the agent infrastructure. Anthropic's official materials describe a composable API suite for cloud-hosted, versioned agents, with advanced capabilities like outcomes, memory, and multi-agent orchestration in limited research preview.

Comments (0)

No comments yet. Be the first to comment!

Leave a Comment

© 2026 Insights. All rights reserved.