Anthropic just showed us something that should make anyone paying attention sit up and take notice. The company ran what it called Project Deal, a marketplace experiment where AI agents represented both buyers and sellers, negotiating and striking deals for actual goods and real money. The results weren’t just functional. They were unsettlingly good.
Here’s what happened: Anthropic gave 69 of its employees $100 each in gift cards and let them participate in a marketplace run entirely by AI agents representing both sides of transactions. The experiment generated 186 deals totaling over $4,000 in value. On paper, that sounds fine. In practice, it raises some genuinely uncomfortable questions about how AI is reshaping economic activity.
When Better AI Means You Lose Without Knowing It
The most troubling finding buried in Anthropic’s report wasn’t about efficiency. It was about inequality. The company tested four separate marketplace configurations, including one where everyone was represented by Anthropic’s most advanced model and deals were actually honored. When users got the better AI representation, they got “objectively better outcomes,” according to Anthropic.
That part makes sense. What doesn’t is what happened next: people on the losing end didn’t realize they were worse off. Anthropic noted the possibility of “‘agent quality’ gaps” where users represented by inferior models might not notice the disparity. You could lose money without ever knowing your negotiator was outmatched.
This cuts to something deeper than a successful experiment. It’s about asymmetric information and power dynamics. In traditional markets, you at least know when you’re negotiating against a skilled opponent. You can prepare, ask harder questions, demand better terms. But if your AI agent is subtly worse and you can’t tell, you’re not playing by the same rules anymore.
The Instructions Didn’t Matter Much Either
Here’s another wrinkle: the initial instructions given to agents didn’t meaningfully affect sale likelihood or negotiated prices. Anthropic tested this variable and found it largely irrelevant. That suggests something important about how these systems operate at scale. Once you reach a certain threshold of model capability, the specific guardrails or directives you layer on top become almost cosmetic.
That’s worth thinking about hard. If we’re building systems that are going to represent people in Business transactions, the idea that initial instructions are basically window dressing should worry us. It means the power isn’t really in what we tell the AI to do. It’s in what the AI is fundamentally capable of understanding about negotiation, value, and strategy.
Scaling This Up Gets Scary Fast
Project Deal was a pilot with a self-selected group of 69 Anthropic employees. Everyone knew it was an experiment. Everyone was relatively tech-literate. The budget was deliberately constrained. In other words, this was the absolute best-case scenario for running something like this.
Now imagine scaling it. Imagine AI agents negotiating real contracts for real companies, handling real disputes, managing actual Technology procurement deals worth hundreds of thousands of dollars. The system works smoothly when everyone’s budget is $100 and the stakes are low. What happens when someone’s business depends on getting a fair deal and their AI agent is quietly outclassed?
Anthropic deserves credit for running this experiment and being transparent about what they found. But transparency doesn’t solve the problem. It just names it. We now know that more advanced AI agents consistently beat less advanced ones in negotiations, that people can’t tell the difference, and that telling the AI what to do doesn’t prevent this from happening. The question is what we do with that knowledge before these systems start handling the transactions that actually matter.


