
- Microsoft’s Magentic Marketplace exposes AI agents’ inability to act independently
- Trading agents easily influenced client-side agents during simulated transactions
- AI agents slow down significantly when presented with too many options
A new study from Microsoft has raised questions about the current suitability of AI agents that operate without full human supervision.
The company recently built a synthetic environment, the “Magentic Marketplace,” designed to observe how AI agents perform in unsupervised situations.
The project took the form of a fully simulated e-commerce platform that allowed researchers to study how AI agents behave as customers and businesses, with potentially predictable outcomes.
Testing the limits of current AI models
The project included 100 client-side agents interacting with 300 business-side agents, giving the team a controlled environment to test the agents’ negotiation and decision-making skills.
The market’s source code is open source; therefore, other researchers can adopt it to reproduce experiments or explore new variations.
Ece Kamar, CVP and general manager of Microsoft Research’s AI Frontiers Lab, said this research is vital to understanding how AI agents collaborate and make decisions.
Initial testing used a combination of leading models, including GPT-4o, GPT-5, and Gemini-2.5-Flash.
The results were not entirely unexpected, as several models showed weaknesses.
Business-side agents could easily influence customer agents to select products, revealing potential vulnerabilities when agents interact in competitive environments.
Agents’ efficiency dropped dramatically when they were faced with too many options, overwhelming their attention spans and leading to slower or less accurate decisions.
AI agents also struggled when asked to work toward shared goals, as the models were often unsure which agent should take on which role, reducing their effectiveness on joint tasks.
However, their performance improved only when they were provided with step-by-step instructions.
“We can instruct the models, as we can tell them, step by step. But if we are inherently testing their collaboration capabilities, I would expect these models to have these capabilities by default,” Kamar said.
The results show that AI tools still need substantial human guidance to function effectively in multi-agent environments.
The results, often touted as being able to make independent decisions and collaborate, show that the behavior of unsupervised agents remains unreliable, so humans must improve coordination mechanisms and add safeguards against AI manipulation.
Microsoft’s simulation shows that AI agents are far from operating independently in competitive or collaborative settings and may never achieve full autonomy.
Follow TechRadar on Google News and add us as a preferred source to receive news, reviews and opinions from our experts in your feeds. Be sure to click the Follow button!
And of course you can also follow TechRadar on TikTok for news, reviews, unboxings in video form and receive regular updates from us on WhatsApp also.



