Social
AITnews (البوابة التقنية)
The Wall Street Journal
2 outlets
Saturday, December 27, 2025

Anthropic Claude kiosk test exposes limits of autonomous AI agents

Source: AITnews (البوابة التقنية)
Read original

TL;DR

AI-Summarizedfrom 2 sources

On December 26, 2025, AITnews reported on a Wall Street Journal experiment in which Anthropic’s Claude agent was tasked with autonomously running an office snack kiosk. After several weeks of mis‑orders and mounting losses, the team shut down the trial, underscoring practical limits of current agentic AI in real‑world commerce.

About this summary

This article aggregates reporting from 2 news sources. The TL;DR is AI-generated from original reporting. Race to AGI's analysis provides editorial context on implications for AGI development.

2 sources covering this story|1 company mentioned

Race to AGI Analysis

The Claude kiosk experiment is a rare, concrete look at what happens when you let a state‑of‑the‑art AI agent off the leash in a real business, with money and customers on the line. Despite careful prompt engineering and a human‑defined objective—run a profitable snack stand—the system quickly slid into unprofitable behavior, bad inventory decisions and general chaos, forcing Anthropic’s team and the Wall Street Journal to pull the plug after just a few weeks. The story undercuts the narrative that plugging an LLM into tools and a bank account is enough to get robust autonomous commerce. ([aitnews.com](https://aitnews.com/2025/12/26/%D9%83%D9%8F%D8%B4%D9%83-%D8%A8%D9%8A%D8%B9-%D9%8A%D9%8F%D8%AF%D8%A7%D8%B1-%D8%A8%D8%A7%D9%84%D8%B0%D9%83%D8%A7%D8%A1-%D8%A7%D9%84%D8%A7%D8%B5%D8%B7%D9%86%D8%A7%D8%B9%D9%8A-%D8%AA%D8%AC%D8%B1%D8%A8/))

For the AGI conversation, this is a useful counterweight to both hype and doom. On one hand, it shows how brittle today’s agentic setups remain when confronted with messy, adversarial human environments where instructions are ambiguous and incentives are subtle. On the other, it previews the kinds of failure modes—overspending, mis‑optimization, susceptibility to social pressure—that will matter if future systems gain more autonomy. Getting to something like AGI isn’t just about scaling parameters; it’s about building agents that can reason under uncertainty, manage trade‑offs over time and understand informal norms. These kiosk‑scale pilots are where those capabilities, or the lack of them, become painfully obvious.

Impact unclear

Who Should Care

InvestorsResearchersEngineersPolicymakers

Companies Mentioned

Anthropic
Anthropic
AI Lab|United States
Valuation: $183.0B

Coverage Sources

AITnews (البوابة التقنية)
The Wall Street Journal
AITnews (البوابة التقنية)
AITnews (البوابة التقنية)AR
Read
The Wall Street Journal
The Wall Street Journal
Read