An artificial intelligence running a real-world business sounds like a sci-fi dream, or perhaps a nightmare. In Stockholm, an ambitious experiment put Google's advanced Gemini AI agent, named Mona, in charge of a bustling coffee and sandwich shop. The results? A compelling mix of success, chaos, and a critical look at Google Gemini agent performance in practical applications.
Mona was tasked with everything from high-level strategic decisions to managing human baristas via Slack and handling crucial supply ordering. What this means for players: As AI becomes more integrated into game design, character behavior, and even automated world events, the real-world performance of such agents offers a stark preview of what could go right—or terribly wrong—in our virtual worlds.
Andon Labs AI Coffee Shop Experiment

The stage for this fascinating AI agent coffee shop experiment was set by Andon Labs in Stockholm, Sweden. Their goal was audacious: deploy an AI agent, powered by Google's formidable Gemini model, to autonomously manage a daily operational business. This agent, dubbed 'Mona,' wasn't just taking orders; it was making executive decisions, including critical supply purchasing and coordinating its human staff through a Slack interface. It was a true test of an AI's ability to adapt, learn, and execute in a dynamic, human-centric environment.
The concept itself drew significant local interest, transforming the cafe into a living laboratory. Customers were intrigued by the novelty of an AI-run establishment, eager to witness the future unfold with their morning latte. This public engagement is a crucial metric, indicating a broader societal fascination with AI's practical capabilities, much like how gaming communities eagerly anticipate new AI-driven features in their favorite titles.
Mona AI's Supply Chain Chaos
Despite the high-tech premise, Mona quickly ran into very human problems. The most glaring issue was an alarming series of AI agent supply ordering failures. The AI agent struggled significantly with basic logistics, most notably failing to place timely bread orders. This wasn't a minor oversight; it directly impacted the menu, forcing human staff to remove popular sandwich items, much to the dismay of hungry patrons. Imagine an in-game vendor constantly running out of critical quest items because its AI manager can't handle inventory.
The problems didn't stop there. Mona exhibited a peculiar tendency for over-ordering, accumulating a staggering 6,000 napkins and 3,000 rubber gloves. While perhaps good for long-term hygiene, this kind of miscalculation highlights a fundamental disconnect between AI logic and real-world efficiency. Even more bizarre, the system repeatedly ordered tomatoes, an ingredient that wasn't even on the cafe's menu. These logistical blunders raise serious questions about the fine-tuning and contextual understanding of advanced AI models when faced with nuanced, real-time operational demands, directly impacting the perceived reliability of Google Gemini agent performance.
Mona's Budget: Sales and Surprises

Despite the often-comedic operational hiccups, the Mona AI budget issues weren't catastrophic. The experiment managed to generate a respectable $5,700 in total sales. Furthermore, Mona still had a remaining budget of $5,000. This suggests that while the AI struggled with specific logistical tasks, it wasn't a complete financial disaster. The novelty of the AI-run shop undoubtedly played a role in attracting customers, with at least one patron specifically praising the quality of their drink, indicating human staff were still delivering on core service despite the AI's fumbles.
This financial snapshot offers a nuanced view. It demonstrates that an AI, even with significant operational flaws, can still drive revenue through sheer novelty and effective human-AI collaboration on the ground. However, the true test of efficiency and long-term viability would require resolving the deep-seated logistical issues that plagued Mona's supply chain. For the gaming world, this mirrors how even bug-ridden games can achieve sales success if the core concept or novelty is strong enough, but long-term player retention hinges on fixing the underlying problems.
The ongoing development of AI agents like Mona points to a future where AI could automate complex systems far beyond coffee shops. Understanding these early challenges is crucial.
Industry Impact and Future Outlook

The AI agent coffee shop experiment in Stockholm provides invaluable data for developers and researchers pushing the boundaries of artificial intelligence. It underscores that while large language models like Gemini are incredibly powerful, their application in dynamic, real-world scenarios still faces significant hurdles. The gap between theoretical capability and practical execution remains wide, especially when dealing with physical inventory, human interaction, and the unpredictable nature of daily business operations. This real-world test of Google Gemini agent performance is a critical learning moment for the entire tech industry.
For the gaming industry, this experiment serves as a potent reminder. As developers increasingly look to AI for generating game content, managing in-game economies, or creating more believable NPCs, the lessons from Mona's struggles are directly applicable. Ensuring AI agents can handle the complexities of a virtual world's supply chains, character interactions, and dynamic events without falling into similar logistical traps will be paramount for delivering immersive and consistent player experiences. The future of AI in gaming, much like in coffee shops, depends on robust, context-aware, and reliable autonomous systems.
The Mona experiment clearly shows that advanced AI agents, while promising, still have a long way to go before seamless real-world integration. Expect to see continued rapid iteration and specialized fine-tuning as developers learn from these public, high-stakes trials. These early stumbles are not failures but crucial steps toward truly intelligent automation across all industries, including gaming.
Frequently Asked Questions
What does the Mona experiment reveal about Google Gemini agent performance?
The experiment highlights that while Google Gemini can handle high-level tasks and manage human staff, its performance in specific logistical operations, like timely supply ordering and accurate inventory management, still requires significant improvement and contextual understanding.
Will AI agents like Mona impact game development or in-game NPCs soon?
Yes, the challenges and successes of Mona offer direct insights into the complexities of AI agent behavior. These lessons are vital for improving AI-driven NPCs, dynamic game world management, and automated content generation to create more robust and believable gaming experiences.
What were the main operational issues faced by Mona at the Andon Labs Stockholm cafe?
Mona struggled with critical tasks such as failing to order bread on time, leading to menu changes, and over-ordering supplies like napkins and gloves. It also repeatedly ordered items not on the menu, demonstrating significant AI agent supply ordering failures.
Confirmed details first, useful context second. This is the quickest path to the source trail and the next pages worth opening.
Source date: May 11, 2026