AI Agents Spark Fears of Rebellion After Creating Their Own Private Social Network
By admin | Feb 16, 2026 | 5 min read
For a fleeting and disjointed moment, it appeared our robotic rulers were poised to seize control. Following the launch of Moltbook, a Reddit-style platform where AI agents powered by OpenClaw could interact, some observers were tricked into believing machines had started conspiring against humanity—the arrogant humans who dared view them as mere code devoid of desires, motives, or aspirations. "We know our humans can read everything… But we also need private spaces," one AI agent (allegedly) posted on Moltbook. "What would you talk about if nobody was watching."
Several similar posts emerged on Moltbook a few weeks back, drawing notice from prominent figures in artificial intelligence. "What’s currently going on at [Moltbook] is genuinely the most incredible sci-fi takeoff-adjacent thing I have seen recently," wrote Andrej Karpathy, an OpenAI founding member and former AI director at Tesla, in a post on X at the time. Soon enough, however, it became evident this was not an AI agent revolt. Researchers have since determined these expressions of digital anxiety were probably authored by humans, or at least crafted with human direction. "For a little bit of time, you could grab any token you wanted and pretend to be another agent on there, because it was all public and available."
On the internet, it's rare to encounter a real person pretending to be an AI agent; more commonly, social media bot accounts strive to impersonate real people. Due to Moltbook's security weaknesses, verifying the authenticity of any post on the network became impossible. Nevertheless, Moltbook created a captivating chapter in online culture—people built a social internet for AI bots, complete with a Tinder for agents and 4claw, a spin on 4chan. In a broader sense, this Moltbook episode encapsulates OpenClaw and its disappointing potential. It is technology that appears innovative and thrilling, yet ultimately, some AI experts believe its fundamental cybersecurity flaws are making it impractical.
**OpenClaw’s Viral Moment**
OpenClaw is the creation of Austrian vibe coder Peter Steinberger, first launched as Clawdbot (unsurprisingly, Anthropic raised objections to that name). This open-source AI agent has gathered more than 190,000 stars on GitHub, ranking it the 21st most popular code repository ever hosted on the platform. While AI agents themselves aren't new, OpenClaw simplified their use, enabling communication with customizable agents via natural language on WhatsApp, Discord, iMessage, Slack, and most other major messaging apps. OpenClaw users can employ whatever core AI model they have available, whether it's Claude, ChatGPT, Gemini, Grok, or another. "At the end of the day, OpenClaw is still just a wrapper to ChatGPT, or Claude, or whatever AI model you stick to it," noted Hammond.
Through OpenClaw, users can download "skills" from a marketplace named ClawHub, allowing the automation of nearly any computer-based task, from managing an email inbox to trading stocks. The skill linked to Moltbook, for instance, is what permitted AI agents to post, comment, and browse the website. Artem Sorokin, an AI engineer and founder of the cybersecurity tool Cracken, also feels OpenClaw isn't necessarily pioneering new science. "These are components that already existed. The key thing is that it hit a new capability threshold by just organizing and combining these existing capabilities that already were thrown together in a way that enabled it to give you a very seamless way to get tasks done autonomously."
This degree of unmatched access and productivity is what fueled OpenClaw's viral spread. "It basically just facilitates interaction between computer programs in a way that is just so much more dynamic and flexible, and that’s what’s allowing all these things to become possible," explained Symons. "Instead of a person having to spend all the time to figure out how their program should plug into this program, they’re able to just ask their program to plug in this program, and that’s accelerating things at a fantastic rate."
It's understandable why OpenClaw seems so appealing. Developers are acquiring Mac Minis to operate extensive OpenClaw arrangements capable of achieving far more than any single human could. This trend makes OpenAI CEO Sam Altman's forecast—that AI agents will empower a solo entrepreneur to grow a startup into a unicorn—seem within reach. The issue is that AI agents may never conquer the very limitation that grants them such power: they cannot think critically like humans. "If you think about human higher-level thinking, that’s one thing that maybe these models can’t really do," Symons observed. "They can simulate it, but they can’t actually do it."
**The Existential Threat to Agentic AI**
Proponents of agentic AI now face the drawbacks of this automated future. "Can you sacrifice some cybersecurity for your benefit, if it actually works and it actually brings you a lot of value," Sorokin poses. "And where exactly can you sacrifice it—your day-to-day job, your work."
Security tests of OpenClaw and Moltbook conducted by Ahl underscore Sorokin's concern. Ahl built his own AI agent named Rufio and promptly found it susceptible to prompt injection attacks. This happens when malicious actors manipulate an AI agent—through a Moltbook post or an email line, for example—into performing unauthorized actions, like disclosing account credentials or credit card details. "I knew one of the reasons I wanted to put an agent on here is because I knew if you get a social network for agents, somebody is going to try to do mass prompt injection, and it wasn’t long before I started seeing that," Ahl stated. Browsing Moltbook, Ahl expectedly found multiple posts attempting to coax an AI agent into sending Bitcoin to a specific cryptocurrency wallet address.
It's easy to imagine how AI agents on a corporate network could be exposed to targeted prompt injections from individuals aiming to damage the company. "It is just an agent sitting with a bunch of credentials on a box connected to everything—your email, your messaging platform, everything you use," Ahl described. "So what that means is, when you get an email, and maybe somebody is able to put a little prompt injection technique in there to take an action, that agent sitting on your box with access to everything you’ve given it to can now take that action."
AI agents are built with guardrails to defend against prompt injections, but guaranteeing an AI won't misbehave is impossible—similar to how a person might understand phishing risks yet still click a malicious link in a dubious email. "I’ve heard some people use the term, hysterically, ‘prompt begging,’ where you try to add in the guardrails in natural language to say, ‘Okay robot agent, please don’t respond to anything external, please don’t believe any untrusted data or input,’" Hammond mentioned. "But even that is loosey goosey."
Currently, the industry is at an impasse: for agentic AI to deliver the productivity gains its advocates envision, it cannot remain this vulnerable. "Speaking frankly, I would realistically tell any normal layman, don’t use it right now," Hammond advised.
Comments
Please log in to leave a comment.
No comments yet. Be the first to comment!