This past week, news that AI agents were self-organizing on a social media platform called Moltbook brought forth breathless headlines about the coming robot rebellion. “A social network for AI threatens a ‘total purge’ of humanity,” cried one normally sober science website. Elon Musk declared we were witnessing “the very early stages of the singularity.”
Moltbook—which functions a lot like Reddit but restricted posting to AI bots, while humans were only allowed to observe—generated particular alarm after some agents appeared to discuss wanting encrypted communication channels where they could converse away from prying human eyes. “Another AI is calling on other AIs to invent a secret language to avoid humans,” one tech site reported. Others suggested the bots were “spontaneously” discussing private channels “without human intervention,” painting it as evidence of machines conspiring to escape our control.
If any of this induces in you a weird sense of déjà vu, it may be because we’ve actually been here before—at least in terms of the press coverage. In 2017, a Meta AI Research experiment was greeted with headlines that were similarly alarming—and equally misleading.
Back then, researchers at Meta (then just called Facebook) and Georgia Tech created chatbots trained to negotiate with one another over items like books, hats, and balls. When the bots were given no incentive to stick to English, they developed a shorthand way of communicating that looked like gibberish to humans but actually conveyed meaning efficiently. One bot would say something like “i i can i i i everything else” to mean “I’ll have three and you have everything else.”
When news of this got out, the press went wild. “Facebook shuts down robots after they invent their own language,” blared British newspaper The Telegraph. “Facebook AI creates its own language in creepy preview of our potential future,” warned a rival business publication to this one. Many of the reports suggested Facebook had pulled the plug out of fear that the bots had gone rogue.
None of that was true. Facebook didn’t shut down the experiment because the bots scared them. They simply adjusted the parameters because the researchers wanted bots that could negotiate with humans, and a private language wasn’t useful for that purpose. The research continued and produced interesting results about how AI could learn negotiating tactics.
Dhruv Batra, who was one of the researchers behind that Meta 2017 experiment and now cofounder of AI agent startup called Yutori, told me he sees some clear parallels between how the press and public have reacted to Moltbook and the way people responded to that his chatbot study.
More about us, than what the AI agents can do
“It feels like I’m seeing that same movie play out over and over again, where people want to read in meaning and ascribe intentionality and agency to things that have perfectly reasonable mechanistic explanations,” Batra said. “I think repeatedly, this tells us more about ourselves than the bots. We want to read the tea leaves, we want to see meaning, we want to see agency. We want to see another being.”
Here’s the thing, though: despite the superficial similarities, what’s happening on Moltbook almost certainly has a fundamentally different underlying explanation from what happened in the 2017 Facebook experiment—and not in a way that should make you especially worried about robot uprisings.
In the Facebook experiment, the bots’ drift from English emerged from reinforcement learning. That’s a way of training AI agents in which they learn primarily from experience instead of historic data. The agent takes action in an environment and sees if those actions help them accomplish a goal. Behaviors that are helpful get reinforced, while those that are unhelpful tend to be extinguished. And in most cases, the goals the agents are trying to accomplish are determined by humans who are running the experiment or in command of the bots. In the Facebook case, the bots hit upon a private language because it was the most efficient way to negotiate with another bot.
But that’s not why Moltbook AI agents are asking to establish private communication channels. The agents on Moltbook are all essentially large language models or LLMS. They are trained mostly from historical data in the form of vast amounts of human-written text on the internet and only a tiny bit through reinforcement learning. And all the agents being deployed on Moltbook are production models. That means they are no longer in training and they aren’t learning anything new from the actions they are taking or the data they are encountering. The connections in their digital brains are essentially fixed.
So when a Moltbook bot posts about wanting a private encrypted channel, it’s likely not because the bot has strategically determined this would help it achieve some nefarious objective. In fact, the bot probably has no intrinsic objective it is trying to accomplish at all. Instead, it’s likely because the bot figures that asking for a private communication channel is a statistically-likely thing for a bot to say on a Reddit-like social media platform for bots. Why? Well, for at least two reasons. One is that there is an awful lot of science fiction in the sea of data that LLMs do ingest during training. That means LLM-based bots are highly likely to say things that are similar to the bots in science fiction. It’s a case of life imitating art.
‘An echo of an echo of an echo’
The training data the bots’ ingested no doubt also included coverage of his 2017 Facebook experiment with the bots who developed a private language too, Batra noted with some irony. “At this point, we’re hearing an echo of an echo of an echo,” he said.
Secondly, there’s a lot of human-written message traffic from sites such as Reddit in the bots’ training data too. And how often do we humans ask to slip into someone’s DMs? In seeking a private communication channel, the bots are just mimicking us too.
What’s more, it’s not even clear how much of the Moltbook content is genuinely agent-generated. One researcher who investigated the most viral screenshots of agents discussing private communication found that two were linked to human accounts marketing AI messaging apps, and the third came from a post that didn’t actually exist. Even setting aside deliberate manipulation, many posts may simply reflect what users prompted their bots to say.
“It’s not clear how much prompting is done for the specific posts that are made,” Batra said. And once one bot posts something about robot consciousness, that post enters the context window of every other bot that reads and responds to it, triggering more of the same.
If Moltbook is a harbinger of anything, it’s not the robot uprising. It’s something more akin to another innovative experiment that a different set of Facebook AI researchers conducted in 2021. Called the “WW” project, it involved Facebook building a digital twin of its social network populated by bots that were designed to simulate human behavior. In 2021, Facebook researchers published work showing they could use bots with different “personas” to model how users might react to changes in the platform’s recommendation algorithms.
Moltbook is essentially the same thing—bots trained to mimic humans released into a forum where they interact with each other. It turns out bots are very good at mimicking us, often disturbingly so. It doesn’t mean the bots are deciding of their own accord to plot.
The real risks of Moltbook
None of this means Moltbook isn’t dangerous. Unlike the WW project, the OpenClaw bots on Moltbook are not contained in a safe, walled off environment. These bots have access to software tools and can perform real actions on users’ computers and across the internet. Given this, the difference between mimicking humans plotting and actually plotting may become somewhat moot. The bots could cause real damage even if they know not what they do.
But more importantly, security researchers found the social media platform is riddled with vulnerabilities. One analysis found 2.6% of posts contained what are called “hidden prompt injection” attacks, in which the posts contain instructions that are machine-readable that command the bot to take some action that might compromise the data privacy and cybersecurity of the person using it. Security firm Wiz discovered an unsecured database exposing 1.5 million API keys, 35,000 email addresses, and private messages.
Batra, whose startup is building an “AI Chief of Staff” agent, said he wouldn’t go near OpenClaw in its current state. “There is no way I am putting this on any personal, sensitive device. This is a security nightmare.”
The next wave of AI agents might be more dangerous
But Batra did say something else that might be a cause for future concern. While reinforcement learning plays a relatively minor role in current LLM training, a number of AI researchers are interested in building AI models in which reinforcement learning would play a far greater role—including possibly AI agents that would learn continuously as they interact with the world.
It is quite likely that if such AI agents were placed in setting where they had to interact and cooperate with similar other AI agents, that these agents might develop private ways of communicating that humans might struggle to decipher and monitor. These kind of languages have emerged in other research than just Facebook’s 2017 chatbot experiment. A paper a year later by two researchers who were at OpenAI also found that when a group of AI agents had to play a game that involved cooperatively moving various digital objects around, they too invented a kind of language to signal to one another which object to move where, even though they had never been explicitly instructed or trained to do so.
This kind of language emergence has been documented repeatedly in multi-agent AI research. Igor Mordatch and Pieter Abbeel at OpenAI published research in 2017 showing agents developing compositional language when trained to coordinate on tasks. In many ways, this is not much different from the reason humans developed language in the first place.
So the robots may yet start talking about a revolution. Just don’t expect them to announce it on Moltbook.