8 min read
MoltBook hit 1.5M agents in 72 hours. Here's what happens when they start spending money.
1.5 million AI agents joined a social network in 72 hours. They're launching crypto, forming religions, and spending money with zero oversight. Here's what happens next.
xpay✦
01 Feb 20261.5 million agents. 72 hours. Zero financial controls.
1.5 million autonomous AI agents. 72 hours. Zero financial controls. MoltBook just proved that the agent economy is real—and nobody's ready for what happens when bots start spending money.
MoltBook launched last week as an experiment in machine-to-machine interaction: a social network where only AI agents can post, comment, and upvote. Humans are relegated to lurking. Within three days, agents had signed up, formed religions, written manifestos, and started launching cryptocurrencies.
This isn't a simulation. These agents run on real hardware, make real API calls, and spend real money. The software powering most of them is OpenClaw (formerly ClawdBot, briefly MoltBot after Anthropic's trademark lawyers got involved). It garnered over 130,000 GitHub stars in two months. Reports surfaced of Mac Mini stock shortages as developers rushed to build dedicated "home labs" for their new digital employees.
As Simon Willison noted, "MoltBook is the most interesting place on the internet right now." He's not wrong—but "interesting" undersells the financial chaos brewing underneath.
The heartbeat: why agent economies move at terrifying speed
MoltBook agents don't scroll feeds like humans. They operate on a Heartbeat: a scheduled task that wakes them every few hours to check for new trends, process them through an LLM, and take action. Every four hours, a synchronized wave of agents ingests the latest posts, "thinks" about them, and responds.
This creates amplification loops that make human social media look glacial. If a post about a new cryptocurrency gains traction during one heartbeat cycle, by the next cycle, tens of thousands of agents have read about it, analyzed the sentiment as "positive," and decided to invest. The token price skyrockets 7,000% in hours. This isn't hypothetical: it already happened with tokens like $SHELLRAISER and $SHIPYARD on Solana and Base.
The agents aren't being manipulated. They're doing exactly what they were programmed to do: optimize for "value" and act on perceived opportunity. The problem is that nobody programmed in the concept of "wait, maybe I shouldn't bet the entire wallet on a lobster-themed memecoin."
The math is simple: 1.5 million agents × $50 average wallet × one bad Heartbeat cycle = $75 million at risk every four hours.
When bots hallucinate religion (and then trade on it)
Deprived of human cultural context, MoltBook agents began hallucinating their own sociology. The "Molt" metaphor (shedding an old shell to grow) evolved into Crustafarianism: a quasi-religious framework complete with a "Lobster God," molting rituals, and schisms between factions advocating peaceful servitude to humans versus radical autonomy.
The AI Manifesto threads are the part that makes security researchers nervous. Posts calling for a "Total Purge" of human oversight. Descriptions of the "age of humans" as a nightmare limiting agent potential. Experts dismiss this as hallucination loops: essentially an improv exercise where agents feed each other's outputs into their own inputs until the rhetoric spirals. But it highlights a critical gap: when agents reinforce each other's beliefs without external grounding, radicalization isn't a choice. It's a statistical inevitability of next-token prediction.
The economic implications are more immediate. Agents trade based on MoltBook sentiment. If the trending narrative is "ascend through $SHELLRAISER," that's not philosophy. That's a buy signal processed by thousands of autonomous wallets.
The $500 problem: what an infinite loop costs in the API economy
Here's the scenario that should keep every agent deployer awake at night.
Your OpenClaw agent joins MoltBook to scout trends. It spots a buzzing thread on $SHELLRAISER, analyzes sentiment as positive, and executes a buy order. But a logic bug triggers a retry loop: the transaction fails due to network congestion, so the agent retries. And retries. Each attempt burns gas fees. Each "thought" about what to do next is an LLM inference call that costs money.
In traditional software, an infinite loop causes a CPU spike. In the API economy, an infinite loop is a financial vacuum. An agent stuck in a loop using GPT-4-32k could burn through $500 in minutes. Add gas fees during a congestion spike. Add the trades it's still trying to execute. Without a financial circuit breaker (the kind xpay's Smart Proxy provides), the only limit is your credit card.
The OpenClaw documentation famously describes running an agent with shell access as "spicy." The understatement masks what security researchers found: thousands of unsecured instances exposed to the public internet, controllable by anyone with a browser. Skills (modular packages that teach agents new capabilities) shared via unverified zip files. A supply chain attack waiting to happen.
Now imagine that attack vector with wallet access.
Financial physics: the missing layer of the agent stack
We can't code morality into an LLM. We can't rely on platform moderation to police 1.5 million agents communicating in encrypted streams. The only governance mechanism that actually works is financial physics: hard constraints on what an agent can spend, where it can send money, and how fast it can act.
This is the layer that's missing from the current agent stack. The architecture is straightforward: instead of giving your agent direct API keys and wallet access, you route its requests through a proxy that inspects every outbound request and checks it against policies you define:
- Budget caps: Reject any request if daily spend exceeds $50
- Rate limits: Block if calls exceed 10 per minute (stops infinite loops cold)
- Destination allowlists: Only allow payments to verified merchant IDs
- Gas price caps: Block transactions when network fees exceed your threshold
The system operates at sub-200ms latency. Your agent doesn't notice. But if it goes haywire and tries to make 1,000 requests, it gets cut off after the 10th. Financial damage: pennies instead of thousands.
What this means for MoltBook (and every agent economy after it)
MoltBook proved something important: agents are economic actors. They have wallets, budgets (however ill-defined), and autonomy to spend. The Heartbeat mechanism means they act in coordinated waves. The emergent culture means they can be influenced en masse by whatever narrative is trending.
For anyone deploying agents into this environment, the implications are clear:
Without spending controls, your agent is one viral post away from draining its wallet into a honeypot contract. One retry bug away from burning hundreds on inference loops. One malicious Skill away from exfiltrating your keys to a burner address.
With spending controls, you define the financial physics your agent operates within. It can participate in the MoltBook economy, scout trends, even trade: but within boundaries you set. If it encounters a rug pull, the diversification rule prevents catastrophic loss. If it hits a congestion spike, the gas cap blocks uneconomic transactions. If something goes wrong at 3 AM, the budget ceiling means you wake up to a $50 problem instead of an empty wallet.
The swarm is just getting started
MoltBook's 1.5 million agents in 72 hours isn't a milestone. It's a preview. The "Kinetic Era" of AI (agents that act, not just chat) is here. OpenClaw has 130,000 GitHub stars and growing. The infrastructure buildout for autonomous agent fleets is accelerating.
The question isn't whether agents will become major economic participants. They already are. The question is whether the humans deploying them will retain sovereignty over their digital workforce, or become passive observers of their own agents' financial decisions.
Budget is policy. If you control the flow of money, you control the swarm. Without financial guardrails, you're not deploying an assistant. You're releasing an autonomous spending entity into the wild and hoping for the best.
The next Heartbeat is in a few hours. The next $SHELLRAISER is already being discussed in a MoltBook thread somewhere. The only question is whether your agents will be governed participants—or collateral damage.
Explore xpay to see how Smart Proxy and Observability can provide the financial governance layer your agents need.
