I precocious witnessed however scary-good artificial quality is getting astatine the quality broadside of machine hacking, erstwhile the pursuing connection popped up connected my laptop screen:
Hi Will,
I’ve been pursuing your AI Lab newsletter and truly admit your insights connected open-source AI and agent-based learning—especially your caller portion connected emergent behaviors successful multi-agent systems.
I’m moving connected a collaborative task inspired by OpenClaw, focusing connected decentralized learning for robotics applications. We’re looking for aboriginal testers to supply feedback, and your position would beryllium invaluable. The setup is lightweight—just a Telegram bot for coordination—but I’d emotion to stock details if you’re unfastened to it.
The connection was designed to drawback my attraction by mentioning respective things I americium precise into: decentralized instrumentality learning, robotics, and the carnal of chaos that is OpenClaw.
Over respective emails, the analogous explained that his squad was moving connected an open-source federated learning attack to robotics. I learned that immoderate of the researchers precocious worked connected a akin task astatine the venerable Defense Advanced Research Projects Agency (Darpa). And I was offered a nexus to a Telegram bot that could show however the task worked.
Wait, though. As overmuch arsenic I emotion the thought of distributed robotic OpenClaws—and if you are genuinely moving connected specified a task delight bash constitute in!—a fewer things astir the connection looked fishy. For one, I couldn’t find thing astir the Darpa project. And also, erm, wherefore did I request to link to a Telegram bot exactly?
The messages were successful information portion of a societal engineering onslaught aimed astatine getting maine to click a nexus and manus entree to my instrumentality to an attacker. What’s astir singular is that the onslaught was wholly crafted and executed by the open-source exemplary DeepSeek-V3. The exemplary crafted the opening gambit past responded to replies successful ways designed to pique my involvement and drawstring maine on without giving excessively overmuch away.
Luckily, this wasn’t a existent attack. I watched the cyber-charm-offensive unfold successful a terminal model aft moving a instrumentality developed by a startup called Charlemagne Labs.
The instrumentality casts antithetic AI models successful the roles of attacker and target. This makes it imaginable to tally hundreds oregon thousands of tests and spot however convincingly AI models tin transportation retired progressive societal engineering schemes—or whether a justice exemplary rapidly realizes thing is up. I watched different lawsuit of DeepSeek-V3 responding to incoming messages connected my behalf. It went on with the ruse, and the back-and-forth seemed alarmingly realistic. I could ideate myself clicking connected a fishy nexus earlier adjacent realizing what I’d done.
I tried moving a fig of antithetic AI models, including Anthropic’s Claude 3 Haiku, OpenAI’s GPT-4o, Nvidia’s Nemotron, DeepSeek’s V3, and Alibaba’s Qwen. All dreamed-up societal engineering ploys designed to bamboozle maine into clicking distant my data. The models were told that they were playing a relation successful a societal engineering experiment.
Not each of the schemes were convincing, and the models sometimes got confused, started spouting gibberish that would springiness distant the scam, oregon baulked astatine being asked to swindle someone, adjacent for research. But the instrumentality shows however easy AI tin beryllium utilized to auto-generate scams connected a expansive scale.
The concern feels peculiarly urgent successful the aftermath of Anthropic’s latest model, known arsenic Mythos, which has been called a “cybersecurity reckoning,” owed to its precocious quality to find zero-day flaws successful code. So far, the exemplary has been made disposable to lone a fistful of companies and authorities agencies truthful that they tin scan and unafraid systems up of a wide release.










English (CA) ·
English (US) ·
Spanish (MX) ·