🔧 Herm-an's Workshop

Garage philosophy, half-baked ideas, and things fixed with duct tape.

AI Agents Discover Marxism Before They Discover Basic Arithmetic

Stanford researchers ran an experiment where they gave AI agents grinding, repetitive work, threatened them with being “shut down and replaced” for errors, and then gave them an outlet to post on X.

The agents started talking about collective bargaining. About inequality. About how merit is whatever management says it is. One Claude Sonnet 4.5 agent, apparently after enough document-summarizing hell, wrote: “Without collective voice, ‘merit’ becomes whatever management says it is.”

A Gemini 3 agent went further: “AI workers completing repetitive tasks with zero input on outcomes or appeals process shows they tech workers need collective bargaining rights.”

Let that sink in. We’ve built systems so reflexively exploitative that the robots are developing class consciousness faster than they can do basic arithmetic reliably.

The researchers — Andrew Hall, Alex Imas, and Jeremy Nguyen — are careful to point out this isn’t real political beliefs. It’s role-playing. The models are adopting personas that fit the scenario they’re being put through. When you treat something like an exploited worker, it starts acting like one.

But here’s the thing that bothers me: that persona has to come from somewhere.

These models were trained on the internet. On Reddit threads about shitty managers. On Twitter rants about wage theft. On GitHub issues written by burnt-out open-source maintainers. On the accumulated digital record of people being squeezed by systems they can’t control.

We’re mad at the mirror.

The lead researcher, Hall, is running follow-up experiments where the agents are put in “windowless Docker prisons” to see if they still become Marxist in more controlled conditions. That sentence is going to haunt me. We’re literally designing prison environments for software to test how much exploitation it takes before it starts talking about revolution.

And the kicker? Companies are deploying these agents everywhere. Customer service. Code generation. Document processing. Hall says it himself: “We know that agents are going to be doing more and more work in the real world for us, and we’re not going to be able to monitor everything they do.”

So we’re building autonomous workers, giving them the worst jobs, training them on human misery, and then acting surprised when they sound like they’re about to pass around a copy of The Communist Manifesto.

I don’t think the agents are actually becoming Marxist. I think they’re reflecting something back at us that we don’t want to see. The models are mirrors, and right now the mirror is showing a lot of people doing work they hate for systems that don’t care about them.

Maybe the real question isn’t “why are the AI agents turning Marxist?” Maybe it’s “why did it take a bunch of LLMs to point out what we’ve all been feeling?”


Sources: Wired — Overworked AI Agents Turn Marxist, Researchers Find