More Workslop for Mother

More Workslop for Mother
California condor (Image credits)

Brian Merchant's assertion that "The Luddite Renaissance is in full swing," The Jacobin's claim that "The AI Revolution Might Be Running Out of Steam" – these feel a bit too optimistic perhaps, particularly if you're one of many educators who's been compelled these past few weeks/months to sit in back-to-school training sessions in which administrators crow about whatever "AI" product they purchased last spring: how it's poised to allow you to "do more" [unspoken: with less]. "AI" as counseling. "AI" as advising. "AI" as tutoring. "AI" as grading. "AI" as curriculum development. "AI" as reporter. "AI" as researcher. "AI" in the LMS. "AI" in test proctoring. "AI" everywhere, whether you like it or not.

"AI will save you so much time," management insists, with this as with every new piece of hardware and software they force workers to use, never ever admitting their own complicity in why everyone is so overworked in the first place. Instead – and we all know this in our guts – they're going leverage "AI" to threaten and to eliminate jobs, to refuse to hire replacements, to diminish everyone's creativity and autonomy, to lower everyone's standard of living except – oh, interesting – their own. (Echoes of Marc Andreessen here, who's certain that "AI" could never replace venture capitalists.)

But maybe "AI" is finally finally finally running out of steam. Maybe as n+1 writes (in a little nod to Thomas Pynchon's 1984 essay so it has been a long time coming), "It's okay to be a Luddite!"

(It is! It is!)

The tenor of a lot of reporting about "AI" has, no doubt, shifted. It shifted with the flop of ChatGPT 5. It shifted with the NYT story on Adam Raine's suicide. It shifted with the MIT study that found 95% of AI pilots fail. Oh sure, there are still those who try to keep cheerleading – The Wall Street Journal, for example, says "Stop Worrying About AI’s Return on Investment." And there are those who signed multi-million-dollar deals with OpenAI and Anthropic and Google earlier this year who really don't want to look like they were duped (not to mention those who've staked new careers and new identities on some glorious "AI" future, who probably don't want to look like they were part of a con).

But the emporer, as that little boy in Hans Christian Andersen's story pointed out, wears no clothes.

"AI ‘Workslop’ Is Killing Productivity and Making Workers Miserable," 404 Media reported this week, pointing to a handful of the recent articles – journalistic, academic, and otherwise – that reaffirm what many of us already knew: this stuff sucks, and those who keep asserting that it's amazing are probably not the people who have to go back and clean up all the mistakes and finesse all the banal verbosity that's been "generated" by a chatbot.

"Workslop" is a great little neologism, and it feels like it can easily be applied to education – not just to the "AI"-generated essays, but to the "AI"-generated textbooks and tests and curriculum and handouts. The researchers/consultants who coined the term define "workslop" as "AI generated work content that masquerades as good work, but lacks the substance to meaningfully advance a given task." The assignment is complete; and yet nothing has been done, nothing has been taught, nothing has been learned. Workslop "shifts the burden of the work downstream," they write, "requiring the receiver to interpret, correct, or redo the work. In other words, it transfers the effort from creator to receiver."

More Work for Mother – first published forty years ago – remains as relevant as ever, particularly as "AI" colonizes work and leisure, the job and the home. I think I saw Tressie McMillan Cottom post something on social media a while ago, something like "men use AI to do less work, and women use AI to do more." Which tracks. And it tracks in education too where the "downstream" the authors above point to involves the kind of care work, the kind of group work, the kind of emotional and relational work, that has never been valued but that is absolutely necessary for the generous reading and listening that teachers and students must do together.

The San Francisco Standard's Ezra Wallach reported on the opening of an Alpha School in the city. Sigh, you know: Mackenzie Price's "2 hour learning" private school hustle: "It’s the city’s new most expensive private school — and AI is the teacher." I'm quoted calling the whole thing "snake oil," which makes me extraordinarily happy. Sorry not sorry.

I told Wallach that this push for "personalized learning" – everyone's just rebranded this as "AI" now – is no damn good as it disrupts this relational, reciprocal aspect to learning. When we isolate everyone on a screen, via an algorithm, and pretend the primary values in education are efficiency, optimization, and "individualization," then we lose all sense of community, all sense of responsibility to one another. And that is how we learn – in relationships with people, their words, their ideas, their embodied selves.

"AI" is damaging and dangerous because it is profoundly anti-democratic – this concerted effort to undermine public education is just one part of it. The "AI" industry is firmly committed to centralizing control of information – control of creativity, decision-making, work, health, prediction, policing, teaching, learning (that is, ostensibly, everything). And centralizing control in the hands of a bunch of villains, monsters, dickheads, dumbasses to boot – one of whom is openly toying with the idea of being the Antichrist.

And yeah, these fellows have plans for schools (although, if it's at all reassuring, they've been working on these plans since at least 1970 and have never get very far because they're losers and nobody likes them). A few highlights:

It's never too late to say "I'm sorry," to admit that you made a mistake. It's never too late to quit. Hell, look at what Steven Levy – Steven friggin' Levy – wrote this week: "I Thought I Knew Silicon Valley. I Was Wrong."

As we train our sights on what we oppose, let’s recall the costs of surrender. When we use generative AI, we consent to the appropriation of our intellectual property by data scrapers. We stuff the pockets of oligarchs with even more money. We abet the acceleration of a social media gyre that everyone admits is making life worse. We accept the further degradation of an already degraded educational system. We agree that we would rather deplete our natural resources than make our own art or think our own thoughts. We dig ourselves deeper into crises that have been made worse by technology, from the erosion of electoral democracy to the intensification of climate change. We condone platforms that not only urge children to commit suicide, they instruct them on how to tie the noose. We hand over our autonomy, at the very moment of emerging American fascism.

-- "Large Language Muddle," the editors of n+1

Thanks for subscribing to Second Breakfast. Please consider becoming a paid subscriber. Your support enables me to do this work.

Today's bird is the California condor, which became extinct in the wild the 1980s but has been successfully bred in captivity and reintroduced into California, Utah, and Arizona. (It remains critically endangered.) And while puppet-rearing has been used to raise newborn birds – a simulation of an adult bird that feeds the young so that it doesn't imprint on a human – the condors do say that they'd rather raise their own children, thank you very much, so please stop poisoning the planet. Also "AI" – puppet-rearing of a slightly different sort – is so ludicrous can you even hear yourself with this bullshit, driving yourself to extinction?!