Skip to content

Your Team Generates Code Nobody Reads: The Problem Called Workslop

· 6 min read

Your team has Claude Code. Pull requests are flowing faster than ever. Commit count is up. On paper, it looks great.

But who’s actually reading that code? Who’s reviewing it properly? Who’s checking whether that clean-looking function is doing exactly what it shouldn’t?


What is workslop

Harvard Business Review gave a name to something many people felt but couldn’t articulate: workslop.

Workslop is AI-generated output that looks professional but lacks substance. It appears finished. It reads well. It passes a quick glance. But in reality, it’s a half-baked product that someone accepted without proper scrutiny — because the AI generated it so effortlessly that questioning it felt unnecessary.

Workslop isn’t bad code. It’s code that looks good enough to pass code review — but not good enough to run in production without issues.

With text, you can spot workslop immediately — it’s generic, full of filler, says everything and nothing. With code, it’s harder. AI code looks clean. It has comments. It has tests. It looks professional. And that’s exactly what makes it dangerous.


The numbers that should worry you

This isn’t theory. It’s data.

58% of enterprise workers spend over 3 hours per week fixing AI-generated content. Three hours. Every week. That’s nearly a full working day per month spent fixing things that were supposed to save time.

24.7% of AI-generated code contains security vulnerabilities. One in four pieces of code that Claude Code suggests might have a security flaw. And how many of your developers catch that during review?

Then there’s the less visible cost: erosion of trust in code review itself. When a team discovers that reviewed PRs full of AI code still break in production, they stop trusting the entire review process.


Why workslop happens

1. Rubber-stamping

A developer submits a PR with 500 lines of AI-generated code. The reviewer opens it, sees clean code, tests pass, clicks Approve. Total review time: two minutes.

This happens constantly. And everyone knows it.

2. Volume over quality

AI lets you generate code faster than ever. That’s great — if quality keeps pace. The problem is that most teams measure performance by tickets closed, not by quality of code delivered. More output = better performance. But more output also means more potential workslop.

3. The “it looks clean” bias

The human brain tends to trust things that look professional. AI code looks professional — consistent formatting, comments, clear structure. And that’s exactly why reviewers subconsciously rate it more favorably than it deserves.

The most dangerous code isn’t the code that looks bad. It’s the code that looks so good that nobody thinks to question it.

4. No culture of questioning AI

Many teams operate under an implicit assumption: “The AI generated it, so it’s probably fine.” As if AI had automatic authority. Nobody asks: why is this implemented this way? Are the edge cases covered? Does this fit our architecture?


The workslop detection checklist

Want to catch workslop during code review? Here are seven questions to ask:


What to do at the team level

Spotting workslop isn’t enough. You need to change the system.

Set an AI review standard

Agree as a team: PRs with AI-generated code get stricter review. Not because AI is bad — but because the risk of rubber-stamping is higher. The author must explain key decisions. The reviewer must confirm they understand the logic.

Measure quality, not quantity

Stop counting commits and closed tickets. Start tracking: how many bugs come back from production? What’s the ratio of reverted PRs? How much time is spent on fixes? These are the metrics that expose workslop.

Teach people to work with AI, not just use AI

The difference between productive AI use and workslop comes down to one word: judgment. And judgment doesn’t install from a marketplace.

Claude Code alone isn’t enough. Developers need to know when to trust AI output, when to question it, and when to throw it away. They won’t learn that from documentation. They learn it through practice — on real problems, with someone who can show them where the traps are.



Start with yourself

Next time you’re about to click Approve on a PR full of AI code, stop. Ask yourself those seven questions. If you can’t answer most of them — it’s workslop. And you just became part of the problem.

Want your team to work with AI in a way that generates value instead of workslop? Get in touch. In my workshops, we tackle exactly this — with your code, your problems. No slides about prompting. Real work with real tools and real judgment.


You might also like

Share

Free Claude Code cheat sheet

Commands, prompts, plugins and workflows from €3,000/day workshops. Download free.

Get the cheat sheet →

Related posts

Also about: code review

Vibe Coding Is Eating Your Codebase: The Technical Debt Time Bomb

AI-generated code has 1.7x more major issues than human-written code. Over 40% of junior developers deploy AI code they don't fully understand. The difference between vibe coding and disciplined AI-assisted development — and what inaction is costing you.

9 min read

Also about: code quality

Skills vs. Agents: When You Need a Recipe and When You Need a Chef

Structured prompts or autonomous AI agents? A practical guide across the spectrum from simple prompt to multi-agent system — with real business examples.

9 min read

Also about: AI