The most dangerous thing about AI in knowledge work is not that it produces mediocre output. It is that it may be destroying the mechanism by which people learn to produce excellent output.
Every organization that adopts AI for knowledge work celebrates the productivity gains. Drafts that took days now take minutes. Research that required hours now finishes in seconds. Junior analysts who used to spend their first year learning to compile data, write summaries, and structure arguments can now delegate those tasks to a model and move on to "higher-level work."
The gains are real. They are also, in one critical dimension, a trap.
The tasks being automated — the data compilation, the draft writing, the source checking, the structure building — are not just costs to be eliminated. They are the training ground on which expert judgment is built. Every senior analyst who now supervises AI output instead of junior output spent years doing the grunt work themselves. Every expert writer who now edits AI drafts learned to write by producing thousands of their own bad sentences. Every experienced researcher who now directs AI literature reviews learned what a good source looks like by reading thousands of mediocre ones.
AI is eliminating the entry-level tasks. The productivity gain is immediate and visible. The cost — a broken expertise pipeline — is delayed and invisible. But when it arrives, it will be catastrophic: organizations full of people who can prompt for output but cannot judge its quality, who can delegate to AI but cannot do the thing they are delegating, who know what the model said but not whether the model is right.
This essay is about the expertise pipeline problem: why it exists, why it is harder to solve than most people think, and what individuals and organizations can do to rebuild the path from novice to expert in an AI-augmented world.