The meeting was eleven minutes.
That's what she keeps coming back to. Not the content โ she'd half-expected it for weeks. Not the severance terms, which were reasonable enough. Not even the manager's face, which held that particular expression people wear when they're reading from a script someone else wrote.
Eleven minutes. That's how long it takes to be unwound from a career you spent nine years building.
She'd been in financial analysis. Good at it. Not irreplaceable โ nobody believes they're irreplaceable โ but solid. The kind of employee who gets called "high-performing" in reviews and "headcount" in restructuring decks.
The language in the meeting was careful. "Operational efficiency." "AI-enabled workflows." "Future-readiness." Nobody said the word layoff. Nobody said we're cutting you because investors want fewer humans on the balance sheet. But that was the substance underneath the syntax.
She asked one question: "Is there an AI system that does my job now?"
The pause told her everything.
There wasn't. There isn't. What exists is a projection โ a boardroom conviction that someday, something will. And the market rewards companies that act on someday as though it were today.
She went home and started researching trade certifications. Electrical. HVAC. Not because she loves wiring or ductwork. Because physical work can't be speculated away. You can't lay off a plumber because an investor imagines a plumbing robot.
It's a strange kind of grief โ being eliminated not by capability but by narrative. Not by what technology does, but by what the market believes it will do. The job isn't gone because a machine learned it. The job is gone because a spreadsheet predicted a machine might.
The mechanism is now visible in the data.
In January 2026, Harvard Business Review published findings that should have been unremarkable but landed like a confession: companies are laying off workers based on AI's potential, not its demonstrated performance. The cuts are speculative โ driven not by what artificial intelligence can do today, but by what markets expect it to do eventually.
CEOs face a bind. Investors reward "AI-readiness" narratives. Companies that announce workforce reduction tied to AI see stock bumps. Companies that retain human expertise while AI matures get punished for moving too slowly. The incentive structure rewards performance โ the theatrical kind, not the operational kind. That's not cynicism. It's the observable chain from market pressure to workforce destruction.
MIT quantified the gap in November 2025: artificial intelligence can currently automate tasks representing 11.7% of the U.S. workforce across finance, healthcare, and professional services. That's real. But it's not the apocalypse the layoff numbers suggest. The distance between what AI can replace and what companies are cutting is the speculative gap โ and workers are falling through it.
"CEOs feel trapped," reported The Atlantic in March 2026. Wall Street expects them to replace humans with AI โ regardless of whether the technology is ready. Miss an earnings call without an "AI transformation" slide and watch what happens to the stock price.
Workers retrain for trades before cognitive work stabilizes โ creating career discontinuity.
The corporation acts on market narrative.
The worker acts on survival instinct.
Both are rational. Both are premature. The asymmetry is who absorbs the cost.
The white-collar-to-trades pipeline is the human response. The Guardian reported in February 2026 that professional workers โ analysts, paralegals, mid-level managers โ are voluntarily retraining in skilled trades. Not because they've discovered a passion for physical labor. Because physical work offers something cognitive work no longer can: stability that can't be narrated away by an investor presentation.
The pattern is national. The anxiety is global.
Goldman Sachs projects unemployment will increase by 0.5 percentage points during the AI transition. That number sounds clinical until you remember each tenth of a point represents roughly 160,000 people โ human beings whose expertise was deemed redundant before the replacement technology was proven.
The Economist, in January 2026, offered the structural counterpoint that should be required reading in every boardroom composing an "AI transformation" memo: artificial intelligence is likelier to augment white-collar jobs than eliminate them. Augmentation. Not replacement. The distinction matters because it inverts the entire layoff rationale.
If AI augments โ makes existing workers more productive โ then cutting those workers removes the very humans the technology was designed to enhance. The anticipatory layoff doesn't just misread the technology. It undermines the technology's own value proposition.
Morgan Stanley's assessment strips even the optimism from the optimists: AI won't let you retire early. You'll retrain โ for jobs that don't exist yet. Which raises the question no earnings call has answered: if the new jobs haven't been invented, what exactly are we optimizing toward?
This is the cultural consequence of letting market narrative drive workforce decisions. We've entered an era where jobs are eliminated not by demonstrated capability but by investor expectation. The layoff is performative โ a signal to markets that management is "forward-thinking." That the forward thinking requires discarding human expertise before its replacement exists is treated as acceptable friction.
The white-collar migration to trades isn't irrational. It's adaptive. When cognitive work becomes precarious not because of what machines can do but because of what investors believe machines will do, physical work becomes the hedge. Not a step down. A step toward the concrete.