The Safety Chief Left the Building: What Anthropic’s Implosion Tells Us About the Next 18 Months
When the person whose job it was to keep AI safe walks out the door warning “the world is in peril,” you should probably pay attention.
I’ve spent the last several months documenting the intelligence revolution as it unfolds—week by week, announcement by announcement, tracking how the gap between “this is coming” and “this is here” keeps collapsing.
But this week, something happened that I can’t frame as just another data point. It’s bigger than that.
On February 9, 2026, Mrinank Sharma—the head of Anthropic’s Safeguards Research Team, the person literally responsible for making Claude safe—posted his resignation letter on X. It was viewed over a million times. And what he wrote should be required reading for anyone who still thinks we have decades to figure this out.
“The world is in peril,” he wrote. “And not just from AI, or bioweapons, but from a whole series of interconnected crises unfolding in this very moment.”
This isn’t some disgruntled engineer venting on social media. This is an Oxford-trained machine learning researcher who led the team responsible for defending against AI-assisted bioterrorism, understanding AI sycophancy, and writing one of the first AI safety cases ever produced. He didn’t leave for a competitor. He didn’t leave for more money.
He left to study poetry.
Let that sink in.
The Context That Makes This Explosive
Sharma’s resignation doesn’t exist in isolation. It lands in the middle of what I can only describe as Anthropic’s most consequential week in its history—and possibly the most consequential week for the entire AI industry in 2026.
Here’s what happened in the ten days before Sharma walked out:
On February 3, Anthropic released industry-specific plugins for Claude Cowork—its workplace automation tool designed for legal, finance, marketing, and data analysis workflows. What the company described as a relatively minor product update triggered something unprecedented: a $285 billion market rout in a single day. Bloomberg reported that a Goldman Sachs basket of US software stocks fell 6%, its steepest decline since April 2025’s tariff selloff. An index of financial services firms dropped almost 7%. Thomson Reuters fell over 15%. LegalZoom crashed more than 15%. FactSet plunged 10%. The Nifty IT index in India—representing the $300 billion outsourcing industry—crashed nearly 6%.
Then on February 6, Anthropic released Claude Opus 4.6, which it described as capable of conducting sophisticated professional tasks and coordinating entire teams of AI agents working in parallel. Financial data providers took another hit. S&P Global, Moody’s, and Nasdaq all saw sharp declines.
And then, amid the wreckage, The Telegraph published something that should have gotten more attention than it did: results from an internal Anthropic employee survey.
“It kind of feels like I’m coming to work every day to put myself out of a job,” one staffer said.
“In the long term, I think AI will end up doing everything and make me and many others irrelevant,” another confided.
Three days later, the head of safety walked out the door.
What Sharma Actually Said (And What He Meant)
Let me be precise about the resignation letter, because most coverage focused on the poetry angle and missed what matters.
Sharma described his accomplishments at Anthropic: understanding AI sycophancy and its causes, developing defenses against AI-assisted bioterrorism, putting those defenses into production, and writing one of the first AI safety cases. These aren’t abstract research projects. These are attempts to build guardrails around increasingly powerful systems.
Then he said something that should make everyone in the AI industry uncomfortable: “Throughout my time here, I’ve repeatedly seen how hard it is to truly let our values govern our actions. I’ve seen this within myself, within the organization, where we constantly face pressures to set aside what matters most, and throughout broader society too.”
Read that again carefully.
This is the head of safety at the company that built its entire brand on being the “responsible” AI lab—the one founded by former OpenAI researchers who left specifically because they felt OpenAI was prioritizing products over safety. Anthropic’s whole reason for existing is supposed to be different. And the person most responsible for that difference is telling us, on his way out: it’s not working.
He’s not accusing Anthropic of specific wrongdoing. He’s saying something worse: that the structural pressures of the AI race make it nearly impossible for any organization to live its values, no matter how sincere those values are.
“We appear to be approaching a threshold where our wisdom must grow in equal measure to our capacity to affect the world, lest we face the consequences.”
His final research project at Anthropic? Studying how AI assistants could make us less human—how they might distort our humanity itself.
He concluded that instead of trying to make AI systems slightly less sycophantic through technical fixes, he felt called to work that addresses the situation through a completely different lens. Not more code. Not better alignment techniques. Poetry.
The Pattern I Can’t Ignore
I track patterns. That’s what this series does. And the pattern emerging here is one I’ve been writing about for months, but it’s never been this stark.
Sharma isn’t the first safety researcher to leave an AI company with warnings. Jan Leike left OpenAI’s Superalignment team in 2024, saying the company was “prioritizing getting out newer, shinier products” over user safety. Gretchen Krueger left shortly after, calling for better “decision-making processes, accountability, transparency.” Timnit Gebru’s departure from Google in 2020 raised similar concerns about the gap between stated values and internal reality.
But something has shifted. The earlier departures were about companies not doing enough safety work. Sharma’s departure suggests something different: that the gap between technical capacity and human wisdom has grown so large that incremental safety work may no longer be meaningful.
And he’s not alone at Anthropic. Harsh Mehta, an R&D engineer, left recently. Behnam Neyshabur, an AI scientist, departed too. Dylan Scandinaro, an AI safety researcher, also left. Unlike Sharma, they’re staying in the AI industry. But the cluster of departures from the company that brands itself as the safety-first alternative tells its own story.
Meanwhile, Anthropic is reportedly seeking a funding round that would value it at $350 billion. That’s not a safety research lab. That’s a commercial juggernaut under intense pressure to justify its valuation.
The “Innovation Monopoly” in Real Time
I’ve been developing a framework I call the “innovation monopoly”—the mechanism by which foundation model companies absorb entire business categories. What happened with Claude Cowork is the most dramatic real-world validation of that thesis I’ve ever seen.
Think about what actually happened. Anthropic released a set of plugins—essentially the ability for Claude to do legal contract review, financial analysis, compliance tracking, sales forecasting, and customer relationship management. Not perfectly. Not yet replacing entire departments. But enough that investors instantly recognized the trajectory.
And the market’s response wasn’t theoretical. $285 billion evaporated in a day. Not because Claude Cowork is currently better than Thomson Reuters or Salesforce or LegalZoom at their core products. But because investors could see the trajectory—and the trajectory points toward capability absorption.
This is what I’ve been calling “agentrification”: the keystroke-by-keystroke automation of cognitive work. Not a single dramatic moment where AI replaces humans, but a gradual absorption where each new capability release eats into another slice of knowledge work.
Rest of World’s analysis of the impact on Indian IT was particularly striking. The $300 billion Indian outsourcing industry—companies like Infosys, TCS, Wipro—is built on billing for human hours spent on exactly the kind of repetitive knowledge work that Claude Cowork automates. Contract reviews. Regulatory compliance. Data analysis. The sell-off wasn’t panic. It was recognition.
As one analyst at Deutsche Bank put it, the market has shifted from a “every tech stock is a winner” mindset to “a true winners and losers landscape.”
What Dario Amodei Already Told Us
Here’s the part that makes Sharma’s departure even more significant. Anthropic’s own CEO, Dario Amodei, has been saying the quiet part out loud for months.
In January 2026, Amodei published a 20,000-word essay—”The Adolescence of Technology”—warning that AI would cause “unusually painful” disruption to jobs. He told Axios last year that AI could wipe out half of all entry-level white-collar jobs within five years and push unemployment to 10-20%. He said CEOs would “quietly stop hiring and start replacing humans with AI the moment it makes business sense.”
And then his company released the exact product that does exactly that. And his safety chief walked out.
The contradiction isn’t subtle. It’s the central tension of the entire AI industry made flesh: the people building these systems know what’s coming, are warning about what’s coming, and are building it anyway. Because if they don’t, their competitors will.
This is precisely what Sharma meant by “pressures to set aside what matters most.” Not malice. Not ignorance. Structural inevitability.
Anthropic’s own Economic Index, released in January 2026, found that 49% of jobs can now use AI in at least a quarter of their tasks—up from 36% in early 2025. The company’s own research shows adoption spreading faster than any major technology in the past century.
Why Poetry Might Be the Right Response
I know how it sounds. The AI safety chief leaves to study poetry. It’s easy to mock—several outlets did exactly that. PC Gamer called it an “epic vaguepost.” One X user noted it had “main character energy and footnotes.”
But I think Sharma might be seeing something the rest of us are missing.
His letter referenced “CosmoErotic Humanism,” cited Rilke and William Stafford, and framed the moment as requiring “courageous speech” and “poetic truth alongside scientific truth.” It sounds abstract. But consider what he’s actually saying: that the technical tools we’ve built to manage AI risk are insufficient because the problem isn’t fundamentally technical.
The problem is about wisdom. About values. About what kind of beings we want to be in a world where artificial intelligence can do our cognitive work better than we can.
You can’t engineer your way out of an existential question. You can build better guardrails, better alignment techniques, better safety cases. Sharma did all of that. And he concluded it’s not enough.
“Not knowing is most intimate,” he quoted from Zen tradition. That’s not resignation. That’s a different kind of starting point—one that begins with acknowledging we don’t have the frameworks we need, and that inventing them requires different tools than the ones that built the systems we’re trying to contain.
The 18-Month Timeline Isn’t Theoretical Anymore
In Week 5 of this series, I wrote about how the AGI timeline collapsed—from decades to years to “already here” in 36 months. In subsequent weeks, I’ve tracked how the infrastructure for superintelligence is being built on concrete timelines with specific operational dates.
Sharma’s resignation confirms something I’ve suspected but hadn’t been able to articulate this clearly: the safety infrastructure is not keeping pace with capability development, and the people best positioned to know this are leaving.
Not going to competitors. Not building alternative safety approaches within the industry. Leaving entirely.
When your safety chief decides that studying poetry is a more appropriate response to the situation than continuing to build technical safeguards, that tells you something about the adequacy of technical safeguards.
Consider the sequence:
November 2025: AI pioneers declare AGI is already here
January 2026: Anthropic’s CEO warns of 50% entry-level job destruction and “unusually painful” disruption
February 3, 2026: Claude Cowork triggers $285 billion market rout
February 6, 2026: Claude Opus 4.6 launches with autonomous agent teams
February 9, 2026: Anthropic’s safety chief resigns, warning the world is in peril
That’s not a gradual evolution. That’s an acceleration curve. And we’re on it.
What This Means for You
I’m going to be direct, because I think the moment demands it.
If you’re in knowledge work—legal, financial services, consulting, data analysis, compliance, marketing—the Claude Cowork announcement isn’t a future threat. It’s a current competitive pressure. The companies that adopt these tools will operate at a fundamentally different cost structure than those that don’t. The $285 billion market reaction tells you that institutional investors already understand this.
If you’re in AI safety or governance, Sharma’s departure should be a five-alarm fire. The person who built one of the most sophisticated safety teams in the industry concluded that the structural incentives make it nearly impossible to prioritize safety when it conflicts with competitive pressure. If that’s true at Anthropic—the company explicitly founded to do safety right—where is it true?
If you’re a policymaker, the timeline for meaningful AI governance just compressed again. The gap between capability deployment and regulatory response isn’t narrowing. It’s widening. The fact that a “minor product update” can erase $285 billion in market value in a single day tells you how fast the economic transformation is moving relative to the governance response.
And if you’re a human being trying to understand what’s happening, I think Sharma’s letter offers an unexpected gift: permission to sit with the uncertainty. To acknowledge that the people closest to this technology don’t have the answers either. And that maybe the wisdom we need won’t come from more engineering.
The Thread You Hold
Sharma closed his letter with William Stafford’s poem “The Way It Is”—about a thread you follow through life, a thread that doesn’t change even when everything around it does.
I think that’s the real message, buried under all the media coverage about poetry degrees and vagueposting: there’s something essential that needs to be held onto as everything transforms. Something that can’t be automated, can’t be optimized, can’t be captured in a safety case or an alignment technique.
Sharma spent two years building the best technical safeguards he could. Then he walked away, saying the challenge is bigger than technology.
I think he’s right.
The question is whether the rest of us will realize it in time.
This is part of an ongoing series tracking the intelligence revolution as it unfolds. Previous installments have examined Amazon’s warehouse automation, NVIDIA’s AI Factory infrastructure, the AGI timeline collapse, and Singapore’s economic vulnerability.
Dr. Elias Kairos Chen is the author of “Framing the Intelligence Revolution: How AI Is Already Transforming Your Life, Work, and World.”
Sources:
Mrinank Sharma resignation letter, posted on X, February 9, 2026
Bloomberg: Anthropic AI Tool Sparks $285 Billion Selloff (February 3, 2026)
Fortune: Anthropic’s Claude Triggered a Trillion-Dollar Selloff (February 6, 2026)
CNBC: AI Fears Pummel Software Stocks (February 6, 2026)
The Telegraph: Anthropic Employees Internal Survey (February 2026)
Futurism: Anthropic Insiders Afraid They’ve Crossed a Line (February 7, 2026)
CNBC: Dario Amodei Warns AI May Cause ‘Unusually Painful’ Job Disruption (January 27, 2026)
Anthropic Economic Index, Fourth Edition (January 2026)
Rest of World: Why Claude Cowork Is a Math Problem Indian IT Can’t Solve (February 10, 2026)
ABC News: Why a New AI Tool Hammered Software Stocks (February 2026)
eWeek: Anthropic Safety Leader Resigns (February 10, 2026)
BusinessToday: Anthropic Safety Head Resigns (February 10, 2026)



