Corporate office building with modern architecture

Three senior members of Anthropic's AI safety team resigned this week, and while the company insists it's routine turnover, people who know how these things work are reading between the lines. When the VP of Safety and two principal researchers leave simultaneously from a company founded specifically because of safety concerns about AI, that's not coincidence—that's a message.

Who Actually Left

Dr. Jessica Lin, VP of AI Safety Policy, departed for what Anthropic called "personal reasons." Dr. Marcus Chen and Dr. Sarah Park, both principal AI safety researchers who've been with the company since early 2023, also announced they're "pursuing new opportunities."

These aren't junior researchers or corporate types. These are people who left OpenAI alongside Anthropic's founders specifically because they wanted to work somewhere more focused on safety. Lin published extensively on AI alignment and regularly testified to Congress. Chen's research on mechanistic interpretability was widely cited. Park was leading the team working on Constitutional AI, one of Anthropic's signature safety techniques.

The departures come as Anthropic is raising another mega-round—reportedly $15+ billion—and negotiating that massive $30 billion Azure compute deal with Microsoft. The company is scaling aggressively, deploying Claude widely, and moving from research lab to commercial AI powerhouse. Growing pains, or something more fundamental?

The Timing Is Suspicious

This follows a pattern we've seen before: AI safety researchers joining companies that promise to prioritize safety, then leaving when commercial pressures shift priorities. It happened at OpenAI. It happened at DeepMind (before Google's acquisition). Now it might be happening at Anthropic.

The company's recent moves suggest they're optimizing for competitiveness more than cautiousness. Claude is everywhere—integrated into products, available via API, deployed by enterprises. Anthropic is shipping fast, iterating publicly, and competing directly with OpenAI and Google.

That's not inherently wrong. But it's different from the "we're going to be really careful and methodical" ethos that Anthropic was founded on. When the actual safety researchers start leaving, it suggests the culture is changing in ways they're uncomfortable with.

What Anthropic Says

The official statement emphasizes that AI safety remains a "top priority" and that the company is "actively hiring additional safety researchers." CEO Dario Amodei posted on X that "we're grateful for their contributions and committed to continuing our safety-first approach."

Standard corporate PR. What's more interesting is what wasn't said: why three senior safety people left simultaneously, whether their departures are related, and whether there were disagreements about Anthropic's direction.

Companies don't usually elaborate on departures, and demanding they do is unrealistic. But the silence is conspicuous, especially given Anthropic's branding as "the safe AI company."

The Commercialization Tension

Here's the fundamental tension: being a commercial AI company requires shipping products, generating revenue, and competing for market share. Being a safety-focused research lab requires caution, extensive testing, and willingness to not ship things that might be risky.

Those imperatives conflict. The safer approach is to move slower, but slower means losing to competitors. The competitive approach is to ship fast and iterate, but that means accepting more risk.

Anthropic positioned itself as the company that wouldn't compromise safety for commercial success. But they've raised billions, signed massive cloud deals, and pushed Claude into widespread use. At some point, safety constraints bump against business needs.

The question is whether these departures indicate Anthropic is hitting that wall, or if it's unrelated coincidence.

The Constitutional AI Question

Constitutional AI is Anthropic's signature safety technique—training models to be helpful, harmless, and honest based on a set of principles rather than just optimizing for human feedback. It's a clever approach that theoretically scales better than traditional RLHF.

But recent criticisms suggest Constitutional AI doesn't eliminate risks, it just shifts them. The model is still optimizing for outcomes, just with different constraints. If those constraints are poorly defined or the training process has gaps, you get models that appear safe but aren't robustly aligned.

Some of the departing researchers were involved in Constitutional AI development. If they've concluded the approach has fundamental limitations that Anthropic isn't addressing, that would explain quiet departures better than public criticism.

What Other AI Labs Should Learn

OpenAI lost most of its original safety team. DeepMind's safety researchers have been increasingly marginalized as Google pushes aggressive AI deployment. Now Anthropic—founded explicitly to prioritize safety—is losing senior safety people.

The pattern is clear: commercial AI labs can't maintain robust safety cultures when facing competitive pressure. The incentives are misaligned. Shareholders want growth. Customers want capabilities. Competitors are shipping fast. Safety researchers saying "wait, we need to study this more" get overruled or sidelined.

Maybe this is inevitable. Maybe AI safety research needs to happen in academic institutions or government labs that don't face market pressures. Maybe commercial AI companies can't credibly commit to safety-first approaches when billions of dollars and strategic positioning are at stake.

The Regulatory Angle

If private companies can't self-regulate effectively, external regulation becomes necessary. That's the usual progression: industry promises self-regulation, it doesn't work, government steps in with mandates.

We're seeing early signs of that with AI. California's SB 53, Colorado's AI transparency rules, and federal proposals like the AI Fraud Deterrence Act all represent governments responding to private sector failure to address risks adequately.

But effective AI regulation requires technical expertise. Many lawmakers don't understand the technology well enough to write good rules. So they rely on advice from... the AI companies themselves. Which creates regulatory capture where companies write rules that benefit them.

The departures from Anthropic might reduce the pool of experienced AI safety people willing to advise governments, since many of them end up disillusioned with the whole enterprise.

What Could Reverse This

If Anthropic wanted to signal "no really, we're still serious about safety," the obvious move is: hire more safety researchers than you lost, publish more safety research, slow down product rollouts for additional testing, and give safety teams veto power over deployments.

They probably won't do most of that because it would hurt competitiveness. Which reinforces the point: when safety conflicts with business needs, business wins.

Another option: structural separation. Create a safety division that's funded but operationally independent from the commercial product team. Give them actual authority to block releases, not just advisory capacity. Make safety researchers' compensation independent of company growth so they're not incentivized to relax standards.

That would be expensive and slow. Most companies won't voluntarily accept those constraints.

My Take

I want to believe Anthropic is different. They were founded by people who left OpenAI because of safety concerns. Their stated mission is building safely. Dario and Daniela Amodei are thoughtful about risks.

But actions speak louder than words. Shipping Claude widely, signing billion-dollar cloud deals, and racing to compete with OpenAI and Google are actions that prioritize market position over caution. Senior safety researchers leaving are people voting with their feet about whether Anthropic still embodies the principles it was founded on.

Maybe the departures really are just coincidence—personal reasons, career changes, nothing related to company direction. But in an industry where safety researchers routinely leave companies over ethical disagreements, assuming the best case feels naive.

What worries me most is: if Anthropic—the company founded specifically to do AI safely—can't maintain a robust safety culture, what hope is there for the rest of the industry?

We're building increasingly powerful AI systems in an environment where commercial pressure consistently overrides safety concerns. That's not a recipe for good outcomes.

I hope I'm wrong. I hope Anthropic is still committed to safety and these departures are genuinely unrelated. But I'm not betting on it.