Anthropic safeguards lead resigns, warns of growing AI safety crisis

by Spencer Haag

Mrinank Sharma, who led safeguards research at Anthropic, resigned from the AI company the day gone by and publicly shared his departure letter.

In the letter posted to X, Sharma cited mounting unease over gaps between acknowledged principles and precise decisions at AI organizations and in society extra broadly. He described a widening disconnect between ethical commitments and operational realities.

This day is my last day at Anthropic. I resigned.

Right here is the letter I shared with my colleagues, explaining my decision. pic.twitter.com/Qe4QyAFmxL

— mrinank (@MrinankSharma) February 9, 2026

“It is a ways definite to me that the time has come to switch on,” Sharma wrote.

Sharma spent two years at the Claude developer, where he labored on defenses in opposition to AI-enabled organic threats, interior accountability instruments, and early frameworks for documenting AI security measures. He additionally studied how chatbots can improve person biases and often reshape human judgment.

The researcher praised used colleagues for their technical capacity and upright seriousness however signaled a shift away from corporate AI work. He announced plans to pursue writing, private educating, and perhaps graduate uncover in poetry.

His departure follows a duration of heightened attention on how leading AI builders arrange interior dissent, show dangers, and balance hasty functionality gains in opposition to security research.

Related Posts