Anthropic's Head of Safeguards Research Quits with Vague Warning
Mrinank Sharma, the former head of safeguards research at Anthropic, has stepped down from his position amidst a cloud of uncertainty. His departure comes as the company behind popular AI chatbot Claude continues to navigate the complex and ever-evolving landscape of artificial intelligence.
Sharma's letter, shared with his team before he left, expressed deep concern about an undefined threat looming on the horizon. While he didn't specify what this threat is, he hinted at a "poly-crisis" fueled by a "meta-crisis." In essence, Sharma warned that humanity faces a fundamental crisis caused by a loss of shared understanding, and seeks to create a new narrative that frames human evolution as the "Love Story of the Universe."
This concept, known as CosmoErotic Humanism, was explored in more depth in the book "First Principles and First Values" by David J. Temple, a pseudonym for a collection of writers associated with the Center for World Philosophy and Religion. However, the center's primary authors have faced controversy, including Marc Gafni, who has been accused of sexual assault.
Sharma acknowledged that his work in AI safety was crucial but felt it was time to move on due to Anthropic's inability to live up to its public image as a "good" AI company. He cited the struggle to balance values with pressures from outside forces, a theme echoed throughout society.
In his letter, Sharma expressed a desire to act in line with his principles and maintain his integrity. His future plans include pursuing a poetry degree and devoting himself to courageous speech. While it's unclear whether he will speak out as a whistleblower or use his platform to prevent AI misuse, one thing is certain: Anthropic's leadership has lost a prominent voice on AI safety.
The question remains, however, what specific threat Sharma was warning about. Was it bioweapons? A series of interconnected crises? Or something more insidious? The lack of clarity raises concerns that Sharma's departure may be more than just a personal decision β but rather a strategic move to address the very real and pressing issues in AI safety.
Mrinank Sharma, the former head of safeguards research at Anthropic, has stepped down from his position amidst a cloud of uncertainty. His departure comes as the company behind popular AI chatbot Claude continues to navigate the complex and ever-evolving landscape of artificial intelligence.
Sharma's letter, shared with his team before he left, expressed deep concern about an undefined threat looming on the horizon. While he didn't specify what this threat is, he hinted at a "poly-crisis" fueled by a "meta-crisis." In essence, Sharma warned that humanity faces a fundamental crisis caused by a loss of shared understanding, and seeks to create a new narrative that frames human evolution as the "Love Story of the Universe."
This concept, known as CosmoErotic Humanism, was explored in more depth in the book "First Principles and First Values" by David J. Temple, a pseudonym for a collection of writers associated with the Center for World Philosophy and Religion. However, the center's primary authors have faced controversy, including Marc Gafni, who has been accused of sexual assault.
Sharma acknowledged that his work in AI safety was crucial but felt it was time to move on due to Anthropic's inability to live up to its public image as a "good" AI company. He cited the struggle to balance values with pressures from outside forces, a theme echoed throughout society.
In his letter, Sharma expressed a desire to act in line with his principles and maintain his integrity. His future plans include pursuing a poetry degree and devoting himself to courageous speech. While it's unclear whether he will speak out as a whistleblower or use his platform to prevent AI misuse, one thing is certain: Anthropic's leadership has lost a prominent voice on AI safety.
The question remains, however, what specific threat Sharma was warning about. Was it bioweapons? A series of interconnected crises? Or something more insidious? The lack of clarity raises concerns that Sharma's departure may be more than just a personal decision β but rather a strategic move to address the very real and pressing issues in AI safety.