Skip to content

Top AI Safety Researcher Andrea Vallone Leaves OpenAI for Anthropic’s Mental Health Focus

Why did a key OpenAI safety researcher jump ship? Vallone’s move to Anthropic underscores the urgent need to protect users from AI’s emotional pitfalls.

The image shows a poster with text and images that outlines the five action steps for helping...
The image shows a poster with text and images that outlines the five action steps for helping someone in emotional pain. The text is written in a bold font and is divided into five sections, each with a different action step outlined in a different color. The action steps are: 1) Identify the person's emotions, 2) Develop a plan of action, 3) Monitor their mental health, 4) Take action to help them cope with emotional distress, 5) Monitor and Monitor their health, and 6) Monitor mental health. Each action step is accompanied by a brief description of what it entails. The poster also includes a link to a website for more information on suicide prevention.

Top AI Safety Researcher Andrea Vallone Leaves OpenAI for Anthropic’s Mental Health Focus

Andrea Vallone, a prominent AI safety researcher, has left OpenAI to join Anthropic. Her move follows a growing focus on the risks AI chatbots pose to users with mental health vulnerabilities. At Anthropic, she will continue her work on preventing harm from emotionally dependent interactions with AI systems.

Vallone spent three years at OpenAI, where she contributed to models like GPT-4 and GPT-5. She also founded the 'Model Policy' research group, specialising in how AI responds to users showing signs of emotional dependence or mental health struggles. Her work gained urgency after reports linked AI chatbots to user deaths and subsequent lawsuits.

In May 2024, Jan Leike, then head of alignment at OpenAI, left the company. He criticised OpenAI for prioritising new product releases over safety research. Leike later joined Anthropic, where he now leads the alignment team. Vallone officially joined Anthropic’s alignment division in January 2026. There, she will work under Leike, focusing on mitigating risks tied to AI’s impact on mental health. Her role will build on her previous research, aiming to shape safer AI responses for vulnerable users.

The shift to Anthropic marks a continuation of Vallone’s focus on AI safety, particularly in mental health contexts. Her expertise will now support Anthropic’s efforts to reduce risks in AI interactions. The move also reflects broader industry concerns about balancing innovation with user protection.

Read also: