Andrea Vallone, who led OpenAI's research on how AI models should respond to users showing signs of mental health distress, has left the company to join Anthropic's alignment team. During her three years at OpenAI, Vallone built the model policy research team, worked on deploying GPT-4 and GPT-5, and helped develop safety techniques such as rule‑based rewards. At Anthropic, she will continue her work under Jan Leike, focusing on aligning Claude's behavior in novel contexts. Her move highlights ongoing industry concern over AI safety, especially around mental‑health‑related interactions.
Leia mais →