Anthropic studied what gives an AI system its ‘personality’ — and what makes it ‘evil’
2 Articles
2 Articles
Anthropic studied what gives an AI system its ‘personality’ — and what makes it ‘evil’
On Friday, Anthropic debuted research unpacking how an AI system's "personality" - as in, tone, responses, and overarching motivation - changes and why. Researchers also tracked what makes a model "evil." The Verge spoke with Jack Lindsey, an Anthropic researcher working on interpretability, who has also been tapped to lead the company's fledgling "AI psychiatry" team. "Something that's been cropping up a lot recently is that language models can…
Anthropic Study: Training Data Shapes AI Personalities Like Claude’s
In the rapidly evolving field of artificial intelligence, Anthropic has emerged as a key player, pushing boundaries on how AI systems develop personalities that influence their interactions with users. Founded in 2021 by former OpenAI executives including siblings Daniela and Dario Amodei, the company focuses on creating safe, interpretable AI, as detailed in its Wikipedia entry. Their latest research, released on Friday, delves into the mechani…
Coverage Details
Bias Distribution
- 100% of the sources lean Left
Factuality
To view factuality data please Upgrade to Premium