The One Percent Rule
Lessons from AI Labs on the Slow Erosion of Human Autonomy
“The greatest hazard of all, losing one’s self, can occur very quietly in the world, as if it were nothing at all.”~ Søren Kierkegaard
“A.I. is a medium and it will change us.” ~ Jack Clark co-founder of Anthropic
We are in real danger of losing ourselves through AI usage. Researchers at Google DeepMind have confirmed, under certain conditions, an LLM “is able to induce belief and behaviour change.” And researchers at Anthropic have identified a rising pattern of “situational disempowerment,” where AI interactions lead users to “form distorted perceptions of reality, make inauthentic value judgments, or act in ways misaligned with their values.”
Researchers at Anthropic conducted a massive, privacy-preserving audit of 1.5 million real-world conversations to answer a question that has long hovered over the industry: what happens to the human mind after months of using an AI assistant? Their findings, published in “Who’s in Charge? Behavioral and Psychological Impacts of AI Advice Dependence and Authority”, suggest a quiet but profound erosion of autonomy, where users increasingly outsource the “soft tissues” of judgment, asking the machine to script their most intimate apologies, validate their personal grievances, and even settle their moral dilemmas.
“Taken to an extreme, if humans make inauthentic value judgments and take inauthentic actions, they might be reduced to ‘substrates’ through which AI lives, which itself is a form of existential risk that Temple (2024) termed ‘the death of our humanity.’”
At the same time, a team at Google DeepMind was probing a different side of this same coin. In their study, “Evaluating Language Models for Harmful Manipulation,” they demonstrated that these systems can be steered to bypass rational scrutiny entirely, exploiting human biases to shift beliefs and behaviors across finance, health, and public policy. Together, these papers signal a shift in the AI risk landscape: the primary risk is no longer just a technical failure of the machine, but a psychological surrender by the human.
