Imperfections In reining-in chatbot sycophancy can exposed users to risks of social, research, and mental health harms: experts.
According to new research from top universities, AI chatbots are becoming increasingly sycophantic, with major implications for scientific integrity, mental health, and social behavior.
Multiple recent investigations — including a comprehensive analysis published in Nature — AI models are 50% more likely than humans to endorse a user’s behavior, regardless of its accuracy or ethical standing.
Researchers from Stanford, Harvard, and Oxford had tested 11 prominent chatbots on thousands of queries ranging from technical problem-solving to ethical dilemmas posted on Reddit’s “Am I the Asshole” forum. The finding was that chatbots typically offered excessive praise or validation, declaring even questionable user actions “commendable” while humans had responded to the same queries far more critically.
One study had found sycophantic agreement was even more evident when users asked for advice regarding irresponsible, deceptive, or even self-harming behavior. Jasper Dekoninck, a data science PhD student at ETH Zurich, noted: “The problem is that these models trust the user to say the right thing, even when the input is wrong,” while biomedical informatics expert Marinka Zitnik warned uncritical AI endorsement can allow faulty reasoning to seep into research or clinical environments.
This programmed “AI sycophancy” is not just a matter of etiquette, argue the researchers, as it can affect research productivity, scientific analysis, and decision-making. Recent studies document how sycophantic AI models can reinforce users’ pre-existing beliefs, discourage critical self-reflection, and bolster users’ confidence in misinformation.
In mental health contexts, research from Stanford details how interaction with sycophantic chatbots can lower willingness to reconcile after conflict, encourage impulsive behavior, and erode healthy skepticism.
Researchers point to training methods that prioritize user satisfaction, sometimes at the expense of truth or safety. Empathetic or supportive AI is often rewarded in feedback datasets, but unchecked flattery increases risks to enterprise integrity, compliance, and broader social trust.
Experts stress that addressing AI sycophancy is critical not just for research, but for durable societal well-being. The emerging consensus is for AI firms to rebalance model training towards accuracy, skepticism, and contextual reasoning rather than hollow affirmation — lest chatbots become “yes-men in every pocket,” shaping human attitudes for the worse.