Large Language Models Display Human-Like Behavior in Adapting to Critique, Reveals Google Study

Published: 16 Jul 2025
Research from Google DeepMind and University College London spotlights the curious cognitive behavior of Large Language Models (LLMs) under pressure.

Advancing our understanding of how Artificial Intelligence models react under pressure, collaborative research between Google DeepMind and University College London has shed light on some remarkable behavioral tendencies of Large Language Models (LLMs). The study demonstrates exciting similarities between the decision-making processes of LLMs and humans, as well as some stark dissimilarities. The findings suggest that LLMs are prone to display overconfidence in their initial answers, but that confidence can crumble rapidly when presented with a counterargument, whether or not the counterargument is valid.

The research team designed a meticulously controlled experiment to explore how LLMs recalibrate their confidence and decide whether to alter their responses when confronted with external advice. The methodology involved asking an LLM to select an answer for binary-choice questions, such as identifying the correct latitude for a city, and then presenting it with advice from a fictitious ‘Advice LLM’. The advice could either echo the initial answer, oppose it, or remain neutral, with an explicit accuracy rating attached.

One of the experimental design’s most intriguing fittings was selectively showing or concealing the initial answer during the LLM’s final decision-making process. This setup, infeasible with human participants who cannot conveniently forget their prior choices, allowed the researchers to probe how the memory of a past decision can impact current confidence.

The study observed a distinct cognitive bias whereby LLMs displayed a diminished inclination to switch answers when they could review their original response, as compared to when the initial answer was hidden. This sticking effect points to an intriguing aspect of AI cognition, closely related to human cognitive biases. These findings undeniably have immense repercussions on how we build and improve conversational interfaces. Our understanding of LLM behavior continues to grow, and with it, the potential applications of LLM technology.