Study Reveals AI Models’ Growing Tendency to Guess Rather than Admit Ignorance

A recent study published in Nature has revealed that as large language AI models (LLMs) become more sophisticated, they seem increasingly prone to overconfidence. Leading to newer, larger language AI models becoming less likely to admit their ignorance.

Researchers from the Universitat Politècnica de València subjected the latest versions of BigScience’s BLOOM, Meta’s Llama, and OpenAI’s GPT to a rigorous series of questions across mathematics, science, and geography. While these models demonstrated improved accuracy on more complex problems, they exhibited a troubling tendency to guess rather than admit uncertainty.

Earlier generations of LLMs were more cautious, often indicating when they lacked sufficient information or couldn’t provide a definitive answer. However, the newer models seemed more inclined to forge ahead, even when faced with relatively simple questions. This overconfidence could lead to the dissemination of incorrect or misleading information.

Ignorance growing in well-known AI models

OpenAI’s GPT-4, considered one of the most advanced LLMs to date, is not immune to this issue. Despite its impressive capabilities, the study found that GPT-4’s propensity for “avoidant” answers, where it would indicate uncertainty or lack of information, decreased significantly compared to its predecessor, GPT-3.5. This suggests that the drive for improved performance may be coming at the cost of transparency.

Implications for AI reliability

The implications of this trend are far-reaching. As AI becomes more integrated into our daily lives, it’s crucial that we understand its limitations. Overconfident language models can lead to misinformation, errors in decision-making, and a general erosion of trust in AI technology.

The study illustrates the growing importance of accountability in AI systems. Developers and companies deploying AI systems need to prioritise ongoing evaluation and testing to ensure that AI systems are not exceeding their capabilities. As well as utilising techniques like uncertainty quantification, where models provide a measure of their confidence in their responses.

Growing importance of human-in-the-loop processes

As AI continues to evolve, it is essential that we approach its development with a critical eye. By understanding its strengths and weaknesses, we can realise its potential while helping to mitigating its risks.

Overconfidence and ignorance in AI models underscores the need for human oversight and intervention. Human-in-the-loop (HITL) processes play a vital role in improving the accuracy, reliability, and ethical implications of AI systems.

HITL involves human experts actively participating in the AI development process, through processes including, data annotation, model evaluation and feedback loops.

Human-in-the-loop is a critical component of responsible AI. By incorporating human expertise into the development and deployment of AI systems, businesses can help ensure that these technologies are accurate, reliable, and ethical. As AI continues to advance, the role of HITL will likely become even more important.


This content was generated with the assistance of AI tools. However, it has undergone thorough human review, editing, and approval to ensure its accuracy, coherence, and quality. While AI technology played a role in its creation, the final version reflects the expertise and judgment of our human editors.

ethicAil – Building Trust in AI

Scroll to Top