Skip to content
PsSeasonO3FactGameHallucinationsChatbotsLairdTechnologyWin

OpenAI's assessments indicate an escalation in ChatGPT's propensity for hallucinatory responses, leaving questions unanswered regarding the root cause.

Advanced robot reasoning capabilities fuel escalating visions of flawed robot aspirations.

OpenAI's assessments indicate an escalation in ChatGPT's propensity for hallucinatory responses, leaving questions unanswered regarding the root cause.

A New Twist in AI's Mysteries: Hallucinations on the Rise

Remember when Anthropic revealed some shocking insights about AI models a while back? They found that the inner workings of these models were vastly different from what the models described as their thought processes. Now, there's a new puzzle to add to the mix—ever-increasing hallucinations. Following investigations by the leading chatbot innovator, OpenAI, it seems that their latest GPT-o3 and GPT-o4-mini large language models are highly prone to hallucinating, or spitting out false information.

According to reports in The New York Times, OpenAI's research findings indicate a substantial increase in hallucinations compared to their previous GPT-o1 model. The new GPT-o3 model hallucinated 33% of the time in PersonQA tests, compared to 51% for the SimpleQA tests. The GPT-o4-mini model even saw hallucinations up to 79%!

So, what's causing this troubling escalation in false information from these AI models? OpenAI is still conducting research to understand the issue better, but some industry experts believe that "reasoning" models may be the culprits.

Reasoning models are a type of language model designed for complex tasks. Unlike standard models that simply spit out text based on probability statistics, reasoning models break down tasks into steps resembling a human thought process.

However, OpenAI has disagreed with the notion that reasoning models are more prone to hallucinations. Gaby Raila of OpenAI told The Times that they are proactively working to reduce the higher rates of hallucination observed in GPT-o3 and GPT-o4-mini.

Whether it's true or not, one fact remains: AI models need to cut back on the nonsense and lies if they are to be as useful as their promoters envision. Given the increasing prevalence of AI, it's crucial to ensure that the output can be trusted without double-checking every statement. While this may be manageable for some tasks, the time-saving benefit of AI quickly disappears when you have to meticulously go through every piece of information it generates.

As of now, it remains to be seen whether OpenAI and the AI industry can tackle this issue and make reliable large language models a reality. AI's potential impact on our lives is immense, and it's crucial to address problems like hallucinations to unleash the full potential of this powerful technology.

Jeremy LairdJeremy has been writing about technology and PCs since the Netburst era (Google it!) and is fascinated by anything that beeps, blinks, or performs calculations faster than a calculator. When not immersed in the latest developments in artificial intelligence and cutting-edge tech, he enjoys tennis, cars, and exploring the intricacies of advanced manufacturing processes.

  1. Despite the rise of hallucinations in AI models, Jeremy Laird continues to investigate the developments in artificial intelligence and PC technology with his usual fascination.
  2. The potential impact of reliable large language models on our lives could be immense, according to Jeremy, who believes it's crucial to address issues like hallucinations to unleash the full potential of AI.
  3. During the Netburst era, Jeremy began writing about technology and PCs and has since shown interest in anything that beeps, blinks, or performs calculations faster than a calculator.
  4. In the midst of increasing hallucinations in AI models, it is vital to ensure AI output can be trusted without double-checking every statement, as Jeremy emphasizes.
  5. AI's complex task-oriented reasoning models, designed like human thought processes, may be contributing to the higher rates of hallucination seen in models like GPT-o3 and GPT-o4-mini, according to some industry experts—a notion that OpenAI has yet to fully accept.
Robots with enhanced reasoning capabilities stir concerns as they may lead to a surge in unproductive futuristic fantasies.
The advancements in robotic reasoning give rise to more harmful visions of artificial intelligence.
Advances in logical thinking among AI lead to an increase in disturbing visions of artificial intelligence.

Read also:

    Latest