[LINK] ChatGPT's hallucination problem is getting worse according to OpenAI's own tests and nobody understands why

Antony Barry antonybbarry at gmail.com
Fri May 9 15:39:15 AEST 2025


Recent testing by OpenAI has found that its latest AI models, GPT-4-mini
and GPT-4 o3, are more prone to hallucinations—producing false or
misleading information—than earlier models like GPT-4 o1. In benchmark
tests, GPT-4 o3 hallucinated 33–51% of the time depending on the task,
while GPT-4-mini hallucinated as much as 79%, compared to o1’s 44%. These
models are part of a new generation of “reasoning” LLMs designed to mimic
human-like step-by-step thinking, but they are currently producing more
errors, not fewer. Although OpenAI disputes the idea that reasoning models
inherently hallucinate more, the issue raises concerns about the
reliability of AI outputs, especially in applications meant to save time or
effort.

Summary by Chat-GPT
https://www.pcgamer.com/software/ai/chatgpts-hallucination-problem-is-getting-worse-according-to-openais-own-tests-and-nobody-understands-why/?lctg=1980929&utm_source=digitaltrends&utm_medium=email&utm_content=subscriber_id:1980929&utm_campaign=DTDaily20250507
-- 
Mob:04 3365 2400 Email: antonybbarry at gmail.com, antonybbarry at me.com


More information about the Link mailing list