7.2 C
New York

Artificial intelligence ‘hallucinates’ at least 3% of the time shows study

A new study suggests that AI systems, specifically chatbots, may ‘hallucinate’ or invent information more often than many realize. The study found that the ChatGPT language model from OpenAI had the lowest hallucination rate, with incorrect answers between 3% and 3.5% of the time. Other language models, such as Llama from Meta and Claude 2 from Anthropic, had hallucination rates of up to 8.5%. The most egregious offender was Google's Palm language model, which had a hallucination rate of between 12.1% and 27.2%.

Artificial Intelligence (AI) has been a game-changer in various fields, but recent studies have shown that AI ‘hallucinates’ or provides incorrect answers at least 3% of the time. This phenomenon is not limited to a single AI model but is observed across various popular language models.

A startup called Vectara, founded by former Google employees, took on the task of investigating this issue. They conducted experiments, tasking different popular language models with summarizing text documents. The results they obtained were nothing short of startling.

OpenAI’s Chat-GPT presented the least hallucinations, with incorrect answers between 3% and 3.5% of the time. Other models like Llama from Meta, Cohere, Claude 2 from Anthropic, Mistral 7B, and Palm from Google showed error rates ranging from 5% to as high as 27.2%.

These hallucinations are not confined to a specific task but can vary depending on the nature of the request. For instance, when requesting a response without specifying a source, the likelihood of hallucinations increases. In the case of Google’s tool, errors are more common due to the longer and contextualized nature of the answers it generates.

Not only chatbots like Google’s and Microsoft’s Bing have been found to spew nonsensical or incorrect information, but they can also fabricate details altogether. For instance, Google’s chatbot provided inaccurate information regarding the James Webb telescope, while Bing’s chatbot offered bogus details on various topics.

In March, ChatGPT cited several fake court cases when writing a legal brief. Even when given straightforward tasks like summarizing news articles, these chatbots persistently invent information.

The implications of these hallucinations can be severe, as demonstrated in a courtroom incident where a lawyer employed ChatGPT to research similar cases for their argument. The AI model generated false cases, which were then used in official documents, potentially leading to legal consequences.

Contrary to popular belief, these AI models lack the capability to reason or conduct concrete searches. They predict the most logical word order based on the user’s question, leading to potential inaccuracies or hallucinations.

This phenomenon extends to image generators, audios, or videos, where abnormalities like strange fingers or missing limbs can occur.

While companies strive to minimize these flaws, some experts question whether they are inherent to the technology itself. They suggest that users of AI language models should remain aware of the risk of hallucinations and employ these models as critical tools, acting as “co-pilots” rather than relying solely on them as replacements for human judgment.

The study’s findings also prompt broader questions about the future of AI. As AI technology continues to advance, it becomes increasingly crucial to consider the ethical implications of its usage and ensure that AI models are utilized responsibly and ethically.

Subscribe

Related articles

AI in 2025: Statistical Breakdown of Trends – The Complete Picture

In 1965, Gordon Moore predicted computing power would double...

10 Types of AI Agents and How They Can Transform Businesses

Businesses gain significant operational benefits through artificial intelligence (AI)...

Ripple vs. SWIFT: A Comparative Analysis and Future Outlook

Ripple and its XRP cryptocurrency pose a threat to...

6 Tech-Based Fuel Management Tools to Help Minimise Costs and Maximise Value

Fuel is a necessary component in almost every industrial...

Serhii Tokarev: Roosh Ventures Propels Jome Forward with New Investment

Ukrainian-founded startup Jome recently announced the successful closure of...
About Author
editorialteam
editorialteam
If you wish to publish a sponsored article or like to get featured in our magazine please reach us at contact@alltechmagazine.com