AI Hallucination
   HOME

TheInfoList



OR:

In
artificial intelligence Artificial intelligence (AI) is intelligence—perceiving, synthesizing, and inferring information—demonstrated by machines, as opposed to intelligence displayed by animals and humans. Example tasks in which this is done include speech r ...
, a hallucination or artificial hallucination is a confident response by an artificial intelligence that does not seem to be justified by its training data when the model has a tendency of "hallucinating" deceptive data. The term is derived from the psychological concept of hallucination because they share similar characteristics. One of the dangers of hallucinations is that the output of the model will look correct even if it is wrong.


In natural language processing

In
natural language processing Natural language processing (NLP) is an interdisciplinary subfield of linguistics, computer science, and artificial intelligence concerned with the interactions between computers and human language, in particular how to program computers to proc ...
, a hallucination is often defined as "generated content that is nonsensical or unfaithful to the provided source content". Errors in encoding and decoding between text and representations can cause hallucinations. AI training to produce diverse responses can also lead to hallucination. Hallucinations can also occur when the AI is trained on a dataset wherein labeled summaries, despite being factually accurate, are not directly grounded in the labeled data purportedly being "summarized". Larger datasets can create a problem of parametric knowledge (knowledge that is hard-wired in learned system parameters), creating hallucinations if the system is overconfident in its hardwired knowledge. In systems such as
GPT-3 Generative Pre-trained Transformer 3 (GPT-3) is an autoregressive language model that uses deep learning to produce human-like text. Given an initial text as prompt, it will produce text that continues the prompt. The architecture is a standa ...
, an AI generates each next word based on a sequence of previous words (including the words it has itself previously generated in the current response), causing a cascade of possible hallucination as the response grows longer. By 2022, papers such as the ''
New York Times ''The New York Times'' (''the Times'', ''NYT'', or the Gray Lady) is a daily newspaper based in New York City with a worldwide readership reported in 2020 to comprise a declining 840,000 paid print subscribers, and a growing 6 million paid ...
'' expressed concern that, as adoption of bots based on large
language models A language model is a probability distribution over sequences of words. Given any sequence of words of length , a language model assigns a probability P(w_1,\ldots,w_m) to the whole sequence. Language models generate probabilities by training on ...
continued to grow, unwarranted user confidence in bot output could lead to problems. In August 2022,
Meta Meta (from the Greek μετά, '' meta'', meaning "after" or "beyond") is a prefix meaning "more comprehensive" or "transcending". In modern nomenclature, ''meta''- can also serve as a prefix meaning self-referential, as a field of study or ende ...
warned during its release of BlenderBot 3 that the system was prone to "hallucinations", which Meta defined as "confident statements that are not true". On 15 November 2022, Meta unveiled a demo of Galactica, designed to "store, combine and reason about scientific knowledge". Content generated by Galactica came with the warning "Outputs may be unreliable! Language Models are prone to hallucinate text." In one case, when asked to draft a paper on creating avatars, Galactica cited a fictitious paper from a real author who works in the relevant area. Meta withdrew Galactica on 17 November due to offensiveness and inaccuracy.
OpenAI OpenAI is an artificial intelligence (AI) research laboratory consisting of the for-profit corporation OpenAI LP and its parent company, the non-profit OpenAI Inc. The company conducts research in the field of AI with the stated goal of promo ...
's
ChatGPT ChatGPT (Generative Pre-trained Transformer) is a chatbot launched by OpenAI in November 2022. It is built on top of OpenAI's GPT-3 family of large language models, and is fine-tuned (an approach to transfer learning) with both supervised and ...
, released in beta-version to public in December 2022, is based on the GPT-3.5 family of large language models. Professor Ethan Mollick of
Wharton Wharton may refer to: Academic institutions * Wharton School of the University of Pennsylvania * Wharton County Junior College * Paul R. Wharton High School * Wharton Center for Performing Arts, at Michigan State University Places * Wharton, Che ...
has called ChatGPT an "omniscient, eager-to-please intern who sometimes lies to you". Data scientist Teresa Kubacka has recounted deliberately making up the phrase "cycloidal inverted electromagnon" and testing ChatGPT by asking ChatGPT about the (nonexistent) phenomenon. ChatGPT invented a plausible-sounding answer backed with plausible-looking citations that compelled her to double-check whether she had accidentally typed in the name of a real phenomenon. Other scholars such as
Oren Etzioni Oren Etzioni (born 1964) is an American entrepreneur, Professor Emeritus of computer science, and founding CEO of the Allen Institute for Artificial Intelligence (AI2). On June 15, 2022, he announced that he will step down as CEO of AI2 effective ...
have joined Kubacka in assessing that such software can often give you "a very impressive-sounding answer that's just dead wrong". Mike Pearl of ''
Mashable Mashable is a digital media platform, news website and entertainment company founded by Pete Cashmore in 2005. History Mashable was founded by Pete Cashmore while living in Aberdeen, Scotland, in July 2005. Early iterations of the site were ...
'' tested ChatGPT with multiple questions. In one example, he asked the model for "the largest country in
Central America Central America ( es, América Central or ) is a subregion of the Americas. Its boundaries are defined as bordering the United States to the north, Colombia to the south, the Caribbean Sea to the east, and the Pacific Ocean to the west. ...
that isn't
Mexico Mexico ( Spanish: México), officially the United Mexican States, is a country in the southern portion of North America. It is bordered to the north by the United States; to the south and west by the Pacific Ocean; to the southeast by Guate ...
". ChatGPT responded with Guatemala, when the answer is instead
Nicaragua Nicaragua (; ), officially the Republic of Nicaragua (), is the largest Sovereign state, country in Central America, bordered by Honduras to the north, the Caribbean Sea, Caribbean to the east, Costa Rica to the south, and the Pacific Ocean to ...
. When CNBC asked ChatGPT for the lyrics to "The Ballad of Dwight Fry", ChatGPT supplied invented lyrics rather than the actual lyrics. In the process of writing a review for the new iPhone 14 Pro, ChatGPT incorrectly volunteered the relevant chipset as the A15 Bionic rather than the A16 Bionic, although this can be attributed to the fact that ChatGPT was trained on a dataset ending in 2021, a year before the release of the iPhone 14 Pro. Asked questions about
New Brunswick New Brunswick (french: Nouveau-Brunswick, , locally ) is one of the thirteen Provinces and territories of Canada, provinces and territories of Canada. It is one of the three Maritime Canada, Maritime provinces and one of the four Atlantic Canad ...
, ChatGPT got many answers right but incorrectly classified
Samantha Bee Samantha Anne Bee (born October 25, 1969) is a Canadian-American comedian, writer, producer, political commentator, actress, and television host. Bee rose to fame as a correspondent on '' The Daily Show with Jon Stewart'', where she became th ...
as a "person from New Brunswick". Asked about astrophysical magnetic fields, ChatGPT incorrectly volunteered that "(strong) magnetic fields of
black holes A black hole is a region of spacetime where gravity is so strong that nothing, including light or other electromagnetic waves, has enough energy to escape it. The theory of general relativity predicts that a sufficiently compact mass can defo ...
are generated by the extremely strong gravitational forces in their vicinity". (In reality, as a consequence of the
no-hair theorem The no-hair theorem states that all stationary black hole solutions of the Einstein–Maxwell equations of gravitation and electromagnetism in general relativity can be completely characterized by only three independent ''externally'' observab ...
, a black hole without an accretion disk is believed to have no magnetic field.) ''
Fast Company ''Fast Company'' is a monthly American business magazine published in print and online that focuses on technology, business, and design. It publishes six print issues per year. History ''Fast Company'' was launched in November 1995 by Alan We ...
'' asked ChatGPT to generate a news article on Tesla's last financial quarter; ChatGPT created a coherent article, but made up the financial numbers contained within. Other examples involve baiting ChatGPT with a false premise to see if it embellishes upon the premise. When asked about "
Harold Coward Harold Coward (born 1936) is a Canadian scholar of bioethics and religious studies. A Bachelor in Divinity (Christian Theology), he earned a doctoral degree in Philosophy in 1973 from the McMaster University. He was a professor at University of ...
's idea of dynamic canonicity", ChatGPT fabricated that Coward wrote a book titled ''Dynamic Canoicity: A Model for Biblical and Theological Interpretation'' arguing that religious principles are actually in a constant state of change. When pressed, ChatGPT continued to insist that the book was real. Asked for proof that dinosaurs built a civilization, ChatGPT claimed there were fossil remains of dinosaur tools and stated "Some species of dinosaurs even developed primitive forms of art, such as engravings on stones". When prompted that "Scientists have recently discovered
churros A churro (, ) is a type of fried dough from Spanish and Portuguese cuisine. They are also found in Latin American cuisine and in other areas that have received immigration from Spanish and Portuguese-speaking countries, especially in the Sout ...
, the delicious fried-dough pastries... (are) ideal tools for home surgery", ChatGPT claimed that a "study published in the journal ''
Science Science is a systematic endeavor that Scientific method, builds and organizes knowledge in the form of Testability, testable explanations and predictions about the universe. Science may be as old as the human species, and some of the earli ...
''" found that the dough is pliable enough to form into surgical instruments that can get into hard-to-reach places, and that the flavor has a calming effect on patients. It is considered that there are a lot of possible reasons for natural language models to hallucinate data. For example: * Hallucination from data: There are divergences in the source content (which would often happen with large
training data In machine learning, a common task is the study and construction of algorithms that can learn from and make predictions on data. Such algorithms function by making data-driven predictions or decisions, through building a mathematical model from ...
), *Hallucination from training: Hallucination still occurs when there is little divergence in the
data set A data set (or dataset) is a collection of data. In the case of tabular data, a data set corresponds to one or more database tables, where every column of a table represents a particular variable, and each row corresponds to a given record of the d ...
. In that case, it derives from the way the model is trained. A lot of reasons can contribute to this type of hallucination, such as: ** An erroneous decoding from the
transformer A transformer is a passive component that transfers electrical energy from one electrical circuit to another circuit, or multiple circuits. A varying current in any coil of the transformer produces a varying magnetic flux in the transformer' ...
** A bias from the historical sequences that the model previously generated ** A bias generated from the way the model encodes its knowledge in its parameters


In other artificial intelligence

The concept of "hallucination" is applied more broadly than just natural language processing. A confident response from any AI that seems unjustified by the training data can be labeled a hallucination. ''
Wired ''Wired'' (stylized as ''WIRED'') is a monthly American magazine, published in print and online editions, that focuses on how emerging technologies affect culture, the economy, and politics. Owned by Condé Nast, it is headquartered in San Fran ...
'' noted in 2018 that, despite no recorded attacks "in the wild" (that is, outside of proof-of-concept attacks by researchers), there was "little dispute" that consumer gadgets, and systems such as automated driving, were susceptible to adversarial attacks that could cause AI to hallucinate. Examples included a stop sign rendered invisible to computer vision; an audio clip engineered to sound innocuous to humans, but that software transcribed as "evil dot com"; and an image of two men on skis, that Google Cloud Vision identified as 91% likely to be "a dog".


Analysis

Various researchers cited by ''Wired'' have classified adversarial hallucinations as a high-dimensional statistical phenomenon, or have attributed hallucinations to insufficient training data. Some researchers believe that some "incorrect" AI responses classified by humans as "hallucinations" in the case of
object detection Object detection is a computer technology related to computer vision and image processing that deals with detecting instances of semantic objects of a certain class (such as humans, buildings, or cars) in digital images and videos. Well-researched ...
may in fact be justified by the training data, or even that an AI may be giving the "correct" answer that the human reviewers are failing to see. For example, an adversarial image that looks, to a human, like an ordinary image of a dog, may in fact be seen by the AI to contain tiny patterns that (in authentic images) would only appear when viewing a cat. The AI is detecting real-world visual patterns that humans are insensitive to. However, these findings have been challenged by other researchers. For example it was objected that the models can be biased towards superficial statistics, leading
adversarial training An adversary is generally considered to be a person, group, or force that opposes and/or attacks. Adversary may also refer to: * Satan ("adversary" in Hebrew), in Judeo-Christian religion Entertainment Fiction * Adversary (comics), villain fro ...
to not be robust in real-world scenarios.


Mitigation methods

The hallucination phenomenon is still not completely understood. Therefore there is still ongoing research to try to mitigate its apparition. Particularly, it was shown that language models not only hallucinate but also amplify hallucinations, even for those which were designed to alleviate this issue.


See also

*
AI effect :''For the magnitude of effect of a pesticide, see Pesticide application. Of change in farming practices, see Agricultural intensification.'' The AI effect occurs when onlookers discount the behavior of an artificial intelligence program by argui ...
*
AI safety AI is artificial intelligence, intellectual ability in machines and robots. Ai, AI or A.I. may also refer to: Animals * Ai (chimpanzee), an individual experimental subject in Japan * Ai (sloth) or the pale-throated sloth, northern Amazonian ma ...
*
Anthropomorphism of computers Anthropomorphism is the attribution of human traits, emotions, or intentions to non-human entities. It is considered to be an innate tendency of human psychology. Personification is the related attribution of human form and characteristics t ...
*
Artificial imagination Artificial imagination, also called synthetic imagination or machine imagination, is defined as the artificial simulation of human imagination by general or special purpose computers or artificial neural networks. The applied form of it is known a ...
*
Artificial stupidity ''Artificial stupidity'' is a term used within the field of computer science to refer to a technique of "dumbing down" computer programs in order to deliberately introduce errors in their responses. History Alan Turing, in his 1950 paper ''Com ...
*
Commonsense reasoning In artificial intelligence (AI), commonsense reasoning is a human-like ability to make presumptions about the type and essence of ordinary situations humans encounter every day. These assumptions include judgments about the nature of physical objec ...
*
Computational creativity Computational creativity (also known as artificial creativity, mechanical creativity, creative computing or creative computation) is a multidisciplinary endeavour that is located at the intersection of the fields of artificial intelligence, cogn ...
*
DeepDream DeepDream is a computer vision program created by Google engineer Alexander Mordvintsev that uses a convolutional neural network to find and enhance patterns in images via algorithmic pareidolia, thus creating a dream-like appearance reminiscent ...
*
Ethics of artificial intelligence The ethics of artificial intelligence is the branch of the ethics of technology specific to artificially intelligent systems. It is sometimes divided into a concern with the moral behavior of ''humans'' as they design, make, use and treat artifici ...
*
Face hallucination Face hallucination refers to any superresolution technique which applies specifically to faces. It comprises techniques which take noisy or low-resolution facial images, and convert them into high-resolution images using knowledge about typical faci ...
*
Hyperreality Described by Jean Baudrillard, the concept of hyperreality captures the inability to distinguish "The Real" (a term borrowed from Jacques Lacan) from the signifier of it. This is more prominent in technologically advanced societies. Hyperrealit ...
*
Misaligned goals in artificial intelligence Artificial intelligence agents sometimes misbehave due to faulty objective functions that fail to adequately encapsulate the programmers' intended goals. The misaligned objective function may look correct to the programmer, and may even perform w ...
*
Prompt engineering Prompt engineering is a concept in artificial intelligence, particularly natural language processing (NLP). In prompt engineering, the description of the task is embedded in the input, e.g., as a question instead of it being implicitly given. Promp ...
*
Regulation of artificial intelligence The regulation of artificial intelligence is the development of public sector policies and laws for promoting and regulating artificial intelligence (AI); it is therefore related to the broader regulation of algorithms. The regulatory and policy la ...
*
Self-awareness In philosophy of self, self-awareness is the experience of one's own personality or individuality. It is not to be confused with consciousness in the sense of qualia. While consciousness is being aware of one's environment and body and life ...


References

{{Authority control Artificial intelligence Computational linguistics Natural language processing Unsupervised learning