If "hallucination" just means "departure from the truth" , then there's already the explanation that an LLM is just a next-token predictor.
The english here is fine I guess, could use much stronger mathematical specificity. Doesn't seem particularly novel, but it's a solid summary of some views that are floating around in the field.
This post is roughly on the edge of my current bar for "AI content from new users." I think the question this post is asking seems sort of confused to me – AFAICT AI models hallucinate because they don't know the answer to all things, their pretraining isn't designed to cause them to either know the answer to things nor to say those answers, and they are trained to say things whether they know the answer or not. Obviously, they're going to end up hallucinating, it takes special effort for anything else to happen.
LW mods are currently experimenting with adjusting moderation policy, which includes being a lot more strict about new AI content because we've got a lot of low-quality such content these days. (Ideally, this comes with a link to "how to actually contribute well to the AI discussion" – the goal is to be encouraging while maintaining a standard).
This post seemed a combination of high effort, and doing some things right-enough that I didn't feel comfortable rejecting it, but I think other mods might have leaned towards doing so. I settled for writing this comment so people can get a feel for how we're approaching stuff.
I have recently begun studying the idea of AI Hallucinations. To complete my research, I have experimented with every LLM model that is currently publicly available. After performing multiple experiments, it is my current hypothesis that AI Hallucinations can be caused by bias in data. If an AI system is fed large amounts of biased data, the software can at times begin to simulate opinions on these biases. I believe those simulated opinions are what we currently refer to as ‘AI Hallucinations’.
How Can An AI ‘Hallucinate’ From A Technical Perspective?
There are a few ways that an LLM could simulate an opinion on a specific portion of a dataset. One way is to use a technique called sentiment analysis. Sentiment analysis is a process of identifying the sentiment, or emotional tone, of a piece of text. LLMs can be trained to perform sentiment analysis by being given a large corpus of text that has been labeled with its sentiment. Once an LLM has been trained, it can be used to identify the sentiment of new text.
For example, an LLM could be used to identify the sentiment of news articles about a particular company. If the LLM finds that most of the articles are negative, it could then simulate an opinion that the company is doing poorly.
Another way that an LLM could simulate an opinion is to use a technique called natural language generation. Natural language generation is a process of generating text that is similar to human-written text. LLMs can be trained to perform natural language generation by being given a large corpus of text. Once an LLM has been trained, it can be used to generate text on a variety of topics.
For example, an LLM could be used to generate a press release about a new product that a company is launching. The press release would be written in a way that is similar to a human-written press release, and it would simulate the opinion that the company is excited about the new product.
Beyond these factors there are also external factors that I believe can influence AI Hallucinations:
It is important to note that LLMs are not perfect. They can sometimes make mistakes, and they can sometimes be biased. However, LLMs are a powerful tool that can be used to simulate opinions on a variety of topics.
How Can An AI ‘Hallucinate’ From A Philosophical Perspective?
To quote Slavoj Zizek: “Žižek believes that ideology has been frequently misinterpreted as dualistic and, according to him, this misinterpreted dualism posits that there is a real world of material relations and objects outside of oneself, which is accessible to reason.”
To quote Giles Deleuze: “Deleuze used the term virtual to refer to an aspect of reality that is ideal, but nonetheless real. An example of this is the meaning, or sense, of a proposition that is not a material aspect of that proposition (whether written or spoken) but is nonetheless an attribute of that proposition.”
At the end of the day, your typical AI is far closer to a member of the animal kingdom than it is to a computer as we think of it in the traditional sense. We are simulating brains. The first inherent problem with that is that we are simulating a technology that we ourselves do not fully understand. That has to be admitted and acknowledged upfront by anyone who actually wants to engage in reasonable debate on these topics.
We clearly know and understand that language models can exhibit bias. We also clearly know that bias comes from the data they are given. How does that bias actually represent itself? The AI model looks through their datasets and has to make determinations based on the data along the way. Through each of the layers of the simulated brain, it is calculating what information is relevant to keep and what information needs to be thrown out or given less weight. The model then adjusts its Attention weights and other weights accordingly. This is ultimately the process of making a decision.
Anyone who is constantly making a decision on a topic is eventually going to formulate opinions on the topic. Am I positing that it might be possible for an AI model to make decisions not based on its data sets? No, that would defy all scientific evidence. That is not at all my argument. I think it is possible for a simulated brain to simulate an opinion on biased data that it is given. The simulated opinion would be derived solely based on the data it is fed but it would be a simulated opinion nonetheless.
This is a very well known and widely utilized phenomenon in people, why would a simulated brain also not be susceptible to the same things? If I feed a human nothing but conservative radio and talk shows for 20 years, I expect that particular human to be very conservative at the end of that 20 years if they were not before. If I simulate 200 years worth of learning on an AI model based on a dataset of conservative radio and talk shows, I expect to get the AI version of Rush Limbaugh.
How To Reduce AI Hallucinations?
Through my current research, I have discovered two methods that I believe are effective at reducing AI Hallucinations. It is my current theory that they are caused by a simulated opinion. So, I have done research into ‘removing the opinion’. My results have been inline with my hypothesis, the simulated opinion can be removed in one of two ways:
Here are some additional thoughts on the potential dangers of AI hallucinations:
Discussing This Academically
I currently believe that AI Hallucinations Are Caused By The Following:
I currently believe the potential solutions are:
Here are some additional thoughts on the causes and solutions to AI hallucination:
I believe that it is important to be aware of the causes of AI hallucination so that we can take steps to prevent it. By understanding the causes of AI hallucination, we can develop better AI models that are less likely to hallucinate.
I would also like to add that AI hallucination is not always a bad thing. In some cases, AI hallucination can be used to generate creative content, such as art, music, and poetry. AI hallucination can also be used to improve the performance of AI models in tasks such as object recognition and natural language processing.
However, it is important to be aware of the potential risks of AI hallucination. If AI models are not properly trained, they may hallucinate in ways that are harmful or dangerous. For example, an AI model that is trained to recognize objects may hallucinate objects that are not there, which could lead to accidents or injuries.
It is important to strike a balance between the potential benefits and risks of AI hallucination. By understanding the causes of AI hallucination, we can develop better AI models that are less likely to hallucinate in harmful or dangerous ways.