Large language models (LLMs) are artificial intelligence systems capable of analyzing and generating human-like text. But they have a problem – LLMs hallucinate, i.e., make stuff up. LLM hallucinations have made researchers worried about the progress in this field because if researchers cannot control the outcome of the models, then they cannot build critical systems to serve humanity. More on this later.
Generally, LLMs use vast amounts of training data and complex learning algorithms to generate realistic outputs. In some cases, in-context learning is used to train these models using only a few examples. LLMs are becoming increasingly popular across various application areas ranging from machine translation, sentiment analysis, virtual AI assistance, image annotation, natural language processing, etc.
Despite the cutting-edge nature of LLMs, they are still prone to biases, errors, and hallucinations. Yann LeCun, current Chief AI Scientist at Meta, recently mentioned the central flaw in LLMs that causes hallucinations: “Large language models have no idea of the underlying reality that language describes. Those systems generate text that sounds fine, grammatically, and semantically, but they don’t really have some sort of objective other than just satisfying statistical consistency with the prompt”.
Hallucinations in LLMs
Hallucinations refer to the model generating outputs that are syntactically and semantically correct but are disconnected from reality, and based on false assumptions. Hallucination is one of the major ethical concerns of LLMs, and it can have harmful consequences as users without adequate domain knowledge start to over-rely on these increasingly convincing language models.
A certain degree of hallucination is inevitable across all autoregressive LLMs. For example, a model can attribute a counterfeit quote to a celebrity that was never said. They may assert something about a particular topic that is factually incorrect or cite non-existent sources in research papers, thus spreading misinformation.
However, getting AI models to hallucinate does not always have adverse effects. For example, a new study suggests scientists are unearthing ‘novel proteins with an unlimited array of properties’ through hallucinating LLMs.
What Causes LLMs Hallucinations?
LLMs can hallucinate due to various factors, ranging from overfitting errors in encoding and decoding to training bias.
Overfitting
Overfitting is an issue where an AI model fits the training data too well. Still, it cannot fully represent the whole range of inputs it may encounter, i.e., it fails to generalize its predictive power to new, unseen data. Overfitting can lead to the model producing hallucinated content.
Encoding and Decoding Errors
If there are errors in the encoding and decoding of text and its subsequent representations, this can also cause the model to generate nonsensical and erroneous outputs.
Training Bias
Another factor is the presence of certain biases in the training data, which can cause the model to give results that represent those biases rather than the actual nature of the data. This is similar to the lack of diversity in the training data, which limits the model’s ability to generalize to new data.
The complex structure of LLMs makes it quite challenging for AI researchers and practitioners to identify, interpret, and correct these underlying causes of hallucinations.
Ethical Concerns of LLM Hallucinations
LLMs can perpetuate and amplify harmful biases through hallucinations and can, in turn, negatively impact the users and have detrimental social consequences. Some of these most important ethical concerns are listed below:
Discriminating and Toxic Content
Since the LLM training data is often full of sociocultural stereotypes due to the inherent biases and lack of diversity. LLMs can, thus, produce and reinforce these harmful ideas against disadvantaged groups in society.
They can generate this discriminating and hateful content based on race, gender, religion, ethnicity, etc.
Privacy Issues
LLMs are trained on a massive training corpus which often includes the personal information of individuals. There have been cases where such models have violated people’s privacy. They can leak specific information such as social security numbers, home addresses, cell phone numbers, and medical details.
Misinformation and Disinformation
Language models can produce human-like content that seems accurate but is, in fact, false and not supported by empirical evidence. This can be accidental, leading to misinformation, or it can have malicious intent behind it to knowingly spread disinformation. If this goes unchecked, it can create adverse social-cultural-economic-political trends.
Preventing LLM Hallucinations
Researchers and practitioners are taking various approaches to address the problem of hallucinations in LLMs. These include improving the diversity of training data, eliminating inherent biases, using better regularization techniques, and employing adversarial training and reinforcement learning, among others:
- Developing better regularization techniques is at the core of tackling hallucinations. They help prevent overfitting and other problems that cause hallucinations.
- Data augmentation can reduce the frequency of hallucinations, as evidenced by a research study. Data augmentation involves augmenting the training set by adding a random token anywhere in the sentence. It doubles the size of the training set and causes a decrease in the frequency of hallucinations.
- OpenAI and Google’s DeepMind developed a technique called reinforcement learning with human feedback (RLHF) to tackle ChatGPT’s hallucination problem. It involves a human evaluator who frequently reviews the model’s responses and picks out the most appropriate for the user prompts. This feedback is then used to adjust the behavior of the model. Ilya Sutskever, OpenAI’s chief scientist, recently mentioned that this approach can potentially resolve hallucinations in ChatGPT: “I’m quite hopeful that by simply improving this subsequent reinforcement learning from the human feedback step, we can teach it to not hallucinate”.
- Identifying hallucinated content to use as an example for future training is also a method used to tackle hallucinations. A novel technique in this regard detects hallucinations at the token level and predicts whether each token in the output is hallucinated. It also includes a method for unsupervised learning of hallucination detectors.
Put simply, LLM hallucinations are a growing concern. And despite the efforts, much work still needs to be done to address the problem. The complexity of these models means it’s generally challenging to identify and rectify the inherent causes of hallucinations correctly.
However, with continued research and development, mitigating hallucinations in LLMs and reducing their ethical consequences is possible.
If you want to learn more about LLMs and the preventive techniques being developed to rectify LLMs hallucinations, check out unite.ai to expand your knowledge.
Credit: Source link