• bitcoinBitcoin(BTC)$77,525.000.03%
  • ethereumEthereum(ETH)$2,315.04-0.12%
  • tetherTether(USDT)$1.00-0.01%
  • rippleXRP(XRP)$1.42-0.92%
  • binancecoinBNB(BNB)$628.59-1.20%
  • usd-coinUSDC(USDC)$1.000.00%
  • solanaSolana(SOL)$86.05-0.24%
  • tronTRON(TRX)$0.3242210.36%
  • Figure HelocFigure Heloc(FIGR_HELOC)$1.02-0.55%
  • dogecoinDogecoin(DOGE)$0.097884-0.59%
  • whitebitWhiteBIT Coin(WBT)$54.850.03%
  • USDSUSDS(USDS)$1.000.00%
  • HyperliquidHyperliquid(HYPE)$41.270.31%
  • leo-tokenLEO Token(LEO)$10.280.30%
  • cardanoCardano(ADA)$0.248952-0.56%
  • bitcoin-cashBitcoin Cash(BCH)$452.40-0.75%
  • moneroMonero(XMR)$373.761.23%
  • chainlinkChainlink(LINK)$9.32-0.81%
  • zcashZcash(ZEC)$356.99-0.20%
  • CantonCanton(CC)$0.151109-1.99%
  • MemeCoreMemeCore(M)$4.35-2.12%
  • stellarStellar(XLM)$0.168902-2.32%
  • daiDai(DAI)$1.00-0.02%
  • USD1USD1(USD1)$1.000.00%
  • litecoinLitecoin(LTC)$55.96-0.92%
  • avalanche-2Avalanche(AVAX)$9.33-0.88%
  • hedera-hashgraphHedera(HBAR)$0.090768-0.21%
  • Ethena USDeEthena USDe(USDE)$1.00-0.01%
  • suiSui(SUI)$0.94-1.34%
  • shiba-inuShiba Inu(SHIB)$0.000006-0.49%
  • paypal-usdPayPal USD(PYUSD)$1.000.00%
  • RainRain(RAIN)$0.007048-6.66%
  • the-open-networkToncoin(TON)$1.30-2.12%
  • crypto-com-chainCronos(CRO)$0.0697080.44%
  • Circle USYCCircle USYC(USYC)$1.120.00%
  • tether-goldTether Gold(XAUT)$4,689.33-0.09%
  • Global DollarGlobal Dollar(USDG)$1.000.00%
  • World Liberty FinancialWorld Liberty Financial(WLFI)$0.074943-1.48%
  • BittensorBittensor(TAO)$246.04-0.50%
  • BlackRock USD Institutional Digital Liquidity FundBlackRock USD Institutional Digital Liquidity Fund(BUIDL)$1.000.00%
  • pax-goldPAX Gold(PAXG)$4,691.25-0.17%
  • mantleMantle(MNT)$0.661.26%
  • polkadotPolkadot(DOT)$1.24-1.35%
  • uniswapUniswap(UNI)$3.24-0.66%
  • SkySky(SKY)$0.0851680.86%
  • nearNEAR Protocol(NEAR)$1.39-0.85%
  • Falcon USDFalcon USD(USDF)$1.00-0.13%
  • Pi NetworkPi Network(PI)$0.1729331.75%
  • okbOKB(OKB)$84.140.45%
  • HTX DAOHTX DAO(HTX)$0.000002-0.57%
TradePoint.io
  • Main
  • AI & Technology
  • Stock Charts
  • Market & News
  • Business
  • Finance Tips
  • Trade Tube
  • Blog
  • Shop
No Result
View All Result
TradePoint.io
No Result
View All Result

What Are LLM Hallucinations? Causes, Ethical Concern, & Prevention

April 29, 2023
in AI & Technology
Reading Time: 7 mins read
A A
What Are LLM Hallucinations? Causes, Ethical Concern, & Prevention
ShareShareShareShareShare

Large language models (LLMs) are artificial intelligence systems capable of analyzing and generating human-like text. But they have a problem – LLMs hallucinate, i.e., make stuff up. LLM hallucinations have made researchers worried about the progress in this field because if researchers cannot control the outcome of the models, then they cannot build critical systems to serve humanity. More on this later.

Generally, LLMs use vast amounts of training data and complex learning algorithms to generate realistic outputs. In some cases, in-context learning is used to train these models using only a few examples. LLMs are becoming increasingly popular across various application areas ranging from machine translation, sentiment analysis, virtual AI assistance, image annotation, natural language processing, etc.

YOU MAY ALSO LIKE

BYD’s next all-electric hypercar is a convertible that’s coming to Europe first

xAI Launches grok-voice-think-fast-1.0: Topping τ-voice Bench at 67.3%, Outperforming Gemini, GPT Realtime, and More

Despite the cutting-edge nature of LLMs, they are still prone to biases, errors, and hallucinations. Yann LeCun, current Chief AI Scientist at Meta, recently mentioned the central flaw in LLMs that causes hallucinations: “Large language models have no idea of the underlying reality that language describes. Those systems generate text that sounds fine, grammatically, and semantically, but they don’t really have some sort of objective other than just satisfying statistical consistency with the prompt”.

Hallucinations in LLMs

Image by Gerd Altmann from Pixabay

Hallucinations refer to the model generating outputs that are syntactically and semantically correct but are disconnected from reality, and based on false assumptions. Hallucination is one of the major ethical concerns of LLMs, and it can have harmful consequences as users without adequate domain knowledge start to over-rely on these increasingly convincing language models.

A certain degree of hallucination is inevitable across all autoregressive LLMs. For example, a model can attribute a counterfeit quote to a celebrity that was never said. They may assert something about a particular topic that is factually incorrect or cite non-existent sources in research papers, thus spreading misinformation.

However, getting AI models to hallucinate does not always have adverse effects. For example, a new study suggests scientists are unearthing ‘novel proteins with an unlimited array of properties’ through hallucinating LLMs.

What Causes LLMs Hallucinations?

LLMs can hallucinate due to various factors, ranging from overfitting errors in encoding and decoding to training bias.

Overfitting

Image by janjf93 from Pixabay

Overfitting is an issue where an AI model fits the training data too well. Still, it cannot fully represent the whole range of inputs it may encounter, i.e., it fails to generalize its predictive power to new, unseen data. Overfitting can lead to the model producing hallucinated content.

Encoding and Decoding Errors

Image by geralt from Pixabay

If there are errors in the encoding and decoding of text and its subsequent representations, this can also cause the model to generate nonsensical and erroneous outputs.

Training Bias

Image by Quince Creative from Pixabay

Another factor is the presence of certain biases in the training data, which can cause the model to give results that represent those biases rather than the actual nature of the data. This is similar to the lack of diversity in the training data, which limits the model’s ability to generalize to new data.

The complex structure of LLMs makes it quite challenging for AI researchers and practitioners to identify, interpret, and correct these underlying causes of hallucinations.

Ethical Concerns of LLM Hallucinations

LLMs can perpetuate and amplify harmful biases through hallucinations and can, in turn, negatively impact the users and have detrimental social consequences. Some of these most important ethical concerns are listed below:

Discriminating and Toxic Content

Image by ar130405 from Pixabay

Since the LLM training data is often full of sociocultural stereotypes due to the inherent biases and lack of diversity. LLMs can, thus, produce and reinforce these harmful ideas against disadvantaged groups in society.

They can generate this discriminating and hateful content based on race, gender, religion, ethnicity, etc.

Privacy Issues

Image by JanBaby from Pixabay

LLMs are trained on a massive training corpus which often includes the personal information of individuals. There have been cases where such models have violated people’s privacy. They can leak specific information such as social security numbers, home addresses, cell phone numbers, and medical details.

Misinformation and Disinformation

Image by geralt from Pixabay

Language models can produce human-like content that seems accurate but is, in fact, false and not supported by empirical evidence. This can be accidental, leading to misinformation, or it can have malicious intent behind it to knowingly spread disinformation. If this goes unchecked, it can create adverse social-cultural-economic-political trends.

Preventing LLM Hallucinations

Image by athree23 from Pixabay

Researchers and practitioners are taking various approaches to address the problem of hallucinations in LLMs. These include improving the diversity of training data, eliminating inherent biases, using better regularization techniques, and employing adversarial training and reinforcement learning, among others:

  • Developing better regularization techniques is at the core of tackling hallucinations. They help prevent overfitting and other problems that cause hallucinations.
  • Data augmentation can reduce the frequency of hallucinations, as evidenced by a research study. Data augmentation involves augmenting the training set by adding a random token anywhere in the sentence. It doubles the size of the training set and causes a decrease in the frequency of hallucinations.
  • OpenAI and Google’s DeepMind developed a technique called reinforcement learning with human feedback (RLHF) to tackle ChatGPT’s hallucination problem. It involves a human evaluator who frequently reviews the model’s responses and picks out the most appropriate for the user prompts. This feedback is then used to adjust the behavior of the model. Ilya Sutskever, OpenAI’s chief scientist, recently mentioned that this approach can potentially resolve hallucinations in ChatGPT: “I’m quite hopeful that by simply improving this subsequent reinforcement learning from the human feedback step, we can teach it to not hallucinate”.
  • Identifying hallucinated content to use as an example for future training is also a method used to tackle hallucinations. A novel technique in this regard detects hallucinations at the token level and predicts whether each token in the output is hallucinated. It also includes a method for unsupervised learning of hallucination detectors.

Token-level Hallucination Detection

Put simply, LLM hallucinations are a growing concern. And despite the efforts, much work still needs to be done to address the problem. The complexity of these models means it’s generally challenging to identify and rectify the inherent causes of hallucinations correctly.

However, with continued research and development, mitigating hallucinations in LLMs and reducing their ethical consequences is possible.

If you want to learn more about LLMs and the preventive techniques being developed to rectify LLMs hallucinations, check out unite.ai to expand your knowledge.

Credit: Source link

ShareTweetSendSharePin

Related Posts

BYD’s next all-electric hypercar is a convertible that’s coming to Europe first
AI & Technology

BYD’s next all-electric hypercar is a convertible that’s coming to Europe first

April 25, 2026
xAI Launches grok-voice-think-fast-1.0: Topping τ-voice Bench at 67.3%, Outperforming Gemini, GPT Realtime, and More
AI & Technology

xAI Launches grok-voice-think-fast-1.0: Topping τ-voice Bench at 67.3%, Outperforming Gemini, GPT Realtime, and More

April 25, 2026
OpenAI’s Sam Altman apologizes for not reporting ChatGPT account of Tumbler Ridge suspect to police
AI & Technology

OpenAI’s Sam Altman apologizes for not reporting ChatGPT account of Tumbler Ridge suspect to police

April 25, 2026
NASA’s initial takeaways from the Artemis II mission, and more science stories
AI & Technology

NASA’s initial takeaways from the Artemis II mission, and more science stories

April 25, 2026
Next Post
How vector databases can revolutionize our relationship with generative AI

How vector databases can revolutionize our relationship with generative AI

Leave a Reply Cancel reply

Your email address will not be published. Required fields are marked *

Search

No Result
View All Result
Mamdani doubles down on Ken Griffin attack despite Citadel’s threat to pull  billion NYC project

Mamdani doubles down on Ken Griffin attack despite Citadel’s threat to pull $6 billion NYC project

April 24, 2026
Adversaries hijacked AI security tools at 90+ organizations. The next wave has write access to the firewall

Adversaries hijacked AI security tools at 90+ organizations. The next wave has write access to the firewall

April 21, 2026
Angry Tesla owners sue Elon Musk’s automaker for ‘misleading’ them about self-driving tech

Angry Tesla owners sue Elon Musk’s automaker for ‘misleading’ them about self-driving tech

April 20, 2026

About

Learn more

Our Services

Legal

Privacy Policy

Terms of Use

Bloggers

Learn more

Article Links

Contact

Advertise

Ask us anything

©2020- TradePoint.io - All rights reserved!

Tradepoint.io, being just a publishing and technology platform, is not a registered broker-dealer or investment adviser. So we do not provide investment advice. Rather, brokerage services are provided to clients of Tradepoint.io by independent SEC-registered broker-dealers and members of FINRA/SIPC. Every form of investing carries some risk and past performance is not a guarantee of future results. “Tradepoint.io“, “Instant Investing” and “My Trading Tools” are registered trademarks of Apperbuild, LLC.

This website is operated by Apperbuild, LLC. We have no link to any brokerage firm and we do not provide investment advice. Every information and resource we provide is solely for the education of our readers. © 2020 Apperbuild, LLC. All rights reserved.

No Result
View All Result
  • Main
  • AI & Technology
  • Stock Charts
  • Market & News
  • Business
  • Finance Tips
  • Trade Tube
  • Blog
  • Shop

© 2023 - TradePoint.io - All Rights Reserved!