• Kinza Babylon Staked BTCKinza Babylon Staked BTC(KBTC)$83,270.000.00%
  • Steakhouse EURCV Morpho VaultSteakhouse EURCV Morpho Vault(STEAKEURCV)$0.000000-100.00%
  • Stride Staked InjectiveStride Staked Injective(STINJ)$16.51-4.18%
  • Vested XORVested XOR(VXOR)$3,404.231,000.00%
  • FibSwap DEXFibSwap DEX(FIBO)$0.0084659.90%
  • ICPanda DAOICPanda DAO(PANDA)$0.003106-39.39%
  • TruFin Staked APTTruFin Staked APT(TRUAPT)$8.020.00%
  • bitcoinBitcoin(BTC)$105,815.001.34%
  • ethereumEthereum(ETH)$2,521.651.04%
  • VNST StablecoinVNST Stablecoin(VNST)$0.0000400.67%
  • tetherTether(USDT)$1.00-0.03%
  • rippleXRP(XRP)$2.180.46%
  • binancecoinBNB(BNB)$652.361.19%
  • Wrapped SOLWrapped SOL(SOL)$143.66-2.32%
  • solanaSolana(SOL)$151.041.22%
  • usd-coinUSDC(USDC)$1.000.00%
  • dogecoinDogecoin(DOGE)$0.1834222.05%
  • tronTRON(TRX)$0.2850232.88%
  • cardanoCardano(ADA)$0.660.25%
  • staked-etherLido Staked Ether(STETH)$2,520.401.03%
  • wrapped-bitcoinWrapped Bitcoin(WBTC)$105,808.001.39%
  • Gaj FinanceGaj Finance(GAJ)$0.0059271.46%
  • Content BitcoinContent Bitcoin(CTB)$24.482.55%
  • USD OneUSD One(USD1)$1.000.11%
  • HyperliquidHyperliquid(HYPE)$34.531.77%
  • SuiSui(SUI)$3.242.21%
  • Wrapped stETHWrapped stETH(WSTETH)$3,038.061.11%
  • UGOLD Inc.UGOLD Inc.(UGOLD)$3,042.460.08%
  • ParkcoinParkcoin(KPK)$1.101.76%
  • chainlinkChainlink(LINK)$13.832.44%
  • avalanche-2Avalanche(AVAX)$20.827.01%
  • leo-tokenLEO Token(LEO)$9.132.22%
  • stellarStellar(XLM)$0.2657380.94%
  • bitcoin-cashBitcoin Cash(BCH)$407.742.87%
  • ToncoinToncoin(TON)$3.181.00%
  • shiba-inuShiba Inu(SHIB)$0.0000132.06%
  • hedera-hashgraphHedera(HBAR)$0.1699173.34%
  • USDSUSDS(USDS)$1.00-0.01%
  • Yay StakeStone EtherYay StakeStone Ether(YAYSTONE)$2,671.07-2.84%
  • litecoinLitecoin(LTC)$88.261.74%
  • wethWETH(WETH)$2,523.511.36%
  • Wrapped eETHWrapped eETH(WEETH)$2,690.670.92%
  • polkadotPolkadot(DOT)$4.033.44%
  • Pundi AIFXPundi AIFX(PUNDIAI)$16.000.00%
  • Binance Bridged USDT (BNB Smart Chain)Binance Bridged USDT (BNB Smart Chain)(BSC-USD)$1.00-0.08%
  • PengPeng(PENG)$0.60-13.59%
  • moneroMonero(XMR)$324.22-0.23%
  • Ethena USDeEthena USDe(USDE)$1.000.03%
  • Bitget TokenBitget Token(BGB)$4.661.48%
  • MurasakiMurasaki(MURA)$4.32-12.46%
TradePoint.io
  • Main
  • AI & Technology
  • Stock Charts
  • Market & News
  • Business
  • Finance Tips
  • Trade Tube
  • Blog
  • Shop
No Result
View All Result
TradePoint.io
No Result
View All Result

What Are LLM Hallucinations? Causes, Ethical Concern, & Prevention

April 29, 2023
in AI & Technology
Reading Time: 7 mins read
A A
What Are LLM Hallucinations? Causes, Ethical Concern, & Prevention
ShareShareShareShareShare

Large language models (LLMs) are artificial intelligence systems capable of analyzing and generating human-like text. But they have a problem – LLMs hallucinate, i.e., make stuff up. LLM hallucinations have made researchers worried about the progress in this field because if researchers cannot control the outcome of the models, then they cannot build critical systems to serve humanity. More on this later.

Generally, LLMs use vast amounts of training data and complex learning algorithms to generate realistic outputs. In some cases, in-context learning is used to train these models using only a few examples. LLMs are becoming increasingly popular across various application areas ranging from machine translation, sentiment analysis, virtual AI assistance, image annotation, natural language processing, etc.

Despite the cutting-edge nature of LLMs, they are still prone to biases, errors, and hallucinations. Yann LeCun, current Chief AI Scientist at Meta, recently mentioned the central flaw in LLMs that causes hallucinations: “Large language models have no idea of the underlying reality that language describes. Those systems generate text that sounds fine, grammatically, and semantically, but they don’t really have some sort of objective other than just satisfying statistical consistency with the prompt”.

Hallucinations in LLMs

Image by Gerd Altmann from Pixabay

Hallucinations refer to the model generating outputs that are syntactically and semantically correct but are disconnected from reality, and based on false assumptions. Hallucination is one of the major ethical concerns of LLMs, and it can have harmful consequences as users without adequate domain knowledge start to over-rely on these increasingly convincing language models.

A certain degree of hallucination is inevitable across all autoregressive LLMs. For example, a model can attribute a counterfeit quote to a celebrity that was never said. They may assert something about a particular topic that is factually incorrect or cite non-existent sources in research papers, thus spreading misinformation.

YOU MAY ALSO LIKE

Make it Home takes interior design on the road

Playdate Season 2 review: The Whiteout and Wheelsprung

However, getting AI models to hallucinate does not always have adverse effects. For example, a new study suggests scientists are unearthing ‘novel proteins with an unlimited array of properties’ through hallucinating LLMs.

What Causes LLMs Hallucinations?

LLMs can hallucinate due to various factors, ranging from overfitting errors in encoding and decoding to training bias.

Overfitting

Image by janjf93 from Pixabay

Overfitting is an issue where an AI model fits the training data too well. Still, it cannot fully represent the whole range of inputs it may encounter, i.e., it fails to generalize its predictive power to new, unseen data. Overfitting can lead to the model producing hallucinated content.

Encoding and Decoding Errors

Image by geralt from Pixabay

If there are errors in the encoding and decoding of text and its subsequent representations, this can also cause the model to generate nonsensical and erroneous outputs.

Training Bias

Image by Quince Creative from Pixabay

Another factor is the presence of certain biases in the training data, which can cause the model to give results that represent those biases rather than the actual nature of the data. This is similar to the lack of diversity in the training data, which limits the model’s ability to generalize to new data.

The complex structure of LLMs makes it quite challenging for AI researchers and practitioners to identify, interpret, and correct these underlying causes of hallucinations.

Ethical Concerns of LLM Hallucinations

LLMs can perpetuate and amplify harmful biases through hallucinations and can, in turn, negatively impact the users and have detrimental social consequences. Some of these most important ethical concerns are listed below:

Discriminating and Toxic Content

Image by ar130405 from Pixabay

Since the LLM training data is often full of sociocultural stereotypes due to the inherent biases and lack of diversity. LLMs can, thus, produce and reinforce these harmful ideas against disadvantaged groups in society.

They can generate this discriminating and hateful content based on race, gender, religion, ethnicity, etc.

Privacy Issues

Image by JanBaby from Pixabay

LLMs are trained on a massive training corpus which often includes the personal information of individuals. There have been cases where such models have violated people’s privacy. They can leak specific information such as social security numbers, home addresses, cell phone numbers, and medical details.

Misinformation and Disinformation

Image by geralt from Pixabay

Language models can produce human-like content that seems accurate but is, in fact, false and not supported by empirical evidence. This can be accidental, leading to misinformation, or it can have malicious intent behind it to knowingly spread disinformation. If this goes unchecked, it can create adverse social-cultural-economic-political trends.

Preventing LLM Hallucinations

Image by athree23 from Pixabay

Researchers and practitioners are taking various approaches to address the problem of hallucinations in LLMs. These include improving the diversity of training data, eliminating inherent biases, using better regularization techniques, and employing adversarial training and reinforcement learning, among others:

  • Developing better regularization techniques is at the core of tackling hallucinations. They help prevent overfitting and other problems that cause hallucinations.
  • Data augmentation can reduce the frequency of hallucinations, as evidenced by a research study. Data augmentation involves augmenting the training set by adding a random token anywhere in the sentence. It doubles the size of the training set and causes a decrease in the frequency of hallucinations.
  • OpenAI and Google’s DeepMind developed a technique called reinforcement learning with human feedback (RLHF) to tackle ChatGPT’s hallucination problem. It involves a human evaluator who frequently reviews the model’s responses and picks out the most appropriate for the user prompts. This feedback is then used to adjust the behavior of the model. Ilya Sutskever, OpenAI’s chief scientist, recently mentioned that this approach can potentially resolve hallucinations in ChatGPT: “I’m quite hopeful that by simply improving this subsequent reinforcement learning from the human feedback step, we can teach it to not hallucinate”.
  • Identifying hallucinated content to use as an example for future training is also a method used to tackle hallucinations. A novel technique in this regard detects hallucinations at the token level and predicts whether each token in the output is hallucinated. It also includes a method for unsupervised learning of hallucination detectors.

Token-level Hallucination Detection

Put simply, LLM hallucinations are a growing concern. And despite the efforts, much work still needs to be done to address the problem. The complexity of these models means it’s generally challenging to identify and rectify the inherent causes of hallucinations correctly.

However, with continued research and development, mitigating hallucinations in LLMs and reducing their ethical consequences is possible.

If you want to learn more about LLMs and the preventive techniques being developed to rectify LLMs hallucinations, check out unite.ai to expand your knowledge.

Credit: Source link

ShareTweetSendSharePin

Related Posts

Make it Home takes interior design on the road
AI & Technology

Make it Home takes interior design on the road

June 7, 2025
Playdate Season 2 review: The Whiteout and Wheelsprung
AI & Technology

Playdate Season 2 review: The Whiteout and Wheelsprung

June 7, 2025
Tire Boy is a wacky open-world adventure game you can tread all over
AI & Technology

Tire Boy is a wacky open-world adventure game you can tread all over

June 7, 2025
RGG’s Project Century is an action game called Stranger Than Heaven
AI & Technology

RGG’s Project Century is an action game called Stranger Than Heaven

June 6, 2025
Next Post
How vector databases can revolutionize our relationship with generative AI

How vector databases can revolutionize our relationship with generative AI

Leave a Reply Cancel reply

Your email address will not be published. Required fields are marked *

Search

No Result
View All Result
Greta Thunberg says the story is not her sailing to Gaza but ‘Israel’s siege on Gaza

Greta Thunberg says the story is not her sailing to Gaza but ‘Israel’s siege on Gaza

June 6, 2025
My Exes Became Friends, and Now I’m Not In My Kids’ Lives

My Exes Became Friends, and Now I’m Not In My Kids’ Lives

June 2, 2025
China eases stranglehold on rare minerals in welcome news for GM, Ford: report

China eases stranglehold on rare minerals in welcome news for GM, Ford: report

June 6, 2025

About

Learn more

Our Services

Legal

Privacy Policy

Terms of Use

Bloggers

Learn more

Article Links

Contact

Advertise

Ask us anything

©2020- TradePoint.io - All rights reserved!

Tradepoint.io, being just a publishing and technology platform, is not a registered broker-dealer or investment adviser. So we do not provide investment advice. Rather, brokerage services are provided to clients of Tradepoint.io by independent SEC-registered broker-dealers and members of FINRA/SIPC. Every form of investing carries some risk and past performance is not a guarantee of future results. “Tradepoint.io“, “Instant Investing” and “My Trading Tools” are registered trademarks of Apperbuild, LLC.

This website is operated by Apperbuild, LLC. We have no link to any brokerage firm and we do not provide investment advice. Every information and resource we provide is solely for the education of our readers. © 2020 Apperbuild, LLC. All rights reserved.

No Result
View All Result
  • Main
  • AI & Technology
  • Stock Charts
  • Market & News
  • Business
  • Finance Tips
  • Trade Tube
  • Blog
  • Shop

© 2023 - TradePoint.io - All Rights Reserved!