

Metaverse
AI models make stuff up. How can hallucinations be controlled? – Crypto News
There are kinder ways to put it. In its instructions to users, OpenAI warns that ChatGPT “can make mistakes”. Anthropic, an American AI company, says that its LLM Claude “may display incorrect or harmful information”; Google’s Gemini warns users to “double-check its responses”. The throughline is this: no matter how fluent and confident AI-generated text sounds, it still cannot be trusted.
Hallucinations make it hard to rely on AI systems in the real world. Mistakes in news-generating algorithms can spread misinformation. Image generators can produce art that infringes on copyright, even when told not to. Customer-service chatbots can promise refunds they shouldn’t. (In 2022 Air Canada’s chatbot concocted a bereavement policy, and this February a Canadian court has confirmed that the airline must foot the bill.) And hallucinations in AI systems that are used for diagnosis or prescription can kill.
All the leaves are brown
The trouble is that the same abilities that allow models to hallucinate are also what make them so useful. For one, LLMs are a form of “generative” AI, which, taken literally, means they make things up to solve new problems. They do this by producing probability distributions for chunks of characters, or tokens, laying out how likely it is for each possible token in its vocabulary to come next. The mathematics dictate that each token must have a non-zero chance of being chosen, giving the model flexibility to learn new patterns, as well as the capacity to generate statements that are incorrect. The fundamental problem is that language models are probabilistic, while truth is not.
This tension manifests itself in a number of ways. One is that LLMs are not built to have perfect recall in the way a search engine or encyclopedia might. Instead, because the size of a model is much smaller than the size of its training data, it learns by compressing. The model becomes a blurry picture of its training data, retaining key features but at much lower resolution. Some facts resist blurring—“Paris”, for example, may always be the highest-probability token following the words “The capital of France is”. But many more facts that are less statistically obvious may be smudged away.
Further distortions are possible when a pretrained LLM is “fine-tuned”. This is a later stage of training in which the model’s weights, which encode statistical relationships between the words and phrases in the training data, are updated for a specific task. Hallucinations can increase if the LLM is fine-tuned, for example, on transcripts of conversations, because the model might make things up to try to be interesting, just as a chatty human might. (Simply including fine-tuning examples where the model says “I don’t know” seems to keep hallucination levels down.)
Tinkering with a model’s weights can reduce hallucinations. One method involves creating a deliberately flawed model trained on data that contradict the prompt or contain information it lacks. Researchers can then subtract the weights of the flawed model, which are in part responsible for its output, from those of the original to create a model which hallucinates less.
It is also possible to change a model’s “temperature”. Lower temperatures make a model more conservative, encouraging it to sample the most likely word. Higher temperatures make it more creative, by increasing the randomness of this selection. If the goal is to reduce hallucinations, the temperature should be set to zero. Another trick is to limit the choice to the top-ranked tokens alone. This reduces the likelihood of poor responses, while also allowing for some randomness and, therefore, variety.
Clever prompting can also reduce hallucinations. Researchers at Google DeepMind found that telling an LLM to “take a deep breath and work on this problem step-by-step” reduced hallucinations and improved problem solving, especially of maths problems. One theory for why this works is that AI models learn patterns. By breaking a problem down into smaller ones, it is more likely that the model will be able to recognise and apply the right one. But, says Edoardo Ponti at the University of Edinburgh, such prompt engineering amounts to treating a symptom, rather than curing the disease.
Perhaps, then, the problem is that accuracy is too much to ask of LLMs alone. Instead, they should be part of a larger system—an engine, rather than the whole car. One solution is retrieval augmented generation (RAG), which splits the job of the AI model into two parts: retrieval and generation. Once a prompt is received, a retriever model bustles around an external source of information, like a newspaper archive, to extract relevant contextual information. This is fed to the generator model alongside the original prompt, prefaced with instructions not to rely on prior knowledge. The generator then acts like a normal LLM and answers. This reduces hallucinations by letting the LLM play to its strengths—summarising and paraphrasing rather than researching. Other external tools, from calculators to search engines, can also be bolted onto an LLM in this way, effectively building it a support system to enhance those skills it lacks.
Even with the best algorithmic and architectural antipsychotics available, however, LLMs still hallucinate. One leaderboard, run by Vectara, an American software company, tracks how often such errors arise. Its data shows that GPT-4 still hallucinates in 3% of its summaries, Claude 2 in 8.5% and Gemini Pro in 4.8%. This has prompted programmers to try detecting, rather than preventing, hallucinations. One clue that a hallucination is under way lies in how an LLM picks words. If the probability distribution of the words is flat, ie many words have similar likelihoods of being chosen, this means that there is less certainty as to which is most likely. That is a clue that it might be guessing, rather than using information it has been prompted with and therefore “knows” to be true.
Another way to detect hallucination is to train a second LLM to fact-check the first. The fact-checker can be given the “ground truth” along with the LLM’s response, and asked whether or not they agree. Alternatively, the fact-checker can be given several versions of the LLM’s answer to the same question, and asked whether they are all consistent. If not, it is more likely to be a hallucination. NVIDIA, a chipmaker, has developed an open-source framework for building guardrails that sit around an LLM to make it more reliable. One of these aims to prevent hallucinations by deploying this fact-checking when needed.
Although such approaches can decrease the hallucination rate, says Ece Kamar, head of the AI frontiers lab at Microsoft, “it is unclear whether any of these techniques is going to completely get rid of hallucinations.” In many cases, that would be akin to self-sabotage. If an LLM is asked to generate ideas for a fantasy novel, for example, its output would be disappointing if limited to the world as it is. Consequently, says Dr Kamar, her research aims not to get rid of all hallucinations, but rather to stop the model from hallucinating when it would be unhelpful.
Safe and warm
The hallucination problem is one facet of the larger “alignment” problem in the field of AI: how do you get AI systems to reliably do what their human users intend and nothing else? Many researchers believe the answer will come in training bigger LLMs on more and better data. Others believe that LLMs, as generative and probabilistic models, will never be completely rid of unwanted hallucinations.
Or, the real problem might be not with the models but with its human users. Producing language used to be a uniquely human capability. LLMs’ convincing textual outputs make it all too easy to anthropomorphise them, to assume that LLMs also operate, reason and understand like humans do. There is still no conclusive evidence that this is the case. LLMs do not learn self-consistent models of the world. And even as models improve and the outputs become more aligned with what humans produce and expect, it is not clear that the insides will become any more human. Any successful real-world deployment of these models will probably require training humans how to use and view AI models as much as it will require training the models themselves.
Curious about the world? To enjoy our mind-expanding science coverage, sign up to Simply Science, our weekly subscriber-only newsletter.
© 2023, The Economist Newspaper Limited. All rights reserved. From The Economist, published under licence. The original content can be found on www.economist.com
-
Blockchain1 week ago
Crypto execs cheer as Australia appoints pro-crypto assistant minister – Crypto News
-
Business6 days ago
How Mid-Sized Treasurers Are Managing Liquidity Amid Uncertainty – Crypto News
-
Blockchain1 week ago
Top Expert Declares It The Best Crypto To Buy Now – Crypto News
-
Business1 week ago
No Truth to Truth Social Memecoin: World Liberty Financial Clarifies – Crypto News
-
others1 week ago
Why Is Crypto Market Down When S&P 500 Flashes Bull Run Ahead? – Crypto News
-
others1 week ago
Hackers Attempting To Extort School Employees via Email After Millions of Students’ Personal Data Leaked in Breach: Report – Crypto News
-
Business1 week ago
XRP Flips Tether’s USDT By Market Cap Reclaiming 3rd Spot, Price Rally To $3? – Crypto News
-
others1 week ago
USD/JPY falls below 148.00 despite persistent uncertainty over BoJ’s policy outlook – Crypto News
-
Business1 week ago
Crypto News: Animoca Brands Eye NYSE Listing Amid Donald Trump’s Crypto Push – Crypto News
-
others1 week ago
MoonX: BYDFi’s On-Chain Trading Engine — A Ticket from CEX to DEX – Crypto News
-
others1 week ago
Analyst Sees Crypto Repeating Dot-Com Bubble, Predicts Rallies for XRP and One Solana Challenger – Crypto News
-
Blockchain1 week ago
10 Signs a Crypto Investment Platform Is a Scam—and How to Avoid It – Crypto News
-
Technology1 week ago
XRP Price Prediction as Binance Data Reveals Early Signs Of Bull Run – Crypto News
-
others1 week ago
SEC Crypto Roundtable: Paul Atkins Vows To Make US Crypto Capital Of The World – Crypto News
-
Business1 week ago
Bitcoin Price Risks Dropping Below $100k As Crypto Liquidations Hit $714M – Crypto News
-
others1 week ago
Investor Kidnapped, Driven to Remote Desert and Robbed of $4,000,000 in Cryptocurrency by Teenagers: Report – Crypto News
-
Metaverse1 week ago
Why AI is central to the new browser wars – Crypto News
-
Business1 week ago
Can WIF Price Hit $2? Pattern Breakout and 100% OI Surge to $445M Signal Major Upside – Crypto News
-
others1 week ago
AUD/USD gains after softer CPI data from the US and trade developments – Crypto News
-
others1 week ago
Breaking: US SEC Delays Decision on Grayscale Spot Solana and Litecoin ETFs – Crypto News
-
Technology1 week ago
Best wireless soundbars in 2025: Top 10 picks to elevate your home audio experience – Crypto News
-
Technology1 week ago
Pi Coin Crashes 33% As Pi Network Community Screams ‘Betrayal’ – Crypto News
-
Business6 days ago
World Liberty Financial Partners Chainlink To Enable USD1 Stablecoin Cross-Chain Transfers – Crypto News
-
Cryptocurrency6 days ago
Ripple’s XRP may enable BRICS to ditch dollar and settle trade in gold – Crypto News
-
others1 week ago
Pepe Coin Price Outperforms DOGE and SHIB, Targets 80% Upside Post-Retest – Crypto News
-
Blockchain1 week ago
Top Expert Declares It The Best Crypto To Buy Now – Crypto News
-
Technology1 week ago
iQOO Neo 10 vs Motorola Edge 60 Pro: Which smartphone to buy under Rs.35000 – Crypto News
-
others1 week ago
Silver trims early gains, holds above 50-day EMA as weak US CPI tempers Fed tightening bets – Crypto News
-
Cryptocurrency1 week ago
Ripple (XRP) Price Analysis: $5.5 Billion XRP Open Interest Signals Positive Reaction to Paul Atkins’ Latest Update – Crypto News
-
Business1 week ago
Ripple (XRP) Price Analysis: $5.5 Billion XRP Open Interest Signals Positive Reaction to Paul Atkins’ Latest Update – Crypto News
-
Technology1 week ago
Ripple (XRP) Price Analysis: $5.5 Billion XRP Open Interest Signals Positive Reaction to Paul Atkins’ Latest Update – Crypto News
-
Cryptocurrency1 week ago
Crypto news today: Bitcoin nears all-time high; ETH, DOGE, PEPE, ATOM show bullish signs – Crypto News
-
Blockchain1 week ago
Alarm bells ring in US over OpenAI’s crypto project World – Crypto News
-
others1 week ago
Tests 100.50 support, with nine-day EMA providing backing – Crypto News
-
Business1 week ago
COIN Stock Soars 23% Ahead of Coinbase’s May 19 Debut on S&P 500 – Crypto News
-
Cryptocurrency1 week ago
Why investors should say ‘no’ more often – Crypto News
-
Cryptocurrency1 week ago
Why investors should say ‘no’ more often – Crypto News
-
others1 week ago
Crypto Market in April: Meme Tokens, Liquidity Battles, and New Market Dynamics – Crypto News
-
Technology1 week ago
Realme GT 7 launching in India on May 27: 7,000mAh Battery, 120W charging, 120FPS gaming confirmed – Crypto News
-
Cryptocurrency1 week ago
Anthropic Accused of Citing AI ‘Hallucination’ in Song Lyrics Lawsuit – Crypto News
-
others1 week ago
Nifty 50 Index Elliott Wave technical analysis [Video] – Crypto News
-
others7 days ago
Crypto Trader Prints 517x Profit on Solana-Based Altcoin That’s Exploded 7,000% in Just One Week: Lookonchain – Crypto News
-
Blockchain7 days ago
Stablecoin bill passes in Northern Marianas as House overrides veto – Crypto News
-
others6 days ago
Dogecoin On-Chain Metrics Hint At DOGE Mega Rally Ahead – Crypto News
-
Blockchain6 days ago
Solana Poised For Upside Move After A Bounce From $168 – Crypto News
-
others6 days ago
XRP Futures ETF Goes Live on May 19: Will It Beat ETH And BTC Debut? – Crypto News
-
Cryptocurrency6 days ago
Top crypto to buy as Saudi Central Bank reveals exposure to MSTR – Crypto News
-
Cryptocurrency6 days ago
UK confirms crypto tax data rules under CARF; first deadline set for May 2027 – Crypto News
-
Technology6 days ago
Coinbase estimates $400M cost after data breach and crypto scam – Crypto News
-
Blockchain6 days ago
Bitcoin Panic Buying? Eric Trump Says the World Is Stockpiling BTC – Crypto News