[TUHS] On the unreliability of LLM-based search results

Alexis flexibeast at gmail.com
Tue May 27 12:21:07 AEST 2025


Jason Bowen via TUHS <tuhs at tuhs.org> writes:

> May 26, 2025 11:57:18 Henry Bent <henry.r.bent at gmail.com>:
>
>> It's like Wikipedia.
>
> No, Wikipedia has (at least historically) human editors who 
> supposedly
> have some knowledge of reality and history.
>
> An LLM response is going to be a series of tokens predicted 
> based on
> probabilities from its training data. The output may correspond 
> to a
> ground truth in the real world, but only because it was trained 
> on
> data which contained that ground truth.
>
> Assuming the sources it cites are real works, it seems fine as a
> search engine, but the text that it outputs should absolutely 
> not be
> thought of as something arrived at by similar means as text 
> produced
> by supposedly knowledgeable and well-intentioned humans.

LLMs are known to hallucinate sources. Here's a database of "legal
decisions in cases where generative AI produced hallucinated 
content":

  https://www.damiencharlotin.com/hallucinations/

Here's a research paper about LLMs hallucinating software 
packages:

  https://arxiv.org/abs/2406.10279

Not to mention about LLMs hallucinating 'facts' about people:

  https://www.abc.net.au/news/2025-03-21/norwegian-man-files-complaint-chatgpt-false-claims-killed-sons/105080604

As a result of what they're trained on, chatbots can be 
"confidently wrong":

  > [T]he chatbots often failed to retrieve the correct
  > articles. Collectively, they provided incorrect answers to 
  > more than
  > 60 percent of queries. Across different platforms, the level 
  > of
  > inaccuracy varied, with Perplexity answering 37 percent of the
  > queries incorrectly, while Grok 3 had a much higher error 
  > rate,
  > answering 94 percent of the queries incorrectly.

-- 
   https://www.cjr.org/tow_center/we-compared-eight-ai-search-engines-theyre-all-bad-at-citing-news.php

and can contradict themselves:

  https://catless.ncl.ac.uk/Risks/34/04/#subj2.1

Coming back to LLMs in the context of software:

  > AI coding tools ‘fix’ bugs by adding bugs
  >
  > ...
  >
  > What happens when you give an LLM buggy code with and tell it 
  > to fix
  > it? It puts in bugs! It might put back the same bug!
  >
  > Worse yet, 44% of the bugs the LLMs make are previously known
  > bugs. That number’s 82% for GPT-4o.
  > 
  > ...
  > 
  > I know good coders who find LLM-based autocomplete quite okay 
  > —
  > because they know what they’re doing. If you don’t know what 
  > you’re
  > doing, you’ll just do it worse. But faster.

-- 
   https://pivot-to-ai.com/2025/03/19/ai-coding-tools-fix-bugs-by-adding-bugs/

A research paper found that:

  > participants who had access to an AI assistant based on 
  > OpenAI's
  > codex-davinci-002 model wrote significantly less secure code 
  > than
  > those without access. Additionally, participants with access 
  > to an
  > AI assistant were more likely to believe they wrote secure 
  > code than
  > those without access to the AI assistant. Furthermore, we find 
  > that
  > participants who trusted the AI less and engaged more with the
  > language and format of their prompts (e.g. re-phrasing, 
  > adjusting
  > temperature) provided code with fewer security 
  > vulnerabilities.

-- https://arxiv.org/abs/2211.03622

As someone who has spent a lot of time and energy working on 
wikis, i would say that a big difference between LLMs and wikis is 
that one can directly fix misinformation on wikis in a way that 
one can't do with LLMs. And wikis typically provide public access 
to the trail of what changes were made to the information, when, 
and by whom (or at least from what IP address), unlike the 
information provided by LLMs.

The LLM cat is well and truly out of the bag. But the combination 
of LLMs hallucinating information, together with the human 
tendency to correlate the confident conveyance of information with 
veracity of that information, means that people should be 
encouraged to take LLM output with a cellar of salt, and to check 
the output against other sources.


Alexis.


More information about the TUHS mailing list