I thought the answer to that is obvious: They don't provide information, they provide noise that looks like information; they simply got really good at making it look like information. Their hallucinations are just noise that doesn't look quite right.
The way they got good at making it look like information is through a big-data, high-dimensional form of what amounts to a high-falutin' form of pattern matching. They look at an enormous amount of human output and try to mimic that output. Having scraped, for example, a court filing with a footnote that says
President and Fellows of Harvard College v. U.S. Dep’t of Health and Human Services et al., No. 25-cv-11048 (D. Mass. Apr. 21, 2025) (the “Funding Case”).
and another citation in a similar context that says
Nat’l Rifle Ass’n of Am. v. Vullo, 602 U.S. 175, 189 (2024) (quoting Bantam Books, Inc. v. Sullivan, 372 U.S. 58, 67 (1963)).
an LLM that's trying to discuss a related subject might invent a citation that says something like
Nat’l Rifle Ass’n of Am. v. U.S. Dep’t of Health and Human Services, 372 US 175, (D. Mass. Mar. 2025)
because, to the LLM, that reads like the sort of thing it has seen humans write. LLMs are more sophisticated than my made-up example, I know (because my made-up example is itself just noise that might look like information), but most AI hallucinations are essentially the sort of BS we often see from humans who don't really know what they're talking about but are trying to pretend they do.
(And those AI hallucinations are themselves mimicking a common human behavior. The ISF archives contain several long threads driven by posters who don't really know what they're talking about but are trying to pretend they do.)
From a conversation with a software developer
A month or two ago, I had an interesting conversation with someone who has been leading a small team of software developers.
Two members of his team are contract programmers. One of those contract programmers asked whether it would be acceptable for him to use an AI tool to help write his code. He was told that would not be acceptable, because using an AI tool to write code would reveal aspects of his code to a third party (the supplier of the AI tool). The contract programmer accepted that explanation, and did not use AI.
The other contract programmer used an AI tool without asking permission first. When that came to light, the team leader had to tell him that was not acceptable, and warned him not to do it again.
At the same time, that particular lead programmer is one of several software developers at his company that have been looking into how the company can use AI tools to help write its software. He told me there are some very good AI tools that specialize in writing software (I did not know that), and they'd like to use them if they could figure out a way to sandbox their use in a way that would ensure none of the resulting code could be revealed to anyone outside the company. He also told me that some of those AI tools learn from working with software developers, and would improve their performance as those tools gain more experience with the particular programming languages and idioms used by the company. There would be mutual benefit from bringing those AI tools into the company's work flow, if only they could solve the problems with intellectual property and trade secrets.
The software developer has also been responsible for interviewing potential hires. Most of those interviews have been done remotely. As part of those interviews, he asks technical questions designed to assess the technical competence of the interviewee. He has the impression that many of those he interviewed attempted to answer some of his technical questions by using LLMs in real time, during the interview.
He also told me that his company is hiring very few junior programmers (i.e. graduates of coding schools up through BS in CS), because the AI tools are now as good (or better!) at the kind of grunt programming they can reasonably entrust to junior-level people. They aren't using those AI tools now, but they expect to be using them in the near future, and they don't want to hire people they'd be hoping to lay off in the near future.
They are still hiring senior-level people, because the AI tools just aren't much good for high-level design or software architecture, and they see no prospect of that changing in the near future.