Article image Logo

Bigger Windows Better Lies

The Industry’s Favorite Bedtime Story

For a while now, the AI industry has been telling a very soothing story. Yes, the chatbot made things up. Yes, it cited fake cases, invented facts, and occasionally behaved like a summer intern with a cocaine habit. But relax. This was temporary. A scaling issue. A retrieval issue. A context issue. Give the model more information, more tokens, more enterprise wrappers, more expensive infrastructure, and the machine would gradually stop acting like a gifted liar at a networking event.

That story was always convenient because it preserved the basic sales pitch. The machine was not fundamentally unreliable. It was merely under-informed. The problem was not the system. The problem was that the system needed more.

Now comes the awkward part. Some of the newer research suggests that when you give these systems more to work with, they do not necessarily become more trustworthy. In some cases, they become more likely to invent. That is a marvelous little twist. The machine is handed a larger pile of evidence and responds by improvising harder.

This is the kind of detail that deserves more public attention because it cuts directly against one of the most popular assumptions in business AI. A lot of companies are not buying chatbots because they enjoy danger. They are buying them because they believe the reliability issue is steadily being engineered away. The promise is always the same. The next version will be more grounded. The next architecture will be more accurate. The next context expansion will bring the machine closer to something resembling judgment.

And yet here we are, watching the high-tech oracle read a thicker stack of documents and somehow drift further from reality.

Give the machine more and watch what happens

The Reuters piece that surfaced this latest round of discussion pulled from a study by JV Roig, who tested large language models on document question-answering tasks across different context lengths. The setup was not some vague philosophical stress test. The source material was known in advance, which meant hallucinations could be measured directly rather than guessed at through vibes, screenshots, or the ancient corporate ritual known as “it seemed fine in the demo.”

At smaller context lengths, some models performed reasonably well. Not perfect, but respectable enough to keep the PowerPoint economy alive. Then the context got larger. Error rates rose. At the highest tested context lengths, some models did not merely wobble. They collapsed into majority-error behavior.

That matters because the enterprise AI pitch is saturated with context inflation. Vendors brag about giant windows the way men in midlife brag about Italian sports cars. Bigger number, bigger promise, bigger implied competence. The assumption is that a model that can ingest more material will be able to reason better across more material. That sounds intuitive right up until the part where it doesn’t.

What the findings suggest is that more context is not the same thing as more truth. It may simply create a larger field in which the model can misread, overgeneralize, confuse signal with noise, or produce an answer that feels smooth enough to pass as knowledge. The machine does not panic when it does not know. It performs.

And that performance is the real seduction. A bad spreadsheet looks bad. A broken printer looks broken. A chatbot hallucination often arrives dressed like competence. It has tone. It has formatting. It has that polished middle-management confidence that makes people think something serious has just happened. In reality, the system may be doing little more than guessing in an expensive accent.

Fluent wrongness is still wrongness

This is where the public conversation often gets weirdly forgiving. People hear the word hallucination and treat it like a quirky side effect, as if the chatbot occasionally sees a ghost and otherwise remains perfectly employable. But hallucination in practice often just means fabrication with style. The model gives you an answer that is false, unsupported, distorted, or entirely invented, and it gives it to you in the same calm voice it would use for a correct one. That is not a cosmetic issue. That is the issue.

If a machine cannot reliably distinguish between “I found the answer in the provided material” and “this sounds like the sort of thing a clever machine would say,” then every high-trust use case starts looking less like automation and more like gambling in business casual. The Reuters piece makes this point through law, accounting, and tax work, where “about right” is often another phrase for “legally dangerous.” The absurdity is that much of the AI economy is still built on selling exactly that kind of almost-rightness as if it were a transitional inconvenience rather than a structural liability.

There is a certain comic beauty to it. Humanity spent decades building machines to reduce error, standardize judgment, and improve reliability. Then we built one that can write in full paragraphs about things that never happened, and the market response was to ask whether it could perhaps do compliance.

The problem may not be a patch away

If the first blow to the comforting narrative is that more context does not reliably cure hallucination, the second is even less cheerful. Reuters also pointed to a Tsinghua paper on so-called hallucination-associated neurons. The significance of that paper is not that researchers found a single magic switch labeled nonsense. It is that they argue some hallucination-linked behavior originates during pretraining, in the basic next-token prediction setup that rewards plausible continuation rather than factual truth.

That distinction matters. If the tendency toward fluent invention is tied to the objective itself, then hallucination stops looking like a bug sitting on top of the model and starts looking like a habit formed deep in its developmental wiring. In other words, the machine may not be broken in the ordinary sense. It may be doing exactly what it was trained to do, just in places where humans wish it would suddenly become something else.

This is one of the funnier and darker features of the whole AI boom. People keep asking these systems to act like databases, witnesses, analysts, and experts, while the systems themselves are built to generate likely continuations of language. Then everyone acts surprised when the result is not truth but plausibility theater.

The machine does not wake up in the morning and decide to betray you. It simply completes the pattern. Unfortunately, human institutions are full of people who mistake pattern completion for knowledge. That is how you get the modern workplace ritual of someone pasting chatbot output into a document, admiring how professional it sounds, and only later discovering that the confidence was entirely decorative.

The business model problem hiding inside the product problem

This is where the story gets bigger than chatbot embarrassment. A lot of generative AI economics depends on the belief that these systems will move from useful assistants to dependable operators. Not merely brainstorming companions, but engines for serious work. Not just drafting fluff, but handling tasks where accuracy has to survive audit, law, money, and blame.

That is where hallucination becomes more than an engineering nuisance. It becomes a business-model problem. If the most lucrative applications require precision and the systems remain structurally prone to polished fabrication, then the ceiling on value may be lower than investors and vendors want to admit. You can absolutely make money selling helpful machine prose. But that is a different business from selling trustworthy machine judgment.

And that gap matters. The expensive part of the market is not built around “kind of useful.” It is built around replacing time, labor, expertise, and decision friction in environments where mistakes cost real money. If hallucination rates stay stubborn, then human review does not disappear. It thickens. Controls multiply. Verification layers return. Suddenly the grand efficiency story begins to look suspiciously like a software subscription attached to a second full-time checking process.

This is why the hallucination question refuses to die. It is not merely a quality issue. It is the tax the whole ecosystem keeps trying to expense somewhere else.

The cultural problem is that people love a confident machine

Of course, none of this would be so dangerous if people treated these systems like weird autocomplete engines with delusions of grandeur. But they do not. They treat them like authorities. Not always officially, not always out loud, but behaviorally. They let the tone do the work. They let fluency substitute for verification. They assume that a machine capable of elegant synthesis must also possess some internal relationship to truth. That assumption is doing an extraordinary amount of damage.

Because the real scandal of hallucination is not merely that the model invents. It is that humans are deeply vulnerable to invention when it arrives in a polished format. The chatbot does not need to force trust. It only needs to make trust feel frictionless.

This is why the phrase “human in the loop” often sounds less like a safeguard and more like a legal prayer. In theory, the human is there to catch the nonsense. In practice, the human is busy, impressed, undertrained, or eager for the machine to be right because the quarterly roadmap already assumed that it would be.

So the fabricated answer travels. It gets copied into a memo, a filing, a slide deck, a customer response, a research note, or a strategic recommendation. By the time someone notices that it was built on air, the problem is no longer technical. It is organizational. The lie has acquired workflow.

The machine is not maturing out of this anytime soon

The most important thing about this story is not that models hallucinate. Everyone knows that by now, at least in the abstract. The important thing is that the latest evidence keeps pointing away from the comforting fantasy that this will soon be solved by sheer iteration.

Maybe future architectures reduce the problem dramatically. Maybe some hybrid system finally forces honesty into the loop. Maybe the long-term answer comes from something other than today’s large language model paradigm. All possible. But that is a different claim from the one the market has been living on.

The market has been living on the idea that current systems are on a relatively smooth road from impressive approximation to dependable cognition. The ugly possibility raised by this research is that they are not. They may remain brilliant at producing language and stubbornly unreliable at staying attached to fact, especially under the very conditions where organizations most want to scale their use.

That would not make them useless. It would make them much easier to understand. These are not truth machines with a few remaining defects. They are probabilistic performance engines that sometimes deliver truth and sometimes deliver a beautifully wrapped counterfeit.

The chatbot era keeps trying to sell us a digital genius. What it may actually be selling, at least much of the time, is a very fast intern who never admits uncertainty, always volunteers an answer, and gets more dangerous the more impressed everyone becomes. That is not a minor product quirk. That is the plot.


©2026 Copyright by Markus Brinsa | Chatbots Behaving Badly™

Sources

  1. Reuters - Does the AI business model have a fatal flaw? reuters.com
  2. arXiv - How Much Do LLMs Hallucinate in Document Q&A Scenarios? A 172-Billion-Token Study Across Temperatures, Context Lengths, and Hardware Platforms arxiv.org
  3. arXiv - H-Neurons: On the Existence, Impact, and Origin of Hallucination-Associated Neurons in LLMs arxiv.org

About the Author