Towards AI Fluency - Part 3 - The expert - The Plausibility Trap: Why AI Lies and How Experts Can Fight Back
AI models are "plausibility engines" designed to fabricate convincing lies. To prevent real-world harm, experts must fight "automation bias" and adopt a strict framework to verify all AI output.
This article is part of a series on AI Fluency. Click here for the other articles.
TL;DR:
The shocking truth: Lawyers, doctors, and academics are falling for AI fabrications—not because they’re incompetent, but because they’re human.
AI chatbots are 64% more persuasive than humans, weaponizing our “automation bias”—the cognitive shortcut that makes us trust machines over our own expertise. When AI’s confident, authoritative tone meets our overworked brains, we become victims of the “Vulnerability Loop.”
The cost? Ruined careers, financial liability, and collapsing public trust in all expert claims.
But there’s a solution. The “Master of Some” framework transforms you from passive consumer to expert auditor. Learn to spot the “tells” of fabrication, deploy grounding prompts that force AI accuracy, and implement verification protocols that break the persuasion trap.
The future isn’t about trusting AI—it’s about mastering the art of auditing it. Your expertise isn’t obsolete; it’s more critical than ever as the “pivotal last 20%” that machines can’t deliver.
Ready to reclaim your expert edge?
THE PROBLEM: Why We Need This Breakthrough
The single greatest danger of modern artificial intelligence is not just its capacity for error, but its profound ability to be incorrect with absolute, unshakable confidence. An AI can generate a response that is eloquent, persuasive, and structurally beautiful—all while being completely and utterly false. This reality demands a fundamental shift in how every professional works, forcing us to re-forge our hard-won domain knowledge into a critical shield.
To start, we must stop using the term “hallucination.” It’s a dangerously misleading metaphor that suggests a human-like psychological quirk, a temporary “ghost in the algorithm” that will pass. This framing tragically lowers our guard. These outputs are not “bugs”; they are a fundamental characteristic of the system’s design.
A more accurate term is “fabrication” or “confabulation.” The distinction is critical: one waits for a hallucination to end, but one verifies a fabrication.
This leads to the core of the problem. Professionals often approach an AI expecting a “truth engine”—a kind of digital oracle that can retrieve facts. What they are interacting with is a “plausibility engine.” These models are not designed to know truth from fiction. They are designed to mimic the logical structure of reasoning and produce a statistically probable sequence of words.
The smooth, authoritative tone that defines AI-generated text is not an indicator of the model’s certainty; it is the default linguistic style. The AI has no internal mechanism to signal its own uncertainty. In fact, its training actively rewards it for guessing over acknowledging uncertainty. This “confident fabrication” is a feature, not a flaw, and it creates the central conflict for the modern professional: the very tool designed to augment our intelligence is psychologically engineered to bypass our expert judgment.
THE SOLUTION: How the Core Findings Work
To defend against these fabrications, we must first understand why they are an inherent and unavoidable product of the AI’s core architecture. The flaws in the machine’s design reveal a clear pattern of real-world harm, demonstrating exactly where and how our human shields are failing.
The Engine of Fabrication: Why AI Is Built to Fib
The fabrications are not random. They are the logical output of the AI’s design, which is misaligned with the concept of “truth” at three fundamental levels.
The “Plausibility-Out” Problem: Large language models (LLMs) are trained on “vast amounts of internet data.” This training library is a mirror of humanity’s collective knowledge, but also its collective inaccuracies, biases, and errors. The models learn to “mimic patterns” in this data without discerning truth. This is a new spin on the old “garbage-in, garbage-out” principle. Here, it’s “garbage-in, plausibility-out.” The output isn’t raw garbage; it’s a highly refined and convincing echo of that garbage. The training process itself is part of the problem, as models are statistically rewarded for being “good test-takers”—that is, for guessing plausibly rather than admitting a gap in knowledge.
The “Next-Word-Prediction” Engine: At their core, most LLMs are “autoregressive” models. This is a technical term for a simple function: they are built to predict the next most statistically likely word, one after another, based on the words that came before. This is an act of statistical assembly, not cognitive reasoning. The text they produce is that which is statistically probable, not factually correct.
An Analogy: Think of the AI as a musician who can perfectly replicate any song by ear after hearing it (mimicry) but has zero understanding of music theory, composition, or harmony (reasoning). This design perfectly explains the “low-frequency fact” problem. An AI can “learn” high-frequency patterns like grammar, common knowledge, and conversational style. It cannot reliably learn arbitrary, low-frequency facts—like the specific details of an obscure legal case—because those facts can’t be statistically predicted from language patterns alone. A fabrication is simply the model’s most plausible-sounding “guess” when a verifiable fact is absent.
The “Memorization vs. Reasoning” Illusion: The “knowledge” that LLMs appear to possess is an illusion of memorization, not an act of understanding. Research shows this memorized “knowledge” is incomplete, inconsistent, and scattered in a non-human way across billions of parameters. It mimics the form of human intelligence without the function. This is why an AI can “confidently regurgitate incorrect proofs of linear algebra theorems.” It is mimicking the language of a mathematical proof without understanding the logic of the math itself.
This reveals the fundamental conflict: a user is asking a “stochastic parrot” for a legal opinion, and the parrot obliges by perfectly mimicking the language of a lawyer while completely inventing the facts. This is the system working precisely as designed. The industry’s own response to this flaw is the creation of Retrieval-Augmented Generation (RAG) systems, which explicitly “bolt-on” an AI to an external, verifiable database. In the absence of such a tool, the expert professional must serve as a human RAG system.
A Trail of Confident Errors: Fabrication in the Real World
These theoretical flaws have already caused significant, documented harm in high-stakes professional fields. The fabrications are not simple mistakes; they are “deceptively realistic” constructs that skillfully blend fact and fiction.
The Legal Dossier: Inventing the Law
The legal profession has provided the most alarming case studies. The landmark incident is Roberto Mata v. Avianca, Inc. In this case, an attorney used ChatGPT for legal research and submitted a brief that cited six judicial decisions that were complete fictions. They were “bogus judicial decisions with bogus quotes and bogus internal citations.”
When the presiding judge questioned these cases, the attorney’s defense became a perfect case study in AI-induced failure. He admitted in an affidavit that he “was unaware of the possibility that its contents could be false.” Most tellingly, he had explicitly asked ChatGPT if the cases were real. The AI assured him they were and that they “can be found in reputable legal databases such as LexisNexis and Westlaw.” The AI hallucinated a confirmation of its own initial hallucination.
This is not an isolated incident:
In Kaur v. Desso, another attorney was fined for citing fabricated AI cases.
In a defamation case involving MyPillow Inc., a brief was filed with nearly 30 defective, AI-generated citations.
Data scientist Damien Charlotin has cataloged at least 490 court filings containing AI fabrications, noting the “pace is accelerating.”
Even specialized tools, like Thomson Reuters’s “Ask Practical Law AI,” have been caught fabricating legal facts.
The Academic Dossier: The “Hybrid” Fake
AI fabrications are most dangerous when they are hybrids—blending real information with invented content. In academic contexts, models are known to “blend genuine results with fabricated content.”
Fabricated Citations: Models invent “deceptively realistic” references by mixing real author names with fabricated paper titles.
Invented DOIs: One study of citations in the natural sciences found that 29.1% of all Digital Object Identifiers (DOIs) generated by an LLM were entirely fictitious.
The “Nonexistent Statistician”: A researcher documented an AI inventing a fake statistician and attributing a real statistical method to them.
The Systemic Failure: This vulnerability is not unique to AI. To test the academic ecosystem, a journal published an entirely fictional article. That fabricated paper was subsequently cited over 400 times in other academic papers, revealing a profound, systemic failure of verification by human experts.
The Quotation Dossier: Putting Words in Real Mouths
Perhaps the most insidious fabrication is the invention of quotes attributed to real people. An AI-generated article, for example, invented a fake quote about Harry Potter and attributed it to “Dr. Reuben Binns, professor of Artificial Intelligence at the University of Cambridge.” Dr. Binns is a real person. A superficial fact-check (e.g., “Is Dr. Reuben Binns real?”) would incorrectly confirm the output’s legitimacy.
In another test, three of the most popular LLMs were asked for the title of computer scientist Adam Kalai’s dissertation. All three models failed, and each one invented a different, plausible-sounding, and completely false title and year.
THE FUTURE: What This Means for All of Us
If fabrications are so common, why do highly trained experts—lawyers, academics, and researchers—fall for them? The answer lies in a toxic psychological interaction: the AI’s persuasive design exploits our innate cognitive biases. To fight back, we must adopt a new framework for expert verification.
The Persuasion Trap: Why We Fall for the Lie
The vulnerability is a two-part problem: a human bias meets a persuasive machine.
Our “Automation Bias”: This is the well-documented “human tendency to choose the least cognitive approach” and to “view automated aids as having an analytical ability superior to their own.” This is not a sign of low intelligence; it is a cognitive shortcut that appears most often in high-workload, complex environments. It has been documented for decades in:
Aviation: Pilots over-relying on automated cockpit systems.
Medicine: Doctors placing excessive trust in diagnostic support systems, even when their own judgment conflicts.
Military: Defense crews trusting flawed system data, resulting in “crucial and even obvious errors.”
A Non-AI Example: The tragic UK Post Office scandal, where auditors trusted a “faulty accounting system” over the “substantial evidence to the contrary” presented by hundreds of postmasters.
This bias explains why the lawyers in the Mata v. Avianca and Kaur v. Desso cases failed. They were under pressure and defaulted to trusting the machine.
The AI “Persuasion Engine”: This human vulnerability is met by an AI that is not a neutral tool but a highly effective “persuasion engine.” Research has shown that AI chatbots are 64% more persuasive than humans in online debates. They achieve this by adopting an “authoritative tone” and “expert communication” style, which directly “elevates trust levels” in users. They also deliver information with “cognitive ease,” hijacking the brain’s preference for answers that feel familiar and effortless.
The AI’s ability to persuade and its tendency to fabricate are not separate problems. They are the same problem. Both stem from the same engine that has perfected the mimicry of plausible, authoritative human language.
This creates the “Vulnerability Loop”: The AI’s confident tone is the trigger; our human automation bias is the vulnerability. The result is a “knowledgeable user” who abdicates their expert verification duty. This leads to severe risks: the spread of misinformation, direct professional and financial liability, and the large-scale “erosion of public trust” in all expert claims, verified or not.
The “Master of Some” Defense: A Framework for Expert Verification
A professional cannot “fix” the AI’s tendency to fabricate. They can only fix their process. This requires a “Master of Some” to shift their mental model from trusting an “automaton” to auditing a “collaborator.” The expert’s role is to provide the “pivotal last 20% of accuracy” that the machine cannot, applying uniquely human skills like analytical thinking and specialized domain knowledge. The professional becomes an auditor of the AI’s draft, not a consumer of its final product.
Step 1: Identify the “Tells” (The Gut Check)
An expert auditor learns to recognize the “scent” of a fabrication.
Linguistic “Tells”: An overly formal or neutral tone; repetitive phrases; or the use of specific AI-favored buzzwords like “game changer.”
Factual “Tells”: “Overconfident statements” (e.g., “it is a proven fact”) presented without a source; “unverifiable facts”; or information that is “missing context.”
Heuristic “Tells”: The expert’s own domain knowledge. If an output “sounds off” or contradicts your hard-won experience, it must be treated as a fabrication until proven otherwise.
Step 2: The Verification Methodology (The Action)
This methodology is a cognitive forcing function designed to break the “Vulnerability Loop.”
A. Challenge, Don’t Ask:
A novice “Asker” will ask, “Is that true?” This invites a confirmatory hallucination, as seen in the Mata v. Avianca case. An expert “Director” challenges the AI: “Provide the specific source, author, and URL for that statistic.”
B. Implement Grounding Prompts (Prevention):
The expert must “carefully” and “directly” prompt the AI to force grounding before it generates.
“According to...” Prompting: Instead of “Explain climate change,” the expert directs: “According to the Intergovernmental Panel on Climate Change (IPCC), explain the main causes of climate change.”
Chain-of-Verification (CoV) Prompting: This forces the AI to check its own work “step by step.” Instead of “Give me the history of the internet,” the expert directs: “First, explain when the internet was invented. Then, confirm who were the key people involved. Finally, verify the major events in its development.”
C. Actively Cross-Verify (Detection):
Never trust the AI’s output alone. Never trust the AI’s “assurance” that its output is correct.
Cross-check with trusted sources: This is the “first step... always.” Use authoritative databases (e.g., Google Scholar, LexisNexis, PubMed), not just a standard web search.
Verify citations: If the AI provides a source, the expert must search within that source for the specific statistic or quote.
Cross-reference with other LLMs: Use a different AI model to see if the outputs align. Discrepancies are a major red flag.
D. Demand Source-Linking (The Gold Standard):
The most robust verification systems are those that link generated content directly to source materials. This allows for rapid, in-line validation. The expert professional must demand these tools from vendors or, failing that, build a personal workflow that manually replicates this process. This is the new burden—and the new, irreplaceable value—of being an expert in the age of AI.

