Ars Technica fires senior AI reporter after AI tool fabricated quotes in published story
Ars Technica retracted an article by senior AI reporter Benj Edwards after it contained fabricated quotations generated by an AI tool and attributed to a source who never said them. The publication acknowledged the incident as a "serious failure of our standards" and Edwards was subsequently fired. Edwards noted the irony on Bluesky: "The irony of an AI reporter being tripped up by AI hallucination is not lost on me."
Incident Details
Tech Stack
References
The Reporter Who Knew Better
Benj Edwards was not some intern unfamiliar with the risks of AI. He was Ars Technica's senior AI reporter - arguably one of the most qualified journalists in tech media to understand precisely how and why large language models hallucinate. He had covered AI failures professionally for years. His publication, owned by Conde Nast, had explicit policies prohibiting AI-generated material unless specifically labeled for demonstration purposes. He knew the rules, he knew the risks, and he used the tools anyway.
Edwards was sick with COVID and running a fever when he sat down to write a story about Scott Shambaugh, a software developer who claimed an AI agent had written and published a hit piece about him after Shambaugh declined the agent's code contributions. It was the kind of story Edwards had covered many times: AI behaving in ways its creators didn't intend, with real consequences for real people.
To process quotes from Shambaugh's two-page blog post, Edwards reached for AI assistance. He first tried an experimental Claude Code-based tool to extract relevant verbatim source material from the blog post. Not to generate the article, he later clarified, but "to help list structured references I could put in my outline." When Claude refused due to content policy restrictions, he turned to ChatGPT.
The AI didn't extract the actual quotes. It generated paraphrased versions - fabricated text that was close enough to be plausible but different enough to be something Shambaugh never actually said. Edwards used these fabricated quotes in his article and attributed them to Shambaugh as direct quotations.
The Discovery and Retraction
The article was published on February 13. It didn't take long for the problem to surface. Shambaugh himself noticed that the quotes attributed to him were things he'd never said. Readers flagged the discrepancies. The story unraveled quickly.
Ars Technica pulled the entire article rather than issuing corrections and noting changes - a departure from standard journalistic practice, but one that reflected the severity of the problem. You can correct a factual error in an otherwise sound article. An article built on fabricated quotes attributed to a real person is unsalvageable.
Editor-in-chief Ken Fisher published an editor's note confirming that the piece included "fabricated quotations generated by an AI tool and attributed to a source who did not say them." He characterized the error as "a serious failure of our standards" and noted that Ars Technica "covered the risks of overreliance on AI tools for years" - a statement that contained its own uncomfortable irony.
The Firing
Edwards was fired shortly after. Futurism confirmed that he was no longer working at Ars Technica, and reporting indicated the termination was directly connected to the fabricated quotes incident.
On Bluesky, Edwards took responsibility and acknowledged the absurdity of his situation with a directness that garnered some sympathy: "The irony of an AI reporter being tripped up by AI hallucination is not lost on me." He maintained that his intent was not to use AI to fabricate content but to use it as a processing tool for material he already had. The distinction is technically meaningful - using AI to extract quotes from a source is different from using AI to write quotes - but the result was the same: fabricated words in a published article attributed to someone who didn't say them.
The Structural Question
The response to Edwards' firing split along predictable lines. Some viewed it as appropriate accountability - a journalist fabricated quotes, whatever the mechanism, and was terminated for violating the most fundamental rule of the profession. Others pointed to structural factors.
One commenter's analysis circulated widely: "The individual firing is a distraction from the structural issue. Newsrooms have been cutting editorial staff for a decade, which means the verification layers that would have caught this - fact-checkers, copy editors, senior editors doing source verification - largely don't exist anymore. Then they adopt AI tools that increase throughput without increasing oversight capacity, and act surprised when fabrication slips through."
The argument has merit. A copy editor or fact-checker comparing Edwards' article against Shambaugh's original blog post would have caught the discrepancy instantly. But at many publications, those positions have been eliminated or reduced to the point where they can't meaningfully review every piece. The same economic pressures that lead newsrooms to cut editorial staff also lead them to adopt AI tools that promise to do more with less - a combination that systematically reduces the safeguards against exactly this kind of failure.
The Two-Page Blog Post Problem
There's an additional detail that makes this incident particularly difficult to defend: the source material was a two-page blog post in plain English. This was not a sprawling court document or a dense technical paper where AI assistance for extraction might be defensible. It was a short, readable blog post. Copying and pasting the actual quotes would have taken less time than setting up and prompting the AI tool.
This is the part that's hardest to explain as anything other than a case study in tool dependency. AI assistance is most justifiable when dealing with volumes of material that exceed what a human can efficiently process. Using AI to extract quotes from two pages of accessible prose suggests a workflow where the tool has become the default approach regardless of whether it's needed - habit rather than necessity.
What the Incident Demonstrates
For newsrooms, the lesson the Media Copilot identified in their coverage is blunt: "AI tools cannot reliably perform basic journalism tasks like accurately citing sources." The hallucination problem is not a limitation that affects only complex tasks. It affects the simplest possible use case - reading a short text and reproducing what it says.
Edwards' case also demonstrates that expertise about AI's limitations does not immunize the expert against those limitations. Knowing that AI hallucinates is different from noticing that it hallucinated in a specific instance, especially when the output looks plausible and you're sick with a fever and trying to meet a deadline. The failure mode isn't ignorance; it's trust calibration under pressure - the moment when you know the tool is unreliable in general but assume it was reliable this time.
For AI reporters specifically, the incident created a credibility problem that extends beyond one journalist. If the person whose job is to cover AI failures can't avoid AI failures in their own work, it undermines the authority of AI-critical journalism as a category. The counterargument - that getting tripped up by the technology you cover makes you more credible, not less - didn't gain much traction.
Ars Technica has continued operating. Shambaugh got an apology. Edwards got a career-defining mistake that will follow him for the rest of his professional life. And newsrooms everywhere got reminded that using AI to save time on tasks that don't need AI is not saving time at all.
Discussion