UK High Court warns lawyers after fake AI citations infected two cases
On June 6, 2025, the High Court of England and Wales issued a joint ruling in two separate matters after lawyers put fake authorities before the court. In one case tied to Qatar National Bank, a filing cited 45 authorities, 18 of which did not exist, while many of the rest were misquoted or irrelevant. In the other, a housing claim against the London Borough of Haringey included five fabricated cases. The Divisional Court, led by Dame Victoria Sharp, said tools such as ChatGPT are not capable of reliable legal research, referred the lawyers involved to their regulators, and warned that more serious future misuse could lead to contempt proceedings or even police referral. The ruling turned individual AI citation blunders into a profession-wide warning.
Incident Details
Tech Stack
References
The United Kingdom finally got its own canonical fake-citation ruling in June 2025. It did not arrive because one junior lawyer quietly embarrassed themselves in a minor filing. It arrived because the High Court had two separate cases on its hands at the same time, both contaminated by bogus authorities, both serious enough to trigger the court's special supervisory jurisdiction, and both offering a depressing amount of evidence that "please verify your sources" was still being treated as optional.
The combined judgment in Ayinde v London Borough of Haringey and Al-Haroun v Qatar National Bank matters because it moved the conversation past individual incompetence. Dame Victoria Sharp and Mr Justice Johnson used the ruling to say, in unusually direct terms, that generative AI tools such as ChatGPT are not reliable legal research engines, that lawyers remain personally responsible for anything filed in their names, and that future misconduct could draw much harsher consequences than a public judicial scolding.
Two cases, same disease
The Ayinde matter began as a housing dispute. Frederick Ayinde was challenging a decision by the London Borough of Haringey related to homelessness support. The legal merits of that dispute are not what made the case famous. What mattered here was that the grounds for judicial review, settled and signed by barrister Sarah Forey, cited five cases that did not exist.
That alone would have been enough for trouble. The part that makes the ruling read like a profession-wide intervention is what came next. When the false citations were pointed out, the response from the legal team was not immediate alarm and withdrawal. The lower court record, which the Divisional Court reviewed, showed a tendency to minimize the issue as citation error and cosmetic defect. Mr Justice Ritchie had already concluded that putting fake cases into a pleading was improper, unreasonable, negligent, and amounted to professional misconduct. The Divisional Court agreed the episode required regulatory attention.
The Al-Haroun matter was even messier. Hamad Al-Haroun was pursuing an £89.4 million claim against Qatar National Bank and QNB Capital. A schedule reviewed by the court listed 45 cited authorities. According to the judgment, 18 of those cases did not exist. Many of the remaining authorities either did not contain the quoted passages, did not support the propositions for which they were cited, or had no relevance to the application at hand. The court's summary of the schedule was devastating: the vast majority of the authorities were made up or misunderstood.
In that case, Al-Haroun accepted responsibility for using publicly available AI tools and online sources and apologized. His solicitor, Abid Hussain, also accepted that his witness statement contained nonexistent authorities, but explained that he had relied on legal research provided by his client. The court's response was about as charitable as one would expect. It called that arrangement extraordinary. Quite right. A client relying on a lawyer for legal accuracy is the point of having a lawyer.
What the court actually said
The ruling is useful because it avoids the dodge that still shows up in AI boosterism: pretending the problem is a few careless prompts rather than a known failure mode of the tools. The court said plainly that large language models are not capable of conducting reliable legal research. They produce text that looks coherent and plausible, but that appearance is not validation. The only real safeguard, the judges said, is checking the output against authoritative sources.
That sounds obvious because it is obvious. Yet the legal profession keeps rediscovering it in the most expensive way possible.
The court stopped short of initiating contempt proceedings in either matter. That detail matters because it keeps the ruling from turning into a simple punishment story. The judges left room for factual uncertainty, especially around intent and knowledge, and they drew a line between incompetence and deliberate deception. But the lack of contempt findings should not be mistaken for leniency. The court referred the relevant lawyers to their professional regulators and used the judgment to warn that more egregious future conduct could justify contempt proceedings, cost sanctions, or referral to the police.
It also sent the judgment to the Bar Council, the Law Society, and the Council of the Inns of Court, effectively telling the profession that whatever guidance existed at the time was not doing the job.
Why this case matters more than another sanctions order
There had already been AI citation fiascos in the United States, Canada, Australia, and other UK proceedings before June 2025. The reason this ruling stands out is institutional level. It was not one trial judge dealing with one bad filing. It was the Divisional Court treating repeated fake-authority incidents as a problem for the administration of justice itself.
That shift matters because the harm from fake citations is not confined to the lawyer who gets caught. The opposing side spends time and money disproving made-up law. Judges and clerks waste time checking whether invented cases exist. Clients pay for advocacy that collapses on contact with a database search. Courts then have to explain, in published judgments, that no, the profession has not quietly agreed that fictional precedent is close enough if the paragraph sounds persuasive.
The Ayinde judgment is especially revealing on this point. One reason the lower court reacted so strongly was not just the presence of fake authorities but the response after they were flagged. Courts can forgive a mistake more easily than they can forgive a refusal to grasp why the mistake matters. Calling nonexistent authorities "minor citation errors" is a good way to convince a judge that the real defect is not clerical but professional.
The Al-Haroun side of the ruling adds another ugly lesson. AI citation failures do not stay contained when lawyers outsource judgment to clients, paralegals, or anyone else in the chain. The court was explicit that responsibility does not disappear because the fake material originated elsewhere. The lawyer who signs off on the filing owns the filing.
The profession's favorite fantasy
One fantasy behind generative AI in law is that it can compress the boring part of legal work without changing the trust model. Let the machine find cases, draft passages, summarize holdings, maybe even pull quotes. The lawyer can then "review" at the end. In reality, once the machine starts generating legal authorities, the review step becomes the work. If you have not checked every authority against an actual source, you have not finished the research. You have only produced text.
That is what makes fake-citation cases so revealing. They expose a workflow in which the appearance of legal work is mistaken for legal work itself. A nice-looking list of authorities, formatted properly and dropped into a neat PDF, is still junk if the cases are imaginary or irrelevant. Generative AI is exceptionally good at producing junk that arrives dressed for court.
The High Court's June 2025 message was not anti-technology. The judges explicitly said AI can be a useful tool. The judgment was anti-delusion. It rejected the idea that legal professionals can treat probabilistic text generation as something close enough to a citator or database if the prose sounds confident.
That should not have required a landmark ruling. But here we are. Two cases, dozens of false authorities, regulators notified, and a formal request for the profession to urgently get a grip. By that point the technology had not merely entered legal practice. It had already started teaching, in published judgments, which lawyers had mistaken autocomplete for research.
Discussion