Senior lawyers should stop using generative AI to prepare their legal arguments!
Or should they?
A High Court judge in the UK has told senior lawyers off for their use of ChatGPT, because it invents citations to cases and laws that don’t exist!
In case in the High Court, Dame Victoria Sharp referred in her judgment to:
actual or suspected use by lawyers of generative artificial intelligence tools to produce written legal arguments or witness statements which are not then checked, so that false information (typically a fake citation or quotation) is put before the court.
— Frederick Ayinde, R (on the application of) v The London Borough of Haringey (2025) EWHC 1040 (Admin)[2]
This is fascinating because we all know about AI hallucinations. Generative AI is notorious for making things up on the spot where data is sparse. But what it’s good at, is reasoning on large amounts of text - provided that enough information is given to it.
In the case R (Ayinde) v Haringey, because of the wasted costs associated with the fake citations, the Court reduced the claimant’s costs from £20,000 to £6,500. Although the lawyer in question denied using AI, the judge ordered the transcript of the case to be sent to the Bar Standards Board and to the Solicitors Regulation Authority. So there is a tangible financial cost to the over-use of generative AI by lawyers for legal work, not to mention the accompanying disruption to a publicly funded and overstretched legal system.
Eugenio Vaccari, Marton Ribary, Miklos Orban, Paul Krause, and I all investigated this when we experimented on using generative AI to answer questions about insolvency in English and Welsh law. We found that off-the-shelf large language models made up cases and statutes which sounded genuine, until we found that they didn’t exist!
Fortunately, a combination of a large language model and a database of law helps us mitigate this. Eugenio Vaccari collected a dataset of cases and statutes in English insolvency law and combined a lookup (finding the right statute due to keywords and text similarity) with a large language model, to get the best of both worlds. We feed the citations (cases and statutes) to GPT as part of the prompt, so we get a response that we know uses real citations.
The response was significantly better than GPT without the extra citations in the prompt. This is called a Retrieval Augmented Generation (“RAG”) system. The size of the performance boost that we found using RAG increased when we used more advanced LLMs
So there is a way to use generative AI and avoid the problem of fake citations - but lawyers shouldn’t be using ChatGPT as-is, without a RAG system or an alternative to keep it on track.
Link to our Insolvency Bot: https://fastdatascience.com/insolvency
Insolvency Bot
We know that US content dominates the training sets of LLMs like ChatGPT. LLMs output text in almost exclusively American spelling and vocabulary. I noticed that UK legal questions often result in US-centric responses. However, the problem of generative AI making fake citations isn’t just confined to the UK. Dame Sharp also mentions the case Mata v Avianca Inc. from the US, where
Two lawyers… used ChatGPT – a large language model AI – to identify relevant caselaw. One prompted the tool to draft a court submission, which they submitted verbatim on behalf of their client. However, unbeknownst to them, the AI-generated legal analysis was faulty and contained fictional citations… …the AI output was entirely fabricated, falsely attributing nonsensical opinions to real judges and embellished with further false citations and docket numbers held by actual cases irrelevant to the matter at hand….
Dame Sharp reproduced a passage of one of the counsel’s submissions to the Court:
Moreover, in R (on the application of Ibrahim) v Waltham Forest LBC [2019] EWHC 1873 (Admin), the court quashed a local authority decision due to its failure to properly consider the applicant’s medical needs, underscoring the necessity for careful evaluation of such evidence in homelessness determinations. The Respondent’s failure to consider the Appellant’s medical conditions in their entirety, despite being presented with comprehensive medical documentation, renders their decision procedurally improper and irrational.
…
The Appellant’s situation mirrors the facts in R (on the application of H) v Ealing LBC [2021] EWHC 939 (Admin), where the court found the local authority’s failure to provide interim accommodation irrational in light of the applicant’s Approved Judgment Ayinde v Haringey Al-Haroun v Qatar vulnerability and the potential consequences of homelessness. The Respondent’s conduct in this case similarly lacks rational basis and demonstrates a failure to properly exercise its discretion.
…
The Respondent’s failure to provide a timely response and its refusal to offer interim accommodation have denied the Appellant a fair opportunity to secure his rights under the homelessness legislation. This breach is further highlighted in R (on the application of KN) v Barnet LBC [2020] EWHC 1066 (Admin), where the court held that procedural fairness includes timely decision-making and the provision of necessary accommodations during the review process. The Respondent’s failure to adhere to these principles constitutes a breach of the duty to act fairly.
The Appellant’s case further aligns with the principles set out in R (on the application of Balogun) v LB Lambeth [2020] EWCA Civ 1442, where the Court of Appeal emphasized that local authorities must ensure fair treatment of applicants in the homelessness review process. The Respondent’s conduct in failing to provide interim accommodation or a timely decision breaches this standard of fairness.
The four cases listed above which I have highlighted do not exist. But the detail is that they look completely genuine - I had no way of knowing that these are not real, until I searched for them online.
We are now working on expanding the Insolvency Bot to cover more UK case law and several other European jurisdictions, in an ambitious international project.
You can read our original paper here:
And you can try the Insolvency Bot here: https://fastdatascience.com/insolvency
Unleash the potential of your NLP projects with the right talent. Post your job with us and attract candidates who are as passionate about natural language processing.
Hire NLP ExpertsListen to the new episode of the Clinical Trial Files podcast, where Karin Avila, Taymeyah Al-Toubah and Thomas Wood of Fast Data Science chat about AI and NLP in pharma, the Clinical Trial Risk Tool, what impact AI can make in clinical trials. This episode commemorates Alan Turing’s 113rd birthday on 23 June 2025.
Fast Data Science at will be presenting at the 4th Annual Conference on the Intersection of Corporate Law and Technology at Nottingham Trent University Join Thomas Wood of Fast Data Science, Marton Ribary and Eugenio Vaccari for their presentation “A Generative AI-Based Legal Advice Tool for Small Businesses in Distress” at the 4th Annual Conference on the Intersection of Corporate Law and Technology at Nottingham Trent University
What we can do for you