The Volokh Conspiracy
Mostly law professors | Sometimes contrarian | Often libertarian | Always independent
Two Federal Judges Apologize For Issuing Opinions With AI Hallucinations
An intern and a law clerk used generative AI, and the judges didn't catch the hallucinations.
In July, I wrote about Judge Julien Xavier Neals of the U.S. District Court for the District of New Jersey, who withdrew an opinion that used generative AI. Judge Henry T. Wingate of the Southern District of Mississippi likewise withdrew an opinion that used generative AI. Both opinions included made-up citations, which were obvious hallucinations.
Senator Chuck Grassley, the Chairman of the Senate Judiciary Committee, wrote to both Neals and Wingate.
Both judges wrote to Judge Robert Conrad, the Director of the Administrative Office of the U.S. Courts.
Judge Neals explained that a law school intern used generative AI, in violation of chambers policy, as well as the student's law school's policy:
As referenced in the Senator's letter, a "temporary assistant," specifically, a law school intern, used CHATGPT to perform legal research in connection with the CorMedix decision. In doing so, the intern acted without authorization, without disclosure, and contrary to not only chambers policy but also the relevant law school policy. My chambers policy prohibits the use of GenAI1 in the legal research for, or drafting of, opinions or orders. . . .
I would be remiss if I did not point out as well that the law school where the intern is a student contacted me after the incident to, among other things, inform me that the student had violated the school's strict policy against the use of GenAI in their internships.
Judge Neals has his chambers in Newark. We can guess which law school the student attends.
Judge Wingate explains that his law clerk used generative AI. However, the draft was published prematurely before it was checked:
In the case of the Court's Order issued July 20, 2025, a law clerk utilized a generative artificial intelligence ("GenAI") tool known as Perplexity strictly as a foundational drafting assistant to synthesize publicly available information on the docket. . . .
The standard practice in my chambers is for every draft opinion to undergo several levels of review before becoming final and being docketed, including the use of cite checking tools.1 In this case, however, the opinion that was docketed on July 20, 2025, was an early draft that had not gone through the standard review process. It was a draft that should have never been docketed. This was a mistake. I have taken steps in my chambers to ensure this mistake will not happen again, as described below
Judge Conrad also sent a letter to Senator Grassley. The AO does not keep statistics on judges who have withdrawn opinions with hallucinations:
We are aware anecdotally of incidents in which judges have taken official action (such as those described above) relating to the integrity of court filings in which the use of AI tools was in question, although we currently do not systematically track such activity at the national level.
We learn that the AO convened a task force on generative AI.
The interim guidance cautions against delegating core judicial functions to AI, including decision-making or case adjudication, and it recommends that users exercise extreme caution especially if using AI to aid in addressing novel legal questions. It recommends that users review and independently verify all AI-generated content or output, and it reminds judges and Judiciary users and those who approve the use of AI that they are accountable for all work performed with the assistance of AI.
I suspect some district court judges will impose the requisite layers of review to detect hallucinations. Other district court judges, who delegate much of their work to law clerks, will not perform these checks.
Litigants should check any adverse decision for hallucinations. This simple step will be cheaper than filing an appeal.
Editor's Note: We invite comments and request that they be civil and on-topic. We do not moderate or assume any responsibility for comments, which are owned by the readers who post them. Comments do not represent the views of Reason.com or Reason Foundation. We reserve the right to delete any comment for any reason at any time. Comments may only be edited within 5 minutes of posting. Report abuses.
Please to post comments
But Mom ... all the kids are doing it!
Is there ancient pre-AI software that will take a text input and simply check that the case citations lead to cases with the right names, or longer versions of short names used in text?
Lexis-Nexis?
How long will it take before a judge fesses up that they did it?
I'm reminded of Spin City and an episode near the end of the show's run. Or it may have been the very last one. There was some colossal fuck up in the mayor's office and everybody knew the mayor wasn't going to be the one who took the fall, it would be some staffer. Everybody spent the episode wondering which one of them would be thrown under the bus, and in the end, Michael J. Fox, the mayor's Chief of Staff (I think?) stepped forward, lied to the public and said he did it. Everybody cheered his act of bravery and self-sacrifice, but looking back on it decades later that episode horrifies me.
I used to watch Spin City regularly.
1) Michael J. Fox's character was actually Deputy Mayor, which — despite its name — is essentially a chief of staff type position in NYC. (It's not elected office, and its role is determined solely by the mayor, as the position isn't defined in law.)
2) MJF left at the end of S4 because of his Parkinsons, and his character was written out of the show in the season finale for essentially the reasons you describe. (But the mayor really hadn't done anything wrong — IIRC, he had unknowingly done business with some guy in the mob — so it shouldn't have been that horrifying.) The show went on for a couple more seasons after that, with Charlie Sheen having replaced MJF. (I stopped watching after that.)
"How long will it take before a judge fesses up that they did it?"
Never. Here is a Federal Judge's response to a motion asking for clarification about how a hallucination-ridded order was issued.
I wonder if characterizing AI hallucinations as "clerical errors" rises to the level of a lie.
"Litigants should check any adverse decision for hallucinations. This simple step will be cheaper than filing an appeal."
Of course anyone considering filing an appeal should first review the order and determine its strengths and witnesses. I suspect the vast majority of lawyers are doing that already. But I disagree with the idea that checking for hallucinated citations is going to obviate the need to appeal.
I suspect in most cases, the bad citations won't be outcome-determinative. The controlling authorities are usually cited and discussed in the parties' memoranda. Hallucinated citations will probably end up as part of a law clerk's research on collateral issues that the parties didn't brief. In other words, removing the bad citations probably won't change the outcome. This is especially so because, being human, most judges are going to be loathe to admit the decision was premised on fake law. They'll probably err on the side of reaching the same result without the bad cases, where it is possible to do so.
So my guess is that in most cases the losing party will discover the bad citations and file a motion to reconsider. The judge will fix the opinion but reach the same result. The losing side will have to appeal anyway.
But I disagree with the idea that checking for hallucinated citations is going to obviate the need to appeal.
He did not say that, as shown in your very own quote of what he said.
Litigants should check any adverse decision for hallucinations. This simple step will be cheaper than filing an appeal.
When someone suggests doing a simple cheap action before a slow expensive one, they are not saying the one replaces the other.
Except they are.
If you do the cheap check and find a hallucinations, that eliminated the need to file an appeal as the decision has to be trashed.
You would only need to file an appeal if you couldn't find such an error.
You could, of course, file both - assuming you found a real flaw in the decision.
>being human, most judges are going to be loathe to admit the decision was premised on fake law.
This is just another way of saying judges decide outcomes and then look for justification of their decision.
It's troubling that judges think (and can) get away with the excuse that it was their underlings' fault. When an attorney gives that as a reason, he or she is told that is too bad because it is your job to review the work of your underlings and make sure that what you submit is quality work.
If this profession is supposed to be rock solid, then ALL players have to abide by these rules, judges included.
My thoughts exactly.
Throwing a young law student under the bus is very bush league.
if the profession is supposed to be rock solid, all players need to face consequences.
I don't want to jump to any conclusions or anything, but I've been starting to wonder if maybe the profession isn't rock-solid.
Ought versus Is.
If judges and lawyers are doing this *now*, think about how crappy their services have been all along - you just did not have it easily brought to your attention.
Kinda undercuts the argument that they need special licensing to protect people from unscrupulous lawyers and judges.
Incunabulum, there's no reason to say "if." Judges and lawyers clearly are presenting falsehoods as truth in their purported analyses of precedent or persuasive legal authorities. And you're correct that we need to "think about how crappy their services have been all along." In some areas of law, falsehoods by judges (and lawyers copying judges' falsehoods) even are commonplace and blatant.
One of the reasons to appreciate the writing of Justice Scalia in SCOTUS opinions was that sometimes he and SCOTUS were willing to call out judges clearly for their violations or misrepresentations of law. The following are examples.
Sometimes, judges (and lawyers) misrepresent the law. Justice Scalia (writing for SCOTUS) had this to say about that in Brogan v. United States, 522 U.S. 398 (1998).
“While communis error facit jus may be a sadly accurate description of reality, it is not the normative basis of" American "jurisprudence. Courts may not create their own limitations on legislation [or constitutions other controlling legal authority], no matter how alluring the policy arguments for doing so, and no matter how widely the blame may be spread” among any quantity or quality of courts. The point of the expression "communis error facit jus" is that many lawyers and judges have violated (and stated falsely) the controlling legal authority.
Sometimes, judges (and lawyers) accurately present controlling legal authority--and then they knowingly violate it. Justice Scalia (writing for SCOTUS) had this to say about that in Allentown Mack Sales and Service, Inc. v. N.L.R.B., 522 U.S. 359 (1998).
"[A]djudication is subject to the requirement of reasoned decisionmaking [and it] is hard to imagine a more violent breach of that requirement than applying a rule" that "is in fact different from the rule or standard formally announced. And the consistent repetition of that breach can hardly mend it."
"Reasoned decisionmaking, in which the rule announced is the rule applied, promotes sound results, and unreasoned decisionmaking the opposite. The evil of a decision that applies a standard other than the one it enunciates spreads in both directions, preventing both consistent application of the law by subordinate" government employees "and effective review of the law by [higher] courts."
"Because reasoned decisionmaking demands it, and because the systemic consequences of any other approach are unacceptable, [every government authority] must be required to apply in fact the clearly understood legal standards that it enunciates in principle [ ]." "Even the most consistent and hence predictable [ ] departure from proper application of [the governing] standards will not alter the legal rule by which the [ ] factfinding is to be judged."
"Judge Neals has his chambers in Newark. We can guess which law school the student attends."
South Texas College of Law Houston?
Why not ask people to guess the law student's race? Josh, does it ever occur to you that maybe, as professor and just a human being, that it is not good to put down people, let alone a student, is such a demeaning way? If you feel that certain law schools should not be in business, just say so. Then we can complain what an elitist you are.
I was aghast at that unnecessary and unbecoming cheap shot.
All of the above and sadly not the first instance of Prof. Blackman being intellectually lazy on this blog in my view. There are two law schools in Newark. His comment is ambiguous.
Smarmy cheap shot, yet not even clear, maybe even to Josh, whom he's trying to insult. So Josh.
Judge Wingate's solution, to require an additional check, is silly in light of the fact that he already had a system of checks to which they failed to adhere. They don't need more checks; they need more compliance.
Where's that log of all docketed items, with the initialed check-offs of requisite reviewers? Who is responsible for assuring the integrity of that log?
All of those checks failed so clearly we needed another check and maybe this new one will work. That's the government solution anyways.
That's exactly what they do with gun control. "None of these other 20,000 plus gun control laws we've [assed have worked. We need more gun control laws!"
I read both of the judge's letters and neither of them contains what I would call an "apology".
And more to AI:
A bag of Doritos is reported as a gun. Of course with fat content ...
https://www.wbrc.com/2025/10/24/police-swarmed-student-after-ai-system-mistook-bag-chips-gun-officials-say/
The interim guidance cautions against delegating core judicial functions to AI, including decision-making or case adjudication, and it recommends that users exercise extreme caution especially if using AI to aid in addressing novel legal questions.
That caution about novel legal questions discloses deep misunderstanding about the causes of AI hallucinations.
As an aging non-lawyer, I don't understand why AI comes up with fake citations at all. Surely double-checking citations to see that they exist is a task perfectly suited to automation?
Both Lexis and Westlaw -- the two big legal research services -- have programs that do just that. These programs can identify all of the citations and quotations in a document and then run those against their database. The program reports the result allowing the lawyer can make corrections as necessary. These sorts of programs will also identify whether a given case has been overruled or questioned by another court.
This is something different. The major "open" AI's are trained on data usually culled from the Internet. So, there is no designated database of cases acknowledged to be real. Moreover, these sorts of AI's are intended to create content; not perform research. So, they will create content whether it exists or not.
Wholly made-up cases are a phenomenon of the new AI models. A more typical issue is that the case exists but doesn't stand for the proposition for which it is cited. The extent to which a case doesn't stand for the proposition for which it is cited can range from legitimate dispute to sloppy research to outright fabrication.
That makes sense. Thank you!