The Volokh Conspiracy
Mostly law professors | Sometimes contrarian | Often libertarian | Always independent
Two Federal Judges Apologize For Issuing Opinions With AI Hallucinations
An intern and a law clerk used generative AI, and the judges didn't catch the hallucinations.
In July, I wrote about Judge Julien Xavier Neals of the U.S. District Court for the District of New Jersey, who withdrew an opinion that used generative AI. Judge Henry T. Wingate of the Southern District of Mississippi likewise withdrew an opinion that used generative AI. Both opinions included made-up citations, which were obvious hallucinations.
Senator Chuck Grassley, the Chairman of the Senate Judiciary Committee, wrote to both Neals and Wingate.
Both judges wrote to Judge Robert Conrad, the Director of the Administrative Office of the U.S. Courts.
Judge Neals explained that a law school intern used generative AI, in violation of chambers policy, as well as the student's law school's policy:
As referenced in the Senator's letter, a "temporary assistant," specifically, a law school intern, used CHATGPT to perform legal research in connection with the CorMedix decision. In doing so, the intern acted without authorization, without disclosure, and contrary to not only chambers policy but also the relevant law school policy. My chambers policy prohibits the use of GenAI1 in the legal research for, or drafting of, opinions or orders. . . .
I would be remiss if I did not point out as well that the law school where the intern is a student contacted me after the incident to, among other things, inform me that the student had violated the school's strict policy against the use of GenAI in their internships.
Judge Neals has his chambers in Newark. We can guess which law school the student attends.
Judge Wingate explains that his law clerk used generative AI. However, the draft was published prematurely before it was checked:
In the case of the Court's Order issued July 20, 2025, a law clerk utilized a generative artificial intelligence ("GenAI") tool known as Perplexity strictly as a foundational drafting assistant to synthesize publicly available information on the docket. . . .
The standard practice in my chambers is for every draft opinion to undergo several levels of review before becoming final and being docketed, including the use of cite checking tools.1 In this case, however, the opinion that was docketed on July 20, 2025, was an early draft that had not gone through the standard review process. It was a draft that should have never been docketed. This was a mistake. I have taken steps in my chambers to ensure this mistake will not happen again, as described below
Judge Conrad also sent a letter to Senator Grassley. The AO does not keep statistics on judges who have withdrawn opinions with hallucinations:
We are aware anecdotally of incidents in which judges have taken official action (such as those described above) relating to the integrity of court filings in which the use of AI tools was in question, although we currently do not systematically track such activity at the national level.
We learn that the AO convened a task force on generative AI.
The interim guidance cautions against delegating core judicial functions to AI, including decision-making or case adjudication, and it recommends that users exercise extreme caution especially if using AI to aid in addressing novel legal questions. It recommends that users review and independently verify all AI-generated content or output, and it reminds judges and Judiciary users and those who approve the use of AI that they are accountable for all work performed with the assistance of AI.
I suspect some district court judges will impose the requisite layers of review to detect hallucinations. Other district court judges, who delegate much of their work to law clerks, will not perform these checks.
Litigants should check any adverse decision for hallucinations. This simple step will be cheaper than filing an appeal.
Editor's Note: We invite comments and request that they be civil and on-topic. We do not moderate or assume any responsibility for comments, which are owned by the readers who post them. Comments do not represent the views of Reason.com or Reason Foundation. We reserve the right to delete any comment for any reason at any time. Comments may only be edited within 5 minutes of posting. Report abuses.
Please to post comments
But Mom ... all the kids are doing it!
Is there ancient pre-AI software that will take a text input and simply check that the case citations lead to cases with the right names, or longer versions of short names used in text?
Lexis-Nexis?
How long will it take before a judge fesses up that they did it?
I'm reminded of Spin City and an episode near the end of the show's run. Or it may have been the very last one. There was some colossal fuck up in the mayor's office and everybody knew the mayor wasn't going to be the one who took the fall, it would be some staffer. Everybody spent the episode wondering which one of them would be thrown under the bus, and in the end, Michael J. Fox, the mayor's Chief of Staff (I think?) stepped forward, lied to the public and said he did it. Everybody cheered his act of bravery and self-sacrifice, but looking back on it decades later that episode horrifies me.
I used to watch Spin City regularly.
1) Michael J. Fox's character was actually Deputy Mayor, which — despite its name — is essentially a chief of staff type position in NYC. (It's not elected office, and its role is determined solely by the mayor, as the position isn't defined in law.)
2) MJF left at the end of S4 because of his Parkinsons, and his character was written out of the show in the season finale for essentially the reasons you describe. (But the mayor really hadn't done anything wrong — IIRC, he had unknowingly done business with some guy in the mob — so it shouldn't have been that horrifying.) The show went on for a couple more seasons after that, with Charlie Sheen having replaced MJF. (I stopped watching after that.)
"How long will it take before a judge fesses up that they did it?"
Never. Here is a Federal Judge's response to a motion asking for clarification about how a hallucination-ridded order was issued.
I wonder if characterizing AI hallucinations as "clerical errors" rises to the level of a lie.
"Litigants should check any adverse decision for hallucinations. This simple step will be cheaper than filing an appeal."
Of course anyone considering filing an appeal should first review the order and determine its strengths and witnesses. I suspect the vast majority of lawyers are doing that already. But I disagree with the idea that checking for hallucinated citations is going to obviate the need to appeal.
I suspect in most cases, the bad citations won't be outcome-determinative. The controlling authorities are usually cited and discussed in the parties' memoranda. Hallucinated citations will probably end up as part of a law clerk's research on collateral issues that the parties didn't brief. In other words, removing the bad citations probably won't change the outcome. This is especially so because, being human, most judges are going to be loathe to admit the decision was premised on fake law. They'll probably err on the side of reaching the same result without the bad cases, where it is possible to do so.
So my guess is that in most cases the losing party will discover the bad citations and file a motion to reconsider. The judge will fix the opinion but reach the same result. The losing side will have to appeal anyway.
But I disagree with the idea that checking for hallucinated citations is going to obviate the need to appeal.
He did not say that, as shown in your very own quote of what he said.
Litigants should check any adverse decision for hallucinations. This simple step will be cheaper than filing an appeal.
When someone suggests doing a simple cheap action before a slow expensive one, they are not saying the one replaces the other.
Except they are.
If you do the cheap check and find a hallucinations, that eliminated the need to file an appeal as the decision has to be trashed.
You would only need to file an appeal if you couldn't find such an error.
You could, of course, file both - assuming you found a real flaw in the decision.
>being human, most judges are going to be loathe to admit the decision was premised on fake law.
This is just another way of saying judges decide outcomes and then look for justification of their decision.
It's troubling that judges think (and can) get away with the excuse that it was their underlings' fault. When an attorney gives that as a reason, he or she is told that is too bad because it is your job to review the work of your underlings and make sure that what you submit is quality work.
If this profession is supposed to be rock solid, then ALL players have to abide by these rules, judges included.
My thoughts exactly.
Throwing a young law student under the bus is very bush league.
if the profession is supposed to be rock solid, all players need to face consequences.
I don't want to jump to any conclusions or anything, but I've been starting to wonder if maybe the profession isn't rock-solid.
Ought versus Is.
If judges and lawyers are doing this *now*, think about how crappy their services have been all along - you just did not have it easily brought to your attention.
Kinda undercuts the argument that they need special licensing to protect people from unscrupulous lawyers and judges.
"Judge Neals has his chambers in Newark. We can guess which law school the student attends."
South Texas College of Law Houston?
Why not ask people to guess the law student's race? Josh, does it ever occur to you that maybe, as professor and just a human being, that it is not good to put down people, let alone a student, is such a demeaning way? If you feel that certain law schools should not be in business, just say so. Then we can complain what an elitist you are.
I was aghast at that unnecessary and unbecoming cheap shot.
Judge Wingate's solution, to require an additional check, is silly in light of the fact that he already had a system of checks to which they failed to adhere. They don't need more checks; they need more compliance.
Where's that log of all docketed items, with the initialed check-offs of requisite reviewers? Who is responsible for assuring the integrity of that log?
All of those checks failed so clearly we needed another check and maybe this new one will work. That's the government solution anyways.
That's exactly what they do with gun control. "None of these other 20,000 plus gun control laws we've [assed have worked. We need more gun control laws!"
I read both of the judge's letters and neither of them contains what I would call an "apology".
And more to AI:
A bag of Doritos is reported as a gun. Of course with fat content ...
https://www.wbrc.com/2025/10/24/police-swarmed-student-after-ai-system-mistook-bag-chips-gun-officials-say/