The Volokh Conspiracy
Mostly law professors | Sometimes contrarian | Often libertarian | Always independent
Lawyer Explains How He Used ChatGPT to Produce Filing "Replete with Citations to Non-Existent Cases"
"Can you show me the courts opinion in Varghese v China Southern Airlines"? "Certainly! ... I hope that helps!"
A month ago, I wrote about Judge Kevin Castel (S.D.N.Y.)'s May 4 order in Mata v. Avianca, Inc.:
The Court is presented with an unprecedented circumstance. A submission filed by plaintiff's counsel in opposition to a motion to dismiss is replete with citations to non-existent cases. When the circumstance was called to the Court's attention by opposing counsel, the Court issued Orders requiring plaintiff's counsel to provide an affidavit annexing copies of certain judicial opinions of courts of record cited in his submission, and he has complied. Six of the submitted cases appear to be bogus judicial decisions with bogus quotes and bogus internal citations. Set forth below is an Order to show cause why plaintiff's counsel ought not be sanctioned.
The Court begins with a more complete description of what is meant by a nonexistent or bogus opinion. In support of his position that there was tolling of the statute of limitation under the Montreal Convention by reason of a bankruptcy stay, the plaintiff's submission leads off with a decision of the United States Court of Appeals for the Eleventh Circuit, Varghese v China South Airlines Ltd, 925 F.3d 1339 (11th Cir. 2019). Plaintiff's counsel, in response to the Court's Order, filed a copy of the decision, or at least an excerpt therefrom.
The Clerk of the United States Court of Appeals for the Eleventh Circuit, in response to this Court's inquiry, has confirmed that there has been no such case before the Eleventh Circuit with a party named Vargese or Varghese at any time since 2010, i.e., the commencement of that Court's present ECF system. He further states that the docket number appearing on the "opinion" furnished by plaintiff's counsel, Docket No. 18-13694, is for a case captioned George Cornea v. U.S. Attorney General, et al. Neither Westlaw nor Lexis has the case, and the case found at 925 F.3d 1339 is A.D. v Azar, 925 F.3d 1291 (D.C. Cir 2019). The bogus "Varghese" decision contains internal citations and quotes, which, in turn, are non-existent: …
The following five decisions submitted by plaintiff's counsel contain similar deficiencies and appear to be fake as well ….
The court therefore ordered plaintiff's counsel to show cause why he shouldn't be sanctioned, and plaintiff's counsel responded that he was relying on the work of another lawyer at his firm, and this second lawyer (who had 30 years of practice experience) he was relying on ChatGPT. The court ordered a further round of explanations, and here's the heart of the filing yesterday from the second lawyer (paragraph numbering removed):
The Opposition to the Motion to Dismiss
After this case was removed, Defendant filed a motion to dismiss, arguing, among other things, that this case was subject to a two year statute of limitations under the Montreal Convention 1999 (the "Montreal Convention"), and that Mr. Mata had somehow filed the Complaint too late. As the partner who had been lead counsel on the case since its inception almost three years before, I took responsibility for preparing a response to the motion.
Our position on the motion would be that the claims were timely filed because either the Montreal Convention's shortened statute of limitations was inapplicable or, in the alternative, that any period of limitations was tolled by Avianca's bankruptcy.
As discussed in our memorandum of law as well as the declaration of Thomas Corvino, the Firm practices primarily in New York state courts as well as New York administrative tribunals. The Firm's primary tool for legal research is a program called Fastcase, which is an online legal research program made available to all lawyers at the Firm.
Based on my experience using Fastcase, I understood that the Firm had access to the New York State database as well as to at least some federal cases.
I first attempted to use Fastcase to perform legal research in this case, however it became apparent that I was not able to search the federal database.
My Use of ChatGPT
In an effort to find other relevant cases for our opposition, I decided to try and use ChatGPT to assist with legal research.
l had never used ChatGPT for any professional purpose before this case. I was familiar with the program from my college-aged children as well as the articles 1 had read about the potential benefits of artificial intelligence (AI) technology for the legal and business sectors.
At the time I used ChatGPT for this case, I understood that it worked essentially like a highly sophisticated search engine where users could enter search queries and ChatGPT would provide answers in natural language based on publicly available information.
I realize now that my understanding of how ChatGPT worked was wrong. Had I understood what ChatGPT is or how it actually worked, I would have never used it to perform legal research.
Based on my erroneous understanding of how ChatGPT worked, I used the program to try and find additional caselaw support for our arguments.
I conducted the search in the same general manner I searched any legal research database. I first asked ChatGPT a broader question about tolling under the Montreal Convention and was provided with a general answer that appeared consistent with my pre-existing understanding of the law. 1 then began to ask ChatGPT to find cases in support of more specific principles applicable to this case, including the tolling of the Montreal Convention's statute of limitations as a result of Defendant's bankruptcy. Each time, ChatGPT provided me with an affirmative answer including case citations that appeared to be genuine. Attached as Exhibit A are copies of my chat history with ChatGPT when I was performing research for our opposition to the motion to dismiss.
In connection with my research, I also asked ChatGPT to provide the actual cases it was citing, not just the summaries. Each time, ChatGPT provided me with what it described as a "brief excerpt" that was complete with a case caption, legal analysis, and other internal citations. I recognize in hindsight that I should have been more skeptical when ChatGPT did not provide the full case I requested. Nevertheless, given the other features of the "cases" it was providing, I attributed this to the program's effort to highlight the relevant language based on my search queries.
As noted above, when I was entering these search queries, I was under the erroneous impression that ChatGPT was a type of search engine, not a piece of technology designed to converse with its users. I simply had no idea that ChatGPT was capable of fabricating entire case citations or judicial opinions, especially in a manner that appeared authentic.
I therefore cited several of the cases that ChatGPT provided in our opposition to the motion to dismiss.
The April 25 Affidavit
In response to Defendant's reply brief—which stated that defense counsel could not locate several of the cases cited in our opposition papers—the Court ordered [the lead lawyer] to file an affidavit annexing nine of the cases cited in our opposition….
I was unable to find one of the cases (which was cited within another case that I had found on ChatGPT). Of the remaining eight cases, I obtained two of those cases from Fastcase. The remaining six, I obtained from ChatGPT (the "ChatGPT Cases").
Similar to how I used ChatGPT when I was preparing the opposition papers, I asked ChatGPT to provide copies of the six ChatGPT Cases. ChatGPT provided me with what appeared to be partial versions of the six cases. Because I did not have another research database with access to the federal reporters available to me, I did not take these citations and obtain full copies of the cases from another source. (I realize now that I could have gone to a bar association library or colleague that had access to Westlaw and Lexis, but it did not occur to me at the time.) However, when I was responding, I still did not believe it was possible that the cases ChatGPT was providing were completely fabricated. I therefore attached the ChatGPT Cases to the April 25 Affidavit….
The First OSC and My Realization That the ChatGPT Cases Were Not Authentic
In response to the April 25 Affidavit as well as a letter from defense counsel representing that it still could not find the ChatGPT Cases, this Court issued an Order to Show Cause on May 4, 2023 …. The First OSC stated that the Court could not find any record of six of the cases that we attached to the April 25 Affidavit (i.e. the ChatGPT Cases).
When I read the First OSC, I realized that I must have made a serious error and that there must be a major flaw with the search aspects of the ChatGPT program. I have since come to realize, following the Order, that the program should not be used for legal research and that it did not operate as a search engine at all, which was my original understanding of how it worked.
Before the First OSC, however, I still could not fathom that ChatGPT could produce multiple fictitious cases, all of which had various indicia of reliability such as case captions, the names of the judges from the correct locations, and detailed fact patterns and legal analysis that sounded authentic. The First OSC caused me to have doubts. As a result, I asked ChatGPT directly whether one of the cases it cited, "Varghese v. China Southern Airlines Co., Ltd., 925 F.3d 1339 (11th Cir. 2009)," was a real case. Based on what I was beginning to realize about ChatGPT, I highly suspected that it was not. However, ChatGPT again responded that Varghese "does indeed exist" and even told me that it was available on Westlaw and LexisNexis, contrary to what the Court and defendant's counsel were saying. This confirmed my suspicion that ChatGPT was not providing accurate information and was instead simply responding to language prompts without regard for the truth of the answers it was providing. However, by this time the cases had already been cited in our opposition papers and provided to the Court.
In an effort to be fully transparent, I provided an affidavit … to submit [to the court]. In my affidavit, I made clear that I was solely responsible for the research and drafting of the opposition and that I used ChatGPT for some of the legal research. I also apologized to the Court and reiterated that it was never my intention to mislead the Court.
The Consequences of this Matter
As detailed above, I deeply regret my decision to use ChatGPT for legal research, and it is certainly not something I will ever do again.
I recognize that if I was having trouble finding cases on the Firm's existing research platform, I should have either asked the Firm to obtain a more comprehensive subscription or used the research resources such as Westlaw and LexisNexis maintained by the law libraries at the local bar associations….
You can see the full ChatGPT transcript here. (Note that this post doesn't discuss the less interesting details related to an error in a date in a notarized document; you can see all the filings here.)
Editor's Note: We invite comments and request that they be civil and on-topic. We do not moderate or assume any responsibility for comments, which are owned by the readers who post them. Comments do not represent the views of Reason.com or Reason Foundation. We reserve the right to delete any comment for any reason at any time. Comments may only be edited within 5 minutes of posting. Report abuses.
Please
to post comments
ChatGPT, AI, etc. are a long way from revolutionizing the practice of law, as some tech bros are claiming. They don't appear to be useful even for the preparation of simple contracts or filling in forms. Kathryn Tewson's exploits with DoNotPay on Twitter have been fun to follow.
I think you could safely use them as a step in legal research, so long as you had an invariable rule of always looking up the cites to make sure they were real and accurate.
Might uncover some stuff you wouldn't easily find, after all.
It's kind of like starting with Wikipedia or a general google search on a legal principle. Not the worst way to start on a project, but definitely a terrible way to finish - you need to follow the citations. Not too different than starting from a secondary source like a treatise or compilation and working back to the primary material. Just more readily accessible (and somewhat more likely to be up-to-date) than going to a library for Bob and Betsy's "Practice Guide for Bankruptcy Tort Contract Disputes, Selection of Law" (Thirty-Fourth Edition).
Generally agree, but treatises I think are better than Google, at least the goods ones. And most of them are now online and updated regularly. Certainly in my area, IP law, the leading treatises are pretty good and sometimes cited by courts as authorities themselves on unclear issues. Ditto Wright and Miller and Moore's on Federal Practice.
I would be surprised to find any strong firm that approved this.
I could imagine areas of legal work where ChatGPT could potentially be incorporated into the process productively. Research is definitely not one of them.
Completely agree. It has no place in research given its current capabilities. It’s primary use for legal work now would be generating first drafts of contracts or demand letters. MAYBE creating an outline for an argument that the lawyer then refines and fills in with the appropriate citations.
But I would never use it as a research step.
I suspect the rate of AI encroachment on the practice of law will surprise many in the profession.
Use cases which interest me more than research per se include various analyses of rulings & statutes (e.g. defensibly-objective ‘motivated reasoning’ scores for federal judges, identification of inconsistencies & redundancies in statutes, readability & gratuitous-complexity scores for statutes, statutory cleanup proposals).
Put some avocado on him, I think he's toast.
Sorry, you don't use an unfamiliar tool that you guess you know how it works from your kids without cite-checking any of it. This attorney deserves to be held in contempt for failing his basic duties to the court and his client.
Yeah if I were the lawyer I would explain myself but also say something along the lines of "I do deserve to be sanctioned for this mistake." Maybe this one did, I don't know.
No, both of the lawyers involved are asking not to be sanctioned.
At this point, his best defense is the stupid defense. "I'm too stupid to be sanctioned."
It once worked in a case of mine. We moved to sanction the lawyers on the other side for making bad faith arguments. The judge denied it, stating there is an alternative possible explanation, namely that the other side were merely engaged in "poor lawyering."
That's essentially what their response to the OSC says.
Agreed. I think his explanations are useful if there's any kind of malice argument, like the client's case is dismissed if they acted with malice. I'd be inclined to go with criminal stupidity and give them a pass on that. But that's a different question from whether the lawyer himself should be sanctioned for his behavior, and I think that's unequivocally yes.
You don't think it's reasonable to read "ChatGPT" and think "search engine" rather than "chat program"?
(I sure don't, and hope the judge thinks hard about whether that claimed misapprehension passes the laugh test.)
I do think that the lawyer genuinely didn't realize that ChatGPT wasn't showing him real cases.
(To be clear, I agree with your implied point that this misapprehension should disqualify him from further practice of law.)
Because I did not have another research database with access to the federal reporters available to me, I did not take these citations and obtain full copies of the cases from another source. (I realize now that I could have gone to a bar association library or colleague that had access to Westlaw and Lexis, but it did not occur to me at the time.)
Must say, this statement left me gobsmacked. How do you practice in federal court without having access to federal case databases?
My thought too. How can you practice at all without searching federal cases? They’re relevant even if you only practice in state court.
He's a local PI lawyer. A NY PI practice does not really involve a whole lot of legal research or legal writing.
Dude, he sued an airline company, Avianca, a Colombia company, for an injury to his client on a plane ride from El Salvador to New York. The motion practice implicated an international convention on aviation liability (the Montreal Convention, which the judge keeps referring to) and the Bankruptcy Code, a federal statute. This is not a routine PI case.
Charitably, he was in over his head.
Absolutely. He thought it was a routine PI case. He filed in state court as an ordinary PI complaint; it reads just like every ordinary slip-and-fall level negligence suit that gets filed in NY courts every day. He almost certainly had no idea that it was even removable; I doubt he'd ever heard of the Montreal Convention. So that was his first surprise. And then he was hit with a motion to dismiss based on both the fact of a bankruptcy filing and the statute of limitations contained in the Montreal Convention. And he didn't have the first clue what to do with that — because you don't get motions to dismiss in ordinary PI cases, and you certainly don't get ones based on international treaties — which is where his trouble began.
The only thing I'm surprised about is that PI firms like this usually just farm out this sort of legal briefing to contract attorneys, rather than trying to wing it themselves.
I noticed his blaming the firm for not providing him LEXIS. Well, who *is* the firm if he is the senior attorney?
I think that the dynamics of that firm would be very interesting to look at from a management perspective.
Makes me think 30 years of practicing has been too much, he needs to start getting serious.
I think this is likely more common than you realize. At my previous firm, I practiced in federal court. But my Westlaw subscription (through my firm) was limited to federal cases within the Tenth Circuit and the Supreme Court. If I wanted to do a broader search for national caselaw, I had to go to other sources (usually Google/Casetext). But it is plausible that one could have easy access to local federal cases but not those from other jurisdictions.
Either way, this guy obviously goofed big and should have found a way to check his sources without the use of ChatGPT.
As we discussed last week, he doesn't practice in federal court. He's not even admitted there.
So he didn't realize that his firm's Fastcase subscription was somehow screwed up so that he didn't have access to federal cases. (Why he didn't just address that, I'm not sure.)
He appeared in a case that was removed to federal court. Not a rare occurrence even for state-court practitioners. So, yes, he does practice in federal court, at least sometimes. How you don't prepare for that blows my mind. (Not really, I have seen so much bad lawyering out there.)
The guy is likely it's Judge Castel. If it were Judge Rakoff, he would have had his head bitten off already.
lucky, not likely.
While there are worse draws than Castel, I would not call that a lucky one. Lucky would've been someone like Oetken.
He's been in practice for 30 years and has never even bothered to get admitted to the SDNY. What does that tell you about where he practices?
https://www.youtube.com/watch?v=1IJC1DQH_sc
Danny Crane....
"Because I did not have another research database with access to the federal reporters available to me"
Google scholar? Free research database that goes back to 1950 on federal non-SCOTUS cases.
The bottom line here is that AI is fabricating citations -- not quoting out of context or misquoting but fabricating out of whole cloth.
The Clerk for the 11th Circuit pointed out the extent of the problem -- not only was the party's name not in the database, but the case number belonged to a second unrelated case while the cite belonged to a third. This isn't even close to something that actually exists.
Forget the idiot lawyer not bright enough to ask his adult children how to use Google, this is Orwellian and the scary part is when it gets good enough not to be this noticeable. There is a fundamental flaw here.
It's fabricating ALL the citations, even the ones that turn out to be real. That's what a lot of people don't understand about these large language models. It's ALL "hallucinations", they've just got it tuned well enough that it mostly "hallucinates" stuff that's real.
But not because it's real. Because there was enough real stuff in the training database that the statistics will mostly generate output that's very likely to be real.
But the model has no concept of "true" and "false", real or fake. It just generates strings of words it calculates would give it a good score from the people who'd been rating its answers.
Thanks,
I thought the fabrication was a flaw, not a feature.
It is even more Orwellian than I thought.
Who do you think Orwell is?
In what possible way could any aspect of "this" be described as Orwellian?
What happens when doctors use it, thinking it’s just a search engine?
A lawyer who uses tools he does not understand should be disbarred.
Just for the record, the letter "A" in AI stands for ARTIFICIAL!
As in 'not real'.
Actually, artificial means "made or produced by human beings rather than occurring naturally."
And "natural" applies to beaver dams, but not similar stick and mud dams built by humans?
What is so unnatural about human-made things?
Can't be the use of tools; animals make tools out of sticks to change their environment.
Yes.
That's a dam good response, damning them with their own words.
You seem very eager about that comment. An eager, umm, water rodent?
Dam excellent!
The explanation doesn't wash.
Even if he thought Chat GPT was like a search engine, you can't just use a search engine like google and rely uncritically on what you find.
If you're using a generic search engine to do any type of research, you need to determine that the results you are getting are from reliable sources, provided in the correct context, etc.
I constantly worry that I will miss an ellipsis or cite the wrong page number or, God forbid, move something around in the editing process and end up with the wrong cite for something. I can't imagine the thought process that leads to filing something in federal court without making sure the case exists -- much less doing so in response to a judge's order saying "I don't think this case exists, can you give me a copy?"
I can confirm that FastCase sucks.
That being said, failure to pull and read full cases before citing them (and failure to Shepherdize them to confirm they're still good law) is basic incompetence regardless of your understanding of how ChatGPT works.
The fact that this lawyer was fooled by what he calls ChatGPT's "indicia of reliability" seems legally significant. Much of the recent erosion of the rules of evidence has been based on creating vague new "indicia of reliability exceptions to the rules, so evidence that would otherwise have been excluded under the old, common law-based rules are let in because they have so-called "indicia of reliability." There are always good-sounding reasons to fiddle with old rules, but these changes have unfortunately allowed a lot of bullsh_t* in the fuse of evidence to flood the courts, for example in the area of "expert opinion" testimony. When bullsh_t* is said by an "expert," it suddenly becomes evidence with so-called indicia of reliability, and courts then rely on it to make dispositive rulings. Now we have ChatGPT, which is an example of "indicia of reliability" being unreliable, and masking pure bullsh_t.*
*I use the term "bullsh_t" to refer to Frankfurtian statements -- that is, statements that meet the definition of bullsh_t proposed in Harry Frankfurt's widely-known treatise ON BULLSH_T. Essentially, statements made without regard to their truth or falsity.
You guys are a little harsh on this man. He used a tool that did not actually do what it appeared to do. That's all. No harm done. The judge got all excited over nothing.
"Who cares that he cited a bunch of cases that don't exist? He got caught!"
It was May 27!
https://reason.com/volokh/2023/05/27/a-lawyers-filing-is-replete-with-citations-to-non-existent-cases-thanks-chatgpt/
Not sure which is worse, that he used ChatGPT to do his “research” or that he revealed that he cites cases he has not read.
It's king kev j. Ausmus thank u
The creative ways in which AI, like ChatGPT, is being used continue to amaze. I'm intrigued by the lawyer's innovative approach to legal work. Experience ChatGPT Portugues completely free without needing an account or login