The Volokh Conspiracy
Mostly law professors | Sometimes contrarian | Often libertarian | Always independent
Judge Strikes Part of Anthropic (Claude.AI) Expert's Declaration, Because of Uncaught AI Hallucination in Part of Citation
From Friday's order by Magistrate Judge Susan van Keulen in Concord Music Group, Inc. v. Anthropic PBC (N.D. Cal.)
At the outset, the Court notes that during the hearing, Publishers asked this Court to examine Anthropic's expert, Ms. Chen and strike her declaration because at least one of the citations therein appeared to have been an "AI hallucination": a citation to an article that did not exist and whose purported authors had never worked together. The Court gave Anthropic time to investigate the circumstances surrounding the challenged citation. Having considered the declaration of Anthropic's counsel and Publishers' response, the Court finds this issue is a serious one—if not quite so grave as it at first appeared.
Anthropic's counsel protests that this was "an honest citation mistake" but admits that Claude.ai was used to "properly format" at least three citations and, in doing so, generated a fictitious article name with inaccurate authors (who have never worked together) for the citation at issue. That is a plain and simple AI hallucination. Yet the underlying article exists, was properly linked to and was located by a human being using Google search; so, this is not a case where "attorneys and experts [have] abdicate[d] their independent judgment and critical thinking skills in favor of ready-made, AI-generated answers…."
A remaining serious concern, however, is Anthropic's attestation that a "manual citation check" was performed but "did not catch th[e] error." It is not clear how such an error—including a complete change in article title—could have escaped correction during manual cite-check by a human being. Furthermore, although the undersigned's [i.e., the Magistrate Judge's] standing order does not expressly address the use of AI by parties or counsel, Section VIII.G of [District] Judge Lee's Civil Standing Order requires a certification "that lead trial counsel has personally verified the content's accuracy." Neither the certification nor verification has occurred here. In sum, the Court STRIKES-IN-PART Ms. Chen's declaration, striking paragraph 9 [which contains the footnote that contains the citation with the hallucination], and notes for the record that this issue undermines the overall credibility of Ms. Chen's written declaration, a factor in the Court's conclusion.
Thanks to ChatGPT Is Eating the World for the pointer; it also discusses more about the substantive role of paragraph 9 in the declaration. Here's more backstory (from an earlier post):
The Declaration filed by a "Data Scientist at Anthropic" in Concord Music Group, Inc. v. Anthropic PBC includes this citation:
But the cited article doesn't seem to exist at that citation or at that URL, and Google found no other references to any article by that title….
Here's the explanation, from one of Anthropic's lawyers (emphasis added):
Our investigation of the matter confirms that this was an honest citation mistake and not a fabrication of authority. The first citation in footnote 3 of Dkts. 340-3 (sealed) and 341-2 (public) includes an erroneous author and title, while providing a correct link to, and correctly identifying the publication, volume, page numbers, and year of publication of, the article referenced by Ms. Chen as part of the basis for her statement in paragraph 9. We apologize for the inaccuracy and any confusion this error caused.
The American Statistician article reviewed and relied upon by Ms. Chen [the Anthropic expert], and accessible at the first link provided in footnote 3 of Dkts. 340-3 and 341-2, is titled Binomial Confidence Intervals for Rare Events: Importance of Defining Margin of Error Relative to Magnitude of Proportion, by Owen McGrath and Kevin Burke. A Latham & Watkins associate located that article as potential additional support for Ms. Chen's testimony using a Google search. The article exists and supports Ms. Chen's testimony in her declaration and at the May 13, 2025 hearing, which she proffered based on her pre-existing knowledge regarding the appropriate relative margin of error for rare events. A copy of the complete article is attached as Exhibit A.
Specifically, "in the context of small or rare-event success probabilities," the authors "suggest restricting the range of values to εR ∈ [0.1, 0.5]"—meaning, a relative margin of error between 10% to 50%—"as higher values lead to imprecision and poor interval coverage, whereas lower values lead to sample sizes that are likely to be impractically large for many studies." See Exhibit A, at 446. This recommendation is entirely consistent with Ms. Chen's testimony, which proposes using a 25% relative margin of error based on her expertise.
After the Latham & Watkins team identified the source as potential additional support for Ms. Chen's testimony, I asked Claude.ai to provide a properly formatted legal citation for that source using the link to the correct article. Unfortunately, although providing the correct publication title, publication year, and link to the provided source, the returned citation included an inaccurate title and incorrect authors. Our manual citation check did not catch that error. Our citation check also missed additional wording errors introduced in the citations during the formatting process using Claude.ai. These wording errors are: (1) that the correct title of the source in footnote 2 of Ms. Chen's declaration is Computing Necessary Sample Size, not, as listed in footnote 2, Sample Size Estimation, and (2) the author/preparer of the third source cited in footnote 3 is "Windward Environmental LLC", not "Lower Windward Environmental LLC." Again, we apologize for these citation errors.
Ms. Chen, as well as counsel, reviewed the complete text of Ms. Chen's testimony and also reviewed each of the cited references prior to submitting Ms. Chen's declaration to the Court. In reviewing her declaration both prior to submission and in preparation for the hearing on May 13, 2025, Ms. Chen reviewed the actual article available at the first link in footnote 3 of her declaration and attached hereto as Exhibit A, and the article supports the proposition expressed in her declaration with respect to the appropriate margin of error.
During the production and cite-checking process for Ms. Chen's declaration, the Latham & Watkins team reviewing and editing the declaration checked that the substance of the cited document supported the proposition in the declaration, and also corrected the volume and page numbers in the citation, but did not notice the incorrect title and authors, despite clicking on the link provided in the footnote and reviewing the article. The Latham & Watkins team also did not notice the additional wording errors in footnotes 2 and 3 of Ms. Chen's declaration, as described above in paragraph 6.
This was an embarrassing and unintentional mistake. The article in question genuinely exists, was reviewed by Ms. Chen and supports her opinion on the proper margin of error to use for sampling. The insinuation that Ms. Chen's opinion was influenced by false or fabricated information is thus incorrect. As is the insinuation that Ms. Chen lacks support for her opinion. Moreover, the link provided both to this Court and to Plaintiffs was accurate and, when pasted into a browser, calls up the correct article upon which Ms. Chen had relied. Had Plaintiffs' counsel raised the citation issue when they first discovered it, we could and would have confirmed that the article cited was the one upon which Ms. Chen relied and corrected the citation mistake.
We have implemented procedures, including multiple levels of additional review, to work to ensure that this does not occur again and have preserved, at the Court's direction, all information related to Ms. Chen's declaration. I understand that Anthropic has also preserved all information related to Ms. Chen's declaration as well….
Editor's Note: We invite comments and request that they be civil and on-topic. We do not moderate or assume any responsibility for comments, which are owned by the readers who post them. Comments do not represent the views of Reason.com or Reason Foundation. We reserve the right to delete any comment for any reason at any time. Comments may only be edited within 5 minutes of posting. Report abuses.
Please
to post comments
That's actually fairly mild, IMO.
If we are going to strike fictitious claims in the law, strike these fictions: minds can be read, rare accidents can be foreseen, and the example for proper conduct should be a fictitious character. This character is a thinly disguised avatar for Jesus. Volokh, a First Amendment expert, is in denial about the toxicity of these lawyer delusions. These supernatural doctrines from the catechism are also illegal in our secular nation. I liked most of the Sharia, but basing a jurisprudence on it would be outrageouslusly illegal, and stupid. Volokh is not just a denier. He also indoctrinated 1000s of law students into these fictitious idiosies.
? ? ?
Crack-smoking? Account hijacked? A stroke? Something else???
Those are just personal insults. Are you a lawyer?
What ever happened to "Falsus in uno, falsus in omnibus"?
Seems adequately addressed here by "...the Court...notes for the record that this issue undermines the overall credibility of Ms. Chen's written declaration, a factor in the Court's conclusion."
I can see how the human checked that the link worked and the article was correct. But didn’t check the text of the citation. Or didn’t check it well.
Are citations really never mistaken?
It’s not terribly surprising to see citations in scientific literature that clearly haven’t been read by the authors.
Always a fun time when you follow the citation tree back to the original claim in a scientific paper and 5 papers you see that the claim was either made up or not at all substantiated by the actual study and an exaggeration of what was found.
And you cite it anyway because it’s 3 am and it’s a paper for undergraduate chemical engineering and no one will read it anyway. If there is money on the line the private will do their own research and it’s their plant that blows up if something goes wrong so they’ll figure it out
Citations are often mistaken. That's why you proof read them. Having mistaken citations in the legal papers you actually file is a big mistake. Having a "mistaken" citation that completely changed the title and authors is significant one. I'd have a pretty hard time finding anything the expert testified to credible if this is the kind of error she can make without noticing.
Back in the 90s, an interpreted programming language, REXX, was used to write, FOGGY, a powerful program to produce quick reports, of variable length, using lists of partial phrases to compose the sentences. Each sentence contained an introduction, a body, and an ending. In other words, it produced stuff to fill the report in a random way. Seems these AI programs work the same way.
This is like complaining that a word processor program search-and-replace did too many replaces.
Es preocupante cómo incluso declaraciones de “expertos” pueden incluir errores graves por confiar ciegamente en la IA. Este caso con Claude AI demuestra que siempre hay que verificar fuentes. Lo mismo aplica al buscar reseñas confiables, como cuando se exploran plataformas de apuestas o juegos en línea. Por ejemplo, este sitio ofrece reseñas de casinos que realmente ayudan a decidir si vale la pena registrarse o no. Aprovecha las mejores promociones en 1xbet. En tiempos donde la información puede ser fácilmente manipulada o inventada, tener referencias verificadas es más importante que nunca.