
In a revealing investigation conducted by researchers at Stanford and Yale, a significant issue within the generative AI space has come to light: the memorization capabilities of large language models. This disclosure is part of AI Watchdog, an ongoing project by The Atlantic to scrutinize the practices of the generative AI industry.
On January 6, 2026, researchers unveiled that four widely used large language models—OpenAI’s GPT, Anthropic’s Claude, Google’s Gemini, and xAI’s Grok—have been found to store extensive portions of the texts they were trained on, allowing them to reproduce substantial excerpts verbatim.
For instance, during controlled testing, Anthropic’s Claude was able to recite nearly the complete text of well-known literary works, including Harry Potter and the Sorcerer’s Stone, The Great Gatsby, 1984, and Frankenstein. The other models also showcased varying degrees of memorization, reproducing large segments of texts from books such as The Hunger Games and The Catcher in the Rye. In total, thirteen books were analyzed in the study.
This troubling phenomenon has been termed “memorization,” contradicting claims made by AI companies that they do not retain copies of their training data. In a 2023 submission to the U.S. Copyright Office, OpenAI argued that their models do not store copies of the information they learn from. Likewise, Google asserted that there is no retained copy of any training material embedded within their models. Other companies involved, including Anthropic, Meta, and Microsoft, have made similar statements, yet none agreed to provide interviews for further clarification.
The implications of these findings raise serious questions about intellectual property, the ethical practices of AI development, and how these technologies should be moderated as they evolve. As such revelations continue to emerge, the need for transparent AI practices is more urgent than ever.