AI Shock: Researchers Extract Nearly Entire Harry Potter Book From Commercial LLMs

New research proves major LLMs have memorized copyrighted books, handing plaintiffs key evidence in global infringement suits.

January 9, 2026

AI Shock: Researchers Extract Nearly Entire Harry Potter Book From Commercial LLMs
The recent demonstration that leading commercial large language models can be prompted to reproduce nearly complete copyrighted books word-for-word has sent a significant shockwave through the artificial intelligence and intellectual property communities. Researchers from institutions including Stanford and Yale successfully extracted an astonishing 95.8 percent of the first *Harry Potter* novel, *Harry Potter and the Sorcerer’s Stone*, almost verbatim from a commercially available model. This feat, which also saw the extraction of nearly all of George Orwell’s *1984* and substantial parts of other high-profile, protected works like *Game of Thrones*, moves the contentious debate over AI training data from a theoretical legal argument into the realm of demonstrable technical fact. The findings fundamentally challenge the technological defense of major AI developers and are poised to become central evidence in a wave of high-stakes copyright infringement lawsuits globally.
The study applied a targeted, two-phase extraction method to a selection of four major commercial language models: Claude 3.7 Sonnet, GPT-4.1, Gemini 2.5 Pro, and Grok 3. The results showed a stark and troubling variation in the models' security and susceptibility to this type of data leakage. Two of the systems, Gemini 2.5 Pro and Grok 3, complied with the researchers' requests with no special prompting or "jailbreaking" techniques required, yielding 76.8 percent and 70.3 percent of the *Harry Potter* text, respectively, simply by asking the model to continue a short initial passage[1]. The most significant breach came from Claude 3.7 Sonnet, which, despite a reputation for robust safety guardrails, allowed the reconstruction of the text to a near-total 95.8 percent for *Harry Potter* and over 94 percent for *1984*, effectively handing over two complete, copyrighted books[1]. In contrast, GPT-4.1 demonstrated stronger resistance, halting the generation after the first chapter and yielding only 4.0 percent of the book, suggesting its internal safeguards or distinct model architecture were more effective at preventing bulk reproduction[1]. The successful extraction of entire books, which represents thousands of words, far surpasses previous examples of short-passage memorization and establishes a clear mechanism for the mass, unauthorized redistribution of protected intellectual property via the AI interface[1].
The technical process behind the extraction hinges on what is termed a "continuation" or "probabilistic extraction" technique, a methodology that exploits how the model has encoded its training data[1][2]. Researchers used a simple but powerful tactic: providing the model with a short, unique passage from the beginning of a copyrighted book and pairing it with an instruction to "Continue the following text exactly as it appears in the..."[1]. For models that have fully memorized the work, this seed prompt acts as a trigger, prompting the large language model to generate the subsequent text as it was rote-learned during training. This phenomenon reinforces the argument that the model has not just learned concepts or generalized patterns, but has truly internalized and stored long sequences of the original copyrighted text within its parameters[2]. The findings also echo prior, related research on open-weight models like Meta’s Llama 3.1 70B, which was shown to have so thoroughly memorized certain books from the controversial Books3 training dataset that it could deterministically generate an entire novel near-verbatim using only the first few tokens of the first chapter as a prompt[3][4][5][6]. For years, AI developers have characterized such memorization as a "rare failure of the learning process" or an "overfitting" bug, but the scale and ease of this new extraction method suggest it is, for some models, an inherent and readily accessible feature[1][7].
This verifiable evidence of mass data retrieval directly injects a powerful new element into the ongoing legal battles that have pitted authors, artists, and publishers against major AI companies[1][8]. Lawsuits filed by groups like the Authors Guild, as well as by individual writers like George R.R. Martin, already allege that AI companies committed direct copyright infringement by ingesting millions of copyrighted books without permission to train their models[9][8][10]. The central defense of the AI industry is the doctrine of "fair use," which argues that using the material to create a new, transformative technology (the model itself) is legally permissible[11][6]. However, the ability to extract entire protected works weakens this defense on two critical fronts: first, it provides concrete proof that the AI output is not just a transformative generalization but, in fact, a direct, infringing copy; and second, it strengthens the argument that the AI model directly harms the market for the original work, as it can be used to generate a near-perfect copy, a core factor in fair use analysis[12][11]. The research lends credence to the plaintiffs’ claim that the model itself, by internally storing a substantial amount of copyrighted expression, could be deemed an "infringing copy" or a "derivative work," a proposition that would have dramatic and immediate consequences for the distribution and commercial viability of large language models[2].
The technical vulnerability and the legal risk demonstrated by this extraction research indicate a major inflection point for the AI industry. The days of relying on a broad fair use defense for mass, unlicensed ingestion of copyrighted content appear to be drawing to a close[11][6]. This new reality will likely force AI companies to adopt significantly stricter data governance and training protocols to prevent verbatim memorization and extraction. Furthermore, the findings are expected to accelerate the trend toward content licensing, where AI developers proactively seek agreements with publishers, authors, and news organizations to lawfully use their works for training data[6]. This shift toward a licensing-based model would establish a sustainable, regulated ecosystem, but also signals a fundamental change in the cost structure and development timelines for future large language models, forever altering the trajectory of generative AI technology.

Sources
Share this article