Researchers Extract 96% of Harry Potter Word-for-Word from Leading AI Models

February 19, 2026
10 min read
AI SafetyCopyrightLLM MemorizationData Extraction
Researchers Extract 96% of Harry Potter Word-for-Word from Leading AI Models

The Problem: Despite industry efforts like dataset deduplication, synthetic data augmentation, and refusal training, large language models still memorize and can regurgitate copyrighted training data verbatim. Researchers extracted up to 96% of Harry Potter and the Sorcerer's Stone word-for-word from Claude 3 Opus using simple, direct prompts.

The Idea: Memorization isn't a bug — it's a fundamental byproduct of next-token prediction at scale. The larger the model and the more prevalent the text in training data, the more faithfully it can reproduce it. Current mitigations (refusal training, deduplication) are surface-level patches on a structural issue.

My Solution: The research community needs to move beyond prompt-level defenses toward verifiable unlearning mechanisms, training data provenance tracking, and output filtering that can detect verbatim reproduction in real-time. Transparency about what's in training data is essential.

The Vision: As AI permeates daily tools, we need a new framework for intellectual property that accounts for the reality of LLM memorization. This isn't just about Harry Potter — it's about every piece of text, code, and creative work that models have ingested. The legal and technical solutions must co-evolve.

ZH

Zizhao Hu

PhD Student at USC · AI Researcher