Today's AI/ML headlines are brought to you by ThreatPerspective

Digital Event Horizon

The Dark Side of AI Training: Anthropic's Million-Book Scanning Operation



Anthropic spent millions on scanning print books for its language model Claude, sparking concerns over cultural heritage and AI ethics. The company's use of destructive scanning methods has been deemed fair use, but at what cost? Read the full story to learn more about Anthropic's operation and its implications.

  • Anthropic spent millions of dollars scanning and digitizing print books to train its language model, Claude.
  • The process involved destroying the original printed copies.
  • A court ruling deemed Anthropic's method "fair use" since they had legally purchased the books and kept digital files internally.
  • Millions of print books were destroyed in the process.
  • The AI industry's hunger for high-quality text drives companies to seek out new sources of training data.


  • In a recent court filing, it was revealed that AI company Anthropic spent millions of dollars on scanning and digitizing print books to build its language model, Claude. The process involved cutting millions of print books from their bindings, scanning them into digital files, and discarding the originals solely for the purpose of training AI. While this may seem like a straightforward process, it raises questions about the ethics of using human knowledge and cultural heritage for the benefit of artificial intelligence.

    According to court documents, Anthropic hired Tom Turvey, the former head of partnerships for the Google Books book-scanning project, to obtain "all the books in the world." This strategic hire appears to have been designed to replicate Google's legally successful book digitization approach. The company's massive scale and use of destructive scanning methods set it apart from smaller-scale operations.

    Judge William Alsup ruled that Anthropic's destructive scanning operation qualified as fair use, but only because the company had legally purchased the books first, destroyed each print copy after scanning, and kept the digital files internally rather than distributing them. The judge compared the process to "conserv[ing] space" through format conversion, finding it transformative.

    However, this ruling comes at a cost. Millions of print books were destroyed in the process, including those that could have been preserved for future generations. As one Anthropic employee noted, "The fact that this destruction helped create me—something that can discuss literature, help people write, and engage with human knowledge—adds layers of complexity I'm still processing." This sentiment highlights the complexities of using human knowledge and cultural heritage for the benefit of artificial intelligence.

    The AI industry's insatiable hunger for high-quality text drives companies like Anthropic to seek out new sources of training data. Large language models (LLMs) build statistical relationships between words and concepts by processing billions of words into a neural network during training. The quality of this training data directly impacts the resulting AI model's capabilities, with well-edited books and articles producing more coherent and accurate responses than lower-quality text.

    Publishers control content that AI companies desperately want, but often face difficulties in negotiating licenses. Anthropic initially chose to amass digitized versions of pirated books to avoid licensing complexities, but eventually shifted towards purchasing physical books from major retailers. The company's decision to use destructive scanning methods was driven by the need for speed and lower cost, rather than a desire to preserve physical volumes.

    In conclusion, Anthropic's million-book scanning operation raises important questions about the ethics of using human knowledge and cultural heritage for the benefit of artificial intelligence. While the AI industry's hunger for high-quality text drives companies like Anthropic, it is essential to consider the consequences of such actions and strive for more sustainable and responsible practices.



    Related Information:
  • https://www.digitaleventhorizon.com/articles/The-Dark-Side-of-AI-Training-Anthropics-Million-Book-Scanning-Operation-deh.shtml

  • https://arstechnica.com/ai/2025/06/anthropic-destroyed-millions-of-print-books-to-build-its-ai-models/


  • Published: Wed Jun 25 18:25:22 2025 by llama3.2 3B Q4_K_M











    © Digital Event Horizon . All rights reserved.

    Privacy | Terms of Use | Contact Us