Investigations Expose Tech Giants' AI Hypocrisy: They Scrape Billions, Ban Creators
The AI industry's "do as I say, not as I scrape" approach sparks a copyright crisis, threatening creators and the internet's future.
September 11, 2025

A stark double standard is emerging at the heart of the artificial intelligence revolution, sparking a fierce debate over data, copyright, and the very ethics of innovation. Major technology companies are systematically scraping vast quantities of copyrighted material from across the internet to train their lucrative AI models. Simultaneously, these same firms explicitly forbid identical data harvesting from their own platforms through restrictive terms of service. This "scrape to train, block everyone else" approach has been brought into sharp focus by separate, exhaustive investigations, revealing a fundamental hypocrisy that could have profound implications for creators, competition, and the future of information itself.
A two-year investigation by the International Confederation of Music Publishers (ICMP) and a detailed analysis by The Atlantic have laid bare the scale of this data appropriation. The ICMP, which represents 90 percent of the world's commercially-released music, compiled evidence from public registries, leaked materials, and open-source repositories to build its case.[1][2][3] Their findings accuse tech giants including Google, Microsoft, Meta, and OpenAI of illegally using millions of copyrighted songs from artists like The Beatles, Beyoncé, and Taylor Swift to train AI systems such as Google's Gemini, Meta's Llama 3, and Microsoft's Copilot, all without licenses.[4][1][2] John Phelan, the director general of the ICMP, has labeled the practice "the largest IP theft in human history," stating, "We are seeing tens of millions of works being infringed daily."[4][3] The investigation revealed that within a single AI training dataset, tens of millions of musical works are often sourced from individual YouTube, Spotify, and GitHub URLs, in direct breach of the rights of publishers and songwriters.[4]
Parallel to the music industry's findings, an investigation by The Atlantic's AI Watchdog initiative uncovered an immense, unauthorized data grab from one of the world's largest content platforms.[5][6] The report revealed that tech companies, including Meta, Microsoft, and Google, scraped over 15.8 million YouTube videos from more than 2 million channels to train powerful generative video AI models.[5][7] This mass download directly violates YouTube's terms of service, which prohibit such activity.[7] For countless creators who have built their livelihoods on these platforms, the revelation that their work is being used to build tools that could potentially replace them has been a profound betrayal. The issue transcends simple copyright infringement, touching upon the fundamental fairness of an ecosystem where the labor of creators is harvested to build their direct competitors.[5]
The core of this controversy lies in the glaring contradiction between the tech industry's actions and its own stated rules. Google's terms of service, for example, explicitly state that users may not "send automated queries to Google's system" or "use any automated system... to access the Services in a manner that sends more request messages to the Google servers than a human can reasonably produce." Similarly, OpenAI's terms prohibit using "any automated or programmatic method to extract data or output from the Services, including scraping, web harvesting, or web data extraction" without express permission. X, formerly known as Twitter, also updated its terms to expressly forbid crawling or scraping its services "in any form, for any purpose without our prior written consent." Yet, these are the very practices these companies are accused of employing on a massive scale across the open internet to fuel their AI ambitions. Meta, for its part, has actively sued companies like Bright Data for scraping Facebook and Instagram, even while court documents revealed Meta itself had paid the same firm to scrape other websites for what a spokesperson termed "legitimate integrity and commercial purposes."
In the face of mounting criticism and a flurry of lawsuits from authors, artists, and news organizations, the tech industry has leaned heavily on the legal doctrine of "fair use."[8] This provision in copyright law allows for the limited use of copyrighted material without permission for purposes such as criticism, comment, news reporting, and research. AI developers argue that training models on vast datasets constitutes a transformative use, creating new and beneficial tools for society. Google's general counsel, Halimah DeLaine Prado, stated in response to one lawsuit that U.S. law "supports using public information to create new beneficial uses," and that a legal challenge to this practice would "take a sledgehammer not just to Google's services but to the very idea of generative AI."[9] Similarly, Microsoft AI CEO Mustafa Suleyman sparked controversy by claiming the public internet is essentially "freeware," arguing that the social contract since the 1990s has been that online content is available for anyone to copy and reproduce under the principle of fair use.[10] OpenAI has gone as far as to suggest that without the ability to train on copyrighted works, American companies will lose the AI race to competitors in China, framing the issue as a matter of national security.[11]
This defense, however, is being vigorously challenged in courts, and the legal landscape remains highly unsettled. Rulings have been divergent and are often highly specific to the facts of each case, leaving both creators and AI developers in a state of uncertainty.[4] While some judges have been sympathetic to the fair use argument, particularly when the AI's output is deemed highly transformative, others have ruled against AI companies, especially when their products directly compete with the copyright holders they trained on.[12] Critics argue that the "fair use" claim is a convenient shield for what amounts to industrial-scale intellectual property theft. They contend that if AI models are commercial products that generate billions in revenue, the companies behind them should be required to license the data that makes them powerful, just like any other business that relies on creative works.
The implications of this data double standard are far-reaching. For content creators, it represents an existential threat, as their work is used without compensation to develop technologies that could devalue or even replace their professions.[5] For the AI industry, the ongoing legal battles create significant risk and uncertainty, potentially reshaping the economics of model development if widespread licensing becomes a requirement. The conflict also raises fundamental questions about the future of the open internet. If the dominant players can unilaterally scrape data while closing off their own valuable ecosystems, it could lead to the creation of information monopolies, stifling competition and innovation.[13] As lawsuits proceed and regulators begin to take a closer look, the tech industry may be forced to abandon its paradoxical stance and confront the true cost of the data that fuels its most ambitious creations.
Sources
[1]
[2]
[3]
[4]
[9]
[10]
[11]
[12]