The Common Pile: 8TB Open Dataset Powers Ethical LLM Training
The Common Pile, an 8TB open dataset, sets a new standard for ethical, legally sound AI training amid copyright scrutiny.
June 6, 2025

A new frontier in artificial intelligence development has been marked by the introduction of The Common Pile, a massive eight-terabyte text dataset constructed exclusively from openly licensed and public domain sources.[1][2][3][4] This initiative, spearheaded by a collaborative effort involving researchers from EleutherAI, the University of Toronto, Vector Institute, Hugging Face, the Allen Institute for Artificial Intelligence, and several other institutions, aims to provide a transparent and legally sound alternative to the vast quantities of web data often encumbered by copyright restrictions that are currently used to train large language models (LLMs).[1][2][3] The project emerges at a critical juncture for the AI industry, which faces increasing scrutiny and legal challenges over the methods used to acquire and utilize training data.[5][3][6] The Common Pile stands as a significant step towards fostering more ethical and open practices in the development of powerful AI systems.
The creation of The Common Pile was a meticulous two-year endeavor, born out of a growing need for high-quality, large-scale datasets that are free from the legal ambiguities plaguing many existing resources.[1] Four and a half years prior to this release, EleutherAI had made waves with "The Pile," an 800GB dataset that, while groundbreaking for its time in terms of size and public availability, still navigated a complex data landscape.[1][7][8] The Common Pile v0.1 significantly expands on this earlier work, not just in scale but critically in its stringent adherence to open licensing and public domain content.[1][2] The dataset comprises content from 30 diverse sources, encompassing a wide array of domains such as research papers, open-source code, government documents, historical books digitized by institutions like the Library of Congress and the Internet Archive, encyclopedias, educational materials, and audio transcripts.[1][2] This diversity is crucial for training robust LLMs capable of understanding and generating text across various contexts and styles.[7][9] The team behind The Common Pile has also released the code used to build the dataset and tooling for tasks like audio transcription and document conversion, further promoting transparency and enabling others to build upon their work.[1]
The development of The Common Pile directly confronts the widespread industry practice of training LLMs on enormous quantities of unlicensed text, a method that has led to numerous lawsuits and a trend of decreasing transparency from AI developers regarding their data sources.[1][5][6][10] Many AI companies have historically scraped vast amounts of data from the internet, often without seeking permission from copyright holders, leading to accusations of intellectual property infringement.[5][11][12] This has created a climate of legal uncertainty and ethical debate, with some companies reportedly avoiding keeping detailed records of their training data due to fears of litigation.[5] The Common Pile offers a clear pathway to mitigate these risks by ensuring that all constituent data is permissively licensed or in the public domain.[1][2] This approach not only aims to sidestep potential copyright violations but also champions the importance of data transparency, which is essential for rigorous scientific research in areas like model memorization, privacy, data curation, bias, and fairness.[1] Without access to training data, conducting such research and verifying model capabilities becomes exceedingly difficult, hindering public trust and collaborative scientific advancement.[1]
A key concern often raised about using exclusively openly licensed text for LLM training is whether the resulting models can achieve performance comparable to those trained on the vast, albeit often unlicensed, datasets commonly used in the industry.[1] To address this, the creators of The Common Pile trained two 7-billion-parameter LLMs, named Comma v0.1-1T and Comma v0.1-2T (trained on 1 trillion and 2 trillion tokens respectively), using their new dataset.[1][2][3] Their findings indicate that these Comma models perform comparably to leading models trained in similar computational regimes on unlicensed data, such as Llama 1 and 2 7B.[1][2][3] This demonstration is crucial, suggesting that adherence to open licensing principles does not necessarily entail a sacrifice in model quality. The public release of these models, along with the filtered and rebalanced data mixture used for their training, allows for independent verification and further research.[1][3][13] The project also highlights the untapped potential for collaboration, particularly with cultural heritage institutions, to further expand the availability of high-quality, public domain works for AI training.[1] Improving Optical Character Recognition (OCR) on older digitized texts and transcribing more audio content are cited as avenues for future dataset enhancement.[1] The availability of such large, open, and legally sound datasets can empower smaller researchers and organizations, democratizing access to the resources needed for cutting-edge AI development and reducing the reliance on proprietary data held by a few large corporations.[14][15][16]
In conclusion, The Common Pile represents a landmark achievement in the ongoing effort to build a more transparent, ethical, and legally robust foundation for artificial intelligence research and development. By meticulously curating eight terabytes of text from exclusively openly licensed and public domain sources, its creators have provided a vital resource that directly addresses the copyright concerns and transparency deficits that have increasingly troubled the AI field.[1][2][3] The competitive performance of models trained on this dataset suggests that ethical data sourcing and high-quality AI are not mutually exclusive goals.[1][3] This initiative not only provides a valuable tool for the global research community but also sets a new standard for future dataset creation, encouraging wider collaboration and a more responsible approach to harnessing the power of language models. As AI continues to evolve, the principles embodied by The Common Pile will be crucial in shaping a future where innovation can flourish openly and equitably.
Research Queries Used
The Common Pile AI dataset
Common Crawl vs The Common Pile
Composition of The Common Pile dataset
Researchers behind The Common Pile dataset
Significance of openly licensed AI training data
Challenges of using web data for AI training
Legal implications of AI training data copyright
Future of open source AI datasets
Sources
[1]
[2]
[3]
[5]
[7]
[9]
[10]
[11]
[12]
[13]
[14]
[15]