Unlocking the Potential of AI Training without Copyright Infringement

Artificial Intelligence (AI) models have long been trained using copyrighted materials, but recent developments suggest that there are alternative ways to train these models without infringing on intellectual property rights. A group of researchers backed by the French government has released a significant AI training dataset composed entirely of text in the public domain. This groundbreaking dataset offers evidence that large language models can be trained without the need for permissionless use of copyrighted materials.

The nonprofit organization Fairly Trained has also made an announcement that it has successfully certified its first large language model, named KL3M. This model was developed by 273 Ventures, a Chicago-based legal tech consultancy startup, using a curated training dataset of legal, financial, and regulatory documents. By adhering to copyright laws and using their own dataset, 273 Ventures has demonstrated that it is possible to build large language models without the contentious issue of copyright infringement.

According to Jillian Bommarito, the co-founder of 273 Ventures, their decision to train KL3M using their own dataset was driven by their risk-averse clients in the legal industry. These clients were concerned about the source of the data and wanted assurances that their AI model was not based on tainted or copyrighted data. By using a carefully curated dataset, Bommarito emphasizes that the size of the model does not need to be overwhelmingly large, and that high-quality data can lead to better performance and specialization.

While datasets like KL3M are currently limited in size compared to those compiled by industry giants like OpenAI, there is hope for the future. Researchers have recently released the Common Corpus, which they claim to be the largest available AI dataset for language models composed solely of public domain content. This dataset, posted on the open-source AI platform Hugging Face, contains text from public domain newspapers digitized by institutions like the US Library of Congress and the National Library of France. Common Corpus aims to provide researchers and startups with a vetted training set that is free from copyright concerns.

While datasets composed of public domain content have their limitations, such as antiquated information, they offer an invaluable resource for training large language models. Projects like Common Corpus and KL3M demonstrate a growing skepticism in the AI community towards the argument of permissionless data scraping. In fact, Fairly Trained recently certified its first company to offer AI voice models, showing that there is a growing trend in the industry towards obtaining proper licensing and respecting intellectual property rights.

Frequently Asked Questions (FAQ)

1. What is Fairly Trained?

Fairly Trained is a nonprofit organization that offers certifications to companies that can prove they have trained their AI models on data they either own, have obtained licenses for, or is in the public domain. The aim of Fairly Trained is to encourage fair and ethical practices in AI development.

2. How does KL3M differ from other large language models?

KL3M is unique because it has been trained using a curated dataset of legal, financial, and regulatory documents that comply with copyright law. Unlike other models, KL3M avoids copyright infringement issues and provides authoritative and trustworthy results, making it ideal for clients in the legal industry.

3. What is the Common Corpus dataset?

Common Corpus is an AI dataset built from public domain content, such as digitized newspapers from institutions like the US Library of Congress and the National Library of France. It aims to offer researchers and startups a vetted training set free from copyright concerns, although it may not contain the most up-to-date information.

4. Why is there a growing trend towards licensing in AI?

As AI technology evolves and becomes more advanced, there is an increasing awareness of the need to respect intellectual property rights. Many organizations, including the Authors Guild and SAG-AFTRA, support Fairly Trained’s mission to promote fair licensing practices in AI development.

Artificial Intelligence (AI) models have seen significant advancements in recent years, but one of the challenges in training these models has been the use of copyrighted materials. However, there are now alternative methods emerging that allow for the training of AI models without infringing on intellectual property rights.

A group of researchers backed by the French government has made a groundbreaking contribution in this area by releasing an AI training dataset composed entirely of text in the public domain. This dataset provides evidence that large language models can be trained without the need for permissionless use of copyrighted materials. This development opens up new possibilities for AI model training, ensuring compliance with copyright laws.

In addition, the nonprofit organization Fairly Trained has achieved a significant milestone by successfully certifying its first large language model called KL3M. Developed by 273 Ventures, a Chicago-based legal tech consultancy startup, KL3M was trained using a curated dataset consisting of legal, financial, and regulatory documents. By adhering to copyright laws and utilizing their own dataset, 273 Ventures has demonstrated that it is possible to build large language models without the contentious issue of copyright infringement.

Jillian Bommarito, co-founder of 273 Ventures, cites the concerns of their risk-averse clients in the legal industry as the motivation behind their decision to train KL3M using their own dataset. Clients wanted assurance that their AI model was not based on copyrighted or tainted data. Bommarito highlights that the size of the model does not need to be overwhelmingly large, and emphasizes the importance of high-quality data in achieving better performance and specialization.

While datasets like KL3M may currently be smaller compared to those compiled by industry giants like OpenAI, there is hope for the future. Researchers have recently released the Common Corpus, claimed to be the largest AI dataset available for language models, composed solely of public domain content. This dataset, hosted on the open-source AI platform Hugging Face, includes text from public domain newspapers digitized by institutions like the US Library of Congress and the National Library of France. Common Corpus aims to provide researchers and startups with a vetted training set that is free from copyright concerns.

Although datasets composed of public domain content have limitations, such as potentially containing antiquated information, they offer an invaluable resource for training large language models. The development of projects like Common Corpus and KL3M reflects a growing skepticism in the AI community towards the argument of permissionless data scraping. Moreover, Fairly Trained has recently certified its first company to offer AI voice models, indicating a growing industry trend towards obtaining proper licensing and respecting intellectual property rights.

These advancements in AI training methodologies and the growing emphasis on fair and ethical practices have led to the rise of organizations like Fairly Trained. Fairly Trained is a nonprofit organization that offers certifications to companies that can prove they have trained their AI models on data they either own or have obtained licenses for, or is in the public domain. Their aim is to encourage fair practices and respect for intellectual property rights in AI development.

Overall, the emergence of alternative methods for training AI models without infringing on copyrights, the availability of curated datasets, and the certification programs by organizations like Fairly Trained demonstrate a shift towards responsible and respectful AI development practices within the industry.

Related Links:
Fairly Trained website
273 Ventures website
Hugging Face platform

Privacy policy
Contact