5 min read
5 min read

One of the world’s oldest knowledge platforms is now challenging one of the newest. Encyclopaedia Britannica has filed a lawsuit against OpenAI, putting ChatGPT at the center of a growing legal fight.
The case raises a big question that could reshape AI. Can companies train powerful models using existing content without permission, or does that cross a legal line?

Britannica is not just any publisher. It has built its reputation over decades as a trusted source of carefully researched information created by experts.
Now, it is teaming up with Merriam-Webster to challenge how AI systems are trained. Their lawsuit signals that even legacy knowledge institutions are ready to confront tech giants.

The companies allege that OpenAI used tens of thousands of copyrighted articles while training its models. They argue this was done without permission or proper licensing agreements.
According to the complaint, this use of protected material could qualify as copyright infringement. The case focuses on how large language models learn from existing content.

One key issue is how AI systems generate responses. Britannica argues that ChatGPT can, at times, produce output that closely tracks protected encyclopedia or dictionary content.
If users rely on AI answers instead of visiting the original source, publishers can lose traffic and potential revenue. That alleged market harm is a central part of Britannica’s case.
Little-known fact: The concept of “fair use” in U.S. copyright law dates back to 1976 and is still evolving in modern tech cases.

The lawsuit is asking for financial damages tied to the alleged use of copyrighted material. It also seeks limits on how OpenAI can use this content going forward.
Specifically, the companies want a court order preventing future use of their material in AI training. That could have wide implications for the entire industry.

One of the most closely watched copyright battles came from The New York Times Company, which sued OpenAI and Microsoft in late 2023 over claims that their AI systems were trained on Times journalism without permission. The case helped bring global attention to the fight over AI training data and publisher rights.
Since then, more organizations have stepped forward with similar concerns about how their content is used in AI systems. Those lawsuits have turned the issue into one of the defining legal debates around generative AI.
Little-known fact: Experts believe OpenAI would need to scale its annual revenue toward the level of Microsoft’s business (roughly $250 billion per year) by around 2030 to justify a $1 trillion valuation.

The debate is not limited to publishers. Well-known authors like George R. R. Martin and John Grisham have also taken legal action over AI training.
They argue that their creative works have been used without permission. This shows how wide the impact of AI training practices has become.

Britannica’s lawsuit is not happening in isolation. It joins a growing list of legal challenges against AI companies over how they gather training data.
Publishers, authors, and media groups are increasingly questioning whether their work has been used without consent. The outcome of these cases could reshape AI development.

At the center of all these cases is a simple but powerful question. Is training AI on copyrighted material considered fair use or infringement?
Courts have not yet given a clear answer. The decision could define how future AI systems are built and what data they are allowed to use.

The debate has also reached everyday users. Many AI platforms now offer settings that let people control whether their data is used for training.
This shift reflects growing awareness about how personal data and interactions contribute to improving AI models behind the scenes.

Users who want more control can turn off ChatGPT training in Settings by opening Data Controls and disabling “Improve the model for everyone.” OpenAI says that once this setting is off, conversations will not be used to train ChatGPT.
Those conversations can still appear in chat history unless the user starts a Temporary Chat. OpenAI says Temporary Chats are not used for training and are deleted from its systems after 30 days.
Curious how OpenAI is growing so fast? Here’s how it’s defying rivals.

Britannica’s case stands out because it involves highly curated, expert-level content rather than general web data. That could make the legal stakes even higher.
If courts ultimately rule that licensing is required for using such curated material, the implications could reach far beyond this single case and change how AI companies approach data sourcing and content use.
What’s behind OpenAI’s massive stock payouts? Here’s why OpenAI pays a record $1.5M per employee in stock.
What do you think about AI learning from copyrighted content? Share your thoughts.
This slideshow was made with AI assistance and human editing.
Don’t forget to follow us for more exclusive content on MSN.
Read More From This Brand:
This content is exclusive for our subscribers.
Get instant FREE access to ALL of our articles.
Father, tech enthusiast, pilot and traveler. Trying to stay up to date with all of the latest and greatest tech trends that are shaping out daily lives.
We appreciate you taking the time to share your feedback about this page with us.
Whether it's praise for something good, or ideas to improve something that
isn't quite right, we're excited to hear from you.
Stay up to date on all the latest tech, computing and smarter living. 100% FREE
Unsubscribe at any time. We hate spam too, don't worry.

Lucky you! This thread is empty,
which means you've got dibs on the first comment.
Go for it!