Microsoft is facing a lawsuit from a group of high-profile authors who allege the company engaged in “unlawful appropriation” of nearly 200,000 pirated books to train its Megatron artificial intelligence. This legal action signifies a growing pushback from the creative community against what they perceive as the widespread unauthorized use of their intellectual property by AI developers. The authors contend that their copyrighted material was illicitly used to enable the AI to generate human-like responses.
The plaintiffs, including acclaimed writers Kai Bird and Jia Tolentino, are seeking significant financial compensation, up to $150,000 per alleged misuse, and a court injunction to prevent further infringement by Microsoft. The complaint highlights how generative AI systems, which produce diverse outputs, are fundamentally dependent on the massive datasets used during their training phase. The authors explicitly detailed how the pirated dataset contributed to an AI that imitates their writing.
Spokespersons for Microsoft have not yet responded to inquiries regarding the lawsuit, nor has an attorney for the authors offered a comment. This legal development arrives shortly after a California federal judge issued a mixed ruling concerning Anthropic’s use of copyrighted material for AI training, and a separate decision favoring Meta in a similar dispute.
The legal front lines between creators and AI developers are constantly shifting, with a growing number of lawsuits targeting various AI applications. From news organizations like The New York Times and Dow Jones challenging text-based AI, to record labels and visual artists suing over AI-generated music and images, the scope of these disputes is broad. Tech companies consistently argue that their use constitutes fair use and that strict copyright enforcement could hinder AI innovation.
