A prominent investigative journalist has initiated a class action lawsuit against the grammar-checking software company Grammarly, alleging the firm used her writing and that of other authors to train its artificial intelligence models without proper consent. The legal action, filed in a United States federal court, centers on claims of privacy and publicity rights violations.
Core Allegations in the Lawsuit
Julia Angwin, a well-known journalist and editor, is the lead plaintiff in the case. The lawsuit contends that Grammarly scraped vast amounts of copyrighted text from the internet, including Angwin’s published articles, to develop and refine its AI-powered writing assistance tools. This alleged data collection and usage occurred without obtaining permission from, or providing compensation to, the original authors.
The legal filing argues this practice violates the plaintiffs’ right of publicity, which protects against the unauthorized commercial use of an individual’s name or likeness. It also raises significant concerns about digital privacy, suggesting the company’s data harvesting methods were overly broad and intrusive.
Grammarly’s Business and the AI Training Process
Grammarly is a widely used digital writing assistant that employs advanced algorithms to check for grammatical errors, suggest style improvements, and enhance clarity. Like many other companies in the generative AI sector, its underlying models require massive datasets of human-written text to learn language patterns and rules.
The core dispute lies in how the company allegedly acquired this training data. The plaintiffs claim the method constitutes an unlawful appropriation of their intellectual property and personal creative output for corporate profit.
Broader Context for Legal Action
This lawsuit emerges amid a growing wave of similar legal challenges against major technology firms. Numerous authors, artists, and media companies have filed suits against AI developers, including OpenAI and Meta, alleging copyright infringement through the unauthorized use of copyrighted works for AI training.
These cases are testing the boundaries of existing copyright and privacy laws in the age of generative AI. The outcomes could establish important legal precedents regarding fair use, data scraping, and creator compensation in the digital economy.
Potential Implications and Next Steps
A ruling in favor of the plaintiffs could compel Grammarly and similar companies to alter their data collection practices significantly. Potential consequences include the requirement to obtain explicit licenses for training data, implement robust opt-out mechanisms for content creators, or potentially delete models trained on disputed data.
For writers and journalists worldwide, the case highlights ongoing tensions between technological innovation and the protection of creative rights. The litigation is expected to proceed through preliminary motions, with both sides presenting arguments on the applicability of copyright and publicity rights to AI training datasets. The court’s decisions on these early motions will likely shape the trajectory and potential settlement discussions of the case.
Source: Adapted from multiple news reports.