AI Lawsuits Roundup

A status update on 24 pending lawsuits against AI companies – what they’re about and what is happening in court – prepared by Minnesota copyright attorney Thomas James.

A very brief summary of where pending AI lawsuits stand as of February 28, 2024. Compiled by Minnesota attorney Thomas James.

Thomson Reuters v. Ross, (D. Del. 2020)

Filed May 6, 2020. Thomson Reuters, owner of Westlaw, claims that Ross Intelligence infringed copyrights in Westlaw headnotes by training AI on copies of them. The judge has granted, in part, and denied, in part, motions for summary judgment. The questions of fair use and whether the headnotes are sufficiently original to merit copyright protection will go to a jury to decide.

Thaler v. Perlmutter (D.D.C. 2022).

Complaint filed June 2, 2022. Thaler created an AI system called the Creativity Machine. He applied to register copyrights in the output he generated with it. The Copyright Office refused registration on the ground that AI output does not meet the “human authorship” requirement. He then sought judicial review. The district court granted summary judgment for the Copyright Office. In October, 2023, he filed an appeal to the District of Columbia Circuit Court of Appeals (Case no. 23-5233).

Doe v. GitHub, Microsoft, and OpenAI (N.D. Cal. 2022)

Complaint filed November 3, 2022. Software developers claim the defendants trained Codex and Copilot on code derived from theirs, which they published on GitHub. Some claims have been dismissed, but claims that GitHub and OpenAI violated the DMCA and breached open source licenses remain. Discovery is ongoing.

Andersen v. Stability AI (N.D. Cal. 2023)

Complaint filed January 13, 1023. Visual artists sued Midjourney, Stability AI and DeviantArt for copyright infringement for allegedly training their generative-AI models on images scraped from the Internet without copyright holders’ permission. Other claims included DMCA violations, publicity rights violations, unfair competition, breach of contract, and a claim that output images are infringing derivative works. On October 30, 2023, the court largely granted motions to dismiss, but granted leave to amend the complaint. Plaintiffs filed an amended complaint on November 29, 2023. Defendants have filed motions to dismiss the amended complaint. Hearing on the motion is set for May 8, 2024.

Getty Images v. StabilityAI (U.K. 2023)

Complaint filed January, 2023. Getty Images claims StabilityAI scraped images without its consent. Getty’s complaint has survived a motion to dismiss and the case appears to be heading to trial.

Getty Images v. Stability AI (D. Del.)

Complaint filed February 3, 2023. Getty Images alleges claims of copyright infringement, DMCA violation and trademark violations against Stability AI. The judge has dismissed without prejudice a motion to dismiss or transfer on jurisdictional grounds. The motion may be re-filed after the conclusion of jurisdictional discovery, which is ongoing.

Flora v. Prisma Labs (N.D. Cal.)

Complaint filed February 15, 2023. Plaintiffs allege violations of the Illinois Biometric Privacy Act in connection with Prisma Labs’ collection and retention of users’ selfies in AI training. The court has granted Prisma’s motion to compel arbitration.

Kyland Young v. NeoCortext (C.D. Cal. 2023)

Complaint filed April 3, 2023. This complaint alleges that AI tool Reface used a person’s image without consent, in violation of the person’s publicity rights under California law. The court has denied a motion to dismiss, ruling that publicity rights claims are not preempted by federal copyright law. The case has been stayed pending appeal.

Walters v. OpenAI (Gwinnett County Super. Ct. 2023), and Walters v. OpenAI (N.D. Ga. 2023)

Gwinnett County complaint filed June 5, 2023.

Federal district court complaint filed July 14, 2023.

Radio talk show host sued OpenAI for defamation. A reporter had used ChatGPT to get information about him. ChatGPT wrongly described him as a person who had been accused of fraud. In October, 2023, the federal court remanded the case to the Superior Court of Gwinnett County, Georgia.  On January 11, 2024, the Gwinnett County Superior Court denied OpenAI’s motion to dismiss.

P.M. v. OpenAI (N.D. Cal. 2023).

Complaint filed June 28, 2023. Users claim OpenAI violated the federal Electronic Communications Privacy Act and California wiretapping laws by collecting their data when they input content into ChatGPT. They also claim violations of the Computer Fraud and Abuse Act. Plaintiffs voluntarily dismissed the case on September 15, 2023. See now A.T. v. OpenAI (N.D. Cal. 2023) (below).

In re OpenAI ChatGPT Litigation (N.D. Cal. 2023)

Complaint filed June 28, 3023. Originally captioned Tremblay v. OpenAI. Book authors sued OpenAI for direct and vicarious copyright infringement, DMCA violations, unfair competition and negligence. Both input (training) and output (derivative works) claims are alleged, as well as state law claims of unfair competition, etc. Most state law and DMCA claims have been dismissed, but claims based on unauthorized copying during the AI training process remain. An amended complaint is likely to come in March. The court has directed the amended complaint to consolidate Tremblay v. OpenAI, Chabon v. OpenAI, and Silverman v. OpenAI.  

Battle v. Microsoft (D. Md. 2023)

Complaint filed July 7, 2023. Pro se defamation complaint against Microsoft alleging that Bing falsely described him as a member of the “Portland Seven,” a group of Americans who tried to join the Taliban after 9/11.

Kadrey v. Meta (N.D. Cal. 2023)

Complaint filed July 7, 2023. Sarah Silverman and other authors allege Meta infringed copyrights in their works by making copies of them while training Meta’s AI model; that the AI model is itself an infringing derivative work; and that outputs are infringing copies of their works. Plaintiffs also allege DMCA violations, unfair competition, unjust enrichment, and negligence. The court granted Meta’s motion to dismiss all claims except the claim that unauthorized copies were made during the AI training process. An amended complaint and answer have been filed.

J.L. v. Google (N.D. Cal. 2023)

Complaint filed July 11, 2023. An author filed a complaint against Google alleging misuse of content posted on social media and Google platforms to train Google’s AI Bard. (Gemini is the successor to Google’s Bard.) Claims include copyright infringement, DMCA violations, and others. J.L. filed an amended complaint and Google has filed a motion to dismiss it. A hearing is scheduled for May 16, 2024.

A.T. v. OpenAI (N.D. Cal. 2023)

Complaint filed September 5, 2023. ChatGPT users claim the company violated the federal Electronic Communications Privacy Act, the Computer Fraud and Abuse Act, and California Penal Code section 631 (wiretapping). The gravamen of the complaint is that ChatGPT allegedly accessed users’ platform access and intercepted their private information without their knowledge or consent. Motions to dismiss and to compel arbitration are pending.

Chabon v. OpenAI (N.D. Cal. 2023)

Complaint filed September 9, 2023. Authors allege that OpenAI infringed copyrights while training ChatGPT, and that ChatGPT is itself an unauthorized derivative work. They also assert claims of DMCA violations, unfair competition, negligence and unjust enrichment. The case has been consolidated with Tremblay v. OpenAI, and the cases are now captioned In re OpenAI ChatGPT Litigation.

Chabon v. Meta Platforms (N.D. Cal. 2023)

Complaint filed September 12, 2023. Authors assert copyright infringement claims against Meta, alleging that Meta trained its AI using their works and that the AI model itself is an unauthorized derivative work. The authors also assert claims for DMCA violations, unfair competition, negligence, and unjust enrichment. In November, 2023, the court issued an Order dismissing all claims except the claim of unauthorized copying in the course of training the AI. The court described the claim that an AI model trained on a work is a derivative of that work as “nonsensical.”

Authors Guild v. OpenAI, Microsoft, et al. (S.D.N.Y. 2023)

Complaint filed September 19, 1023. Book and fiction writers filed a complaint for copyright infringement in connection with defendants’ training AI on copies of their works without permission. A motion to dismiss has been filed.

Huckabee v. Bloomberg, Meta Platforms, Microsoft, and EleutherAI Institute (S.D.N.Y. 2023)

Complaint filed October 17, 2023. Political figure Mike Huckabee and others allege that the defendants trained AI tools on their works without permission when they used Books3, a text dataset compiled by developers; that their tools are themselves unauthorized derivative works; and that every output of their tools is an infringing derivative work.  Claims against EleutherAI have been voluntarily dismissed. Claims against Meta and Microsoft have been transferred to the Northern District of California. Bloomberg is expected to file a motion to dismiss soon.

Huckabee v. Meta Platforms and Microsoft (N.D. Cal. 2023)

Complaint filed October 17, 2023. Political figure Mike Huckabee and others allege that the defendants trained AI tools on their works without permission when they used Books3, a text dataset compiled by developers; that their tools are themselves unauthorized derivative works; and that every output of their tools is an infringing derivative work. Plaintiffs have filed an amended complaint. Plaintiffs have stipulated to dismissal of claims against Microsoft without prejudice.

Concord Music Group v. Anthropic (M.D. Tenn. 2023)

Complaint filed October 18, 2023. Music publishers claim that Anthropic infringed publisher-owned copyrights in song lyrics when they allegedly were copied as part of an AI training process (Claude) and when lyrics were reproduced and distributed in response to prompts. They have also made claims of contributory and vicarious infringement. Motions to dismiss and for a preliminary injunction are pending.

Alter v. OpenAI and Microsoft (S.D.N.Y. 2023)

Complaint filed November 21, 2023. Nonfiction author alleges claims of copyright infringement and contributory copyright infringement against OpenAI and Microsoft, alleging that reproducing copies of their works in datasets used to train AI infringed copyrights. The court has ordered consolidation of Author’s Guild (23-cv-8292) and Alter (23-cv-10211). On February 12,2024, plaintiffs in other cases filed a motion to intervene and dismiss.

New York Times v. Microsoft and OpenAI (S.D.N.Y. 2023)

Complaint filed December 27, 2023. The New York Times alleges that their news stories were used to train AI without a license or permission, in violation of their exclusive rights of reproduction and public display, as copyright owners. The complaint also alleges vicarious and contributory copyright infringement, DMCA violations, unfair competition, and trademark dilution. The Times seeks damages, an injunction against further infringing conduct, and a Section 503(b) order for the destruction of “all GPT or other LLM models and training sets that incorporate Times Works.” On February 23, 2024, plaintiffs in other cases filed a motion to intervene and dismiss this case.  

Basbanes and Ngagoyeanes v. Microsoft and OpenAI (S.D.N.Y. 2024)

Complaint filed January 5, 2024. Nonfiction authors assert copyright claims against Microsoft and OpenAI. On February 6, 2024, the court consolidated this case with Authors Guild (23-cv-08292) and Alter v. Open AI (23-cv-10211), for pretrial purposes.  

Caveat

This list is not exhaustive. There may be other cases involving AI that are not included here. For a discussion of bias issues in Google’s Gemini, have a look at Scraping Bias on Medium.com.

Nontransformative Nuge

A reversal in the 4th Circuit Court demonstrates the impact the Supreme Court’s decision in Andy Warhol Foundation for the Arts v. Goldsmith is already having on the application of copyright fair use doctrine in federal courts.

Philpot v. Independent Journal Review, No. 21-2021 (4th Circ., Feb. 6, 2024)

Philpot, a concert photographer, registered his photograph of Ted Nugent as part of a group of unpublished works. Prior to registration, he entered into a license agreement giving AXS TV the right to inspect his photographs for the purpose of selecting ones to curate. The agreement provided that the license would become effective upon delivery of photographs for inspection. After registration, Philpot delivered a set of photographs, including the Nugent photograph, to AXS TV. He also published the Nugent photograph to Wikimedia Commons under a Creative Commons (“CC”) license. The CC license allows free use on the condition that attribution is given. LJR published an article called “15 Signs Your Daddy Was a Conservative.” Sign #5 was He hearts the Nuge. LJR used Philpot’s photograph of Ted Nugent as an illustration for the article, without providing an attribution of credit to Philpot.

Philpot sued IJR for copyright infringement.  IJR asserted two defenses: (1) invalid copyright registration; and (2) fair use. The trial court did not decide whether the registration was valid or not, but it granted summary judgment for IJR based on its opinion that the news service’s publication of the photograph was fair use. The Fourth Circuit Court of Appeals reversed, ruling in Philpot’s favor on both issues. The Court held that the copyright registration was valid and that publication of the photograph without permission was not fair use.

The copyright registration

Published and unpublished works cannot be registered together. Including a published work in an application for registration of a group of unpublished works is an inaccuracy that might invalidate the registration, if the applicant was aware of the inaccuracy at the time of applying. Cf. Unicolors v. H&M Hennes & Mauritz, 595 U.S. 178 (2022). LJR argued that Philpot’s pre-registration agreement to send photographs to AJX TV to inspect for possible curation constituted “publication” of them so characterizing them as “unpublished” in the registration application was an inaccuracy known to Philpot.

17 U.S.C. § 101 defines publication as “the distribution of copies . . . to the public” or “offering to distribute copies . . . to a group of persons for purposes of further distribution . . . or public display.” The Court of Appeals held that merely entering into an agreement to furnish copies to a distributor for possible curation does not come within that definition. Sending copies to a limited class of people without concomitantly granting an unrestricted right to further distribute them to the public does not amount to “publication.”

Philpot’s arrangement with AXS TV is analogous to an author submitting a manuscript to a publisher for review for possible future distribution to the public. The U.S. Copyright Office has addressed this. “Sending copies of a manuscript to prospective publishers in an effort to secure a book contract does not [constitute publication].” U.S. Copyright Office, Compendium of U.S. Copyright Office Practices § 1905.1 (3d ed. 2021). Philpot had provided copies of his work for the limited purpose of examination, without a present grant of a right of further distribution. Therefore, the photographs were, in fact, unpublished at the time of the application for registration. Since no inaccuracy existed, the registration was valid.

Fair use

The Court applied the four-factor test for fair use set out in 17 U.S.C. § 107.

(1) Purpose and character of the use. Citing Andy Warhol Found. For the Visual Arts v. Goldsmith, 598 U.S. 508 , 527–33 (2023), the Court held that when, as here, a use is neither transformative nor noncommercial, this factor weighs against a fair use determination. LJR used the photograph for the same purpose as Philpot intended to use it (as a depiction of Mr. Nugent), and it was a commercial purpose.

(2) Nature of the work. Photographs taken by humans are acts of creative expression that receive what courts have described as “thick” copyright protection.” Therefore, this factor weighed against a fair use determination.

(3) Amount and substantiality of the portion used. Since all of the expressive features of the work were used, this factor also weighed against a fair use determination.

(4) Effect on the market for the work. Finally, the Court determined that allowing free use of a copyrighted work for commercial purposes without the copyright owner’s permission could potentially have a negative impact on the author’s market for the work. Therefore, this factor, too, weighed against a fair use determination.

Since all four factors weighed against a fair use determination, the Court reversed the trial court’s grant of summary judgment to IJR and remanded the case for further proceedings.

Conclusion

This decision demonstrates the impact the Warhol decision is having on copyright fair use analysis in the courts. Previously, courts had been interpreting transformativeness very broadly. In many cases, they were ending fair use inquiry as soon as some sort of transformative use could be articulated. As the Court of Appeals decision in this case illustrates, trial courts now need to alter their approach in two ways: (1) They need to return to considering all four fair use factors rather than ending the inquiry upon a defendant’s articulation of some “transformative use;” and (2) They need to apply a much narrower definition of transformativeness than they have been. If both the original work and an unauthorized reproduction of it are used for the purpose of depicting a particular person or scene (as distinguished from parodying or commenting on a work, for example), for commercial gain, then it would no longer appear to be prudent to count on the first of the four fair use factors supporting a fair use determination.


Photo: Photograph published in a July, 1848 edition of L’Illustration. Believed to be the first instance of photojournalism, it is now in the public domain.

A copyright win in the Internet Archive lawsuit

Book publishers have won their lawsuit against the Internet Archive. What does it mean for copyright owners? Cokato Copyright Attorney Tom James explains.

New York Public Library public domain image; this is NOT the library that is involved in this lawsuit.

In a previous blog post, I wrote about a lawsuit that book publishers Hachette Book Group et al. have filed against Internet Archive (“IA”) et al. The lawsuit alleges that IA scans copyright-protected printed books into a digital format, uploads them to its servers, and distributes these digital copies to members of the public via a website – all without a license and without paying the authors and publishers. The lawsuit asserts claims of copyright infringement.

A permanent injunction

Judge John Koeltl has now approved a consent judgment providing for a permanent injunction that prohibits Internet Archive from scanning and distributing copyrighted books. It applies only to books that copyright owners have already published and made available in e-book format. As Judge Koeltl put it: “The Court has narrowly tailored the injunctive relief in this case to cover only copyrighted works, like the Works in Suit, that are available from the Publishers in electronic form.”

IA reportedly plans to appeal.

A new lawsuit

This month, Sony Music Entertainment et al. filed a similar copyright infringement lawsuit against IA. This complaint alleges that IA digitized and distributed digital copies of 78 rpm records by Frank Sinatra, Billie Holiday, and other recording artists in violation of the rights of copyright owners.

Generative-AI: The Top 12 Lawsuits

Artificial intelligence (“AI”) is generating more than content; it is generating lawsuits. Here is a brief chronology of what I believe are the most significant lawsuits that have been filed so far.

Artificial intelligence (“AI”) is generating more than content; it is generating lawsuits. Here is a brief chronology of what I believe are the most significant lawsuits that have been filed so far.

Most of these allege copyright infringement, but some make additional or other kinds of claims, such as trademark, privacy or publicity right violations, defamation, unfair competition, and breach of contract, among others. So far, the suits primarily target the developers and purveyors of generative AI chatbots and similar technology. They focus more on what I call “input” than on “output” copyright infringement. That is to say, they allege that copyright infringement is involved in the way particular AI tools are trained.

Thomson Reuters Enterprise Centre GmbH et al. v. ROSS Intelligence (May, 2020)

Thomson Reuters Enterprise Centre GmbH et al. v. ROSS Intelligence Inc., No. 20-cv-613 (D. Del. 2020)

Thomson Reuters alleges that ROSS Intelligence copied its Westlaw database without permission and used it to train a competing AI-driven legal research platform. In defense, ROSS has asserted that it only copied ideas and facts from the Westlaw database of legal research materials. (Facts and ideas are not protected by copyright.) ROSS also argues that its use of content in the Westlaw database is fair use.

One difference between this case and subsequent generative-AI copyright infringement cases is that the defendant in this case is alleged to have induced a third party with a Westlaw license to obtain allegedly proprietary content for the defendant after the defendant had been denied a license of its own. Other cases involve generative AI technologies that operate by scraping publicly available content.

The parties filed cross-motions for summary judgment. While those motions were pending, the U.S. Supreme Court issued its decision in Andy Warhol Found. for the Visual Arts, Inc. v. Goldsmith, 598 U.S. ___, 143 S. Ct. 1258 (2023). The parties have now filed supplemental briefs asserting competing arguments about whether and how the Court’s treatment of transformative use in that case should be interpreted and applied in this case. A decision on the motions is expected soon.

Doe 1 et al. v. GitHub et al. (November, 2022)

Doe 1 et al. v. GitHub, Inc. et al., No. 22-cv-06823 (N.D. Calif. November 3, 2022)

This is a class action lawsuit against GitHub, Microsoft, and OpenAI that was filed in November, 2022. It involves GitHub’s CoPilot, an AI-powered tool that suggests lines of programming code based on what a programmer has written. The complaint alleges that Copilot copies code from publicly available software repositories without complying with the terms of applicable open-source licenses. The complaint also alleges removal of copyright management information in violation of 17 U.S.C. § 1202, unfair competition, and other tort claims.

Andersen et al. v. Stability AI et al. (January 13, 2023)

Andersen et al. v. Stability AI et al., No. 23-cv-00201 (N.D. Calif. Jan. 13, 2023)

Artists Sarah Andersen, Kelly McKernan, and Karla Ortiz filed this class action lawsuit against generative-AI companies Stability AI, Midjourney, and DeviantArt on January 13, 2023. The lawsuit alleges that the defendants infringed their copyrights by using their artwork without permission to train AI-powered image generators to create allegedly infringing derivative works.  The lawsuit also alleges violations of 17 U.S.C. § 1202 and publicity rights, breach of contract, and unfair competition.

Getty Images v. Stability AI (February 3, 2023)

Getty Images v. Stability AI, No. 23-cv-00135-UNA (D. Del. February 23, 2023)

Getty Images has filed two lawsuits against Stability AI, one in the United Kingdom and one in the United States, each alleging both input and output copyright infringement. Getty Images owns the rights to millions of images. It is in the business of licensing rights to use copies of the images to others. The lawsuit also accuses Stability AI of falsifying, removing or altering copyright management information, trademark infringement, trademark dilution, unfair competition, and deceptive trade practices.

Stability AI has moved to dismiss the complaint filed in the U.S. for lack of jurisdiction.

Flora et al. v. Prisma Labs (February 15, 2023)

Flora et al. v. Prisma Labs, Inc., No. 23-cv-00680 (N.D. Calif. February 15, 2023)

Jack Flora and others filed a class action lawsuit against Prisma Labs for invasion of privacy. The complaint alleges, among other things, that the defendant’s Lensa app generates sexualized images from images of fully-clothed people, and that the company failed to notify users about the biometric data it collects and how it will be stored and/or destroyed, in violation of Illinois’s data privacy laws.

Young v. NeoCortext (April 3, 2023)

Young v. NeoCortext, Inc., 2023-cv-02496 (C.D. Calif. April 3, 2023)

This is a publicity rights case. NeoCortext’s Reface app allows users to paste images of their own faces over those of celebrities in photographs and videos. Kyland Young, a former cast member of the Big Brother reality television show, has sued NeoCortext for allegedly violating his publicity rights. The complaint alleges that NeoCortext has “commercially exploit[ed] his and thousands of other actors, musicians, athletes, celebrities, and other well-known individuals’ names, voices, photographs, or likenesses to sell paid subscriptions to its smartphone application, Refacewithout their permission.”

NeoCortext has asserted a First Amendment defense, among others.

Walters v. Open AI (June 5, 2023)

Walters v. OpenAI, LLC, No. 2023-cv-03122 (N.D. Ga. July 14, 2023) (Complaint originally filed in Gwinnett County, Georgia Superior Court on June 5, 2023; subsequently removed to federal court)

This is a defamation action against OpenAI, the company responsible for ChatGPT. The lawsuit was brought by Mark Walters. He alleges that ChatGPT provided false and defamatory misinformation about him to journalist Fred Riehl in connection with a federal civil rights lawsuit against Washington Attorney General Bob Ferguson and members of his staff. ChatGPT allegedly stated that the lawsuit was one for fraud and embezzlement on the part of Mr. Walters. The complaint alleges that Mr. Walters was “neither a plaintiff nor a defendant in the lawsuit,” and “every statement of fact” pertaining to him in the summary of the federal lawsuit that ChatGPT prepared is false. A New York court recently addressed the questions of sanctions for attorneys who submit briefs containing citations to non-existent “precedents” that were entirely made up by ChatGPT. This is the first case to address tort liability for ChatGPT’s notorious creation of “hallucinatory facts.”

In July, 2023, Jeffery Battle filed a complaint against Microsoft in Maryland alleging that he, too, has been defamed as a result of AI-generated “hallucinatory facts.”

P.M. et al. v. OpenAI et al. (June 28, 2023)

P.M. et al. v. OpenAI LP et al., No. 2023-cv-03199 (N.D. Calif. June 28, 2023)

This lawsuit has been brought by underage individuals against OpenAI and Microsoft. The complaint alleges the defendants’ generative-AI products ChatGPT, Dall-E and Vall-E collect private and personally identifiable information from children without their knowledge or informed consent. The complaint sets out claims for alleged violations of the Electronic Communications Privacy Act; the Computer Fraud and Abuse Act; California’s Invasion of Privacy Act and unfair competition law; Illinois’s Biometric Information Privacy Act, Consumer Fraud and Deceptive Business Practices Act, and Consumer Fraud and Deceptive Business Practices Act; New York General Business Law § 349 (deceptive trade practices); and negligence, invasion of privacy, conversion, unjust enrichment, and breach of duty to warn.

Tremblay v. OpenAI (June 28, 2023)

Tremblay v. OpenAI, Inc., No. 23-cv-03223 (N.D. Calif. June 28, 2023)

Another copyright infringement lawsuit against OpenAI relating to its ChatGPT tool. In this one, authors allege that ChatGPT is trained on the text of books they and other proposed class members authored, and facilitates output copyright infringement. The complaint sets forth claims of copyright infringement, DMCA violations, and unfair competition.

Silverman et al. v. OpenAI (July 7, 2023)

Silverman et al. v. OpenAI, No. 23-cv-03416 (N.D. Calif. July 7, 2023)

Sarah Silverman (comedian/actress/writer) and others allege that OpenAI, by using copyright-protected works without permission to train ChatGPT, committed direct and vicarious copyright infringement, violated section 17 U.S.C. 1202(b), and their rights under unfair competition, negligence, and unjust enrichment law.

Kadrey et al. v. Meta Platforms (July 7, 2023)

Kadrey et al. v. Meta Platforms, No. 2023-cv-03417 (N.D. Calif. July 7, 2023)

The same kinds of allegations as are made in Silverman v. OpenAI, but this time against Meta Platforms, Inc.

J.L. et al. v. Alphabet (July 11, 2023)

J.L. et al. v. Alphabet, Inc. et al. (N.D. Calif. July 11, 2023)

This is a lawsuit against Google and its owner Alphabet, Inc. for allegedly scraping and harvesting private and personal user information, copyright-protected works, and emails, without notice or consent. The complaint alleges claims for invasion of privacy, unfair competition, negligence, copyright infringement, and other causes of action.

On the Regulatory Front

The U.S. Copyright Office is examining the problems associated with registering copyrights in works that rely, in whole or in part, on artificial intelligence. The U.S. Federal Trade Commission (FTC) has suggested that generative-AI implicates “competition concerns.”. Lawmakers in the United States and the European Union are considering legislation to regulate AI in various ways.

Copyright owners prevail in Internet Archive lawsuit

A federal district court has ruled in favor of book publishers in their copyright infringement lawsuit against Internet Archives

In June, 2020 four book publishers filed a copyright infringement lawsuit against Internet Archive. The publishers asserted that the practice of scanning books and lending digital copies of them to online users infringed their copyrights in the books. On Friday, March 24, 2023, a federal district court judge agreed, granting the publishers’ motion for summary judgment.

The Internet Archive operation

Internet Archive is a nonprofit organization that has undertaken several archiving projects. For example, it created the “Wayback Machine,” an online archive of public webpages. This lawsuit involves another of its projects, namely, the creation of a digital archive of books. Some of these are in the public domain. Also included in this archive, however, are over 3 million books that are protected by copyright. The judge determined that 33,000 of them belong to the plaintiffs in the lawsuit.

According to the Order granting summary judgment, after scanning the books, Internet Archive made them publicly available online for free, without the permission of the copyright owners.

“Fair Use”

According to the Order, Internet Archive did not dispute that it violated copyright owners’ exclusive rights to reproduce the works, to make derivative works based on them, to distribute their works, to publicly perform them (Internet Archive offered a “read aloud” function on it website), and to display them (in this case, on a user’s browser.) In short, the Order determined that the operation violated all five of the exclusive rights of copyright owners protected by the United States Copyright Act (17 U.S.C. sec. 106).

Internet Archive asserted a “fair use” defense.

In previous cases involving massive operations to scan and digitize millions of books, Authors Guild v. Google., Inc. and Authors Guild v. HathiTrust, judicial analyses resulted in “fair use” determinations unfavorable to copyright owners. Internet Archive, of course, invited the judge to do the same thing here. The judge declined the invitation.

The judge distinguished this case from its predecessors by ruling that unlike the uses made of copyrighted works in those cases, the use in this case was not transformative. For example, Google had digitized the entire text of books in order to create a searchable index of books. “There is nothing transformative,” however, about copying and distributing the entire texts of books to the public, the judge declared.

The judge observed that Google reproduces and displays to the public only enough context surrounding the searched term to help a reader evaluate whether the book falls within the range of the reader’s interest. The Court of Appeals in Google had warned that “[i]f Plaintiff’s claim were based on Google’s converting their books into a digitized form and making that digitized version accessible to the public,” then the “claim [of copyright infringement] would be strong.”

The judge also determined that the alleged benefit to the public of having access to the entire text of books without having to pay for them “cannot outweigh the market harm to the publishers.”

Ultimately, the judge concluded that all four “fair use” factors (character and purpose of the use, nature of the work, amount and substantiality of the portion copied, and the effect on the market for the work) weighed against a finding of fair use.

What’s next?

Internet Archive apparently intends to appeal the decision. In the meantime, it appears that it will continue other kinds of digitized book services, such as interlibrary loans, citation linking, access for the print-disabled , text and data mining, purchasing e-books, and receiving and preserving books.

Why Machine Training AI with Protected Works is Not Fair Use

… if the underlying goal of copyright’s exclusive rights and the fair use exception is to promote new “authorship,” this is doctrinally fatal to the proposal that training AIs on volumes of protected works favors a finding of fair use.

Guest blogger David Newhoff lays out the argument against the claim that training AI systems with copyright-protected works is fair use. David is the author of Who Invented Oscar Wilde? The Photograph at the Center of Modern American Copyright (Potomac Books 2020) and is a copyright advocate/writer at The Illusion of More.


As most copyright watchers already know, two lawsuits were filed at the start of the new year against AI visual works companies. In the U.S., a class-action was filed by visual artists against DeviantArt, Midjourney, and Stability AI; and in the UK, Getty Images is suing Stability AI. Both cases allege infringing use of large volumes of protected works fed into the systems to “train” the algorithms. Regardless of how these two lawsuits might unfold, I want to address the broad defense, already being argued in the blogosphere, that training generative AIs with volumes of protected works is fair use. I don’t think so.

Copyright advocates, skeptics, and even outright antagonists generally agree that the fair use exception, correctly applied, supports the broad aim of copyright law to promote more creative work. In the language of the Constitution, copyright “promotes the progress of science,” but a more accurate, modern description would be that copyright promotes new “authorship” because we do not tend to describe literature, visual arts, music, etc. as “science.”

The fair use doctrine, codified in the federal statute in 1976, originated as judge-made law, and from the seminal Folsom v. Marsh to the contemporary Andy Warhol Foundation v. Goldsmith, the courts have restated, in one way or another, their responsibility to balance the first author’s exclusive rights with a follow-on author’s interest in creating new expression. And as a matter of general principle, it is held that the public benefits from this balancing act because the result is a more diverse market of creative and cultural works.

Fair use defenses are case-by-case considerations and while there may be specific instances in which an AI purpose may be fair use, there are no blanket exceptions. More broadly, though, if the underlying goal of copyright’s exclusive rights and the fair use exception is to promote new “authorship,” this is doctrinally fatal to the proposal that training AIs on volumes of protected works favors a finding of fair use. Even if a court holds that other limiting doctrines render this activity by certain defendants to be non-infringing, a fair use defense should be rejected at summary judgment—at least for the current state of the technology, in which the schematic encompassing AI machine, AI developer, and AI user does nothing to promote new “authorship” as a matter of law.

The definition of “author” in U.S. copyright law means “human author,” and there are no exceptions to this anywhere in our history. The mere existence of a work we might describe as “creative” is not evidence of an author/owner of that work unless there is a valid nexus between a human’s vision and the resulting work fixed in a tangible medium. If you find an anonymous work of art on the street, absent further research, it has no legal author who can assert a claim of copyright in the work that would hold up in any court. And this hypothetical emphasizes the point that the legal meaning of “author” is more rigorous than the philosophical view that art without humans is oxymoronic. (Although it is plausible to find authorship in a work that combines human creativity with AI, I address that subject below.)

As a matter of law, the AI machine itself is disqualified as an “author” full stop. And although the AI owner/developer and AI user/customer are presumably both human, neither is defensibly an “author” of the expressions output by the AI. At least with the current state of technologies making headlines, nowhere in the process—from training the AI, to developing the algorithm, to entering prompts into the system—is there an essential link between those contributions and the individual expressions output by the machine. Consequently, nothing about the process of ingesting protected works to develop these systems in the first place can plausibly claim to serve the purpose of promoting new “authorship.”

But What About the Google Books Case?

Indeed. In the fair use defenses AI developers will present, we should expect to see them lean substantially on the holding in Authors Guild v. Google Books—a decision which arguably exceeds the purpose of fair use to promote new authorship. The Second Circuit, while acknowledging that it was pushing the boundaries of fair use, found the Google Books tool to be “transformative” for its novel utility in presenting snippets of books; and because that utility necessitates scanning whole books into its database, a defendant AI developer will presumably want to make the comparison. But a fair use defense applied to training AIs with volumes of protected works should fail, even under the highly utilitarian holding in Google Books.

While people of good intent can debate the legal merits of that decision, the utility of the Google Books search engine does broadly serve the interest of new authorship with a useful research tool—one I have used many times myself. Google Books provides a new means by which one author may research the works of another author, and this is immediately distinguishable from the generative AI which may be trained to “write books” without authors. Thus, not only does the generative AI fail to promote authorship of the individual works output by the system, but it fails to promote authorship in general.

Although the technology is primitive for the moment, these AIs are expected to “learn” exponentially and grow in complexity such that AIs will presumably compete with or replace at least some human creators in various fields and disciplines. Thus, an enterprise which proposes to diminish the number of working authors, whether intentionally or unintentionally, should only be viewed as devastating to the purpose of copyright law, including the fair use exception.

AI proponents may argue that “democratizing” creativity (i.e., putting these tools in every hand) promotes authorship by making everyone an author. But aside from the cultural vacuum this illusion of more would create, the user prompting the AI has a high burden to prove authorship, and it would really depend on what he is contributing relative to the AI. As mentioned above, some AIs may evolve as tools such that the human in some way “collaborates” with the machine to produce a work of authorship. But this hypothetical points to the reason why fair use is a fact-specific, case-by-case consideration. AI Alpha, which autonomously creates, or creates mostly without human direction, should not benefit from the potential fair use defense of AI Beta, which produces a tool designed to aid, but not replace, human creativity.

Broadly Transformative? Don’t Even Go There

Returning to the constitutional purpose of copyright law to “promote science,” the argument has already been floated as a talking point that training AI systems with protected works promotes computer science in general and is, therefore, “transformative” under fair use factor one for this reason. But this argument should find no purchase in court. To the extent that one of these neural networks might eventually spawn revolutionary utility in medicine or finance etc., it would be unsuitable to ask a court to hold that such voyages of general discovery fit the purpose of copyright, to say nothing of the likelihood that the adventure strays inevitably into patent law. Even the most elastic fair use findings to date reject such a broad defense.

It may be shown that no work(s) output by a particular AI infringes (copies) any of the works that went into its training. It may also be determined that the corpus of works fed into an AI is so rapidly atomized into data that even fleeting “reproduction” is found not to exist, and, thus, the 106(1) right is not infringed. Those questions are going to be raised in court before long, and we shall see where they lead. But to presume fair use as a broad defense for AI “training” is existentially offensive to the purpose of copyright, and perhaps to law in general, because it asks the courts to vest rights in non-humans, which is itself anathema to caselaw in other areas.[1]

It is my oft-stated opinion that creative expression without humans is meaningless as a cultural enterprise, but it is a matter of law to say that copyright is meaningless without “authors” and that there is no such thing as non-human “authors.” For this reason, the argument that training AIs on protected works is inherently fair use should be denied with prejudice.


[1] Cetaceans v. Bush holding that animals do not have standing in court was the basis for rejecting PETA’S complaint against photographer Slater for infringing the copyright rights of the monkey in the “Monkey Selfie” fiasco.


Getty Images Litigation Update

Getty Images has now filed a lawsuit for copyright infringement in the United States.

In a previous post, I reported on a lawsuit that Getty Images had filed in the United Kingdom against Stability AI. Now the company has filed similar claims against the company in the United States.

The complaint, which has been filed in federal district court in Delaware, alleges claims of copyright infringement; providing false copyright management information; removal or alteration of copyright management information; trademark infringement; trademark dilution; unfair competition; and deceptive trade practices. Both monetary damages and injunctive relief are being sought.

An interesting twist in the Getty litigation is that AI-generated works allegedly have included the Getty Images trademark.

Getty Images logo on AI-generated image
(Reproduction of a portion of the Complaint filed in Getty Images v. Stability AI, Inc. in U.S. district court for the district of Delaware, case no. Case 1:23-cv-00135-UNA (2023). The image has been cropped to avoid reproducing the likenesses of persons appearing in the image and to display only what is needed here for purposes of news reporting and commentary,)

Getty Images, which is in the business of collecting and licensing quality images, alleges (among other things) that affixing its trademark to poor quality AI-generated images tarnishes the company’s reputation. If proven, this could constitute trademark dilution, which is prohibited by the Lanham Act.

Court agrees to hear parody goods case

The U.S. Supreme Court will hear Jack Daniel’s v. VIP Products, the “dog toy” trademark case. Cokato Copyright Attorney Thomas James explains.

In my last blog post (“MSCHF Testing the Limits of Free Speech“) I wrote about the Wavy Baby Shoes case in the Second Circuit Court of Appeals. Now, in a different case raising similar issues, the United States Supreme Court will have an opportunity to resolve a circuit split on the parody goods question. On Monday, the Court granted certiorari in Jack Daniel’s Properties v. VIP Products.

The “Bad Spaniels” Dog Toy

Jack Daniel’s Products claims trademark rights, including trade dress, in the distinctive shape and label of its whiskey product. VIP Products has made a dog toy called “Bad Spaniels.” It has a very similar shape and label. Jack Daniel’s sued VIP, asserting trademark infringement and dilution claims.

The district court ruled in favor of Jack Daniel’s, finding that the dog toy was likely to confuse consumers about the source of the product and tarnish the Jack Daniel’s brand. The Ninth Circuit Court of Appeals however, reversed. The dog toys, the Court ruled, are parody goods protected as expression by the First Amendment.

The Second, Seventh and Eighth Circuit Courts of Appeals have taken different approaches regarding the scope of First Amendment protection for parody goods. The Supreme Court granted certiorari to resolve the split in the circuits.

The Trademark Dilution Revision Act

The Trademark Dilution Revision Act expressly excludes parody from dilution liability. It applies, however, only when the challenged use is “other than as a designation of source for the person’s own goods or services.” 15 U.S.C. § 1125(c)(3)(A)(ii). Thus, the Act covers parody advertisements or other parodic references to a product in a magazine, movie or other traditional form of artistic or literary expression, but it does not reach situations where a parody mark is used as a designation of source.

Rogers v. Grimaldi

As discussed in a previous blog post, the Second Circuit Court of Appeals ruled, in Rogers v. Grimaldi, 875 F.2d 994 (2d Cir. 1989), that the use of a trademark in the title of a film or other artistic work is not actionable unless “the title has no artistic relevance to the underlying work whatsoever, or, if it has some artistic relevance, unless the title explicitly misleads as to the source or the content of the work.” Id. at 999.

In the Jack Daniel’s case, the district court held that the rule articulated in Rogers is limited to the use of a trademark in the title of a film or other artistic work. The Ninth Circuit Court of Appeals did not agree that the rule is so limited.

The case squarely raises the question whether and to what extent the First Amendment insulates makers and sellers of parody goods from trademark liability.

Conclusion

A decision in this case will be coming in a few months. It is difficult to predict what it will be. Many legal scholars, I am sure, will frame the issue as whether the Court will choose to extend Grimaldi to parody goods, on one hand, or to limit the case to its facts, on the other. Another possibility, however, is that the Court might choose to disapprove Grimaldi altogether. As usual, we will just have to wait and see.

The Top Copyright Cases of 2021

by Minnesota attorney Thomas James

I initially had set out to put together a “Top 10” list. Really, though, I think the list can be boiled down to three. Admittedly, this is only my personal opinion. Time will tell. Nevertheless, for what it’s worth, here is my list of the 3 Top Copyright Cases of 2021.

Google v. Oracle America

Google v. Oracle America, __ U.S. __, 141 S. Ct. 1183 (2021)

This United States Supreme Court decision is the culmination of many years of litigation between tech giants Google and Oracle.

At issue was Google’s copying of 11,500 lines of code of the Java SE API. Illustrating the murkiness of the “fair use” concept, the United States Supreme Court declared that this was fair use.

The case highlights the relatively weak protection that copyright offers for computer programs. The functional aspects of a computer program are better protected by patent than copyright.

It is dangerous to read too much into the decision, though. It does not mean that computer program copyrights are worthless To the contrary, the case was decided on the basis of fair use. Google’s copying of the code was infringement. “Fair use” simply means that a court came to the conclusion that a particular defendant should not be held liable for a particular kind or instance of infringement. Another court could come to a different conclusion in a different case involving different parties, a different kind of computer program, and a different kind of use of it.

Warhol v. Goldsmith

Andy Warhol Foundation for the Visual Arts v. Goldsmith, No. 19-2420 (2nd Cir. 2021).

This case is notable primarily because of the celebrities involved. Lynn Goldsmith took a photograph of Prince in her studio in 1981. Andy Warhol created a series of silkscreen prints and pencil illustrations based on it. Goldsmith sued for infringement of the copyright in the photograph. The district court found in favor of Warhol, citing the transformative use doctrine. The Court of Appeals reversed, asserting that the district court misapplied the four “fair use” factors.

Reversals of “fair use” findings on appeal are not uncommon. They illustrate the nebulous nature of the four-factor test that courts use to evaluate fair use claims.

Design Basics v. Signature Construction

Design Basics v. Signature Construction, No. 19-2716 (7th Cir. 2021).

Design Basics holds registered copyrights in thousands of floor plans for single-family homes. The company attempts to secure “prompt settlements” of infringement claims. The court ruled against the company on an infringement claim, finding that these designs consisted mainly of unprotectable stock elements, much of which were dictated by functional considerations and existing design considerations.

Architectural designs are protected by copyright, but the protection is thin. Only a “strikingly similar” work can give risk to an infringement claim. In other words, infringement of an architectural work requires a showing of extremely close copying.

Need help with a copyright registration or a copyright matter? Contact the Cokato Copyright Attorney Tom James.