Fair Use and Model Training, Getty Images vs. Stable Diffusion, Adobe’s Regulatory Capture Bet

Good morning,

Last week I guest-hosted Sharp China with Bill Bishop; we discussed the meeting between Taiwan President Tsai Ing-wen and Speaker Kevin McCarthy, China’s military exercises in response, and gauged the mood on the ground in Taiwan. We also looked ahead to the upcoming Taiwan presidential election, and discussed China’s proposed AI rules.

Yesterday on Sharp Tech we reacted to my interview with Adobe Chief Strategy Office Scott Belsky and discussed the future of content in an AI world. Meanwhile, Greatest of All Talk covered the opening weekend of the NBA playoffs. You can listen to all of these podcasts at the above links, or by adding the shows to your podcast player using the links at the bottom of this email.

On to the update:

Fair Use and Model Training

I apologize that yesterday’s Article AI, NIL, and Zero Trust Authenticity came out so late: this was one of those Articles that sprawled to something massively larger than I originally intended, and I had to cut it down substantially; after all, I have today’s Update to touch on some of the topics I originally intended to include.

To that end, I wanted to go back to the Financial Times article I quoted in yesterday’s piece about music labels:

Universal Music Group has told streaming platforms, including Spotify and Apple, to block artificial intelligence services from scraping melodies and lyrics from their copyrighted songs, according to emails viewed by the Financial Times. UMG, which controls about a third of the global music market, has become increasingly concerned about AI bots using their songs to train themselves to churn out music that sounds like popular artists. AI-generated songs have been popping up on streaming services and UMG has been sending takedown requests “left and right”, said a person familiar with the matter.

The company is asking streaming companies to cut off access to their music catalogue for developers using it to train AI technology. “We will not hesitate to take steps to protect our rights and those of our artists,” UMG wrote to online platforms in March, in emails viewed by the FT. “This next generation of technology poses significant issues,” said a person close to the situation. “Much of [generative AI] is trained on popular music. You could say: compose a song that has the lyrics to be like Taylor Swift, but the vocals to be in the style of Bruno Mars, but I want the theme to be more Harry Styles. The output you get is due to the fact the AI has been trained on those artists’ intellectual property.”

It’s not at all clear to me that there is anything illegal about using copyrighted material for model training, as it very well may fall under the fair use exception to copyright. From the Stanford Library:

The only way to get a definitive answer on whether a particular use is a fair use is to have it resolved in federal court. Judges use four factors to resolve fair use disputes, as discussed in detail below. It’s important to understand that these factors are only guidelines that courts are free to adapt to particular situations on a case‑by‑case basis. In other words, a judge has a great deal of freedom when making a fair use determination, so the outcome in any given case can be hard to predict.

The four factors judges consider are:

  • The purpose and character of your use
  • The nature of the copyrighted work
  • The amount and substantiality of the portion taken, and
  • The effect of the use upon the potential market.

AI-generation models are transformative in my mind (i.e. the purpose and character of use is different); AI-generated art, for example, is not some sort of collage; it’s a uniquely new image that is pulled out of (literal) noise. Sure, you can get oddities like a facsimile of the Getty Images watermark if you prompt the model to include said watermark, but that is a function of the user, not the tool (will Getty Images sue an image editor because its users add a much more perfect copy of their watermark?).

The amount and substantiality question is more interesting: models take the entire image, but they don’t “copy” any of it into the final work; what matters more, the taking or the copying? The effect on the potential market, meanwhile, is the most challenging question: does the fact that ingesting an artists’ work means that said artist will face a more challenging market for future commissions bear on the fact that there is no actual reproduction in terms of the final product?

Getty Images vs. Stable Diffusion

From The Verge in February:

Getty Images has filed a lawsuit in the US against Stability AI, creators of open-source AI art generator Stable Diffusion, escalating its legal battle against the firm. The stock photography company is accusing Stability AI of “brazen infringement of Getty Images’ intellectual property on a staggering scale.” It claims that Stability AI copied more than 12 million images from its database “without permission … or compensation … as part of its efforts to build a competing business,” and that the startup has infringed on both the company’s copyright and trademark protections.

The lawsuit is the latest volley in the ongoing legal struggle between the creators of AI art generators and rights-holders. AI art tools require illustrations, artwork, and photographs to use as training data, and often scrape it from the web without the creator’s consent.

Getty Images lawsuit is here; the first thing you will note about the lawsuit is that the phrase “fair use” does not appear a single time. That noted, Getty Images does try to address the four factors at various places in its lawsuit, particularly here:

Upon information and belief, Stability AI followed links included in LAION’s dataset to access specific pages on Getty Images’ websites and copied many millions of copyrighted images and associated text. Such copying was done without Getty Images’ authorization and in violation of the express prohibitions against such conduct contained in its websites’ terms of use…Stability AI then created another copy of the content to encode it into a form its model could interpret…Stability AI then created yet additional copies with visual noise added, while retaining encoded copies of the original images without noise for comparison to help train its model.

Upon information and belief, the unauthorized copies of Getty Images’ content made by Stability AI are neither transitory nor ephemeral, and they were made with the express aim of enabling Stability AI to supplant Getty Images as a source of creative visual imagery.

This takes on all four factors: Getty Images is accusing Stability AI of taking its images and captions as is (the purpose and character of use), precisely because it is unique and descriptive (the nature of the copyrighted work), in totality (the amount and substantiality of the portion taken), with the explicit goal of disrupting Getty Images (the effect of the use upon the potential market).

I do for the most part side with Stability AI in the case; I do think that using images for model training, which does not entail reproducing a single pixel in terms of the final product, is fair use. Where I do think that Getty Images has a legitimate complaint is in the oddity I mentioned above, in which Stability AI reproduces garbled versions of the Getty Images watermark. That is, per yesterday’s Article, Getty Images’ name, image, and likeness, and Stability AI should have taken care to exclude it from output. Still, the point I made above holds: the way to get that watermark is to ask for it, which raises the question of who is responsible — the toolmaker or the user?

Adobe’s Regulatory Capture Bet

Adobe, meanwhile, is betting that courts will come down in favor of entities like Getty Images, or are, at a minimum, eager that their customers believe the risk is real. From the company’s press release for Firefly, its new set of generative AI models:

Adobe Firefly will be made up of multiple models, tailored to serve customers with a wide array of skillsets and technical backgrounds, working across a variety of different use cases. Adobe’s first model, trained on Adobe Stock images, openly licensed content and public domain content where copyright has expired, will focus on images and text effects and is designed to generate content safe for commercial use. Adobe Stock’s hundreds of millions of professional-grade, licensed images are among the highest quality in the market and help ensure Adobe Firefly won’t generate content based on other people’s or brands’ IP. Future Adobe Firefly models will leverage a variety of assets, technology and training data from Adobe and others.

My takeaway from this paragraph is that Adobe, as owners of a large stock photo library and longstanding relationships with risk-averse corporations, would be a major beneficiary of a very narrow fair use definition that excludes AI models; I put this exact question to Adobe Chief Strategy Officer Scott Belsky in that Stratechery Interview:

Obviously, a huge focus, and you mentioned it right here, is about this authenticity bit, and you’ve talked about that. Well, there’s two angles to that. We’ll get to the authenticity program, but as far as the training goes, you are pushing that it’s only on public domain images and Adobe Stock photo library. Now, I’m not sure if I can extract a truly honest answer from you because you are an executive at Adobe and so you have a motivation here. I will say, looking at it from the outside, number one, it’s not really clear to me why using photos is not fair use, it seems like a pretty textbook example of transformative work. There’s a bit where Adobe is leaning into the use of these photos for training as not being ok, and it kind of feels like a weird form of regulatory capture in a way, where it’s in Adobe’s interest to have actually a very narrow view of copyright and that this is not advanced by fair use, because “Hey, we have a huge stock photo library and we have the capability of attesting that this is the case”. I want to give you a good question that you can talk about here, but is this just your view? My guess would be that “Look, that’s what the big companies want, so we’re going to give it to them”, is that the way to put it?

Scott Belsky: It was interesting, I was having a debate earlier today, actually, with a friend around — does this go the route of the Uber’s and Airbnb’s of the world, where they, on a regional level, were pushing the laws and then they just had to fight region-by-region and they were able to just get the galvanization of their users to allow them to become, in most areas, legal and accepted? Or is this more the Napster and LimeWire world, where there was an era where it was just a free-for-all and then suddenly, Congress and I guess on a federal level, there were just determinations about what the outcome of this would be, and then it actually all went away pretty quickly and then you had iTunes and Spotify and whatever else.

Here’s the thing, Ben, I understand today, it seems like it could go either way, but if you just play this out six-plus months from now, where you can actually make unauthorized sequels to movies and you can, say, you can watch Succession and be like, “You know what? I’m going to make my own script out of prompts with ChatGPT-5,” or whatever we have by then…Then I’m going to feed it into a text-to-video model and I’m basically going to make an unauthorized episode using the exact actors, their likeness, their voices, and everything. I seriously believe this will all be achievable very soon and it’s just going to be a copyright dumpster fire. What is going to happen? There’s no way they’re going to be able to commercially monetize that content, there’s just no way, you’re using Spider-Man, you’re using The Matrix, you’re using whatever. If you believe that that’s an inevitability, which I do, then I think that there will have to be clean models that you can rely on commercially. Sure, you’ll make fun memes and unauthorized things that will creep around the internet, but from a commercial perspective, I don’t think it’s just a big company party line, I actually believe this is going to be a requirement for this technology to work commercially.

I respect Belsky’s skill in providing an answer to a question I didn’t ask: to my mind issues like producing a homemade version of a copyrighted work are already solved — it’s illegal — and if Congress wishes to strengthen the law around questions of name, image, and likeness then that’s fine and reasonable. The key thing to note, though, is that this is a question of output; what Adobe is implicitly attacking is the question of input, and its argument just happens to one that heavily favors company’s with large pre-existing well-labeled stock photo libraries, like Adobe!

Still, I will note that Adobe and Getty Images have spent a lot of time and money building up those libraries; I am cognizant of the argument that that means they ought to have exclusive access to their property. I am concerned, though, about the implications of a world where fair use does not apply to AI, which means a massive advantage for companies that are already large. Facebook, for example, has a massive walled garden of content; Google is already crawling the whole web with the permission of site owners who see exclusion from the search index as death. I am, particularly relative to just a year ago, far more optimistic that AI can be something that benefits everyone without being intermediated by a centralized entity, which is very much in the spirit of why we have the fair use exception in the first place.

That said, Adobe’s approach certainly makes strategic sense; it would be helpful if the Getty Images case moved more quickly than expected to determine to what extent Adobe’s insinuations hold water.


This Update will be available as a podcast later today. To receive it in your podcast player, visit Stratechery.

The Stratechery Update is intended for a single recipient, but occasional forwarding is totally fine! If you would like to order multiple subscriptions for your team with a group discount (minimum 5), please contact me directly.

Thanks for being a subscriber, and have a great day!