Featured Article

The current legal cases against generative AI are just the beginning

AI that can generate art, text and more is in for a reckoning

Comment

Image Credits: Fry Design Ltd / Getty Images

As generative AI enters the mainstream, each new day brings a new lawsuit.

Microsoft, GitHub and OpenAI are currently being sued in a class action motion that accuses them of violating copyright law by allowing Copilot, a code-generating AI system trained on billions of lines of public code, to regurgitate licensed code snippets without providing credit.

Two companies behind popular AI art tools, Midjourney and Stability AI, are in the crosshairs of a legal case that alleges they infringed on the rights of millions of artists by training their tools on web-scraped images.

And just last week, stock image supplier Getty Images took Stability AI to court for reportedly using millions of images from its site without permission to train Stable Diffusion, an art-generating AI.

At issue, mainly, is generative AI’s tendency to replicate images, text and more — including copyrighted content — from the data that was used to train it. In a recent example, an AI tool used by CNET to write explanatory articles was found to have plagiarized articles written by humans — articles presumably swept up in its training dataset. Meanwhile, an academic study published in December found that image-generating AI models like DALL-E 2 and Stable Diffusion can and do replicate aspects of images from their training data.

The generative AI space remains healthy — it raised $1.3 billion in venture funding through November 2022, according to PitchBook, up 15% from the year prior. But the legal questions are beginning to affect business.

Some image-hosting platforms have banned AI-generated content for fear of legal blowback. And several legal experts have cautioned generative AI tools could put companies at risk if they were to unwittingly incorporate copyrighted content generated by the tools into any of products they sell.

“Unfortunately, I expect a flood of litigation for almost all generative AI products,” Heather Meeker, a legal expert on open source software licensing and a general partner at OSS Capital, told TechCrunch via email. “The copyright law needs to be clarified.”

Content creators such as Polish artist Greg Rutkowski, known for creating fantasy landscapes, have become the face of campaigns protesting the treatment of artists by generative AI startups. Rutkowski has complained about the fact that typing text like “Wizard with sword and a glowing orb of magic fire fights a fierce dragon Greg Rutkowski” will create an image that looks very similar to his original work — threatening his income.

Given generative AI isn’t going anywhere, what comes next? Which legal cases have merit and what court battles lie on the horizon?

Eliana Torres, an intellectual property attorney with Nixon Peabody, says that the allegations of the class action suit against Stability AI, Midjourney, and DeviantArt will be challenging to prove in court. In particular, she thinks it’ll be difficult to ascertain which images were used to train the AI systems because the art the systems generate won’t necessarily look exactly like any of the training images.

State-of-the-art image-generating systems like Stable Diffusion are what’s known as “diffusion” models. Diffusion models learn to create images from text prompts (e.g. “a sketch of a bird perched on a windowsill”) as they work their way through massive training datasets. The models are trained to “re-create” images as opposed to drawing them from scratch, starting with pure noise and refining the image over time to make it incrementally closer to the text prompt.

Image-generating AI can copy and paste from training data, raising IP concerns

Perfect recreations don’t occur often, to Torres’ point. As for images in the style of a particular artist, style has proven nearly impossible to shield with copyright.

“It will … be challenging to get a general acceptance of the definition of ‘in style of’ as ‘a work that others would accept as a work created by that artist whose style was called upon,’ which is mentioned in the complaint [i.e. against Stability AI et al.],” Torres told TechCrunch in an email interview. 

Torres also believes the suit should be directed not at the creators of these AI systems, but at the party responsible for compiling the images used to train them: Large-scale Artificial Intelligence Open Network (LAION), a nonprofit organization. Midjourney, DeviantArt and Stability AI use training data from LAION’s datasets, which span billions of images from around the web.

“If LAION created the dataset, then the alleged infringement occurred at that point, not once the dataset was used to train the models,” Torres said. “It’s the same way a human can walk into a gallery and look at paintings but is not allowed to take photos.”

Companies like Stability AI and OpenAI, the company behind ChatGPT, have long claimed that “fair use” protects them in the event that their systems were trained on licensed content. This doctrine enshrined in U.S. law permits limited use of copyrighted material without first having to obtain permission from the rightsholder.

Supporters point to cases like Authors Guild v. Google, in which the New York-based U.S. Court of Appeals for the Second Circuit ruled that Google manually scanning millions of copyrighted books without a license to create its book search project was fair use. What constitutes fair use is constantly being challenged and revised, but in the generative AI realm, it’s an especially untested theory.

A recent article in Bloomberg Law asserts that the success of a fair use defense will depend on whether the works generated by the AI are considered transformative — in other words, whether they use the copyrighted works in a way that significantly varies from the originals. Previous case law, particularly the Supreme Court’s 2021 Google v. Oracle decision, suggests that using collected data to create new works can be transformative. In that case, Google’s use of portions of Java SE code to create its Android operating system was found to be fair use.

Interestingly, other countries have signaled a move toward more permissive use of publicly available content — copyrighted or not. For example, the U.K. is planning to tweak an existing law to allow text and data mining “for any purpose,” moving the balance of power away from rightsholders and heavily toward businesses and other commercial entities. There’s been no appetite to embrace such a shift in the U.S., however, and Torres doesn’t expect that to change anytime soon — if ever.

The Getty case is slightly more nuanced. Getty — which Torres notes hasn’t yet filed a formal complaint — must show damages and connect any infringement it alleges to specific images. But Getty’s statement mentions that it has no interest in financial damages and is merely looking for a “new legal status quo.” 

Andrew Burt, one of the founders of AI-focused law firm BNH.ai, disagrees with Torres to the extent that he believes generative AI lawsuits focused on intellectual property issues will be “relatively straightforward.” In his view, if copyrighted data was used to train AI systems — whether because of intellectual property or privacy restrictions — those systems should and will be subject to fines or other penalties.

Burt noted that the Federal Trade Commission (FTC) is already pursuing this path with what it calls “algorithmic disgorgement,” where it forces tech firms to kill problematic algorithms along with any ill-gotten data that they used to train them. In a recent example, the FTC used the remedy of algorithmic disgorgement to force Everalbum, the maker of a now-defunct mobile app called Ever, to delete facial recognition algorithms the company developed using content uploaded by people who used its app. (Everalbum didn’t make it clear that the users’ data was being used for this purpose.)

“I would expect generative AI systems to be no different from traditional AI systems in this way,” Burt said.

What are companies to do, then, in the absence of precedent and guidance? Torres and Burt concur that there’s no obvious answer.

For her part, Torres recommends looking closely at the terms of use for each commercial generative AI system. She notes that Midjourney has different rights for paid versus unpaid users, while OpenAI’s DALL-E assigns rights around generated art to users while also warning them of “similar content” and encouraging due diligence to avoid infringement.

“Businesses should be aware of the terms of use and do their due diligence, such as using reverse image searches of the generated work intended to be used commercially,” she added.

Burt recommends that companies adopt risk management frameworks such as the AI Risk Management Framework released by National Institute of Standards and Technology, which gives guidance on how to address and mitigate risks in the design and use of AI systems. He also suggests that companies continuously test and monitor their systems for potential legal liabilities.

“While generative AI systems make AI risk management harder — it is, to be fair, much more straightforward to monitor an AI system that makes binary predictions for risks — there are concrete actions that can be taken,” Burt said.

Some firms, under pressure from activists and content creators, have taken steps in the right direction. Stability AI plans to allow artists to opt out of the dataset used to train the next-generation Stable Diffusion model. Through the website HaveIBeenTrained.com, rightsholders will be able to request opt-outs before training begins in a few weeks’ time. Rival OpenAI offers no such opt-out mechanism, but the firm has partnered with organizations like Shutterstock to license portions of their image galleries.

For Copilot, GitHub introduced a filter that checks code suggestions with their surrounding code of about 150 characters against public GitHub code and hides suggestions if there’s a match or “near match.” It’s an imperfect measure — enabling the filter can cause Copilot to omit key pieces of attribution and license text — but GitHub has said that it plans to introduce additional features in 2023 aimed at helping developers make informed decisions about whether to use Copilot’s suggestions.

Taking the ten-thousand-foot view, Burt believes that generative AI is being deployed more and more without an understanding of how to address its dangers. He praises efforts to combat the obvious problems, like copyrighted works being used to train content generators. But he cautions that the opacity of the systems will put pressure on businesses to prevent the systems from wreaking havoc — and having a plan to address the systems’ risks before they’re put into the wild.

“Generative AI models are among the most exciting and novel uses of AI — with the clear potential to transform the ‘knowledge economy’,” he said. “Just as with AI in many other areas, the technology is largely there and ready for use. What isn’t yet mature are the ways to manage all of its risks. Without thoughtful, mature evaluation and management of these systems’ harms, we risk deploying a technology before we understand how to stop it from causing damage.”

Meeker is more pessimistic, arguing that not all businesses — regardless of the mitigations they undertake — will be able to shoulder the legal costs associated with generative AI. This points to the urgent need for clarification or changes in copyright law, she says.

“If AI developers don’t know what data they can use to train models, the technology could be set back by years,” Meeker said. “In a sense, there is nothing they can do, because if businesses are unable to lawfully train models on freely available materials, they won’t have enough data to train the models. There are only various long-term solutions like opt-in or opt-out models, or systems that aggregate royalties for payment to all authors … The suits against AI businesses for ingesting copyrightable material to train models are potentially crippling to the industry, [and] could cause consolidation that would limit innovation.”

More TechCrunch

When the developers replied to the July 19 email, Yelp sent a deck of pricing tiers with base pricing starting from $229 per month for a limit of 1,000 API…

Yelp’s lack of transparency around API charges angers developers

Featured Article

Cloud infrastructure revenue approached $80 billion this quarter

The cloud infrastructure market has put the doldrums of 2023 firmly behind it with another big quarter. Revenue continues to grow at a brisk pace, fueled by interest in AI. Synergy Research reports revenue totaled $79 billion for the quarter, up $14.1 billion or 22% from last year. This marked…

Cloud infrastructure revenue approached $80 billion this quarter

The pharma giant won’t say how many patients were affected by its February data breach. A count by TechCrunch confirms that over a million people are affected.

Pharma giant Cencora is alerting millions about its data breach

Self-driving technology company Aurora Innovation is looking to raise hundreds of millions in additional capital as it races toward a driverless commercial launch by the end of 2024.  Aurora is…

Self-driving truck startup Aurora Innovation to sell up to $420M in shares ahead of commercial launch

Payments infrastructure firm Infibeam Avenues has acquired a majority 54% stake in Rediff.com for up to $3 million, a dramatic twist of fate for the 28-year-old business that was the…

Rediff, once an internet pioneer in India, sells majority stake for $3M

The ruling confirmed an earlier decision in April from the High Court of Podgorica which rejected a request to extradite the crypto fugitive to the United States.

Terraform Labs co-founder and crypto fugitive Do Kwon set for extradition to South Korea

A day after Meta CEO Mark Zuckerberg talked about his newest social media experiment Threads reaching “almost” 200 million users on the company’s Q2 2024 earnings call, the platform has…

Meta’s Threads crosses 200 million active users

TechCrunch Disrupt 2024 will be in San Francisco on October 28–30, and we’re already excited! Disrupt brings innovation for every stage of your startup journey, and we could not bring you this…

Connect with Google Cloud, Aerospace, Qualcomm and more at Disrupt 2024

Featured Article

A comprehensive list of 2024 tech layoffs

The tech layoff wave is still going strong in 2024. Following significant workforce reductions in 2022 and 2023, this year has already seen 60,000 job cuts across 254 companies, according to independent layoffs tracker Layoffs.fyi. Companies like Tesla, Amazon, Google, TikTok, Snap and Microsoft have conducted sizable layoffs in the…

A comprehensive list of 2024 tech layoffs

Intel announced it would layoff more than 15% of its staff, or 15,000 employees, in a memo to employees on Thursday. The massive headcount is part of a large plan…

Intel to lay off 15,000 employees

Following the recent lawsuit filed by the Recording Industry Association of America (RIAA) against music generation startups Udio and Suno, Suno admitted in a court filing on Thursday that it did, in…

AI music startup Suno claims training model on copyrighted music is ‘fair use’

In spite of a drop for the quarter, iPhone remained Apple’s most important category by a wide margin.

iPad sales help bail out Apple amid a continued iPhone slide

Molly Alter wears a lot of hats. She’s a mocumentary filmmaker working on a project about an alternate reality where charades is big business. She’s a caesar salad connoisseur and…

How filming a cappella concerts and dance recitals led Northzone’s newest partner Molly Alter to a career in VC

Microsoft has a long and tangled history with OpenAI, having invested a reported $13 billion in the ChatGPT maker as part of a long-term partnership. As part of the deal,…

Microsoft now lists OpenAI as a competitor in AI and search

The San Jose-based startup raised $60 million in a round that values it lower than the $500 million valuation it garnered in its most recent round, according to multiple sources.

Sequoia-backed Knowde raises Series C at a valuation cut

X (formerly Twitter) can no longer be accessed in the Mac App Store, suggesting that it has been officially delisted.  Searches for both “Twitter” and “X” on Apple’s platform no…

Twitter disappears from Mac App Store

Google Thursday said that it is introducing new Gemini-powered features for Chrome’s desktop version, including Lens for desktop, tab compare for shopping assistance, and natural language integration for search history.…

Google brings Gemini-powered search history and Lens to Chrome desktop

When Xiaoyin Qu was growing up in China, she was obsessed with learning how to build paper airplanes that could do flips in the air. Her parents, though, didn’t have…

Heeyo built an AI chatbot to be a billion kids’ interactive tutor and friend

While the company was awarded a massive, $4.2 billion contract to accelerate Starliner development in 2014, it was structured as a “fixed-price” model.

Boeing bleeds another $125M on Starliner program, bringing total losses to $1.6B

Welcome back to TechCrunch Mobility — your central hub for news and insights on the future of transportation. Sign up here for free — just click TechCrunch Mobility! Summer road…

Anthony Levandowski bets on off-road autonomy, Nuro plots a comeback and Applied Intuition gets more investor love

Google’s new features include Gemini in BigQuery and Looker to help users with data engineering and analysis.

Google Cloud expands its database portfolio with new AI capabilities

Rad Power Bikes, the Seattle-based e-bike startup that has raised more than $300 million from investors, went through another round of layoffs in July, TechCrunch has exclusively learned. This is…

VC darling Rad Power Bikes hit with another round of layoffs

Five years ago, as robotaxis and self-driving truck startups were still raking in millions in venture capital, Anthony Levandowski turned to off-road autonomy. Now, that decision — which brought the…

Why Anthony Levandowski returned to his off-road autonomous vehicle roots with AV startup Pronto

Commercial space station company Vast is building a private microgravity research lab as part of its wider Haven-1 station plans. The module is set to launch no earlier than the…

Vast plans microgravity lab on its Haven-1 private space station

Google Cloud is giving Y Combinator startups access to a dedicated, subsidized cluster of Nvidia graphics processing units and Google tensor processing units to build AI models. It’s part of…

Google Cloud now has a dedicated cluster of Nvidia GPUs for Y Combinator startups

StackShare is one of the more popular platforms for developers to discuss, track, and share the tools they use to build applications.

Open source startup FOSSA is buying StackShare, a site used by 1.5M developers

Featured Article

Indian startups gut valuations ahead of IPO push

Ola Electric and FirstCry are set to test investor appetite with public listing, both pricing their shares below their previous valuation asks.

Indian startups gut valuations ahead of IPO push

The European Union’s risk-based regulation for applications of artificial intelligence has come into force starting from today.

The EU’s AI Act is now in force

The company also said it has received regulatory clearance to start Phase 2 clinical trials for a new drug in the U.S. later this year.

Healx, an AI-enabled drug discovery platform for rare diseases, raises $47M

The European Commission (EC) has given the go-ahead to HPE’s planned megabucks acquisition of Juniper Networks.

EU greenlights HPE’s $14B Juniper Networks acquisition