2 Sources
2 Sources
[1]
The Hypocrisy at the Heart of the AI Industry
Tech companies believe in intellectual property, but not yours. In April 2024, Eric Schmidt, the former Google CEO and a current AI evangelist, gave a closed-door lecture to a group of Stanford students. If these young people hoped to be Silicon Valley entrepreneurs, Schmidt explained, then they should be prepared to breach some ethical boundaries. At that point, 19 lawsuits had been filed against generative-AI companies for copyright infringement, alleging that Anthropic, OpenAI, and others had stolen books and other media to train their generative models. Yet Schmidt told the students to go ahead and download whatever they need to build an accurate "test" version of their AI product. If the product takes off, "then you hire a whole bunch of lawyers to go clean the mess up," he said. "If nobody uses your product, then it doesn't matter that you stole all the content." Stanford posted a video of the talk on YouTube in August 2024, but it was removed a day later. (Stanford did not respond to my request for comment about the removal.) When I recently obtained a copy, I was struck by Schmidt's readiness to say the quiet part out loud. He was articulating an attitude that is common in Silicon Valley but is usually stated as a legal or philosophical argument. When I reached one of Schmidt's spokespeople, they defended his position by telling me that Schmidt believes that the "fair use" of copyrighted work drives innovation. Others in the industry have cited the techno-libertarian idea that "information wants to be free," a frequently misunderstood credo that portrays information as a natural resource that should flow without restriction to whoever can use it. But the credo never seems to apply to Silicon Valley's own information, whether it's the troves of personal data that companies have collected about us or the software they write. Photoshop, for example, doesn't want to be free. In fact, Photoshop is one of thousands of tech-industry products that are protected by patents. Inventions such as Google's original search algorithm and even design details, such as the "rounded rectangle" shape of Apple's iPhone, have also been patented, and companies employ teams of high-end attorneys to prosecute infringements. The industry has long been a kind of intellectual-property battle zone, where damages in lawsuits frequently exceed nine figures. In 2017, for example, Waymo, Google's self-driving-car company, alleged that a former employee had stolen "confidential files and trade secrets, including blueprints, design files and testing documentation" for self-driving cars that were eventually shared with Uber. The case was settled for roughly $245 million. In the 2010s, Apple sued Samsung for copying elements of the iPhone and was initially awarded more than $1 billion in a patent-infringement battle that lasted seven years. Apple and Qualcomm have sued each other over IP in so many jurisdictions that it's hard to track. In the pursuit of generative AI, tech companies have recently turned their aggressive strategies toward less prepared industries. As my reporting has shown, many top AI models have been trained on data sets containing massive numbers of copyrighted books, videos, and other works. This large-scale piracy has been excused in a number of ways: OpenAI (which has a corporate partnership with The Atlantic's business team) has claimed that the company uses "publicly available information" to train its models; Anthropic has said that it has used books, but not in any commercial products; and Meta admits that it has used books in commercial products, but that doing so was "quintessential fair use." Even as they claim the right to train their models on work belonging to other people, the AI companies have rejected similar reasoning when it comes to their own products. Consider OpenAI's terms of service for ChatGPT, which forbid use of the bot's "output to develop models that compete with OpenAI." Anthropic, Google, and xAI have similar clauses forbidding people from using the material generated by their chatbots to train competing products. In other words: We can train on your work, but you can't train on ours. In the current economic environment, it's not surprising that companies vying for market dominance would operate with standards that serve their bottom line. But it's striking nonetheless how sharply their actions can contradict their professed values. Meta apparently does not want copies of its models on the web, even though it claims those models are "open," a word that typically means software is free and publicly available, and that implies a degree of goodwill or generosity on the part of the creator. It has reportedly sent notices demanding the deletion of such copies from online platforms. (Meta did not respond to a request for comment.) Companies also know the value of training data, and at least one of them foresaw the backlash that taking such data might create. In 2021, one year before OpenAI released ChatGPT and two years before my reporting first revealed what was being used as AI-training data, Anthropic CEO Dario Amodei wrote an internal memo titled "An Economic Model for Compensating Data Producers." (It was recently unsealed in a copyright-infringement lawsuit against the company.) In the document, Amodei acknowledges that AI could be "an increasingly extractive concentrator of wealth" and that creators might eventually "grumble" or "get mad" as this fact becomes apparent. Resistance from creators might slow down AI progress, Amodei writes, and for this reason, he suggests compensating them "with a fraction of the profits from the model produced." Giving creators equity in the company could be a "great fit" for Anthropic's "public benefit orientation," Amodei wrote. Today, Anthropic still claims to provide a public benefit, but it has argued in court that using copyrighted books is "fair use" -- meaning, essentially, that the authors are entitled to nothing. Anthropic declined to comment when I reached out for this article. Companies argue that AI training is fair use because their AI models produce original work that is not derived from the sources they use for training. This is not necessarily true: My reporting has shown that chatbots and image generators can produce near-exact copies of media they were trained on, spitting out near-complete copies of Harry Potter and the Sorcerer's Stone, for example, or rendering images that are fuzzy copies of existing artwork. But companies have tried to downplay this fact and focus the copyright discussion elsewhere, even invoking geopolitics and the idea of an international "AI race" as a sort of trump card. "Without fair use access, the race for AI is effectively over. America loses," OpenAI wrote to the Office of Science and Technology Policy last year. Not everyone in the AI industry is on the same page. Ed Newton-Rex, a former VP of audio at Stability AI, quit his job in November 2023 and wrote on X that, regardless of fair use, which "wasn't designed with generative AI in mind," he didn't see how current AI-training practices "can be acceptable in a society that has set up the economics of the creative arts such that creators rely on copyright." Newton-Rex started a nonprofit called Fairly Trained, which certifies AI models that are trained on properly acquired data. It's worth noting that Silicon Valley has itself regularly been a victim of IP theft, in the form of software piracy. Partially in response to that problem, major companies have changed how software is distributed. Today, you cannot just buy Adobe Photoshop: Instead, you pay a rental fee to access the program, which verifies your license every time you use it. Microsoft has taken a similar approach with the 365 version of its Office suite, and Google's office software can't be downloaded at all. These companies have made their IP harder to steal by developing new methods of controlling access -- an option that is not realistically available to the artists, authors, and open-source-software developers they take material from. Given the double standard, it's difficult to tell whether Silicon Valley's arguments about fair use are genuine or just legally expedient. On one hand, generative AI is a new technology that raises new questions about the use of copyrighted work. On the other hand, the AI industry's aggressive approach is business as usual for Silicon Valley: moving fast and breaking things. And betting that the lawyers can "clean the mess up."
[2]
The CEO of Patreon blasts AI companies for the 'bogus excuse' they're using to not pay artists | Fortune
Patreon's CEO Jack Conte is tired of watching AI companies strike deals with huge corporations like Disney while ignoring the myriad of smaller creators who contribute to their models. Speaking at the South by Southwest conference this week, Conte, whose company allows people to pay their favorite creators directly, argued AI companies should view creators' work in the same way it views that of Disney, Conde Nast, or Warner Music, aiming to reach agreements with them rather than use their content without permission. He attacked the legal doctrine of "fair use," which allows someone to use copyrighted material without permission or payment depending on the purpose and character of the use, the nature of the original work, how much of the work was used, and whether the use harms the market. AI companies have cited fair use to justify using content to train or contribute to their models without paying. These companies often argue they are using copyrighted content in a "transformative" way and not just regurgitating it verbatim. For Conte, this legal "fair use" loophole is utter quackery. "The AI companies are claiming fair use, but this argument is bogus," Conte said during the conference. "It's bogus because while they claim it's fair to use the work of creators as training data, they do multimillion-dollar deals with rights holders and publishers like Disney, and CondΓ© Nast, and Vox, and Warner Music." Conte pointed out the large licensing deals these AI companies have reached with intellectual property owners in recent years demonstrate the double standard of these companies. While AI companies recognize some copyrighted content requires permission and agreements, the same doesn't seem to be true for creator-made content. In the past several years, AI companies like OpenAI have made waves for the deals they have struck with some content owners while staving off lawsuits from others like the New York Times, which in 2023 accused OpenAI of training ChatGPT on millions of its articles without permission. In December, OpenAI, the AI giant led by CEO Sam Altman, struck a deal that saw Disney invest $1 billion in the company and licensed more than 200 characters to OpenAI so they could be featured in the company's video app, Sora. OpenAI has also signed licensing deals with CondΓ© Nast, which owns The New Yorker and also with Vox Media, which owns New York Magazine. In November, Warner Music Group struck two separate licensing deals with music-focused AI companies Suno and Udio, after settling copyright suits with the companies. Conte mentioned these deals specifically to highlight the hypocrisy demonstrated by AI companies when deciding who gets a licensing agreement and who doesn't. Smaller creators, he claims, are being left out. "If it's legal to just use it, why pay?" Conte asked the crowd, according to TechCrunch. "Why pay them and not creators -- not the millions of illustrators and musicians and writers -- whose work has been consumed by these models to build hundreds of billions of dollars of value for these companies?" The AI companies' fair use claims have been called into question several times as AI models have become increasingly more popular. The New York Times filed a lawsuit in 2023 claiming OpenAI used millions of its articles without permission and that its large language model ChatGPT was in some cases regurgitating entire Times articles, potentially striking a blow to OpenAI's fair use argument. A date for the trial has not yet been set, but if the Times wins it could be owed billions in damages. More recently, dictionary makers Encyclopaedia Britannica and Merriam-Webster sued OpenAI after it rebuffed the companies' offer of a licensing agreement in 2024. The publishers claimed in the lawsuit that OpenAI's ChatGPT is cutting into their search traffic and ad revenue by absorbing the content created by their hundreds of human writers and editors. OpenAI rival Anthropic also settled a class action lawsuit by a group of authors to the tune of $1.5 billion in September. As a result of the case, the judge ruled that training an AI model on pirated books -- as the authors accused Anthropic of doing -- did not qualify as "fair use," but that training an AI model on purchased books qualified as legal transformative use. While Conte said he was not against AI, generally, and noted that change is inevitable, humans will continue to enjoy human-created content long into the future, he said. "Still, the AI companies should pay creators for our work, not because the tech is bad -- but because a lot of it is good, or it will be soon -- and it's going to be the future. And when we plan for humanity's future, we should plan for society's artists, too, not just for their sake, but for the sake of all of us. Societies that value and incentivize creativity are better for it," he said.
Share
Share
Copy Link
Former Google CEO Eric Schmidt advised Stanford students to steal content for AI training and hire lawyers later. Meanwhile, Patreon CEO Jack Conte blasts AI companies for signing multimillion-dollar licensing deals with Disney and Warner Music while claiming fair use to avoid paying smaller creators whose work trains their models.
The AI industry faces mounting criticism over its approach to intellectual property, highlighted by revelations from a closed-door lecture by Eric Schmidt, former Google CEO. In April 2024, Schmidt told Stanford students aspiring to be Silicon Valley entrepreneurs to download whatever content they need to build AI products, then "hire a whole bunch of lawyers to go clean the mess up" if the product succeeds
1
. His candid advice came as 19 copyright lawsuits had already been filed against generative-AI companies including Anthropic and OpenAI for allegedly stealing books and media to train their models. Stanford posted the lecture video in August 2024 but removed it within a day, though Schmidt's spokesperson later defended his position by claiming he believes fair use of copyrighted work drives innovation1
.
Source: The Atlantic
The contradiction at the heart of the AI industry becomes stark when examining how AI companies protect their own intellectual property while freely using others' work. OpenAI's terms of service for ChatGPT explicitly forbid using the bot's output "to develop models that compete with OpenAI," with Anthropic, Google, and xAI maintaining similar restrictions
1
. This double standard in AI extends beyond terms of service. Tech companies have long operated as an intellectual property battle zone where damages frequently exceed nine figures. Waymo settled with Uber for roughly $245 million in 2017 over stolen self-driving car trade secrets, while Apple's patent-infringement battle with Samsung over iPhone design elements lasted seven years and initially awarded Apple more than $1 billion1
.Jack Conte, CEO of Patreon, directly challenged the AI industry's selective approach to paying for content at South by Southwest this week. Conte attacked the fair use doctrine that AI companies cite to justify using copyrighted material without permission or payment, calling it "bogus" and highlighting the hypocrisy of AI companies striking multimillion-dollar licensing deals with corporations like Disney, CondΓ© Nast, Vox, and Warner Music while ignoring smaller creators
2
. "If it's legal to just use it, why pay?" Conte asked, questioning why AI companies pay major rights holders but not the millions of illustrators, musicians, and writers whose work trains AI models that have built hundreds of billions of dollars in value2
.
Source: Fortune
Related Stories
The legal landscape around AI training data has become increasingly contentious as copyright infringement cases mount against major AI companies. The New York Times filed a lawsuit in 2023 claiming OpenAI used millions of its articles without permission, with ChatGPT allegedly regurgitating entire Times articles in some cases
2
. If successful, the Times could be owed billions in damages, though no trial date has been set. Dictionary makers Encyclopaedia Britannica and Merriam-Webster also sued OpenAI after the company rebuffed their licensing agreement offer in 2024, claiming ChatGPT cuts into their search traffic and ad revenue by absorbing content from their human writers and editors2
. Anthropic settled a class action lawsuit by authors for $1.5 billion in September, with the judge ruling that training AI models on pirated books does not qualify as fair use, though training on purchased books constitutes legal transformative use2
.The licensing deals AI companies have pursued with major content owners directly contradict their fair use arguments. In December, OpenAI secured a deal where Disney invested $1 billion and licensed more than 200 characters for use in OpenAI's video app, Sora
2
. OpenAI has also signed licensing agreements with CondΓ© Nast and Vox Media, while Warner Music Group struck two separate licensing deals with music-focused AI companies Suno and Udio in November after settling copyright suits2
. These deals demonstrate that AI companies recognize when copyrighted content requires permission and payment, yet smaller creators receive no such consideration. The question of paying artists and creators has become central to the debate, with Conte arguing that societies valuing and incentivizing creativity benefit everyone, not just artists themselves2
.Summarized by
Navi
[1]
30 May 2025β’Policy and Regulation

15 Jul 2025β’Policy and Regulation

02 Apr 2025β’Policy and Regulation
