The New York Times’s OpenAI lawsuit could put a damper on AI’s 2024 ambitions
Welcome to AI Decoded, Fast Company’s weekly LinkedIn newsletter that breaks down the most important news in the world of AI. I’m Mark Sullivan, a senior writer at Fast Company, covering emerging tech, AI, and tech policy.
This week, I’m diving into The New York Times’s lawsuit against OpenAI and Microsoft. Plus, there’s a new contender in the AI wearable race, and a new survey sheds light on businesses’ adoption of AI.
If a friend or colleague shared this newsletter with you, you can sign up to receive it every week here. And if you have comments on this issue and/or ideas for future ones, drop me a line at sullivan@fastcompany.com, and follow me on X (formerly Twitter) @thesullivan.
What The New York Times suit against OpenAI could mean for AI
The New York Times filed a lawsuit against OpenAI and Microsoft late last month, alleging the companies used its content to train their respective AI models without permission or compensation. Developers of large language models have routinely scraped huge batches of data from the internet, and then allowed their models to train by processing and finding patterns in the data. But the Times’s claims go deeper: the suit says OpenAI and Microsoft encoded the newspaper’s articles into their language models’ memory so that ChatGPT and Bing Chat (now called Copilot) could access and regurgitate the information—in some cases verbatim, and without proper citation (the suit contains numerous examples of this). The lawsuit demands that any chatbot trained using the data be taken offline. The lawsuit came as a surprise to OpenAI: A company spokesperson told Axios that the two sides had been discussing content licensing terms.
The suit marked a sobering coda to 2023, a year in which the AI industry sprinted forward unrestrainedly, and mostly without regulation. Many in the tech industry had hoped that 2024 would bring far wider application of AI systems. But lawsuits over copyrights could slow everything down, as legal exposure concerns
The OpenAI lawyers may argue that an AI model isn’t much different from a human who ingests a bunch of information from the web then uses it as a basis for their own thoughts. That whole debate may be moot if the Times can prove that it was financially harmed when OpenAI’s and Microsoft’s AI models spat out line-for-line text lifted from the paper’s coverage. But the main issue is that this is all uncharted legal territory; a high-profile trial may begin to establish how copyright law applies to the training of AI models. Even if OpenAI ends up paying damages, the two parties may still come to an accommodation allowing the AI company to use Times content for training.
News publishers’ posture toward AI companies runs the gamut: The Wall Street Journal, News Corp, and Gannett want to license their stories to AI developers, while others such as Reuters and CNN have begun blocking AI companies from accessing their content. Meanwhile, it's still not outside the realm of possibility that the courts or the Federal Trade Commission could order AI companies to delete training data they’ve already scraped from the web. (The FTC did, after all, open an inquiry on OpenAI’s training data acquisition practices last summer.)
“In the months ahead, we’ll continue to see additional licensing agreements between credible publishers and AI companies,” says Alon Yamin, the cofounder and CEO of Copyleaks, which makes an AI plagiarism detection tool. “And yes, additional lawsuits.”
Ready for another buzzy AI smartphone killer?
First there was Humane’s Ai Pin, an AI device you can wear on your lapel. Now, another company, L.A.-based Rabbit, is set to reveal its own AI-centered device, called the r1, during next week’s CES trade show. The demo video shows people instructing a device to order an Uber, find a new podcast, and “tell everybody I’m going to be a little late.”
The r1 (which appears as a mysterious pixelated blob in the video) uses a large language model to understand spoken requests in both content and context. But that’s just the front door. Rabbit’s big idea is a foundation model (called the Large Action Model, or LAM) that orchestrates the actions of a number of different apps (including Uber) in order to meet a user’s demand.
I’m skeptical of the work of companies like Humane and Rabbit only because we’re still in the early days of foundation models that are actually useful in everyday life. But I also love where this is all headed. These companies are early players in an AI-powered evolution away from smartphones to something far more personal and functional. As the models that power these devices improve, personal AI devices will only get better.
Vinod Khosla, who is a major investor in Rabbit, says the startup’s concept points to a future of autonomous AI agents. "In a decade, we could have tens of billions more agents than people on the planet running around the net doing things on our behalf," he says in a statement. "The Rabbit team is bringing powerful new consumer experiences for every human to have an agent in their pocket."
What IT decision-makers think about generative AI
The following insights come from new research conducted by the research and consulting firm Creative Strategies:
MORE AI COVERAGE FROM FAST COMPANY:
FROM AROUND THE WEB:
Student at North South University
1yThe New York Times' lawsuit against OpenAI and Microsoft raises copyright concerns in AI training data, potentially slowing advancements. Rabbit's r1 wearable device represents a trend towards personal and functional AI. Creative Strategies survey shows significant engagement with generative AI technologies, with business analytics as a leading application. However, concerns about integrating AI with existing systems, skilled personnel shortage, and security issues remain
Helper
1yGarbage in garbage out.
Interesting. That is same as take consumers private information and using it for publication to make a profit in journalism news 📰 articles. The consumers are not being paid for their private information data is being ♻️