The alleged theft at the heart of ChatGPT
When best-selling thriller writer Douglas Preston began playing around with OpenAI's new chatbot, ChatGPT, he was, at first, impressed. But then he realized how much in-depth knowledge GPT had of the books he had written. When prompted, it supplied detailed plot summaries and descriptions of even minor characters. He was convinced it could only pull that off if it had read his books.
Large language models, the kind of artificial intelligence underlying programs like ChatGPT, do not come into the world fully formed. They first have to be trained on incredibly large amounts of text. Douglas Preston, and 16 other authors, including George R.R. Martin, Jodi Piccoult, and Jonathan Franzen, were convinced that their novels had been used to train GPT without their permission. So, in September, they sued OpenAI for copyright infringement.
This sort of thing seems to be happening a lot lately–one giant tech company or another "moves fast and breaks things," exploring the edges of what might or might not be allowed without first asking permission. On today's show, we try to make sense of what OpenAI allegedly did by training its AI on massive amounts of copyrighted material. Was that good? Was it bad? Was it legal?
This episode was hosted by Keith Romer and Erika Beras and produced by Willa Rubin and Sam Yellowhorse Kesler. It was edited by Kenny Malone and fact-checked by Sierra Juarez. Engineering by Robert Rodriguez. Alex Goldmark is Planet Money's executive producer.
Help support Planet Money and get bonus episodes by subscribing to Planet Money+ in Apple Podcasts or at plus.npr.org/planetmoney.
Always free at these links: Apple Podcasts, Spotify, Google Podcasts, NPR One or anywhere you get podcasts.
Find more Planet Money: Facebook / Instagram / TikTok / Our weekly Newsletter.
Music: Elias Music - "Elevated," Universal Music Production - "Don't Cross the Line," and "This is Not Goodbye"