Skip Navigation

Thousands of authors demand payment from AI companies for use of copyrighted works

Thousands of authors demand payment from AI companies for use of copyrighted works::Thousands of published authors are requesting payment from tech companies for the use of their copyrighted works in training artificial intelligence tools, marking the latest intellectual property critique to target AI development.

357 comments
  • I don't know how I feel about this honestly. AI took a look at the book and added the statistics of all of its words into its giant statistic database. It doesn't have a copy of the book. It's not capable of rewriting the book word for word.

    This is basically what humans do. A person reads 10 books on a subject, studies become somewhat of a subject matter expert and writes their own book.

    Artists use reference art all the time. As long as they don't get too close to the original reference nobody calls any flags.

    These people are scared for their viability in their user space and they should be, but I don't think trying to put this genie back in the bottle or extra charging people for reading their stuff for reference is going to make much difference.

    • It’s not at all like what humans do. It has no understanding of any concepts whatsoever, it learns nothing. It doesn’t know that it doesn’t know anything even. It’s literally incapable of basic reasoning. It’s essentially taken words and converted them to numbers, and then it examines which string is likely to follow each previous string. When people are writing, they aren’t looking at a huge database of information and determining the most likely word to come next, they’re synthesizing concepts together to create new ones, or building a narrative based on their notes. They understand concepts, they understand definitions. An AI doesn’t, it doesn’t have any conceptual framework, it doesn’t even know what a word is, much less the definition of any of them.

      • How can you tell that our thoughts don't come from a biological LLM? Maybe what we conceive as "understanding" is just a feeling emerging from a more fondamental mechanism like temperature emerges from the movement of particles.

      • When people are writing, they aren’t looking at a huge database of information and determining the most likely word to come next, they’re synthesizing concepts together to create new ones, or building a narrative based on their notes. They understand concepts, they understand definitions.

        A huge part of what we do is like drawing from a huge mashup of accumulated patterns though. When an image or phrase pops into your head fully formed, on the basis of things that you have seen and remembered, isn't that the same sort of thing as what AI does? Even though there are (poorly understood) differences between how humans think and what machine learning models do, the latter seems similar enough to me that most uses should be treated by the same standard for plagiarism; only considered violating if the end product is excessively similar to a specific copyrighted work, and not merely because you saw a copyrighted work and that pattern being in your brain affected what stuff you spontaneously think of.

      • I don't think this is true.

        The models (or maybe the characters in the conversations simulated by the models) can be spectacularly bad at basic reasoning, and misunderstand basic concepts on a regular basis. They are of course completely insane; the way they think is barely recognizable.

        But they also, when asked, are often able to manipulate concepts or do reasoning and get right answers. Ask it to explain the water cycle like a pirate, and you get that. You can find the weights that make the Eifel Tower be in Paris and move it to Rome, and then ask for a train itinerary to get there, and it will tell you to take the train to Rome.

        I don't know what "understanding" something is other than to be able to get right answers when asked to think about it. There's some understanding of the water cycle in there, and some of pirates, and some of European geography. Maybe not a lot. Maybe it's not robust. Maybe it's superficial. Maybe there are still several differences in kind between whatever's there and the understanding a human can get with a brain that isn't 100% the stream of consciousness generator. But not literally zero.

      • I didn't say what you said, that's a lot of words and concepts you're attributing to me that I didn't say.

        I'm saying, LLM ingests data in a way it can average it out, in essence it learns it. It's not wrote memorization, but it's not truly reasoning either, though it's approaching it if you consider we might be overestimating human comprehension. It pulls in the data from all the places and uses the data to create new things.

        People pull in data over a decade or two, we learn it, then end up writing books, or applying the information to work. They're smart and valuable people and we're glad they read everyone's books.

        The LLM ingests the data and uses the statistics behind it to do work, the world is ending.

      • I think you underestimate the reasoning power of these AIs. They can write code, they can teach math, they can even learn math.

        I've been using GPT4 as a math tutor while learning linear algebra, and I also use a text book. The text book told me that (to write it out) "the column space of matrix A is equal to the column space of matrix A times its own transpose". So I asked GPT4 if that was true and it said no, GPT disagreed with the text book. This was apparently something that GPT did not memorize and it was not just regurgitating sentences. I told GPT I saw it in a text book, the AI said "sorry, the textbook must be wrong". I then explained the mathematical proof to the AI, and the AI apologized, admitted it had been wrong, and agreed with the proof. Only after hearing the proof did the AI agree with the text book. This is some pretty advanced reasoning.

        I performed that experiment a few times and it played out mostly the same. I experimented with giving the AI a flawed proof (I purposely made mistakes in the mathematical proofs), and the AI would call out my mistakes and would not be convinced by faulty proofs.

        A standard that judged this AI to have "no understanding of any concepts whatsoever", would also conclude the same thing if applied to most humans.

  • This is tough. I believe there is a lot of unfair wealth concentration in our society, especially in the tech companies. On the other hand, I don't want AI to be stifled by bad laws.

    If we try to stop AI, it will only take it away from the public. The military will still secretly use it, companies might still secretly use it. Other countries will use it and their populations will benefit while we languish.

    Our only hope for a happy ending is to let this technology be free and let it go into the hands of many companies and many individuals (there are already decent models you can run on your own computer).

    • So, in your "only hope for a happy ending" scenario, how do the artists get paid? Or will we no longer need them after AI runs everything ;)

      • I don't know. I only believe that things will be worse if individuals cannot control these AIs.

        Maybe these AI have reached a peak (at least for now), and so they aren't good enough to write a compelling novel. In that case, writers who produce good novels and get lucky will still get paid, because people will want to buy their work and read it.

        Or maybe AI will quickly surpass all humans in writing ability, in which case, there's not much we can do. If the AI produces books that are better, then people will want AI produced books. They might have to get those from other countries, or they might have to get them from a secret AI someone is running on a beefy computer in their basement. If AI surpasses humans then that's not a happy day for writers, no way around it. Still, an AI that surpasses humans might help people in other ways, but only if we allow everyone to have and control their own AI.

        As the industrial revolution threatened to swallow society Carl Marx wrote about how important it was that regular people be able to control "the means of production". At least that part of his philosophy has always resonated with me, because I want to be empowered as an individual, I want the power to create and compete in our society. It's the same now, AI threatens to swallow society and I want to be able to control my own AI for my own purposes.

        If strong AI is coming, it's coming. If AI is going to be the source of power in society then I want regular people to have access to that power. It's not yet clear whether this is the case, but if strong AI is coming it's going to be big, and writers complaining about pay isn't going to stop it.

        All that said, I believe we do a terrible job of caring for individuals in our society. We need more social safety nets, we need to change to provide people better and happier lives. So I'm not saying "forget the writers, let them stave".

      • To be honest, I don't think AI is going to get good enough to replace human creativity. Sure, some of the things that AI can do is pretty nice, but these things are mostly already solved problems - sure, AI can make passable art, but so can humans - and they can go further than art, they can create logical connections between art pieces, and extrapolate art in new reasonably justified ways, instead of the direction-less, grasping in the dark methods that AI seems to do it in.

        Sure, AI can make art a thousand times faster than a person, but if only one in thousand is tolerably good, then what's the problem?

  • Isn’t learning the basic act of reading text? I’m not sure what the AI companies are doing is completely right but also, if your position is that only humans can learn and adapt text, that broadly rules out any AI ever.

    • Isn’t learning the basic act of reading text?

      not even close. that's not how AI training models work, either.

      if your position is that only humans can learn and adapt text

      nope-- their demands are right at the top of the article and in the summary for this post:

      Thousands of authors demand payment from AI companies for use of copyrighted works::Thousands of published authors are requesting payment from tech companies for the use of their copyrighted works in training artificial intelligence tools

      that broadly rules out any AI ever

      only if the companies training AI refuse to pay

      • Isn’t learning the basic act of reading text?

        not even close. that’s not how AI training models work, either.

        Of course it is. It's not a 1:1 comparison, but the way generative AI works and the we incorporate styles and patterns are more similar than not. Besides, if a tensorflow script more closely emulated a human's learning process, would that matter for you? I doubt that very much.

        Thousands of authors demand payment from AI companies for use of copyrighted works::Thousands of published authors are requesting payment from tech companies for the use of >> their copyrighted works in training artificial intelligence tools

        Having to individually license each unit of work for a LLM would be as ridiculous as trying to run a university where you have to individually license each student reading each textbook. It would never work.

        What we're broadly talking about is generative work. That is, by absorbing one a body of work, the model incorporates it into an overall corpus of learned patterns. That's not materially different from how anyone learns to write. Even my use of the word "materially" in the last sentence is, surely, based on seeing it used in similar patterns of text.

        The difference is that a human's ability to absorb information is finite and bounded by the constraints of our experience. If I read 100 science fiction books, I can probably write a new science fiction book in a similar style. The difference is that I can only do that a handful of times in a lifetime. A LLM can do it almost infinitely and then have that ability reused by any number of other consumers.

        There's a case here that the renumeration process we have for original work doesn't fit well into the AI training models, and maybe Congress should remedy that, but on its face I don't think it's feasible to just shut it all down. Something of a compulsory license model, with the understanding that AI training is automatically fair use, seems more reasonable.

      • Okay, given that AI models need to look over hundreds of thousands if not millions of documents to get to a decent level of usefulness, how much should the author of each individual work get paid out?

        Even if we say we are going to pay out a measly dollar for every work it looks over, you’re immediately talking millions of dollars in operating costs. Doesn’t this just box out anyone who can’t afford to spend tens or even hundreds of millions of dollars on AI development? Maybe good if you’ve always wanted big companies like Google and Microsoft to be the only ones able to develop these world-altering tools.

        Another issue, who decides which works are more valuable, or how? Is a Shel Silverstein book worth less than a Mark Twain novel because it contains less words? If I self publish a book, is it worth as much as Mark Twains? Sure his is more popular but maybe mine is longer and contains more content, whats my payout in this scenario?

    • A key point is that intellectual property law was written to balance the limitations of human memory and intelligence, public interest, and economic incentives. It's certainly never been in perfect balance. But the possibility of a machine being able to consume enormous amounts of information in a very short period of time has never been a variable for legislators. It throws the balance off completely in another direction.

      There's no good way to resolve this without amending both our common understanding of how intellectual property should work and serve both producers and consumers fairly, as well as our legal framework. The current laws are simply not fit for purpose in this domain.

357 comments